We provide IT Staff Augmentation Services!

Associate Technical Consultant Resume

Tampa, FL

PROFESSIONAL SUMMARY:

  • Overall 6+ years of experience in Software applications development including Analysis, Design, Development, Integration, Testing and Maintenance.
  • Work experience in Big data/ Hadoop Development and Ecosystem Analytics using programming languages like Java and Scala.
  • Experienced developing big data applications in cloud, specifically amazon web services.
  • Have good experience of Spark SQL and Spark using Scala.
  • Configured Spark Streaming to receive real time data from the Apache Kafka and store the stream data to HDFS using Scala.
  • Experienced in building highly scalable Big - data solutions using Hadoop and multiple distributions i.e., Cloudera, Hortonworks and NoSQL platforms.
  • Expertise in Big data architecture with Hadoop File system and its eco system tools MapReduce, HBase, Hive, Agile, Apache spark and Spark Streaming and Spark SQL.
  • Hands on experience in Apache Sqoop, Apache Storm and Apache Hive integration.
  • Hands on experience working with different File Formats like TEXTFILE, JSON, AVROFILE, ORC for HIVE querying and processing.
  • Expertise in migration data from different databases (i.e. SQL, Oracle, DB2, Teradata) to HDFS.
  • Experience in designing and coding web applications using Core Java & Web Technologies- JSP, Servlets and JDBC.
  • Experience in designing the User Interfaces using HTML, CSS, JavaScript and JSP.
  • Experience in version control tools like Git
  • Good working Experience in Bigdata (Spark) in Azure/on-premises project Planning, Analysis, Design, Programming, Unit Testing of new applications and troubleshooting existing programs in production.
  • Developed Scala 2.10+ applications on Hadoop and Spark SQL for high-volume and real-time data processing.
  • Knowledge in Spark APIs to cleanse, explores, aggregate, transform, and store data.
  • Experience with RDBMS and writing SQL and PL/SQL scripts used in stored procedures.
  • Familiar with Spark, Kafka and Event Hubs.
  • Knowledge in Spark APIs to cleanse, explores, aggregate, transform, and store data.
  • Experience with RDBMS and writing SQL and PL/SQL scripts used in stored procedures.
  • Strengths include good team player, excellent communication interpersonal and analytical skills, flexible to work with new technologies and ability to work effectively in a fast-paced, high volume, deadline-driven environment.

TECHNICAL EXPERIENCE:

Big Data Technologies: HDFS, MapReduce, Hive, Pig, Sqoop, Oozie, Storm, Zookeeper, Kafka, Event Hubs, Apache Spark, Spark Streaming, PySpark, Spark SQL, HBase and Cassandra

Hadoop Distributions: Azure, Cloudera, Hortonworks.

Databases: DB Oracle, MySQL, NoSql, Teradata

Programming Languages: SQL, Scala, Java, Python, Unix Shell Scripting

Java Technologies: JDBC, Servlets, JSP, Spring and Hibernate

Operating System: Windows XP/7/8/10, Linux Distro (Ubuntu), Cent OS.

Tools: & Utilities: HP Quality Center, Git, Maven.

EXPERIENCE SUMMARY:

Confidential, Tampa, FL

Associate Technical Consultant

Responsibilities:

  • Interacting with Business Analysts to understand the requirements behind BRD/FRD/SR.
  • Low level/high level design documents for the application/framework components.
  • Design, develop and enhance enterprise applications in risk technology area using Big Technologies using Spark.
  • Develop Spark framework components to process data and integrate with Hadoop, Oozie and Hive systems to perform CECL reserves Calculation and Aggregation.
  • Develop Batch processing for Model output validations.
  • Develop data layer to persist account and non-account level data.
  • Participate in Unit Testing, Integration Testing and UAT/SIT support.
  • Responsible for fixing high priority issues in production environment and support in all application activities.
  • Develop integration component to interact with Genesis systems used to exchange the data between CECL and Genesis for account level and non-account level calculations.
  • Develop workflow to process control and data files for Account and non-account model outputs.
  • Implement batch processing mechanism using Spring and Google Guise framework.
  • Implemented Data processing through IBatis for interacting with Oracle database.
  • Develop components to fetch aggregation data from HDFS.
  • Develop oozie workflow for scheduling the jobs.
  • Implement File Watcher, trigger batch processing when the files are available in landing zone.
  • Develop code to store aggregation data into Hive systems.
  • Check in and maintain the code components in Git Repository.
  • Implement AutoSys jobs for scheduling batch jobs.

Environment: Apache Spark, Java, Hive, HDFS, Jira, Oracle PLSQL, Autosys, Tectia, Git, Maven, Eclipse.

Confidential, Redmond, WA

Spark/Azure Developer

Responsibilities:

  • Evaluated Business requirements and prepared detailed specifications that follow project guidelines required to develop written programs.
  • Worked on analyzing Hadoop cluster and different big data analytical and processing tools including Pig, Hive, Spark and Spark Streaming.
  • Design solution for various system components using Confidential Azure
  • Experience developing multi-user cloud hosted software with one or more public cloud platforms such as Azure
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Performed Code Optimizations to improve the performance of the applications.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and exported the data from HDFS to MYSQL using Sqoop.
  • Experience with Azure relational and no-SQL experience, HDInsights, Apache Storm, Spark
  • Configured Spark Streaming to receive real time data from the Event Hubs and store the stream data to HDFS using Scala.
  • Developed numerous Spark jobs in Scala 2.10.x for Data Cleansing and Analyzing Data in Impala 2.1.0.
  • Hands on experience in Spark and Spark Streaming creating RDD's, applying operations -Transformation and Actions.
  • Used HIVE to analyze the partitioned and bucketed data and compute various metrics for reporting.
  • Develop cutting edge cloud platform solutions using the latest tools and platforms like Azure, etc
  • Experience with Azure PaaS services such as web sites, SQL, Stream Analytics, IoT Hubs, Event Hubs, Data Lake, Azure Data Factory
  • Hands on experience in working on Spark SQL queries, Data frames, and import data from Data sources, perform transformations; perform read/write operations, save the results to output directory into HDFS.
  • Responsible for Ingestion of Data from Blob to Kusto and maintaining the PPE and PROD pipelines.
  • Collaborated with the infrastructure, network, database, application and teams to ensure data quality and availability.
  • Experienced in running query using Impala and used BI tools to run ad-hoc queries directly on Hadoop.
  • Responsible for creating Hive tables, partitions, loading data and writing hive queries.
  • Imported and exported the data using Sqoop between Hadoop Distributed File System (HDFS) and Relational Database systems.
  • Hands on experience in Spark and Spark Streaming creating RDD's, applying operations -Transformation and Actions.

Environment: MapReduce, Azure, Python, HDFS, Hive, Pig, Spark, Spark-Streaming, Spark SQL, Sqoop, Java, Scala, Eclipse, Git, Shell Scripting and Cassandra.

Confidential, Fort Wayne, IN

Hadoop/Spark Developer

Responsibilities:

  • Involved in Cluster Setup, monitoring and administration tasks like commission and decommission nodes.
  • Good at working on Hadoop, MapReduce, and Yarn/MRv2 developed multiple MapReduce jobs for structured, semi-structured and unstructured data in java.
  • Developed MapReduce programs in Java for parsing the raw data and populating staging Tables.
  • Created Hive queries to compare the raw data with EDW tables and performing aggregates
  • Experienced in developing custom input formats and data types to parse and process unstructured and semi structured input data and mapped them into key value pairs to implement business logic in Map-Reduce.
  • Experience in implementing custom sterilizer, interceptor, source and sink as per the requirement in Flume to ingest data from multiple sources.
  • Perform big data processing using Hadoop, MapReduce, Sqoop, Oozie, and Impala
  • Involved in developing Hive DDLs to create, alter and drop Hive tables and storm, & Kafka.
  • Experience in setting up Fan-out workflow in flume to design v shaped architecture to take data from many sources and ingest into single sink.
  • Implemented extensive Impala 2.7.0 queries and creating views for adhoc and business processing.
  • Used Spark Streaming APIs to perform transformations and actions on the fly for building common learner data model which gets the data from Kafka in near real time and persist it to Cassandra.
  • Consumed JSON messages using Kafka and processed the JSON file using Spark Streaming to capture UI updates
  • Performance tuning of SQOOP, Hive and Spark jobs.
  • Worked with .Net and C# to create dash board according to the client requirements.
  • Experienced in writing live Real-time Processing and core jobs using Spark Streaming with Kafka as a data pipe-line system.
  • Implemented OLAP multi-dimensional cube functionality using Azure SQL Data Warehouse.
  • Wrote AZURE POWERSHELL scripts to copy or move data from local file system to HDFS Blob storage
  • Used Spark Streaming APIs to perform transformations and actions on the fly for building common learner data model which gets the data from Kafka in near real time and persist it to Cassandra.
  • Experienced in writing live Real-time Processing and core jobs using Spark Streaming with Kafka as a data pipe-line system.
  • Used Kafka Streams to Configure Spark streaming to get information and then store it in HDFS.
  • Experienced in managing and reviewing Hadoop log files.
  • Involved in Data Migration process using Azure by integrating with Github repository and Jenkins.
  • Experienced in writing live Real-time Processing using Spark Streaming with Kafka.
  • Developed custom mappers in python script and Hive UDFs and UDAFs based on the given requirement.
  • Used HiveQL to analyze the partitioned and bucketed data and compute various metrics for reporting.
  • Experienced in querying data using SparkSQL on top of Spark engine.
  • Experience in managing and monitoring Hadoop cluster using Cloudera Manager.
  • Implemented analytical platform that used Hive Functions and different kind of join operations like Map joins, Bucketed Map joins.
  • Unit tested a sample of raw data and improved performance and turned over to production.

Environment: CDH, Java (JDK1.7), Impala, Hadoop, Azure, MapReduce, HDFS, Hive, Sqoop, Flume, NFS, Cassandra, Pig, Oozie, Kerberos, Scala, SparkSQL, Spark Streaming, Kafka, Linux, Shell Scripting, MySQL Oracle 11g, SQL*PLUS, C++, C#

Confidential, Madison, WI

Hadoop Developer

Responsibilities:

  • Understand the exact requirement of report from the Business groups and users.
  • Imported trading and derivatives data in Hadoop Distributed File System using Eco System components MapReduce, Pig, Hive, Sqoop.
  • Was part of activity to setup Hadoop ecosystem at development & QA Environment.
  • Managed and reviewed Hadoop Log files.
  • Responsible writing PIG Script and Hive queries for data processing.
  • Running Sqoop for importing data from Oracle & Oracle & Another Database.
  • Creation of shell script to collect raw logs from different machines.
  • Created Partitions in Hive as static and dynamic.
  • Implemented Pig Latin scripts using operators such as LOAD, STORE, DUMP, FILTER, DISTINCT, FOREACH, GENERATE, GROUP, COGROUP, ORDER, LIMIT and UNION.
  • Defined some PIG UDFs for some functions such as swap, hedging, Speculation and arbitrage.
  • Coded MapReduce program to process unstructured logs file.
  • Worked on Import and export data into HDFS and Hive using Sqoop.
  • Used parameterize Pig Script and optimized script using illustrate and explain.
  • Involved in the process of configuring HA, Kerberos security issues and name node failure restoration activity time to time as a part of zero downtime.
  • Implemented FAIR Scheduler as well.
  • Used Spring framework that handles application logic and makes calls to business, make them as Spring Beans.
  • Implemented, configured data sources, session factory and used Hibernate Template to integrate Spring framework with Hibernate.
  • Developed JUNIT test cases for application unit testing.
  • Used SVN as version control to check in the code, created branches and tagged the code in SVN.
  • Used RESTFUL Services to interact with the Client by providing the RESTFUL URL mapping.
  • Used Log4j framework to log/track application and debugging.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Shell Scripting, Sqoop, Java, Eclipse, Spring, Hibernate, SOAP, REST, SVN, Log4j

Confidential

Java/J2EE Developer

Responsibilities:

  • Involved in the complete SDLC software development life cycle of the application from requirement analysis to testing.
  • Developed the modules based on Struts MVC Architecture.
  • Followed AGILE methodology (Scrum Stand-ups, Sprint Planning, Sprint Review, Sprint Showcase and Sprint Retrospective meetings).
  • Developed business components using Core Java concepts and classes like Inheritance, Polymorphism, Collections, Serialization and Multithreading etc.
  • Developed the Web Interface using Servlets, Java Server Pages, HTML and CSS.
  • Developed the DAO objects using JDBC.
  • Business Services using the Servlets and Java.
  • Used Spring Framework for Dependency injection and integrated with the Struts Framework and Hibernate.
  • Developed JUnit test cases for all the developed modules.
  • Used Log4j to capture the log that includes runtime exceptions, monitored error logs and fixed the problems.
  • Performed Unit Testing, System Testing and Integration Testing.
  • Provided technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects. Resolved more priority defects as per the schedule.

Environment: Java, JDBC, spring, Hibernate, HTML, CSS, Java Script, Log4j, Oracle, Struts and Eclipse.

Confidential

Jr. Java/J2EE Developer

Responsibilities:

  • Developed User Interfaces module using JSP, Java Script, DHTML and form beans for presentation layer.
  • Developed Servlets and Java Server Pages (JSP).
  • Developed PL/SQL queries, and wrote stored procedures and JDBC routines to generate reports based on client requirements.
  • Enhancement of the System according to the customer requirements.
  • Involved in the customization of the available functionalities of the software for an NBFC (Non-Banking Financial Company).
  • Involved in putting proper review processes and documentation for functionality development.
  • Providing support and guidance for Production and Implementation Issues.
  • Used Java Script validation in JSP.
  • Used Hibernate framework to access the data from back-end SQL Server database.
  • Used AJAX (Asynchronous JavaScript and XML) to implement user friendly and efficient client interface.
  • Used MDB for consuming messages from JMS queue/topic.
  • Designed and developed Web Application using Struts Framework.
  • ANT to compile and generate EAR, WAR, and JAR files.
  • Created tes1t case scenarios for Functional Testing and wrote Unit test cases with JUnit.
  • Responsible for Integration, unit testing, system testing and stress testing for all the phases of project.

Environment: Java, J2EE, JSP 1.2, Performance Tuning, Spring1.2,Hibernate 2.0, JSF1.2,EJB 1.2, IBM WebSphere6.0, Servlets, JDBC, XML, XSLT, DOM, CSS, HTML, DHTML, SQL, Java Script, Log4J, ANT1.6, WSAD6.0, Oracle 9i, Windows 2000.

Hire Now