We provide IT Staff Augmentation Services!

Sr. Hadoop/ Spark Developer Resume

3.00/5 (Submit Your Rating)

O Fallon, MO

PROFESSIONAL SUMMARY:

  • 7+ years IT experience in design, development, testing, deployment & post production maintenance.
  • 3+ years of experience in Hadoop Ecosystem (MapReduce,Hive, Pig, Flume, Sqoop,Spark, HBase,Oozie, Yarn, Kafka,Impala, Zookeeper) as Developer.
  • Experience in analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
  • Extending Hive and Pig core functionality by writing custom UDFs.
  • Hands on Experience with installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster.
  • Experience in managing and reviewing Hadoop log files.
  • Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Extensive experience in Unix Shell scripting.
  • Experience with build tools like Jenkins.
  • Automated several batch process to eliminate manual intervention.
  • Good understanding of NoSQL databases and hands on experience in writing applications on NoSQL databases like HBase and Cassandra.
  • Experience in J2EE technologies including Servlets, JSP, spring, Struts, JMS, Hibernate, JDBC, XML, XSLT and JNDI for building various client-server applications and Unix Shell Scripting.
  • Working experience withOracle, MYSQL, DB2 databases and writing SQL, PLSQLStored Procedures, Triggers, Functions and Packages.
  • Experience in Middle Tier Development using Design Patterns, Servlets, JSPs, Custom Tag Libraries, XML, and EJB.
  • Experience in working with application servers like JBoss, Weblogic, IBM Websphere, and Apache Tomcat.
  • Experience in data processing and interfacing (Back-end Front-end connectivity) using Oracle 10g/11g (SQL, Stored procedures), and MySQL.
  • Hands on experience in using JUnit, Log4J for debugging and testing the applications.
  • Good knowledge in Financial, HealthCare and Insurance domains.
  • Solid experience in communicating with all levels of enterprise.
  • Following Agile methodology and SCRUM meetings to track optimize and resolve impediments to meet deadlines on time.
  • Effective leadership quality with good skills in strategy, business development, clientmanagement and project management.
  • Experience in Onsite - Offshore model.
  • Strong Problem Solving and Analytical skills and abilities to make Balanced and Independent Decisions.
  • Motivation, Initiative and Project Management Attributes.

TECHNICAL SKILLS:

Hadoop Ecosystem: Map Reduce, Hive, Yarn, Pig, Flume, Scoop, Oozie,Hbase,Spark

Java Technologies: Java, J2EE, Servlets, Custom Tag Libraries, EJB, JDBC, JMS, JSF, RMI, JNDI, JSTL.

Web Technologies: HTML/DHTML, JSPs, CSS, AJAX, JavaScript

XML Related: XML, XSLT, JAXP, JAXB, JAXRPC, DOM, SAX, XMLSchema, DTD, XMLBeans

Frameworks: Jakarta Struts, Spring, Enterprise Java Beans (EJB)

Languages: Java,Scala, Python, C, C++, SQL

Tools/Software: Apache Axis, Log4j, Ant, Toad, AltovaXMLSpy, Maven, UML, Rational Rose, MVC architecture, Unix Shell Script

Databases: Oracle, PostGresSQL, SQL Server 2005

ORM Tool: Hibernate

Application/Web Servers: WebLogic, IBM Websphere, Apache Tomcat, JBoss

Development tools: RAD, Eclipse, Net Beans

Operating Systems: Windows 2000/XP, Linux, UNIX

PROFESSIONAL EXPERIENCE:

Confidential, O’Fallon, MO

Sr. Hadoop/ Spark Developer

Responsibilities:

  • Involved in requirement sessions along with all the business stakeholders.
  • Developed multiple MapReduce programs in Java.
  • Developed UDF in Java
  • Defining Job flows and dependencies
  • Written Unix shell scripts to integrate all the job flows
  • Experience with continuous integration build tools like Jenkins
  • Involved in Unit testing, System integration testing and provided support for UAT testing.
  • Involved in deployments in Pre-prod & Production.
  • Responsible for ingesting data into HDFS and HBASE from different data sources.
  • Developed data ingestion scripts using Sqoop.
  • Handled initial loads and delta during the ingestion process
  • Wrote shell scripts to invoke MapReduce programs, Hive scripts, Sqoop scripts etc.
  • Worked on performance tunings.
  • Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala.
  • Experience in streaming the data using Spark with Kafka.
  • Import data using Sqoopto load data from RDBMS(Oracle, PostGresSQL) to HDFS on regular basis.
  • Played a key role in installation and configuration of the various Hadoop ecosystem tools such as Solr, Kafka, Pig,HBase.
  • Implemented Daily Oozie jobs that automate parallel tasks of loading the data into HDFS and pre-processing with Pig using Oozie co-coordinator jobs.
  • Created Hive tables and wrote queries to fetch the data from HDFS.
  • Created the developer Unit test plans and executed unit testing in the development cluster.
  • Involved in automating the test flows and integrating them in the continuous integration tools like Jenkins
  • Participated in code review sessions.

Environment:Hadoop, HDFS, HBASE, Spark, Spark-SQL, Talend, Oozie, Flume, Sqoop, Pig, Hive,SQL,XML, Eclipse, Java Regex, Jira, Shell Scripting.

Confidential, Dallas, TX

Sr.Hadoop/ Java Developer

Responsibilities:

  • Involved in functional requirement review. Worked closely with Risk & Compliance Team and BA.
  • Developed complex Map Reduce programs in Java for Data Analysis on different data formats.
  • Developed Unix shell scripts to invoke all the process flows.
  • Designed and configured Flume servers to collect data from the network proxy servers and store to HDFS and HBASE.
  • Involved in Unit testing, System integration testing and provided support for UAT testing.
  • Used agent E2E Chain for reliability and failover in flume.
  • Worked on documentation of all Extract, Transform and Load, designed, developed, validated and deploy the Talend ETL Processes for Data ware house team using PIG and HIVE.
  • Worked on Sequence files, RC files, Map side joins, partitioning, bucketing for Hive performance enhancement and storage improvement.
  • Designed and implemented Restful APIs to retrieve the data from Hadoop Platform to Employee Portal Web Application.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Extensively used Sqoop to get data from RDBMS sources like Teradata and Netezza.
  • Installed Oozieworkflow engine to run multiple Hive and Pig Jobs.
  • Wrote MRUnit tests for unit testing the Map Reduce jobs.
  • Implemented Hive, Pig Generic UDF's to implement business logic.
  • Worked on installing cluster, commissioning & decommissioning of DataNodes, NameNode recovery, capacity planning, and slots configuration.
  • ImplementedNameNode backup using NFS. This was done for High availability.
  • Involved in the installation of HDP 2.3 and up-gradation from HDP 2.3to HDP 2.4.0
  • Responsible for developing data pipeline using HDInsight, flume, Sqoop and pig to extract the data from weblogs and store in HDFS.
  • Used Apache Log4J for logging.
  • Facilitated Knowledge transfer sessions.
  • Use TDD and Worked in an agile environment.

Environment:Hadoop, Yarn, HDFS,HDP,Flume, Sqoop, Hive, Pig,Hbase,Java,Scala, solr, Teradata, Netezza, Spring, Eclipse, Maven, XML, GIT, JSON, Linux, Avro.

Confidential, Atlanta, GA

Hadoop Developer

Responsibilities:

  • Developed Map Reduce programs in java for parsing the raw data and populating staging tables.
  • Implemented Generic writable to incorporate multiple data sources into reducer to implement recommendation based reports using Map Reduce programs.
  • Responsible for ingesting data into HDFS and HBASE from different data sources.
  • Developed data ingestion scripts using Sqoop.
  • Handled initial loads and delta during the ingestion process
  • Importing and exporting data into HDFS and Hive using Sqoop
  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures
  • Load and transform large sets of structured, semi structured and unstructured data
  • Load log data into HDFS using Flume.
  • Wrote shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions
  • Involved in loading data from UNIX file system to HDFS, configuring Hive and writing Hive UDFs
  • Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process, etc.

Environment: java, Hadoop, Linux, MapReduce, HDFS, Hive,Sqoop, Flume, pig, J2EE, MySQL, Shell Scripting, Eclipse, SVN, Maven, JIRA.

Confidential, Richardson, TX

Java/J2EE Developer

Responsibilities:

  • Involved in quickly understanding the requirements and reporting the manager about the daily status of this project as this was a very urgent need for the company.
  • Extensively used design pattern like Application façade, Data Access object and MVC.
  • Worked with different layers like Java beans, Dao, transactions, service and presentation.
  • Developed a presentation layer using JSP, CSS and JavaScript.
  • Understood the existing code and involved in helping and explaining the team members about the flow of the code.Worked on Server Tuning for increase the heap size for UNIX machines.
  • Handled offshore team for implement Customer gateway design/architecture.
  • Worked on Pl/SQL for query and fetch the data from database.
  • Implemented web layer using JSF and Ice faces.
  • Implemented business layer using Spring MVC.
  • Involved in writing the POJO code and making it consistent with the existing code.
  • Involved in interacting and arranging a live meeting session with the team member to be clear and to be on the same page.
  • Involved in testing the code locally and fixing the bugs reported by the QA.
  • Involved in debugging the code and replicating the problems reported by the tester.
  • Wrote different UTPs for the team and the tester and also worked-through the project in the live meeting session.
  • Worked with the SERENA dimension to check-in and check-out the project.
  • Used RAD 7.0 to develop the Web Application and deployed on WebSphere application server.

Environment: RAD 7.0, WebSphere 6.1, JAVA, JSP, HTML, JavaScript, CSS, Serena Dimensions, DB2, JavaBeans, Design Pattern, ANT.

Confidential 

Java/J2EE Developer

Responsibilities:

  • Used WebSphere, which has high performance and full-integrated Java platform for Enterprise Applications.
  • Actively involved in component development, deployment for the application interface.
  • Strongly followed the coding standards and implemented MVC Design Patterns.
  • Involved in creating EJBs that handle business logic and persistence of data.
  • Understood the existing code and involved in helping and explaining the team members about the flow of the code.
  • Involved in impact analysis of Change requests and Bug fixes.
  • Unit testing and integration testing of the modules.
  • Involved in testing the code locally and fixing the bugs reported by the QA.
  • Involved in the configuration management using ClearCase.
  • Detecting and resolving errors/defects in the quality control environment.
  • Using Ibatis for mapping Java classes with database.
  • Involved in Code review and integration testing.
  • Integrated the modules with the other modules of the system.
  • Wrote different UTPs for the team and the tester and also worked-through the project in the live meeting session.
  • Java Naming/Directory Interface (JNDI) to support transparent access to distributed components.

Environment: Sybase, WebSphere Studio Application Developer WSAD, Enterprise Java Beans (EJB), Struts, WebSphere Application Server, HTML, Java.

We'd love your feedback!