Sr. Hadoop/ Spark Developer Resume
O Fallon, MO
PROFESSIONAL SUMMARY:
- 7+ years IT experience in design, development, testing, deployment & post production maintenance.
- 3+ years of experience in Hadoop Ecosystem (MapReduce,Hive, Pig, Flume, Sqoop,Spark, HBase,Oozie, Yarn, Kafka,Impala, Zookeeper) as Developer.
- Experience in analyzing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
- Extending Hive and Pig core functionality by writing custom UDFs.
- Hands on Experience with installation, configuration, supporting and managing of Big Data and underlying infrastructure of Hadoop Cluster.
- Experience in managing and reviewing Hadoop log files.
- Involved in converting Hive/SQL queries into Spark transformations using Spark RDDs
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice - versa.
- Extensive experience in Unix Shell scripting.
- Experience with build tools like Jenkins.
- Automated several batch process to eliminate manual intervention.
- Good understanding of NoSQL databases and hands on experience in writing applications on NoSQL databases like HBase and Cassandra.
- Experience in J2EE technologies including Servlets, JSP, spring, Struts, JMS, Hibernate, JDBC, XML, XSLT and JNDI for building various client-server applications and Unix Shell Scripting.
- Working experience withOracle, MYSQL, DB2 databases and writing SQL, PLSQLStored Procedures, Triggers, Functions and Packages.
- Experience in Middle Tier Development using Design Patterns, Servlets, JSPs, Custom Tag Libraries, XML, and EJB.
- Experience in working with application servers like JBoss, Weblogic, IBM Websphere, and Apache Tomcat.
- Experience in data processing and interfacing (Back-end Front-end connectivity) using Oracle 10g/11g (SQL, Stored procedures), and MySQL.
- Hands on experience in using JUnit, Log4J for debugging and testing the applications.
- Good knowledge in Financial, HealthCare and Insurance domains.
- Solid experience in communicating with all levels of enterprise.
- Following Agile methodology and SCRUM meetings to track optimize and resolve impediments to meet deadlines on time.
- Effective leadership quality with good skills in strategy, business development, clientmanagement and project management.
- Experience in Onsite - Offshore model.
- Strong Problem Solving and Analytical skills and abilities to make Balanced and Independent Decisions.
- Motivation, Initiative and Project Management Attributes.
TECHNICAL SKILLS:
Hadoop Ecosystem: Map Reduce, Hive, Yarn, Pig, Flume, Scoop, Oozie,Hbase,Spark
Java Technologies: Java, J2EE, Servlets, Custom Tag Libraries, EJB, JDBC, JMS, JSF, RMI, JNDI, JSTL.
Web Technologies: HTML/DHTML, JSPs, CSS, AJAX, JavaScript
XML Related: XML, XSLT, JAXP, JAXB, JAXRPC, DOM, SAX, XMLSchema, DTD, XMLBeans
Frameworks: Jakarta Struts, Spring, Enterprise Java Beans (EJB)
Languages: Java,Scala, Python, C, C++, SQL
Tools/Software: Apache Axis, Log4j, Ant, Toad, AltovaXMLSpy, Maven, UML, Rational Rose, MVC architecture, Unix Shell Script
Databases: Oracle, PostGresSQL, SQL Server 2005
ORM Tool: Hibernate
Application/Web Servers: WebLogic, IBM Websphere, Apache Tomcat, JBoss
Development tools: RAD, Eclipse, Net Beans
Operating Systems: Windows 2000/XP, Linux, UNIX
PROFESSIONAL EXPERIENCE:
Confidential, O’Fallon, MO
Sr. Hadoop/ Spark Developer
Responsibilities:
- Involved in requirement sessions along with all the business stakeholders.
- Developed multiple MapReduce programs in Java.
- Developed UDF in Java
- Defining Job flows and dependencies
- Written Unix shell scripts to integrate all the job flows
- Experience with continuous integration build tools like Jenkins
- Involved in Unit testing, System integration testing and provided support for UAT testing.
- Involved in deployments in Pre-prod & Production.
- Responsible for ingesting data into HDFS and HBASE from different data sources.
- Developed data ingestion scripts using Sqoop.
- Handled initial loads and delta during the ingestion process
- Wrote shell scripts to invoke MapReduce programs, Hive scripts, Sqoop scripts etc.
- Worked on performance tunings.
- Performed advanced procedures like text analytics and processing, using the in-memory computing capabilities of Spark using Scala.
- Experience in streaming the data using Spark with Kafka.
- Import data using Sqoopto load data from RDBMS(Oracle, PostGresSQL) to HDFS on regular basis.
- Played a key role in installation and configuration of the various Hadoop ecosystem tools such as Solr, Kafka, Pig,HBase.
- Implemented Daily Oozie jobs that automate parallel tasks of loading the data into HDFS and pre-processing with Pig using Oozie co-coordinator jobs.
- Created Hive tables and wrote queries to fetch the data from HDFS.
- Created the developer Unit test plans and executed unit testing in the development cluster.
- Involved in automating the test flows and integrating them in the continuous integration tools like Jenkins
- Participated in code review sessions.
Environment:Hadoop, HDFS, HBASE, Spark, Spark-SQL, Talend, Oozie, Flume, Sqoop, Pig, Hive,SQL,XML, Eclipse, Java Regex, Jira, Shell Scripting.
Confidential, Dallas, TX
Sr.Hadoop/ Java Developer
Responsibilities:
- Involved in functional requirement review. Worked closely with Risk & Compliance Team and BA.
- Developed complex Map Reduce programs in Java for Data Analysis on different data formats.
- Developed Unix shell scripts to invoke all the process flows.
- Designed and configured Flume servers to collect data from the network proxy servers and store to HDFS and HBASE.
- Involved in Unit testing, System integration testing and provided support for UAT testing.
- Used agent E2E Chain for reliability and failover in flume.
- Worked on documentation of all Extract, Transform and Load, designed, developed, validated and deploy the Talend ETL Processes for Data ware house team using PIG and HIVE.
- Worked on Sequence files, RC files, Map side joins, partitioning, bucketing for Hive performance enhancement and storage improvement.
- Designed and implemented Restful APIs to retrieve the data from Hadoop Platform to Employee Portal Web Application.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Extensively used Sqoop to get data from RDBMS sources like Teradata and Netezza.
- Installed Oozieworkflow engine to run multiple Hive and Pig Jobs.
- Wrote MRUnit tests for unit testing the Map Reduce jobs.
- Implemented Hive, Pig Generic UDF's to implement business logic.
- Worked on installing cluster, commissioning & decommissioning of DataNodes, NameNode recovery, capacity planning, and slots configuration.
- ImplementedNameNode backup using NFS. This was done for High availability.
- Involved in the installation of HDP 2.3 and up-gradation from HDP 2.3to HDP 2.4.0
- Responsible for developing data pipeline using HDInsight, flume, Sqoop and pig to extract the data from weblogs and store in HDFS.
- Used Apache Log4J for logging.
- Facilitated Knowledge transfer sessions.
- Use TDD and Worked in an agile environment.
Environment:Hadoop, Yarn, HDFS,HDP,Flume, Sqoop, Hive, Pig,Hbase,Java,Scala, solr, Teradata, Netezza, Spring, Eclipse, Maven, XML, GIT, JSON, Linux, Avro.
Confidential, Atlanta, GA
Hadoop Developer
Responsibilities:
- Developed Map Reduce programs in java for parsing the raw data and populating staging tables.
- Implemented Generic writable to incorporate multiple data sources into reducer to implement recommendation based reports using Map Reduce programs.
- Responsible for ingesting data into HDFS and HBASE from different data sources.
- Developed data ingestion scripts using Sqoop.
- Handled initial loads and delta during the ingestion process
- Importing and exporting data into HDFS and Hive using Sqoop
- Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures
- Load and transform large sets of structured, semi structured and unstructured data
- Load log data into HDFS using Flume.
- Wrote shell scripts to monitor the health check of Hadoop daemon services and respond accordingly to any warning or failure conditions
- Involved in loading data from UNIX file system to HDFS, configuring Hive and writing Hive UDFs
- Gained very good business knowledge on health insurance, claim processing, fraud suspect identification, appeals process, etc.
Environment: java, Hadoop, Linux, MapReduce, HDFS, Hive,Sqoop, Flume, pig, J2EE, MySQL, Shell Scripting, Eclipse, SVN, Maven, JIRA.
Confidential, Richardson, TX
Java/J2EE Developer
Responsibilities:
- Involved in quickly understanding the requirements and reporting the manager about the daily status of this project as this was a very urgent need for the company.
- Extensively used design pattern like Application façade, Data Access object and MVC.
- Worked with different layers like Java beans, Dao, transactions, service and presentation.
- Developed a presentation layer using JSP, CSS and JavaScript.
- Understood the existing code and involved in helping and explaining the team members about the flow of the code.Worked on Server Tuning for increase the heap size for UNIX machines.
- Handled offshore team for implement Customer gateway design/architecture.
- Worked on Pl/SQL for query and fetch the data from database.
- Implemented web layer using JSF and Ice faces.
- Implemented business layer using Spring MVC.
- Involved in writing the POJO code and making it consistent with the existing code.
- Involved in interacting and arranging a live meeting session with the team member to be clear and to be on the same page.
- Involved in testing the code locally and fixing the bugs reported by the QA.
- Involved in debugging the code and replicating the problems reported by the tester.
- Wrote different UTPs for the team and the tester and also worked-through the project in the live meeting session.
- Worked with the SERENA dimension to check-in and check-out the project.
- Used RAD 7.0 to develop the Web Application and deployed on WebSphere application server.
Environment: RAD 7.0, WebSphere 6.1, JAVA, JSP, HTML, JavaScript, CSS, Serena Dimensions, DB2, JavaBeans, Design Pattern, ANT.
Confidential
Java/J2EE Developer
Responsibilities:
- Used WebSphere, which has high performance and full-integrated Java platform for Enterprise Applications.
- Actively involved in component development, deployment for the application interface.
- Strongly followed the coding standards and implemented MVC Design Patterns.
- Involved in creating EJBs that handle business logic and persistence of data.
- Understood the existing code and involved in helping and explaining the team members about the flow of the code.
- Involved in impact analysis of Change requests and Bug fixes.
- Unit testing and integration testing of the modules.
- Involved in testing the code locally and fixing the bugs reported by the QA.
- Involved in the configuration management using ClearCase.
- Detecting and resolving errors/defects in the quality control environment.
- Using Ibatis for mapping Java classes with database.
- Involved in Code review and integration testing.
- Integrated the modules with the other modules of the system.
- Wrote different UTPs for the team and the tester and also worked-through the project in the live meeting session.
- Java Naming/Directory Interface (JNDI) to support transparent access to distributed components.
Environment: Sybase, WebSphere Studio Application Developer WSAD, Enterprise Java Beans (EJB), Struts, WebSphere Application Server, HTML, Java.