We provide IT Staff Augmentation Services!

Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Bellevue, WA

SUMMARY

  • 8+ years of overall IT experience and 4+ Years of comprehensive experience as a Apache Hadoop Developer. Expertise in writing Hadoop Jobs for analyzing data using Hive, Pig and oozie.
  • Java Programmer with over 5+ years of Extensive programming experience in developing web based applications and Client - Server technologies using Java, J2EE.
  • Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts.
  • Experience in working with MapReduce programs using Hadoopfor working with Big Data.
  • Experience in analyzing data using Hive QL, Pig Latin and custom MapReduce programs in Java.
  • Build a continuous ETL pipeline by using Kafka, Spark Streaming and HDFS.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Working experience on designing and implementing complete end-to-end Hadoop Infrastructure including PIG, HIVE, Sqoop, Oozie, Flume and zookeeper.
  • Experience in providing support to data analyst in running Pig and Hive queries.
  • Developed Map Reduce programs to perform analysis.
  • Performed Importing and exporting data into HDFS and Hive using Sqoop.
  • Experience in writing shell scripts to dump the Sharded data from MySQL servers to HDFS.
  • Experience in designing both time driven and data driven automated workflows using Oozie.
  • Experience in setting up Infiniband network and build Hadoop cluster to improve the map reduce performance.
  • Experience in performance tuning the Hadoop cluster by gathering and analyzing the existing infrastructure.
  • Experience in automating the Hadoop Installation, configuration and maintaining the cluster by using the tools like Puppet.
  • Experience with Spark and Scala.
  • Experience with reporting tool Tableau.
  • Extensively used ETL methodology for supporting Data Extraction, transformations and loading processing.
  • Ability to design and develop Extract Transform Load (ETL) packages in Oracle for manipulating data as per the requirement of the project.
  • Strong debugging and problem solving skills with excellent understanding of system development methodologies, techniques and tools.
  • Worked in complete Software Development Life Cycle (analysis, design, development, testing, implementation and support) in different application domain involving different technologies varying from object oriented technology to Internet programming on Windows NT, Linux and UNIX/ Solaris platforms and RUP methodologies.
  • Experience in implementing Big data solution with cloud based technologies like Amazon web services.
  • Strong with RDBMS concepts and worked on Oracle 8i/9i, SQLServer 7.0., DB2 8.x/7.x
  • Involved in writing shell scripts, Ant scripts for Unix OS for application deployments to production region.
  • Exceptional ability to quickly master new concepts and capable of working in-group as well as independently with excellent communication skills.

TECHNICAL SKILLS

Languages/Tools: Java, C, C++, VB, XML, HTML/XHTML, HDML, DHTML.

Big Data: Hadoop, Map Reduce, Scala, Hive, Pig, Sqoop and MRUnit

J2EE Standards: JDBC, JNDI, JMS, Java Mail & XML Deployment Descriptors.

Web/Distributed: J2EE, Servlets 2.1/2.2, JSP 2.0, Struts 1.1, Hibernate 3.0, JSF

Technologies: JSTL1.1,EJB 1.1/2.0, RMI,JNI, XML, JAXP, XSL, XSLT, UML, MVC, STRUTS, Spring 2.0, Corba, Java Threads.

Operating System: Windows 95/98/NT/2000/XP, MS-DOS, UNIX, Linux6.2

Databases: Oracle 8i/9i, MS SQL Server 2000, DB2, MS Access & MySql.

Browser Languages: HTML, XHTML, CSS, XML, XSL, XSD, XSLT.

Browser Scripting: Java script, HTML DOM, DHTML, AJAX.

App/Web Servers: IBM Websphere 5.1.2/5.0/4.0/3.5 , BEA Weblogic 5.1/7.0, Jdeveloper, Apache Tomcat, JBoss.

GUI Environment: Swing, AWT.

Messaging &: SOAP, WSDL,UDDI, XML, SOA, JAX-RPC, IBM WebSphere MQ v5.3, Web Services Technology JMS.

Networking Protocols: HTTP, HTTPS, FTP, UDP, TCP/IP, SNMP, SMTP, and POP3.

Testing &Case Tools: JUnit, Log4j, Rational Clear case, CVS, ANT, JBuilder.

PROFESSIONAL EXPERIENCE

Confidential, Bellevue, WA

Hadoop Developer

Responsibilities:

  • Worked in a fast paced Scrum, dev-ops and cross functional core development team.
  • Worked extensively on designing and implementing complete end-to-end Hadoop solution starting from data ingestion to target table dispatch using PIG, HIVE, Sqoop, Oozie and zookeeper.
  • Worked with Teradata to understand and create mock data for data virtualization and SAP BO reports.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the IDW.
  • Developed custom artifacts for data cleansing, preparation and dispatch jobs using technologies like Hive, HBase, Pig, and MapReduce.
  • Installed and configured Pig and also written Pig Latin scripts for complex data transformations according to the Source to Target Mapping.
  • Worked moderately on source system analysis, attribute mapping and Data modeling as part of the cross functional dev-ops team.
  • Worked extensively on performance of queries by initiating comprehensive dialogue with the business users.
  • Developing scripts and batch Jobs to schedule various Hadoop Program by using the Control M workflow automation tool.
  • Involved in dev-testing the Hadoop deliverables in both the Dev and QA environments.
  • Performed the role of a Scrum master on a time to time basis and well versed with Agile methodology.
  • Worked in two week sprints in a Scrum team using Rally as an enterprise-scale agile platform.

Environment: Java 1.5, J2EE, Servlets, JSP 2.5, JUNIT4.1, SVN, Struts 1.2, Hibernate 3.2.x, Oracle10g, Apache, Weblogic 9.2, Eclipse IDE 3.6, Log4J, Ajax, HTML 3, Design Patterns.

Confidential, Chicago, IL

Hadoop Developer

Responsibilities:

  • Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Installed and configured Pig and also written PigLatin scripts.
  • Wrote MapReduce job using Pig Latin.
  • Have solid understanding of REST architecture style and its application to well performing web sites for global usage.
  • Involved in ETL, Data Integration and Migration. Imported data using Sqoop to load data from Oracle to HDFS on regular basis.
  • Developing Scripts and Batch Job to schedule various Hadoop Program.
  • Written Hive queries for data analysis to meet the business requirements.
  • Creating Hive tables and working on them using Hive QL. Importing and exporting data into HDFS from Oracle Database and vice versa using Sqoop.
  • Implemented test scripts to support test driven development and continuous integration.
  • Responsible to manage data coming from different sources.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Experience in managing and reviewing Hadoop log files.
  • Worked on Hive for exposing data for further analysis and for generating transforming files from different analytical formats to text files.
  • Managing and scheduling Jobs on a Hadoop cluster.
  • Installed and configured Hadoop MapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Involved in creating Hive tables, loading with data and writing hive queries, which will run internally in map, reduce way.
  • Used Pig as ETL tool to do transformations, event joins, filter bot traffic and some pre-aggregations before storing the data onto HDFS.
  • Written Hive queries for data analysis to meet the business requirements.
  • Involved in writing Hive scripts to extract, transform and load the data into Database.
  • Used JIRA for bug tracking.
  • Used CVS for version control.

Environment: Hadoop, Hive, Linux, MapReduce, HDFS, Hive, Pig, Sqoop, Shell Scripting, Java (JDK 1.6), Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Toad 9.6, Linux, JIRA 5.1, CVS, JIRA 5.2.

Confidential, Pittsburgh

Java/J2EE Developer

Responsibilities:

  • Involved in Business requirements gathering, Design, Development and unit testing of BillPay Account Accelerator (BPAA) & Alphanumeric ID prjects.
  • Involved in maintenance & development of pnc.com and their related web sites like pncvirtualwallet, Wealth Management & Mutual Funds etc.
  • Responsible for developing use cases, class and sequence diagram for the modules using UML and Rational Rose.
  • Involved in preparation of docs like Functional Specification document and Deployment Instruction documents.
  • Set up the deployment environment on WebSphere 6.1 Developed system preferences UI screens using JSP2.0 and HTML.
  • Used Java Script for Client side validations.
  • Code and Unit Test according to client standards. Provide production support and quickly resolving the issues until Integration Test is passed.
  • Fix defects as needed during the QA phase, support QA testing, troubleshoot defects and identify the source of defects.
  • Used JMS for Point-to-Point asynchronous messaging for high transactional Banking operation.
  • Involved in preparation of unit and system test cases and testing of the module in 3 phases named unit testing and system testing and regression testing.
  • Involved in writing shell scripts, Ant scripts for Unix OS for application deployments on production region.
  • Developed core banking business components as a Web Service for enterprise-wide SOA Architecture strategy.
  • Used Rational ClearCase as source control management system.
  • Implemented SOA architecture with web services using SOAP, WSDL, UDDI and XML.
  • Involved in deployments in all environments like Dev, Test,UAT and prod respctively.
  • Involved in design Credit Card Service layer on mainframe with MQ series and WBI. Provide XML based messaging service to front-end applications.
  • Extensively used IBM RAD 7.1 IDE for building, testing, and deploying applications.
  • Responsible for developing use cases, class and sequence diagram for the modules using UML and Rational Rose 2000
  • Worked with Single Sign-On (SSO) using SAML for retrieving data from third party applications like Yodlee.

Environment: Java (jdk1.5), J2EE, WebSphere 6.1, IBM RAD 7.5, Rational Clearcase 7.0, XML, JAXP, XSL, XSLT, XML Schema(XSD), WSDL 2.0, SAML 2.0, AJAX 1.0, Web Services, SOA, JSP 2.2, CSS, Servlets, JProfiler, Struts 2.0, Spring, Rational HATS, JavaScript, JCF, HTML, IBM DB2, JMS, AXIS 2, Swing, MQ, Open source technologies (ANT, LOG4j and Junit), Oracle 10g, UNIX

Confidential

Java Developer

Responsibilities:

  • Implemented new features like creating highly preferment, multi-threaded transforms to process incoming messages into trading object model using Java, Struts 1.2.
  • Conducted client side validations using JavaScript.
  • Coded JDBC calls in the servlets to access the Oracle database tables. Used JMS/MQ for service implementation. Also invoked EJB 2.1 Stateless Session beans for business service implementation.
  • Designed and coded the java class that will handle errors and will log the errors in a file.
  • Involved in creating User Authentication page using Java Servlets.
  • Responsible for Integration, unit testing, system testing and stress testing for all the phases of project.
  • Implemented the project using JSP and Servlets based tag libraries.
  • Generate SQL Scripts to update the parsed message into Database.
  • Completed proof of concepts using Java, JSP, Servlets and XML using JBoss server for selecting different standards.

Environment: JDK 1.3, Java, JMS/MQ, J2EE, Struts 1.2, EJB 2.1, Servlets, Oracle, XML, JDBC, Eclipse 3.3, JUnit, JBoss Application Server

We'd love your feedback!