We provide IT Staff Augmentation Services!

Hadoop Developer Resume

3.00/5 (Submit Your Rating)

San Francisco, CA

SUMMARY

  • 7+ years of IT experience in Analysis, Design, Development, Implementation, Integration and testing of Application Software in web - based environments, distributed n-tier products and Client/Server architectures.
  • Experience in SDLC and Object Oriented Application Design and Programming.
  • Strong programming skills in Core Java technology. Strong programming skills in advanced frameworks like Spring, Struts and Hibernate.
  • Experience in OO Design using IBM Rational Rose and UML.
  • Strong experience with SOA and Web Services.
  • Strong hands on experience of 3 years with Big Data Technologies including Hadoop (HDFS & Map Reduce), PIG,HIVE, HBASE, ZOOKEEPER, SQOOP
  • Excellent knowledge on Hadoop Architecture; as in HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Experience with distributed systems, large scale non-relational data stores, MapReduce systems, data modeling, and big data systems.
  • Involved in developing solutions to analyze large data sets efficiently
  • Excellent hands on with importing and exporting data from different Relational Database Systems like MySQL and Oracle into HDFS and Hive and vice-versa, using Sqoop.
  • Hands-on experience in writing Pig Latin scripts, working with grunt shells and job scheduling with Oozie.
  • Experience in analysing data using Hive QL, Pig Latin, and custom MapReduce programs in Java.
  • Experience with web-based UI development using jQuery, ExtJS, CSS, HTML, HTML5, XHTML and JavaScript.
  • Knowledge of job workflow scheduling and monitoring tools like Oozie and Zookeeper
  • Experience with databases like DB2, Oracle 9i, Oracle 10g, MySQL, SQL Server and MS Access.
  • Experience in creating complex SQL Queries and SQL tuning, writing PL/SQL blocks like stored procedures, Functions, Cursors, Index, triggers and packages.
  • Very good understanding on NOSQL databases like MongoDB and HBase.
  • Have good Knowledge in ETL and hands on experience in Informatica ETL.
  • Experienced in SDLC, Agile (SCRUM) Methodology, Iterative Waterfall
  • Experience in developing test cases, performing Unit Testing, Integration Testing, experience in QA with test methodologies and skills for manual/automated testing using tools like WinRunner, JUnit.
  • Experience with various version control systems Clear Case, CVS, SVN.
  • Expertise in extending Hive and Pig core functionality by writing custom UDFs.
  • Development Experience with all aspects of software engineering and the development life cycle
  • Good Experience on SDLC (Software Development Life cycle).
  • Possess strong Communication skills of written, oral, interpersonal and presentation.

TECHNICAL SKILLS

Big Data Ecosystem: Hadoop, MapReduce, HDFS, HBase, Zoo Keeper, Hive, Pig, Sqoop

Java/J2EE: Java 6, Ajax, Log4j, JSP 2.1 Servlets 2.3, JDBC 2.0, XML, Java Beans

Methodologies: Agile, UML, Design Patterns

Frameworks: Struts, Hibernate, Spring

Database: Oracle 10g, PL/SQL, MySQL

Application Server: Apache Tomcat 5.x 6.0, JBoss 4.0

Web Tools: HTML, Java Script, XML, XSL, XSLT, XPath, DOM

IDE/ Testing Tools: NetBeans, Eclipse

Scripts: Bash, ANT, SQL, HiveQL, Shell Scripting

Testing API: JUNIT

PROFESSIONAL EXPERIENCE

Hadoop Developer

Confidential - San Francisco, CA

Responsibilities:

  • Involved in gathering the requirements, designing, development, testing & implement.
  • Developed multiple Map Reduce jobs in java for data cleaning and preprocessing.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Responsible for developing data pipeline using flume, Sqoop and pig to extract the data from weblogs and store in HDFS.
  • Used Oozie workflow engine and UC4 scheduling to run multiple Hive and Pig Jobs.
  • Experienced in managing and reviewingHadooplog files.
  • Used Sqoop to import and export data from HDFS to RDBMS and vice-versa.
  • Created Hive tables and involved in data loading and writing Hive UDFs.
  • Experienced in runningHadoopstreaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.
  • Worked with systems to accept events from Kafka producer and emit into DB.
  • Involved in developingHadoopMapReduce jobs for merging and appending the repository data.
  • Experience in optimization of Map reduce algorithm using combiners and partitions to deliver the best results and worked on Application performance optimization for a HDFS cluster.
  • Hands on experience in setting up HBase Column based storage repository for archiving and retro data.
  • Working experience with NOSQL database.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Experience working with off-shore teams and communicating daily status on issues, road-blocks.

Environment: Hadoop, MapReduce, HDFS, Hive, PIG, Sqoop, Oozie, UC4, Kafka, Cloudera, Flume, HBase, ZooKeeper, Oracle, NoSQL and Unix/Linux, Java (JDK 1.6), Eclipse

Hadoop Developer

Confidential - Chevy Chase, MD

Responsibilities:

  • Developed Managed, External and partition tables as per the requirement.
  • Experience in loading and transforming of large sets of structured, semi structured and unstructured data.
  • Ingested structured data into appropriate schemas and tables to support the rule and analytics.
  • Developed custom User Defined Function (UDF's) in Hive to transform the large volumes of data with respect to business requirement.
  • Responsible for building scalable distributed data solutions usingHadoop.
  • Involved in loading data from edge node to HDFS using shell scripting.
  • Implemented scripts for loading data from UNIX file system to HDFS.
  • Implemented a script to transmit sysprint information from Oracle to HBase using Sqoop.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Automated workflow using Shell Scripts.
  • Good experience in Hive partitioning, bucketing and perform different types of joins on Hive tables and implementing Hive series like REGEX, JSON and Avro.
  • Used various compression techniques like grip, bzip2, snappy.
  • Developed Pig Scripts, Pig UDFs and Hive Scripts, Hive UDFs to load data files.
  • Used Kafka for messaging services instead of message broker.
  • Experience inHadoop2.x with spark and Scale.
  • ManagedHadoopjobs using Oozie workflow scheduler system for Map Reduce, Hive, Pig and Sqoop actions.
  • Good knowledge on Data Ingestion and Data Processing.
  • Sound knowledge on Python and R.
  • Experience in managing and reviewingHadooplog files.
  • Used Oozie workflow engine to run multiple Hive and pig jobs.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Responsible to manage the test data coming from different sources.
  • Responsible for developing batch process using Unix Shell Scripting.

Environment: ApacheHadoop, HDFS, Hive, Pig, Sqoop, HBase, UNIX, Shell Scripting, Spark, Scale, Kafka, Oozie, Zookeeper, CDH5

Hadoop Developer

Confidential - Denver, CO

Responsibilities:

  • Installed and configured Hadoop Map Reduce, HDFS, developed multiple Map Reduce jobs in Java for data cleaning and preprocessing.
  • Good understanding and related experience with Hadoop stack - internals, Hive, Pig and Map/Reduce
  • Wrote Map Reduce jobs to discover trends in data usage by users.
  • Involved in defining job flows
  • Involved in managing and reviewing Hadoop log files
  • Involved in running Hadoop streaming jobs to process terabytes of text data
  • Load and transform large sets of structured, semi structured and unstructured data
  • Responsible to manage data coming from different sources
  • Supported Map Reduce Programs those are running on the cluster
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive QL scripts.
  • Responsible to manage data coming from different sources
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in map reduce way
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.
  • Used ClearCase for version control.
  • Used JUnit for unit testing and Continuum for integration testing.

Environment: Hadoop, Map Reduce, HDFS, Hive, Java (jdk1.6), Hadoop distribution of Horton Works, Cloudera, Map, Flat files, Oracle 11g/10g, UNIX Shell Scripting, ClearCase, JUnit.

Java/J2EE Developer

Confidential - Dallas, TX

Responsibilities:

  • Designing UML diagrams for the entire application using Rational Rose products.
  • Used Struts 1.2 framework for implementing the frontend of the reporting module.
  • Developed web components using Struts framework Incorporated design patterns like Factory pattern, Session Façade, and Singleton pattern.
  • Also used Gang of four (Goff) design patterns.
  • Developed MDB in EJB 2.0 for handling offline requests through a WebSphere MQ.
  • Used JDBC to manage Profile-related tables in My Profile data store.
  • XPath was used for parsing SOAP requests.
  • We used XSL&XSLT for extracting and transforming the data between these systems.
  • Involved plugging in other tools that talk to Mainframe systems.
  • Implemented SOAP functionality for messaging across various types of systems - POS (.Net) in the Front-end, and GDS Mainframes in the backend, and parsing of XSL content.
  • Developed JSPs and JavaBeans.
  • Created EJB session and message-driven beans.
  • Tracing and Logging frameworks implemented using Log4j.
  • Usage of Exception handling framework to wrap the exceptions appropriately and show user-friendly error messages to the end user.
  • Use of Spring's backend plug-in for Hibernate to support the My Profile database application which consisted of user profiles.
  • Used MVC architecture (Jakarta Struts framework) for Web tier Debugging was an important part of the development and maintenance phase.
  • Used CVS as the version-control tool for managing the development cycle. Unit testing with JUnit at various stages of application development.

Environment: J2SE 1.4, J2EE 1.4 (Servlets 2.4, JavaBeans 1.0, JSP 2.0, EJB 2.1, JDBC 3.0),Struts 1.2, WebSphere MQ, Web sphere Application Server 5.1, WSAD 5.1, JUnit 3.8.1, XML, XSL, XSLT, SAX parser, SOAP, Spring 1.2, Hibernate 2.1, CVS, Log4j, Windows XP, Oracle 9i, Mainframes.

Java/J2EE Developer

Confidential - Rockford, IL

Responsibilities:

  • Analyzed Business Requirements and Identified mapping documents required for system and functional testing efforts for all test scenarios.
  • Performed Requirement Gathering & Analysis by actively soliciting, analyzing and negotiating customer requirements and prepared the requirements specification document for the application using Microsoft Word.
  • Developed Use Case diagrams, business flow diagrams, Activity/State diagrams.
  • Adopted J2EE design patterns like Service Locator, Session Facade and Singleton.
  • Configuration of application using spring, Hibernate, DAO’s, Actions Classes, Java Server Pages.
  • Configuring Hibernate and Tiles related XML files.
  • Developed presentation layer using Java Server Faces (JSF) MVC framework.
  • Used JSP, HTML and CSS, JQuery as view components in MVC.
  • Extensively used Spring IOC for Dependency Injection and worked on Custom MVC Frameworks loosely based on Struts.
  • Developed Servlets and Java Server Pages (JSP), to route the submittals to the EJB components and render-retrieved information using Session Facade, Service Locator (design pattern).
  • Developed J2EE components on Eclipse IDE.
  • Used JDBC to invoke Stored Procedures and also used JDBC for database connectivity to SQL.
  • Deployed the applications on Web sphere Application Server
  • Developed Web services using Restful and JSON.
  • Used Oracle11g database for tables creation and involved in writing SQL queries using Joins and Stored Procedures.
  • Used Toad database tool to develop oracle quires.
  • Writing complex SQL queries and reviewing SQL queries for the other team members.
  • Developed JUnit Test Cases for Code unit test.
  • Worked with configuration management groups for providing various deployment environments set up including System Integration testing, Quality Control testing etc.

Environment: J2EE, Hibernate, RAD, RSA, SQL Developer, Oracle 11g, Rational Clear Case, Rational Clear Quest, Rational Requisite Pro, UML, MS Visio, MS Office.

Java/J2EE Developer

Confidential

Responsibilities:

  • Developed Use Case Diagrams, Object Diagrams and Class Diagrams in UML using Rational Rose As a team member involved in Development using JSP, Servlets and an internal tool called Content Management to organize the JSPs created.
  • Provide a login screen for the employees with validation, added client side validations using java script and server side validation using Struts Validationframework.
  • Developed Add/Edit KYC Details, Enhanced KYC, Risk Score/EDD and Comments modules using J2EE architecture in Spring MVC Framework by implementing MVC model, which included invocation of Web Services/EJB for integration with external system.
  • Used JavaScript for client-side validations, and AJAX to create interactive front-end GUI.
  • Used Rational Clear Case for version control and log4j for logging purposes.
  • Designed WSDL's and elaborated type of definitions; schemas based on corporate standards, integration and security specifications Tracked errors or unexpected results/bugs with Bugzilla and coordinated with help desk staff.
  • Maintained version control using VSS.
  • Performed Unit Test using Junit, Integration Testing, and System Testing.

Environment: Java, JSP, Servlets, Struts 1.2, JavaScript, Eclipse and Oracle as backend.

We'd love your feedback!