We provide IT Staff Augmentation Services!

Hadoop Developer Resume

5.00/5 (Submit Your Rating)

Norfolk, VA

SUMMARY

  • Around 8 years of overall IT experience in a variety of industries, which includes hands on experience of 2+ years in Big Data technologies and extensive experience of 5+ years in Java.
  • More than +2 years’ experience with the tools in Hadoop Ecosystem including Pig, Hive, HDFS, MapReduce, Sqoop, Oozie, Zookeeper.
  • Experience in migrating the data using Sqoop from HDFS to Relational Database System and vice - versa according to client’s requirement.
  • Experience in analysing data using HiveQL, Pig Latin, HBase and custom Map Reduce programs in Java.
  • Good knowledge of Data warehousing concepts and ETL processes.
  • Knowledge of manipulating/analyzing large datasets and finding patterns and insights within structured and unstructured data.
  • Knowledgeable of Spark and Scala mainly in framework exploration for transition from Hadoop/MapReduce to Spark.
  • Expertise in working with HDP 2.x (HortonWorks).
  • Strong experience in Object-Oriented Design, Analysis, Development, Testing and Maintenance.
  • Excellent implementation knowledge of Enterprise/Web/Client Server using Java, J2EE.
  • Strong experience in designed, developed, deployed, XQuery and XSLT transformations.
  • Experience in XML, XSLT, XSD, XQuery.
  • Implemented SOA architecture with web services using SOAP, WSDL, UDDI and XML.
  • Experienced in using agile approaches, including Extreme Programming, Test-Driven Development and Agile Scrum.
  • Represented platform in work identification, scoping, allocation and prioritization meetings with business users, application owners, architects including VPs.
  • Identified several new business streams for the application including online presentation of certain instruments, simplified handling of change requests, simplified content changes etc.
  • Identified application performance parameters using industry standard tools and documented them for future applications built in the same space.
  • Identified as a major contributor to Customer Acquisitions for work in letters processing. Worked extensively on Visual Compuset. Built several reusable styles for different needs.
  • Loading of data into HDFS, processing of data using PIG scripts and loading of data into Hive.
  • Worked in large and small teams for systems requirement, design & development.
  • Key participant in all phases of software development life cycle with Analysis, Design, Development, Integration, Implementation, Debugging, and Testing of Software Applications in client server environment, Object Oriented Technology and Web based applications.
  • Experience in using various IDEs Eclipse, My Eclipse and repositories SVN and CVS.
  • Experience of using build tools Ant and Maven.
  • Experience in implementing the declarative enterprise services using Spring AOP and implementation knowledge on Spring web flow to manage the application page flows.
  • Preparation of Standard Code guidelines, analysis and testing documentations.
  • Self-motivated, responsible and proper time management with good Written, Verbal and Listening skill, commitment to co-operative teamwork.
  • Good team player with reputation of integrity and an ability to work in multiple areas.

TECHNICAL SKILLS

Hadoop eco system components: Hadoop, HDFS, Map reduce, yarn, hive, pig, sqoop, HBase and flume.

Hadoop Clusters: Hortonworks and Cloudera

Cloud Environments: AWS, Apache Tomcat

Programming Languages: Java, J2EE

Scripting Languages: Shell script, SQL, Java script, JQuery, Ajax, Html, CSS

Frameworks: Spring, Struts, Hibernate

Tools: Eclipse, Maven, SVN, CVS

PROFESSIONAL EXPERIENCE

Confidential, Norfolk, VA

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig.
  • Provide mentorship and guidance to other architects to help them become independent.
  • Provide review and feedback for existing physical architecture, data architecture and individual code.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Involved in complete Implementation lifecycle, specialized in writing custom MapReduce, Pig and Hive programs. Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Extensively used Hive/HQL or Hive queries to query or search for a particular string in Hive tables in HDFS.
  • Did various performance optimizations like using distributed cache for small datasets, Partition, Bucketing in hive and Map Side joins.
  • Experience in developing customized UDF's in java to extend Hive and Pig Latin functionality.
  • Created HBase tables to store various data formats of data coming from different portfolios.
  • Managing and scheduling Jobs to remove the duplicate log data files in HDFS using Oozie.
  • Implemented test scripts to support test driven development and continuous integration.
  • Responsible to manage data coming from different sources.
  • Wrote MapReduce jobs to discover trends in data usage by users.
  • Involved in running Hadoop streaming jobs to process terabytes of text data.
  • Analyzed large data sets by running Hive queries and Pig scripts.
  • Develop Core Framework based on Hadoop to Migrate Existing ETL (RDBMS) Solution.
  • Wrote Pig Scripts to generate Map Reduce jobs and performed ETL procedures on the data in HDFS.
  • A deep and thorough understanding of ETL tools and how they can be applied in a Big Data environment.
  • Responsible for smooth error-free configuration of DWH-ETL solution and Integration with Hadoop.
  • Worked extensively with Sqoop for importing metadata from Oracle.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries.
  • Responsible for managing data from multiple source.
  • Designed, developed and did maintenance of data integration programs in a Hadoop and RDBMS environment with both traditional and non-traditional source systems as we as RDBMS and NoSQL data stores for data access and analysis. Experienced in running Hadoop streaming jobs to process terabytes of xml format data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources.
  • Assisted in exporting analyzed data to relational databases using Sqoop.
  • Expert knowledge developing and debugging in Java/J2EE.
  • Wrote Hive Queries and UDF’s.
  • Developed Hive queries to process the data and generate the data cubes for visualizing.
  • Created Pig Latin scripts to sort, group, join and filter the enterprise wise data.
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Gained experience in managing and reviewing Hadoop log files.

Environment: HDP2.x, Hadoop, MapReduce, HDFS, Pig, Hive, HBase, Java, Oracle 10g, MySQL, RedHat

Confidential, Minneapolis, MN

Hadoop/Big Data Developer

Responsibilities:

  • Involved in defining job flows, managing and reviewing log files.
  • Supported Map Reduce Programs those are running on the cluster.
  • Designed and implemented Incremental Imports into Hive tables.
  • Involved in Unit testing and delivered Unit test plans and results documents using Junit and MRUnit.
  • Involved in collecting, aggregating and moving data from servers to HDFS using Apache Flume.
  • Written Hive jobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data.
  • Wrote multiple java programs to pull data from Hbase.
  • Involved with File Processing using Pig Latin.
  • Involved in creating Hive tables, loading with data and writing hive queries that will run internally in map reduce way.
  • Experience in optimization of Map reduce algorithm using combiners and partitions to deliver the best results and worked on Application performance optimization for a HDFS cluster.
  • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
  • As a Big Data Developer, implemented solutions for ingesting data from various sources and processing the Data-at-Rest utilizing Big Data technologies such as Hadoop, MapReduce Frameworks, HBase, Hive, Oozie, Flume, Sqoop etc.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it.
  • Imported Bulk Data into HBase Using Map Reduce programs.
  • Developed and written Apache PIG scripts and HIVE scripts to process the HDFS data.
  • Perform analytics on Time Series Data exists in HBase using HBase API.
  • Worked on debugging, performance tuning of Hive & Pig Jobs
  • Used Hive to find correlations between customer's browser logs in different sites and analyzed them to build risk profile for such sites.
  • Created and maintained Technical documentation for launching HADOOP Clusters and for executing Hive queries and Pig Scripts.

Environment: Java, Hadoop 2.1.0, Map Reduce2, Pig 0.12.0, Hive 0.13.0, Linux, Sqoop 1.4.2, Flume 1.3.1, Eclipse, AWS EC2, and Cloudera CDH 4.

Confidential, Dallas, TX

Java Developer

Responsibilities:

  • Designed and developed the application using Agile methodology.
  • Implementation of new module development, new change requirement, fixes the code. Defect fixing for defects identified in pre-production environments and production environment.
  • Wrote technical design document with class, sequence, and activity diagrams in each use case.
  • Created Wiki pages using Confluence Documentation.
  • Developed various reusable helper and utility classes which were used across all modules of the application.
  • Involved in developing XML compilers using XQuery.
  • Developed the Application using Spring MVC Framework by implementing Controller, Service classes.
  • Involved in writing Spring Configuration XML file that contains declarations and other dependent objects declaration.
  • Used Hibernate for persistence framework, Involved in creating DAO's and used Hibernate for ORM mapping.
  • Written Java classes to test UI and Web services through JUnit.
  • Performed functional and integration testing, Extensively involved in release/deployment related critical activities. Responsible for designing Rich user Interface Applications using JSP, JSP Tag libraries, Spring Tag libraries, JavaScript, CSS, HTML.
  • Used SVN for version control. Log4J was used to log both User Interface and Domain Level Messages.
  • Used SoapUI for testing the Web services.
  • Use of MAVEN for dependency management and structure of the project
  • Create the deployment document on various environments such as Test, QC, and UAT.
  • Involved in system wide enhancements supporting the entire system and fixing reported bugs.
  • Explored Spring MVC, Spring IOC, Spring AOP, Hibernate in creating the POC.
  • Done data manipulation on front end using JavaScript and JSON.

Environment: Java, J2EE, JSP, Spring, Hibernate, CSS, JavaScript, Oracle, Jboss, Maven, Eclipse, JUnit, Log4J, AJAX, Web services, JNDI, JMS, HTML, XML, XSD, XML Schema .

Confidential, Springfield, IL

Java Developer

Responsibilities:

  • Responsible for Business Analyst activities for critical functionality for the business interfacing projects.
  • Responsible for successful deployment of all Major and Minor Releases.
  • Developed key modules of the Application using Frameworks / Languages / Tools as JMS, Spring 2.5, Oracle 9i. And Hibernate 3.0.
  • As application developer I was actively involved in designing of various Business Layer and Data Management components of this web based system over J2EE architecture using JSTL, JSP, Servlets, and JavaScript.
  • Implemented MVC architecture using Spring Framework, coding involved writing Action Classes/Custom Tag Libraries, JSP
  • Have been involved in the design and key component(Matching Engine) of the system using PL/SQL procedure on Oracle Database
  • Built packages and procedures for designing business rules for the applications on the database side.
  • Provided request and reports and Supporting Client Services for all customer requests being an SME of the application Experience developing for Unix/Linux based systems Development of Tools and Value adds to assist Performance Testing.
  • Used AGILE methodology.
  • Was involved in creating front-end pages using HTML 4.01, CSS and JavaScript.
  • Used log4j to log exception and modified the xml to roll up, zip and archive logs.

Environment: Java 1.5, JSP, JSTL, JMS, Spring 2.5, Servlets, JavaScript, Oracle 9i, HTML 4.01, CSS, Log4j, Hibernate 3.0, PL/SQL, AGIL

Confidential

Java Developer

Responsibilities:

  • Responsible for Design and Development of business and presentation layers.
  • Developed the application using the MVC design pattern and encapsulated front end from the underlying business logic, and data model.
  • Responsible for generation of XSLT.
  • Worked on Struts Framework at later stages of the project to support ongoing development by the developers.
  • Created Dynamic web pages using HTML, and XSLT.
  • Developed front end templates and hand-code pages using HTML, JavaScript, and CSS.
  • Developed EJB’s which invokes the Unit-Tel web services using apache axis SOA
  • Implemented different EJBs like State full session and stateless session beans.
  • Developed the JSPs as web based user interface validated all client Connect screens using Java Scripts.
  • Used the J2EE components like, JSPs including tag libraries, Servlets, EJBs.
  • Conducted Code review with team members.
  • Implemented STRUTS Framework including, creating templates for Action Form classes.
  • Implemented internationalization concepts using Struts Framework.
  • Wrote Session Beans to handle Business Logic.
  • Developed XML file for each server. The XML files have information of all the components reporting to that server. XSL was used to transform XML document. GUI will use this data to draw Tables and Menus with specified attributes.
  • Created unit test cases, tested application using J Unit testing tool.

Environment: Core Java, J2EE, CSS, Servlets, JSP, Struts, Tiles, XML, Tomcat, and MySQL, EJB

We'd love your feedback!