We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

Charlotte, NC

SUMMARY

  • 7+ years of professional experience inHadoopand Java technologies like HDFS, MapReduce,Apache Pig, Hive, Hbase, Sqoop, Oracle, JSP, JDBC and Spring
  • 3 years of experience in Client - Server and Web based application development using Java technologies Java, J2EE, JSP, Java Script, Servlets, Springs.
  • 4+ years of working experience inHadoopeco-system technologies like Apache Pig, Apache Hive, Apache Sqoop, Apache Flume and Apache Hbase.
  • Experience in analyzing data using HiveQL, Pig latin and custom MapReduce programs in Java.
  • Hands on experience in writing Pig UDFs, Hive UDFs and UDAFs in the analysis of data.
  • Worked with NoSQL database like HBase.
  • Experience in importing and exporting data from relational database into HDFS using Sqoop.
  • Developed MapReduce jobs, Used different optimization techniques to improve performance in Map Reduce Programs.
  • Good knowledge in using job scheduling and monitoring tools like Oozie and ZooKeeper
  • Extensive experience in configuring Flume to stream data into HDFS.
  • Experience in implementing Spark in Scala and SparkSQL for faster analyzing and processing of data.
  • Excellent understanding ofHadooparchitecture and its components like HDFS, MapReduce, Job Tracker, Task Tracker, Name Node and Data Node.
  • Used Apache Impala to read, write and query theHadoopdata in HDFS, Hbase and Cassandra.
  • Hands on experience in Application Development using Java,Hadoop, RDBMS and Linux shell scripting.
  • Extensive experience in developing and deploying Java based applications.
  • Involved in working on all phases of software development life cycle (SDLC) from requirements gathering to programming, testing and maintenance.
  • Strong skills in designing, developing and testing of Client-Server products and distributed programming using Java, J2EE and Web related technologies.
  • Expertise in J2EE Application development using JSP, Servlets, JDBC, XML, Spring.
  • Strong experience in handling different Web Servers like Tomcat, and Application Servers like Web logic, Websphere and JBOSS.
  • Experienced in GUI design. Extensively used HTML, XML, Java Script, and JSP.
  • Involved in the development of Spring JDBC DAO Support for data base interactions.
  • Involved in developing Static and Dynamic pages using JSP and Servlets.
  • Setup Struts framework in Web logic Server, Tomcat.
  • Involved in Software development Lifecycle.
  • Functional skills include, project management, Leadership skills, Quality control, L&D
  • Good knowledge with PL/SQL stored procedures using Oracle.
  • Ability to master new Concepts.
  • Excellent problem solving capabilities and communication skills.
  • Have proficient time-management skills, able to deliver assignments with superior quality on schedule and within budget.

TECHNICAL SKILLS:

Big Data Technologies: MapReduce, Pig, Hive, Impala, Sqoop, Oozie, Zookeeper, Kafka, Flume, Spark, Yarn, Storm

Hadoop Distributed Systems: Cloudera, Hortonworks, MapR

Java & J2EE Technologies: Spring, Hibernate, Servlets, JDBC, JSP, Struts, JMS, EJB

IDE Tools: Eclipse, Net Beans

Web Technologies: HTML, CSS, Javascript, Jquery, Bootstrap, XML, JSON

Databases: Oracle, MySQL, DB2, PostgreSQL, SQL Server

Programming Languages: Java, C++, Python, Scala

NoSQL Databases: HBase, Cassandra, MongoDB

Operating Systems: Windows, Unix, Linux

Version Control: SVN, CVS, Tortoise, GITOther Tools: Talend, Maven, Ant, Jenkins

PROFESSIONAL EXPERIENCE

Confidential , Charlotte, NC

Hadoop Developer

Responsibilities:

  • Built an ETL system in Hadoop to process raw data arriving from hospitals using custom Spark jobs and applying complex transformations to store it on HDFS.
  • Built a process based on RabbitMQ, Spark and Spring-boot to send/receive data to Symedical for Reference Data Management.
  • Worked on optimizing ETL jobs based on Spark in Scala to improve performance in order to meet demanding client requirements.
  • Worked on Proof of concepts for emerging big data technologies
  • Wrote shell scripts to automate the data pipeline
  • Interacted with architects and business teams to gather requirements
  • Followed agile best practices.
  • Designed and Maintained Oozie workflows to manage the flow of jobs in the cluster.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig.
  • Provide mentorship and guidance to other architects to help them become independent.
  • Provide review and feedback for existing physical architecture, data architecture and individual code.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Involved in Hadoop cluster task like commissioning & decommissioning Nodes without any effect to running jobs and data.
  • Wrote MapReduce jobs to discover trends in data usage by users.
  • Involved in running Hadoop streaming jobs to process terabytes of text data.
  • Analyzed large data sets by running Hive queries and Pig scripts.
  • Involved in ETL, Data Integration and Migration.
  • Experience in using Pig Scripts to implement ad-hoc Mapreduce programs.
  • Job management using Fair scheduler.
  • Develop Core Framework based on Hadoop to Migrate Existing ETL (RDBMS) Solution.
  • Wrote Pig Scripts to generate Map Reduce jobs and performed ETL procedures on the data in HDFS.
  • A deep and thorough understanding of ETL tools and how they can be applied in a Big Data environment.
  • Worked extensively with Sqoop for importing metadata from Oracle.
  • Involved in creating Hive tables, and loading and analyzing data using hive queries.
  • Responsible for managing data from multiple sources.

Environment: Hadoop, Hive, MapReduce, Hdfs, Pig, Spark, Scala, Oozie, Impala, Distribution Cloudera

Confidential - Phoenix, AZ

Hadoop Developer

Responsibilities:

  • Extensively implemented various Big Data strategies in all stages of SDLC by following Agile.
  • Developed Pig Scripts for validating and cleansing the data.
  • Developed MapReduce programs to pharse the raw data, and stored the refined data in Hbase.
  • Created Hive queries for extracting data from Cornerstone (Data Lake) to HDFS locations.
  • Managed and reviewedHadooplogs.
  • Tested raw data and executed performance scripts.
  • Shared responsibility for administration ofHadoop, Hive and Pig.
  • Exported data from HDFS to RDBMS for visualization and user report generation using Tableau.
  • Involved in the process of load, transform and analyze Transactions data from various providers intoHadoopon an on-going basis.
  • Filtered, transformed and combined data which came from Cornerstone (Data Lake) based on business requirements using custom Pig Scripts and stored in Cognition (downstream DB)
  • Responsible for design and creation of Test cases (in Rally) and tested the Tableau dashboards using Functional testing, system testing, Integration testing, regression testing and UAT.
  • Participated and conducted Issue Log weekly status meetings, Report status meetings and Project status meetings to discuss issues and workarounds.
  • Communicated withdevelopers(On-shore and Off-shore) throughout all the phases of Development to eliminate Roadblocks.
  • Generated daily progress report and represented in daily Agile Scrum meetings.
  • Hands on experience on Apache Pig.
  • Creating Unit Test plans and Test cases.

Environment: ApacheHadoop, PIG, Hive, SQL, Map Reduce, Core Java, Rally, MAPR, UNIX/LINUX, Tableau, Windows, MS Office, Microsoft Outlook.

Confidential, Hartford, CT

Hadoop Developer

Responsibilities:

  • Extracted and updated the data into HDFS using Sqoop import and export command line utility interface.
  • Responsible for developing data pipeline using Flume, Sqoop, and Pig to extract the data from weblogs and store in HDFS.
  • Involved in using HCATLOG to access Hive table metadata from MapReduce and Pig code.
  • Involved in developing Hive UDFs for the needed functionality.
  • Involved in creating Hive tables, loading with data and writing Hive queries.
  • Managed works including indexing data, tuning relevance, developing custom tokenizes and filters, adding functionality includes playlist, custom sorting and regionalization with Solr search engine.
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting.
  • Used pig to do transformations, event joins, filter boot traffic and some pre-aggregations before storing the data onto HDFS.
  • Implemented advanced procedures like text analytics and processing using the in-memory computing capabilities like spark.
  • Enhanced and optimized product Spark code to aggregate, group and run data mining tasks using the Spark framework.
  • Extending Hive and Pig core functionality by writing custom UDFs.
  • Experience in managing and reviewing Hadoop log files
  • Developed data pipeline using Flume,Sqoop, pig and java Mapreduce to ingest customer behavioral data and financial histories into HDFS for analysis.
  • Involved in emitting processed data from Hadoop to relational databases and external file systems using Sqoop.
  • Orchestrated hundreds of Sqoop scripts, pig scripts, Hive queries using Oozie workflows and sub-workflows.
  • Loaded cache data into Hbase using Sqoop.
  • Experience in custom talend jobs to ingest, entich and distribute data in MapR, Cloudera Hadoop ecosystem.
  • Created lots of external tables on Hive pointed to Hbase tables.
  • Analyzed Hbase data in Hive by creating external partitioned and bucketed tables.
  • Worked with cache data stored in Cassandra.
  • Injected the data from External and Internals Flow Organizations.
  • Used the external tables in Impala for data analysis.
  • Supported MapReduce Programs those are running on the cluster.
  • Participated in apache Spark POCS for analyzing the sales data based on several business factors
  • Participated in daily scrum meetings and iterative development.

Environment: Hadoop, MapReduce, Hdfs, Pig, Hive, Hbase, Impala, Sqoop, Flume, Oozie, Apache Spark, Java, Linux, SQL Server, Zookeeper, Autosys, Tableau, Cassandra.

Confidential

Java Developer

Responsibilities:

  • Gathered requirements, designed and implemented the application that utilizes Struts, Spring, JSP and Oracle database.
  • Implemented J2EE design patterns like MVC and Front Controller.
  • Implemented Static and Dynamic web pages using JSP, Java Script, CSS.
  • Involved in Requirement analysis, design and provide the estimation.
  • Responsibilities include designing and delivering web based J2EE solutions. Used JavaScript for Client Side validations.
  • Involved in writing PL/SQL queries and stored procedures.
  • Responsible for setup the environment and Production Environments in Server and Database level.
  • Involved in developing portlets and deploying in Weblogic Portal Server.
  • Involved in writing of release notes to deploy in various environments and production.
  • Monitored the Server load average and prepare status report on it.
  • Point of Contact to the client for all technical aspects.
  • Prepared status reports.

Environment: JSP, Struts, Spring, Tomcat, JavaScript, Oracle10G, Weblogic, XML, NeatBeans

Confidential

Responsibilities:

  • Developed Code Generator GUI tool used to create Java files.
  • Involved in gathering requirements, design and implemented the application using Java/J2EE technologies.
  • Created Static and dynamic web pages using Struts, JSP and Servlets.
  • As a part of bug fixing team played a key role fixing issues.
  • Implemented presentation tier on Servlets, JSP, and HTML for custom business requirements.
  • Played an active role in testing the application for different scenarios and extensively involved in bug fixing.
  • Involved in deploying the application in different environments.
  • Involved in the implementation of design using vital phases of the Software development life cycle (SDLC) that includes Development, Testing, Implementation and Maintenance Support.
  • Used J2EE for the development of business layer services.
  • Gathered business requirements from the end users and wrote functional specifications and design documents.
  • Involved backend-stored procedures, functions and triggers.
  • Handled the issues that came during the Production time and resolved them.
  • JDBC framework has been used to connect the application with the Database.
  • Used Net Beans for the Development, Testing and Debugging of the application.
  • SQL Developer was used as a database client.
  • Involved in gathering requirements, design and implemented the application using Java/J2EE technologies.
  • Created Static and dynamic web pages using JSP and Servlets.
  • As a part of bug fixing team played a key role fixing issues.
  • Implemented presentation tier on Servlets, JSP, and HTML for custom business requirements.
  • Handled the issues that came during the Production time and resolved them.
  • Involved in Design, Development and Support phases of Software Development Life Cycle (SDLC)
  • Extensively worked on Core java.
  • Supported, Testing and coding issues in Production/QA environment.

Environment: JSP, HTML, CoreJava, Servlets, JavaScript, SQL Developer, Eclipse and Apache Tomcat.

We'd love your feedback!