We provide IT Staff Augmentation Services!

Hadoop Developer Resume



  • Six years of professional experience in Hadoop, Big Data and Java technologies such as HDFS, MapReduce, Apache Pig, Impala, Flume, Hive, HBase, Sqoop, Spark, Zookeeper, Oracle, MySQL, JSP, JDBC and Spring.
  • Good working knowledge on creating Hive tables and worked using Hive QL for data analysis to meet the business requirements.
  • Understanding of AWS Ecosystem, services like EC2, S3, EMR, DynamoDB, Redshift, Cloud formation, CloudWatch, CloutFront, etc.
  • Excellent knowledge of Hadoop Architecture and its related components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Experienced in managing and reviewing Hadoop log files.
  • Hands on experience in importing and exporting data from relational databases to HDFS and vice versa using Sqoop.
  • Strong knowledge on implementing SPARK core - SPARK SQL and Spark streaming.
  • Experience in developing both Front End and Back End applications using Java, Servlet’s, JSP, Web Services, Spring, Hibernate, JDBC, JavaScript, HTML, and XML.
  • Exceptional ability to learn new technologies and deliver outputs in short deadlines.


AWS Services: Amazon EMR, S3, DynamoDB, RDS, Amazon Redshift, EC2 CloudFront, CloudWatch

Big Data: Hadoop, MapReduce, HDFS, HBase, Hive, Pig, Oozie, Sqoop Spark, Impala, Kafka, Zookeeper, Flume

Java/J2EE Frameworks: Spring, Hibernate, Struts, Tiles

Languages: JAVA,C, C++, PL/SQL, Shell scripting

Web Technologies: HTML, XML, Java Script, CSS

Web Services: SOAP,REST

Methodologies & Concepts: Agile, Waterfall system analysis and design

Web/Application Servers: Apache Tomcat, JBOSS, WebSphere

Build Tools: ANT, Maven

IDE Tools: Eclipse, NetBeans

RDBMS: Oracle Database, MySQL Database, Cassandra


Confidential, NY

Hadoop Developer


  • Created Hive tables as per requirement as internal or external tables, intended for efficiency.
  • Implemented partitioning, bucketing in Hive for better organization of the data.
  • Reflected the same tables which were created in Hive on Impala.
  • Developed UDF’s in hive for various hive scripts achieving various functionalities.
  • Moved all flat data files generated from various sources to HDFS for further processing.
  • Used MapReduce paradigm to process the data coming from HDFS.
  • Used Amazon Redshift data warehouse for loading data and also examined the distribution and compression of data.
  • Used different file formats like AVRO, Parquet, XML, JSON and flat files.
  • Loaded and processed data on Amazon EMR cluster to benchmark the performance of Spark vs MapReduce.
  • Worked on creating Oozie workflows for scheduling different jobs of hive, map reduce and shell scripts.
  • Worked extensively with Sqoop for importing and exporting the data from HDFS to Relational Database systems.
  • Following agile methodology (SCRUM) during development of the project and oversee the software development in Sprints by attending daily stand-ups.
  • Worked with Version control tools like GIT and SVN.

Environment: Java 7, Eclipse IDE, Hive, HBase, Map Reduce, HDFS, Oozie, Sqoop, Pig, Spark, flume, Impala, MySQL, AWS, GIT, Maven, Linux.

Confidential, Plano, TX

Hadoop Developer


  • Involved in requirement gathering phase of the SDLC and helped team by breaking up the complete project into modules with the help of my team lead.
  • Installed and configured Hadoop and Hadoop stack on a 4 node cluster.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables.
  • Involved in data ingestion into HDFS using Sqoop and Flume from variety of sources.
  • Responsible for managing data from various sources.
  • Worked on Kafka to produce the streamed data into topics and consumed that data.
  • Got good experience with NoSQL database HBase.
  • Worked with NoSQL databases like HBase in creating HBase tables to load large sets of semi structured data coming from various sources.
  • Installed and configured Hive and also wrote Hive UDAFs that helped spot market trends.
  • Involved in loading data from UNIX file system to HDFS.
  • Involved in creating Hive tables, loading the data using it and in writing Hive queries to analyze the data.
  • Used wide range of services around Big Data, NoSQL and Hadoop including Hadoop installation, performance tuning & scalability.
  • Operational services using Cloudera Manager, Oozie, and MapReduce v2.0 YARN development.

Environment: HDFS, Pig, Hive, HBase, Sqoop, Spark, Oozie, Sqoop, flume, AWS, Servlets, JSPs, HTML, JavaScript, CSS, Linux Shell Scripting.

Confidential, Dallas, TX

Jr.Hadoop Developer


  • Providing the technical support to the analysts during requirement gathering.
  • Identified, designed and coded different modules for the application.
  • Configure & Modify Hadoop XML parameters according to hardware/storage requirements.
  • Working with development team on various updates and implement changes to the systems and Map Reduce framework.
  • Created simple hive tables and loaded data from HDFS.
  • Worked on UNIX shell commands for moving data.
  • Implemented partitioning and bucketing concepts to easily perform querying on data.
  • Importing data from relational databases to hive using Sqoop.
  • Worked on basic HDFS file operations.
  • Wrote simple MapReduce programs and executed them by using EclipseIDE.
  • Developed web services using XML, REST and SOAPAPIs for providing business services to other applications.
  • Mentored developers regarding business knowledge and application framework knowledge.
  • Developed JDBC code to connect the java application to interact with data base server.
  • Deployed the application on Apache Tomcat Application Server.
  • Involved in documentation of the complete project.

Environment: JDK, JDBC, XML, servlet, Eclipse IDE, oracle 11g, WebSphere, Linux, Putty, WinSCP, Hive, HBase, Eclipse, Linux.


Jr.Java Developer


  • Performed analysis for the client requirements based on the developed detailed design documents.
  • Developed Use Cases, Class Diagrams, Sequence Diagrams and Data Models.
  • Implemented this module on an existing framework with the help of my lead.
  • Developed the data access classes using JDBC and SQL queries.
  • Worked on developing various HTML files and integrating them CSS.
  • Developed UML diagrams for the development of the project and mentored their implementation to the team members.
  • Developed servlets for the complete project.
  • Designed and documented REST/HTTP APIs, including JSON data formats.
  • Involved in writing Java Server Pages (JSP) for the complete project.
  • Following agile methodology (SCRUM) during development of the project and oversee the software development in Sprints by attending daily stand-ups.
  • Designed and modified User Interfaces using JSP, JavaScript, and CSS.
  • Involved in writing Junit test cases.
  • Extensively worked with Servlets and spring based applications in developing J2EE Components.

Environment: Java, J2EE, JDBC, servlet, JSON, Junit, Eclipse IDE, oracle 11g, Apache Tomcat, Spring, html, CSS, JSP.

Hire Now