We provide IT Staff Augmentation Services!

Sr. Hadoop Developer Resume



  • Over 7+ years of experience in the field of information technology by managing and leading software development, data management and providing high quality solutions.
  • Having 4 years of experience as a Hadoop Developer and 3 years of experience in the field of Mainframe, JAVA and web Applications Design, Development, Support and improving quality of final deliverables to meet Business Goals.
  • Experience in installation, configuration and management of Hadoop Clusters.
  • In depth knowledge of Job Tracker, Task Tracker, Name Node, Data Nodes and MapReduce concepts
  • Hands on experience in installing, configuring, and usingHadoopecosystem components likeHadoopMapReduce(MR), HDFS, HBase, Oozie, Hive, Sqoop, Pig, Flume, HBase, Hue, Zookeeper.
  • Experience in .
  • Configured Pseudo - distributed and Fully distributed Hadoop Clusters.
  • Experience in importing and exporting the data using Sqoop from HDFS to Relational Database systems/ mainframe and vice-versa.
  • Having experience of managing columnar databases like ParAccel, Amazon Redshift.
  • Experience with AquaData for visual administration of data.
  • Experienced in developing MapReduce programs using Apache Hadoop for working with Big Data.
  • Experience inHadoopShell commands. Expert in using Sqoop to import/export data into HDFS from RDBMS.
  • Having knowledge about Hcatalog, Zookeeper, Cassandra, MongoDB and Neo4j.
  • Having basic knowledge about real-time processing tools Kafka, Spark 1.6.
  • Integrating SPARK with HADOOP ecosystem and different data sources(HDFS,Hive,HBase,Cassandra)
  • Making RDD transformations using Python.
  • Worked on different file formats like JSON, AVRO, and ORC.
  • Experience in installation, configuration, supporting and managing - Cloud Era’sHadoopplatform along with CDH 4&5 versions.
  • Experienced in managing and reviewingHadoop log files and Loading log data directly into HDFS using Flume.
  • Setup alerts with Cloudera Manager about memory and disk usage on the cluster.
  • Having knowledge on VMware installation and usage. Worked on Tableau data visualization tools.
  • Managing and scheduling batch Jobs on aHadoopCluster using Oozie.
  • Strong experience in Java/ J2EE/ Web Technologies like HTML, Java Script, XML, XSD, CSS, J2EE 1.3/1.4, JDBC, Servlets, JSP, Java Beans, EJB, JNDI, JAXP, JAXB, SOAP, WSDL and Struts1.3.8, iBatis 3.0
  • Used Zookeeper to provide coordination services to the cluster.
  • Written Hive queries for data analysis and to process the data for visualization.
  • Strong understanding of the software life cycle methodologies Agile Scrum, Waterfall and various Maintenance, Development, Testing, Production Support and improving quality of final deliverables to meet Business Goals.
  • Worked on SQL, MySQL and Oracle 10g.
  • Designing and deploying Service Oriented Architecture (SOA) thru web services.
  • Experience working individually and in team environment.
  • Excellent ability to communicate with people who have varying levels of understanding of Application development, production support includes writing/executing test cases.
  • Excellent interpersonal and communication skills, creative, research-minded, technically competent and result-oriented with problem solving and leadership skills.


Big Data Ecosystems: Hadoop, HDFS, YARN, MapReduce, Hive, Pig, HBase, Zookeeper, Sqoop, Oozie, Flume, Kafka, Apache Spark

Frameworks: JPA,J2EE, JSP, Servlets, Struts, Hibernate, .NET Framework 4.5

Methodology: Agile software development

Languages: Java, Hive QL, Pig Latin, R, Python, Advanced PL/SQL, SQL, VBA, C++, C, Shell

Scripting Languages: HTML, CSS, JavaScript, DHTML, XML, JQuery

Web Technologies: Java, J2EE, Servlets, JSP, JDBC, XML, AJAX, SOAP, Restful

Architectures: SOA, Cloud Computing(AWS, EC2)

Application Server: Apache Tomcat, Glassfish 4.0, Web Logic

Database Systems: Oracle 11g/10g/9i, DB2, MS-SQL Server, MySQL, MS-Access

Development Tools(IDEs/): JIRA, Clear case, Tableau, Splunk, RStudio, Eclipse/Net Beans, Toad, SQL Developer, AWK

Platforms: UNIX, Windows XP / 7, Ubuntu(Linux), CentOS


Confidential, NY

Sr. Hadoop Developer


  • Gathering business requirements from the Business Partners and Subject Matter Experts.
  • Leverage data mining skills, including data auditing, aggregation, validation and reconciliation.
  • Testing, extracting, analyzing and explaining data results in clear and concise reports.
  • Working with SQL database(s), Excel and a plus with data oriented programming languages in a UNIX environment.
  • Involved in implementing nine node CDH4Hadoopcluster on Redhat LINUX.
  • Importing and exporting data into HDFS from RDBMS and vice versa using Sqoop.
  • Hands on experience in storing large volumes of data on Cassandra for High availability of analytical data.
  • Developed custom MapReduce jobs in java for preprocessing and data cleaning.
  • Creating Hive tables and working on them using Hive QL.
  • Written Hive queries for data analysis to meet the business requirements.
  • Hands on experience in Python for streaming MapReduce programs.
  • Responsible for developing PIG Latin scripts.
  • Managing and scheduling batch Jobs on a Hadoop Cluster using Oozie.
  • Experienced in loading and transforming large sets of structured, semi-structured and unstructured data.
  • Migration of ETL processes from Oracle to Hive to test the easy data manipulation.
  • Written Hive queries to process the data for visualization.
  • Assisted in monitoring Hadoop cluster using Ganglia.
  • Involved in daily SCRUM meetings to discuss the development/progress of Sprints and was active in making scrum meetings more productive.
  • Worked on installing and configuring EC2 instances on Amazon Web Services (AWS) for establishing clusters on cloud.

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Cassandra, Java, Hadoop distribution of Cloudera, Pig, AWS, Linux, Python, XML, Eclipse, Oracle 10g, PL/SQL.

Confidential, Middletown, NJ

Sr. Application Developer (Big/Data)


  • Involved in review of functional and non-functional requirements.
  • Installed and configured Hadoop MapReduce and HDFS.
  • Acquired good understanding and experience of NoSQL databases such as HBase and Cassandra.
  • Installed and configured Hive and also implemented various business requirements by writing Hive UDFs.
  • Extensively worked on user interface for few modules using HTML, JSP’s, JavaScript, Python and Ajax.
  • Generated Business Logic using Servlets, Session beans and deployed them on Web logic server.
  • Created complex SQL queries and stored procedures.
  • Developed the XML schema and Web services for the data support and structures.
  • Implemented the Web service client for login verification, credit reports and applicant information using Apache Axis 2 web service.
  • Managing streaming data using Kafka and running analytics over that data.
  • Used Hibernate ORM framework with spring framework for data persistence and transaction management.
  • Used struts validation framework for form level validations.
  • Wrote test cases in JUnit for unit testing of classes.
  • Provided technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects.
  • Built and deployed Java application into multiple UNIX based environments and produced both unit and functional test results along with release notes.

Environment: Hadoop, HBase, Hive, Java, Eclipse, J2EE 1.4, Struts 1.3, JSP, Servlets 2.5, WebSphere 6.1, HTML, XML, ANT 1.6, Python, JavaScript, Junit 3.8.

Confidential, New York, NY

Sr. Hadoop Developer


  • Responsible for analyzing and understanding of data Sources like iTunes, Spotify, YouTube and Facebook data.
  • Developed a multithreaded framework to grab data for playback, Traffic source, Social, device, and Demographic reports from YouTube.
  • Developed reusable component in java to load data from Hadoop distributed file system to ParaAccel.
  • Developed Map Reduce jobs to process the Music metric data. Scripts for uploading the data in ParAccel server.
  • Developed Map Reduce codes for data manipulation.
  • Implemented POC using Spark and Spark SQL.
  • Working as an Architect for providing solutions
  • Have been involved in designing & creating hive tables to upload data in Hadoop.
  • Experienced in migrating all historical data from ParAccel to AWS S3 file System with help of SQOOP for feeds like iTunes Preorders, Radio Monitor and etc.
  • Responsible for all the data flow and quality of data.
  • Responsible for end to end development for the client.
  • Involved in Designing, development, coding, Unit testing.

Environment: Hadoop, MapReduce, HDFS, Hive, Java (jdk1.6), Hadoop distribution of Horton works, Cloudera, MapR, DataStax, Spring 2.5, Hibernate 3.0, JSF, Servlets, JDBC, JSP,JSTL, JPA, JavaScript, Eclipse 3.4, log4j,Oracle 10g, CVS, CSS, Xml, XSLT, SMTP, Windows-XP.

Confidential, Monroe, MI

Hadoop/Java Developer


  • Involved in designing and developingHadoop MapReduce jobs Using JAVA Runtime Environment for the batchprocessingto search and match the scores.
  • Used Rational Rose for developing Use case diagrams, Activity flow diagrams, Class diagrams and Object diagrams in the design phase.
  • Used Struts with Tiles in the MVC framework for the application.
  • Extensively worked on Servlets, JSPs, Struts 1.3 and Tiles, JavaScript, Expression language, JSTL, JSP custom tags.
  • Involved in using Solr Cloud implementation to provide real time search capabilities on the repository with terabytes of data.
  • Involved in developing Hadoop MapReduce jobs for merging and appending therepository data.
  • Hands on experience insetting up H-base Column based storage repositoryforarchivingand retro data.
  • Developed XML schema and DOM parser for all the XML document used for Data Transfers and also developed XSLT code for them.
  • Configured Struts-config.xml, ejb-jar.xml and web.xml on the application.
  • Used Apache CXF web service stack for developing web services and SOAP UI and XML-SPY for testing web services.
  • Used Hibernate 3.0 in data access layer to access and update information in the database.
  • Used Java Message Service (JMS 1.1) for reliable and asynchronous exchange of important
  • Involved in agile SCRUM methodology implementation. Involved in various performance projects to increase the response time of the application.
  • Involved in integration of Legacy Scoring and Analytical Models like SMG3 into thenew application using Web Services.
  • Involved in development of batch processing application using Multi-threaded executorpools for faster processing.
  • Responsible for writing Pig UDFs and Hive UDFs.
  • Handled importing of data from various data sources, performance transformation using Hive.
  • Experience in optimization of Map reduce algorithm using combiners and partitions to deliver the best results and worked on Application performance optimization for a HDFS cluster.
  • Created various calculated fields and created various visualizations and dashboards using tableau desktop.
  • Published the dashboards created on Tableau desktop onto Tableau server.
  • Experience working with off-shore teams and communicating daily status on issues, road-blocks.

Environment: Java, J2EE, Tableau Desktop, Tableau Server, Hadoop, Hbase, Kettle, Zookeeper, Solr cloud, Pig Latin, Oozie scheduler, JavaBeans, Agile SCRUM, IBM Data Power, JProfiler, Spring, Struts1.3, Hibernate3.0, Jboss Application Server, Eclipse, Rational Clear case, CXF 2.2.4, JNDI, Java Script, Servlet 2.3, JUnit, Maven, SVN, Jboss, XML Web services, HTML DB2, JDBC, ANT, UML, Unix, Windows NT/2000.

Confidential, Webster, MA

JAVA Developer


  • Involved in the analysis, design, and development and testing phases of Software Development Life Cycle (SDLC).
  • Used Rational Rose for developing Use case diagrams, Activity flow diagrams, Class diagrams and Object diagrams in the design phase.
  • Analysis, design and development of Application based onJ2EEusingStruts and Tiles,Spring 2.0 andHibernate 3.0.
  • Involved in interacting with the Business Analyst and Architect during the Sprint Planning Sessions.
  • Used XML Web Services for transferring data between different applications.
  • Used Apache CXF web service stack for developing web services and SOAP UI and XML-SPY for testing web services.
  • Used JaxB for binding XML to Java. Used SAX and DOM parsers to parse xml data. Used Xpath to parse XML documents
  • Hibernate was used forObject Relational mappingwith Oracle database.
  • Worked with Spring IOCfor injecting the beans and reduced the coupling between the classes.
  • Involved in developing the user interface usingStruts.
  • Implemented Spring IOC (Inversion of Control)/DI (Dependency Injection)for wiring the object dependencies across the application.
  • Implemented spring transactionmanagement for implementing transaction's for the application.
  • Implemented design patterns for Service Locator.
  • Performed unit testing usingJunit 3, EasyMock Testing Framework for performing Unit testing.Worked onPL/SQLstored procedures usingPL/SQL Developer.
  • Involved in Fixing theproductionDefects for the application.UsedEclipseas IDE for application development.
  • Used ANT as build-tool for building J2EE applications.
  • UsedTomcat5.5 for application deployment.
  • Participated in SCRUM software development process as part of agile software development methodology.

Environment: Java 1.6, Struts, PL/SQL, Spring IOC, Spring Transaction Management, Hibernate 3.0, Springs2.0 JSP 2.0, Oracle 11g, Eclipse, JUnit 3, PL/SQL Developer, Application Server, JDBC, Maven, CVS, Harvest, UML Struts 1.2.3, XML Web Services.

Confidential, Stamford, CT

Software Engineer


  • Involved in the Design, Coding, Testing and Implementation of the web application.
  • Developed JSP Java Server Pages starting from HTMLs and detailed technical design specification documents. Pages included HTML, CSS, JavaScript, Hibernate and JSTL.
  • Developed SOAP based requests for communicating with Web Services.
  • Used agile systems and strategies to provide quick and feasible solutions, based on agile system, to the organization.
  • Implemented HTTP Modules for different applications in Struts Framework that uses Servlets, JSP, ActionForm, ActionClass and ActionMapping.
  • Developing web applications using MVC Framework, spring, Struts, Hibernate.
  • Involved in the creation of custom interceptors for Validation purposes.
  • Analyzed and fixed defects in the Login application.
  • Involved in dynamic creation of error elements on demand when there is an error.
  • Involved in Ajax - based Rich Browser User Interfaces.
  • Ensured design consistency with client’s development standards and guidelines.
  • Improved user experience by designing and creating new web components and features.

Environment: Java, J2EE, Struts, SOAP web services, SOA, Spring, Hibernate, JavaScript, jQuery, Oracle, AJAX, JSP, Servlets, Eclipse, CVS Source control, Linux.

Hire Now