We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

Hartford, CT

SUMMARY

  • 8+ years of overall IT experience in a variety of industries, which includes hands on experience of 3 years in Big Data technologies and extensive experience of 5+ years in Java.
  • Good understanding/knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
  • Strong knowledge on creating and monitoring Hadoop clusters on Amazon EC2, VM, Hortonworks Data Platform 2.1 & 2.2, CDH3, CDH4Cloudera Manager on Linux, Ubuntu OS etc.
  • Have a good understanding to ETL,SQL, Data Warehousing Technologies concepts.
  • Analyzed large data sets using Hive queries and Pig Scripts.
  • Hands on experience in writing MapReduce jobs in Java.
  • Hands on experience in installing, configuring and using ecosystem components like Hadoop Map Reduce, Hive, Spark, Sqoop, Pig, HDFS, YARN, ZooKeeper, Oozie, and Flume.
  • Hands on knowledge on RDD transformations, DataFrame transformations in Spark.
  • In - depth understanding of Data Structures and Algorithms.
  • Experienced with NoSQL databases like Hbase and Cassandra.
  • Developed Map Reduce Programs in Perl, Python using Apache Hadoop for analysing big data.
  • Extensive experience in creating Class Diagrams, Activity Diagrams, Sequence Diagrams using Unified Modeling Language(UML)
  • Hands on experience in building and performance tuning of core Java, JDBC, Servlets, JSP, JavaScript, web services, SQL and Stored Procedures.
  • Experience in deploying applications using Web Servers like Apache Tomcat and Application Servers like Web Logic.
  • Built data quality checklist and implemented same using HIVE queries to identify and notify the source systems to reprocess the invalid data.
  • Conducted Quality Assurance(QA) and made sure quality rules are followed.
  • Developed Java applications by using Rational Application Developer and IDE’s such as Eclipse.
  • Experience in using RDBMS concepts and worked with Oracle 10g/11g, SQL server and good experience in writing stored procedures, Functions and Triggers using PL/SQL.
  • Strong development experience in Java, J2EE, Servlets, JSP, Structs, JDBC, Java Beans, JMS, JNDI, XML, XML Schema, Web Services, Junit, ANT, HTML, Java Script.
  • Experience in Big Data platforms like Hortonworks, Cloudera, Amazon AWS and Apache.
  • Good problem-solving skills, quick learner, effective individual and team player, excellent communication and presentation skills.

TECHNICAL SKILLS

Technologies: Hadoop, MR, Pig, Hive, Hbase, Scoop, Flume, J2EE, Java 6, Servlets, JSP, XML, UML, Spring, Hibernate, REST API, Webservices, Quality Assurance(QA)

BigData Technologies: Hadoop MapReduce, Pig, Hive, Spark, Kafka, YARN and Hbase.

Big Data platforms: Hortonworks, Cloudera, Amazon AWS and Apache.

Databases: Oracle 10g/11g, DB2, MySQL, MS Access, Cassandra

ETL: Talend

Mark-up Lang: HTML, CSS, DHTML, and XML.

App Servers: Web Logic.

Web Servers: Apache.

Operating Systems: Windows, Unix.

Scripting Lang: Shell, Java Script, Python

PROFESSIONAL EXPERIENCE

Confidential, Hartford, CT

Hadoop Developer

Responsibilities:

  • Good understanding and related experience with Hadoop stack-internals, Hive, Pig and Map/Reduce.
  • Deep understanding of schedulers, workload management, availability, scalability and distributed data platforms.
  • Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Involved in loading data from UNIX file system to HDFS.
  • Wrote MapReduce jobs to discover trends in data usage by users.
  • Used Map Reduce JUnit for unit testing.
  • Involved in managing and reviewing Hadoop log files.
  • Involved in running Hadoop streaming jobs to process terabytes of text data.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Wrote pig UDF’s.
  • Develop HIVE queries for the analysts.
  • Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
  • Exported the result set from HIVE to MySQL using Shell scripts.
  • Worked with SPARK for quick analytics on object relationships.
  • Used Zookeeper for various types of centralized configurations.
  • Involved in maintaining various Unix Shell scripts.
  • Implemented data quality and cleansing process using Map reduce.
  • Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
  • Automated all the jobs starting from pulling the Data from different Data Sources like MySQL to pushing the result set Data to Hadoop Distributed File System using Sqoop.
  • Used SVN for version control.
  • Worked on defect resolution in production environment.
  • Worked on production issues and logged them in Quality Center.
  • Helped the team to increase Cluster from 25 Nodes to 40 Nodes.
  • Used GraphX API of Apache Spark for developing graphs to analyse data.
  • Maintain System integrity of all sub-components (primarily HDFS, MR, HBase, and Flume).
  • Monitor System health and logs and respond accordingly to any warning or failure conditions.

Environment: Hadoop, HDFS, Map Reduce, Hive, Pig, Spark, Yarn, Sqoop, Java 1.6, Apache Spark, GraphX, UNIX Shell Scripting.

Confidential, Salt Lake City, UT

Hadoop Developer

Responsibilities:

  • Involved in architecture development of Hadoop Clusters with CDH4 on Linux, using Cloudera Manager
  • Created design specifications to get rid of system limitations and create platform convergence by migrating data from various sources to a Hadoop cluster.
  • Created Hive External tables and designed the loading of the data(overwrite/append) into tables and query data using HQL.
  • Created a road map of Systems/Processes industrialization initiatives to decrease the operational costs of Confidential by 35% over a period of 5 years that included ideas of platform convergence (Ex: SAP migration, Mainframe Rehosting, Portals creation) and automation.
  • Involved in writing set of multiple Map Reduce jobs to parse the files in Python and Perl.
  • Data model documentation at source systems to identify data sources for migration.
  • Used Kafka as a messaging system to get data from different sources.
  • Migrate data to/from sources to/from HDFS using SQOOP.
  • Design of data transformation layer using PIG.
  • Design of the DW (upon data transformation using PIG) on HDFS using HIVE.
  • Support and maintain a repository of map - reduce programs used for creating metrics for regulatory reporting.
  • Worked on Sequence files, Map side joins, bucketing, partitioning for hive performance enhancement and storage improvement.
  • Built strong data quality rules and coded the same for quality assurance (QA) process.
  • Design Oozie workflows containing map-reduce and PIG scripts.
  • Support and maintain Oozie workflows (On call Support).
  • Performed Graph processing on the data collected and anlyse using Apache Giraph Graph
  • Liasoning with the BI team to enable data migration of metrics to Sybase server for BO reporting.
  • Create support documentation and co-ordinate to form a COE (Centre of Excellence) for the Hadoop framework.

Environment: CDH4 Cloudera, Hadoop, PIG, HIVE, SQOOP, Kafka, Map Reduce, Mainframe, DB2, Sybase, SAP BO, COBOL, Apache Giraph Graph, Python, QA, Syncort DMX-h & Oozie

Confidential, CA

Hadoop Developer

Responsibilities:

  • Translation of functional and technical requirements into detailed architecture and design.
  • Involved in configuring multi-nodes fully distributed Hadoop cluster.
  • Supported Map Reduce Programs those are running on the cluster.
  • Responsible to manage data coming from different sources.
  • Importing and exporting data into HDFS using Flume.
  • Experienced in analyzing data with Hive and Pig.
  • Worked with Talend to perform ETL operations.
  • Experienced in defining job flows.
  • Responsible for operational support of Production system.
  • Loading log data directly into HDFS using Flume.
  • Experienced in managing and reviewing Hadoop log files.
  • Worked on performance tuning.
  • Involved in creating Best Practice checklist, data quality and governance check lists and coding standards.
  • Analyzing data with Hive, Pig and Hadoop Streaming.

Environment: Hadoop, Linux, HDFS, Hive, Sqoop, Flume, Kafka, Zookeeper, Talend and HBase

Confidential, Stamford, CT

Java/J2ee Developer

Responsibilities:

  • Performed requirement analysis, designed, developed the system using iterative software engineering processes and extreme programming practices.
  • Designed the application using UML use case diagrams, class diagrams, and sequence diagrams.
  • Designed and developed the front-end using JSP’s for the entire application using Struts framework as centralized controller of the application with the help of JSTL and Struts tag libraries.
  • Used different spring components like Spring AOP, Spring JDBC, Spring MVC, Spring Tiles etc to manage business layer.
  • Developed the application using JSF for front end, spring to manage business layer components and Hibernate for performing database transactions.
  • With Data Integration projected user views into single web application
  • Developed Web-Services module for integration using SOAP, WSDL, JAX-RPC, SAAJ and JAXP.
  • Programmed in PL/SQL to debug queries and stored procedures for business logic and batch jobs.
  • Also created Oracle Stored Procedures for processing the carrier’s response and saving the response to database
  • Used CVS for Version Control.
  • Modified Struts configuration files as per application requirements.
  • Developed multiple formatting, validation utilities in Java, JavaScript functions and CSS Style Sheets so that they can be reused across the application.
  • Worked on multithread applications
  • Developed UI for Customer Service Modules and Reports using JSF, JSP’s and My Faces Components.
  • Worked on DB2 thoroughly to do database transactions, which helps in attaining DML and DDL operations
  • Performed multiple code reviews, unit & system testing of application to ensure that the QA and application is working as per the business requirement.
  • Wrote build files to configure build process using ANT.
  • Developed and deployed the application on Web Sphere Application Server.

Environment: Java, JDBC, Servlets, JSF, JSTL, Eclipse, Spring, EJB, JNDI, XML, JavaScript, CSS, JAXP, SOAP, WSDL, WSAD, Clear Quest, Clear Case, Junit, CVS, Multithreading Web Logic, Db2 Windows XP, Unix

Confidential, El Segundo, CA

Java/J2ee Developer

Responsibilities:

  • Extensively written JAVA code.
  • Developed SOAP Web services using WSDL on SOA Environment.
  • Extensively worked on J2EE components and Services.
  • Extensively written CORE JAVA code in application
  • Developed Java, JSP, Java Script, HTML, and CSS for UI & business logic.
  • Implemented Validator framework to implement Client side and server side for Validations.
  • Written JDBC statements, prepared statements, and callable statements in Java, JSPs and Servlets.
  • Followed Scrum approach for the development process
  • Extensively worked on Struts with Spring Integration for communicating with business components and also worked on Spring with Hibernate Integration for ORM mappings.
  • Modified and added database functions, procedures and triggers pertaining to business logic of the application.
  • Used TOAD to check and verify all the database turnaround times and also tested the connections for response times and query round trip behaviour.
  • Used ANT Builder to build the code for production line.
  • Used Eclipse IDE for all recoding in Java, Servlets and JSPs.
  • Used IBM Clear Case for Versioning and Maintenance.
  • Involved in discussions with the business analysts for bug validation and fixing.
  • Modified technical design document, functional design document to accommodate change requests.
  • Wrote JUnit test cases for system testing, Used Log4j for logging.
  • Used JIRA as a bug-reporting tool for updating the bug report.
  • Involved in performance tuning where there was a latency or delay in execution of code

Environment: Java, J2EE, Servlets, JSP, Struts, Spring, Hibernate, JDBC, JNDI, JMS, JavaScript, XSLT, DTD, SAX, DOM, XML, UML, TOAD, Jasper Reports, Oracle10g, SQL, Eclipse RCP, IBM Clear Case, WebSphere, Unix/Windows.

Confidential, Columbus, OH

Java/J2ee Developer

Responsibilities:

  • Understanding and analyzing the project requirements.
  • Analysis and Design with UML and Rational Rose.
  • Created Class Diagrams, Sequence diagrams and Collaboration Diagrams
  • Used the MVC architecture.
  • Worked on Jakarta Struts open framework.
  • Developed Servlets in order to deal with requests for account activity,
  • Developed Controller Servlets and Action Servlets to handle the requests and responses.
  • Developed Servlets and created JSP pages for viewing on a HTML page.
  • Developed the front end using JSP.
  • Developed various EJB's to handle business logic.
  • Designed and developed numerous Session Beans deployed on Web logic Application Server.
  • Implemented Database interactions using JDBC with back-end Oracle.
  • Worked on Database designing, Stored Procedures, and PL/SQL.
  • Created triggers and stored procedures using PL/SQL.
  • Written queries to get the data from the Oracle database using SQL.

Environment: J2EE, Servlets, JSP, Struts, Spring, Hibernate, Oracle, TOAD, Weblogic Server

Confidential

Associate JAVA Developer

Responsibilities:

  • Involved in the complete SDLC software development life cycle of the application from requirement analysis to testing.
  • Developed the modules based on struts MVC Architecture.
  • Developed The UI using JavaScript, JSP, HTML, and CSS for interactive cross browser functionality and complex user interface.
  • Created Business Logic using Servlets, Session beans and deployed them on Weblogic server.
  • Used MVC struts framework for application design.
  • Created complex SQL Queries, PL/SQL Stored procedures, Functions for back end.
  • Prepared the Functional, Design and Test case specifications.
  • Involved in writing Stored Procedures in Oracle to do some database side validations.
  • Performed unit testing, system testing and integration testing so as to ensure quality.
  • Developed Unit Test Cases. Used JUNIT for unit testing of the application.
  • Provided Technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects. Resolved more priority defects as per the schedule.

Environment: Java, JSP, Servlets, Weblogic, Oracle, JUnit, SQL, XML, Toad, Testing, QA

We'd love your feedback!