Hadoop Developer Resume
Hartford, CT
SUMMARY
- 8+ years of overall IT experience in a variety of industries, which includes hands on experience of 3 years in Big Data technologies and extensive experience of 5+ years in Java.
- Good understanding/knowledge of Hadoop architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and Map Reduce programming paradigm.
- Strong knowledge on creating and monitoring Hadoop clusters on Amazon EC2, VM, Hortonworks Data Platform 2.1 & 2.2, CDH3, CDH4Cloudera Manager on Linux, Ubuntu OS etc.
- Have a good understanding to ETL,SQL, Data Warehousing Technologies concepts.
- Analyzed large data sets using Hive queries and Pig Scripts.
- Hands on experience in writing MapReduce jobs in Java.
- Hands on experience in installing, configuring and using ecosystem components like Hadoop Map Reduce, Hive, Spark, Sqoop, Pig, HDFS, YARN, ZooKeeper, Oozie, and Flume.
- Hands on knowledge on RDD transformations, DataFrame transformations in Spark.
- In - depth understanding of Data Structures and Algorithms.
- Experienced with NoSQL databases like Hbase and Cassandra.
- Developed Map Reduce Programs in Perl, Python using Apache Hadoop for analysing big data.
- Extensive experience in creating Class Diagrams, Activity Diagrams, Sequence Diagrams using Unified Modeling Language(UML)
- Hands on experience in building and performance tuning of core Java, JDBC, Servlets, JSP, JavaScript, web services, SQL and Stored Procedures.
- Experience in deploying applications using Web Servers like Apache Tomcat and Application Servers like Web Logic.
- Built data quality checklist and implemented same using HIVE queries to identify and notify the source systems to reprocess the invalid data.
- Conducted Quality Assurance(QA) and made sure quality rules are followed.
- Developed Java applications by using Rational Application Developer and IDE’s such as Eclipse.
- Experience in using RDBMS concepts and worked with Oracle 10g/11g, SQL server and good experience in writing stored procedures, Functions and Triggers using PL/SQL.
- Strong development experience in Java, J2EE, Servlets, JSP, Structs, JDBC, Java Beans, JMS, JNDI, XML, XML Schema, Web Services, Junit, ANT, HTML, Java Script.
- Experience in Big Data platforms like Hortonworks, Cloudera, Amazon AWS and Apache.
- Good problem-solving skills, quick learner, effective individual and team player, excellent communication and presentation skills.
TECHNICAL SKILLS
Technologies: Hadoop, MR, Pig, Hive, Hbase, Scoop, Flume, J2EE, Java 6, Servlets, JSP, XML, UML, Spring, Hibernate, REST API, Webservices, Quality Assurance(QA)
BigData Technologies: Hadoop MapReduce, Pig, Hive, Spark, Kafka, YARN and Hbase.
Big Data platforms: Hortonworks, Cloudera, Amazon AWS and Apache.
Databases: Oracle 10g/11g, DB2, MySQL, MS Access, Cassandra
ETL: Talend
Mark-up Lang: HTML, CSS, DHTML, and XML.
App Servers: Web Logic.
Web Servers: Apache.
Operating Systems: Windows, Unix.
Scripting Lang: Shell, Java Script, Python
PROFESSIONAL EXPERIENCE
Confidential, Hartford, CT
Hadoop Developer
Responsibilities:
- Good understanding and related experience with Hadoop stack-internals, Hive, Pig and Map/Reduce.
- Deep understanding of schedulers, workload management, availability, scalability and distributed data platforms.
- Installed and configured Hadoop MapReduce, HDFS, developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
- Involved in loading data from UNIX file system to HDFS.
- Wrote MapReduce jobs to discover trends in data usage by users.
- Used Map Reduce JUnit for unit testing.
- Involved in managing and reviewing Hadoop log files.
- Involved in running Hadoop streaming jobs to process terabytes of text data.
- Load and transform large sets of structured, semi structured and unstructured data.
- Wrote pig UDF’s.
- Develop HIVE queries for the analysts.
- Implemented Partitioning, Dynamic Partitions, Buckets in HIVE.
- Exported the result set from HIVE to MySQL using Shell scripts.
- Worked with SPARK for quick analytics on object relationships.
- Used Zookeeper for various types of centralized configurations.
- Involved in maintaining various Unix Shell scripts.
- Implemented data quality and cleansing process using Map reduce.
- Implemented Fair schedulers on the Job tracker to share the resources of the Cluster for the Map Reduce jobs given by the users.
- Automated all the jobs starting from pulling the Data from different Data Sources like MySQL to pushing the result set Data to Hadoop Distributed File System using Sqoop.
- Used SVN for version control.
- Worked on defect resolution in production environment.
- Worked on production issues and logged them in Quality Center.
- Helped the team to increase Cluster from 25 Nodes to 40 Nodes.
- Used GraphX API of Apache Spark for developing graphs to analyse data.
- Maintain System integrity of all sub-components (primarily HDFS, MR, HBase, and Flume).
- Monitor System health and logs and respond accordingly to any warning or failure conditions.
Environment: Hadoop, HDFS, Map Reduce, Hive, Pig, Spark, Yarn, Sqoop, Java 1.6, Apache Spark, GraphX, UNIX Shell Scripting.
Confidential, Salt Lake City, UT
Hadoop Developer
Responsibilities:
- Involved in architecture development of Hadoop Clusters with CDH4 on Linux, using Cloudera Manager
- Created design specifications to get rid of system limitations and create platform convergence by migrating data from various sources to a Hadoop cluster.
- Created Hive External tables and designed the loading of the data(overwrite/append) into tables and query data using HQL.
- Created a road map of Systems/Processes industrialization initiatives to decrease the operational costs of Confidential by 35% over a period of 5 years that included ideas of platform convergence (Ex: SAP migration, Mainframe Rehosting, Portals creation) and automation.
- Involved in writing set of multiple Map Reduce jobs to parse the files in Python and Perl.
- Data model documentation at source systems to identify data sources for migration.
- Used Kafka as a messaging system to get data from different sources.
- Migrate data to/from sources to/from HDFS using SQOOP.
- Design of data transformation layer using PIG.
- Design of the DW (upon data transformation using PIG) on HDFS using HIVE.
- Support and maintain a repository of map - reduce programs used for creating metrics for regulatory reporting.
- Worked on Sequence files, Map side joins, bucketing, partitioning for hive performance enhancement and storage improvement.
- Built strong data quality rules and coded the same for quality assurance (QA) process.
- Design Oozie workflows containing map-reduce and PIG scripts.
- Support and maintain Oozie workflows (On call Support).
- Performed Graph processing on the data collected and anlyse using Apache Giraph Graph
- Liasoning with the BI team to enable data migration of metrics to Sybase server for BO reporting.
- Create support documentation and co-ordinate to form a COE (Centre of Excellence) for the Hadoop framework.
Environment: CDH4 Cloudera, Hadoop, PIG, HIVE, SQOOP, Kafka, Map Reduce, Mainframe, DB2, Sybase, SAP BO, COBOL, Apache Giraph Graph, Python, QA, Syncort DMX-h & Oozie
Confidential, CA
Hadoop Developer
Responsibilities:
- Translation of functional and technical requirements into detailed architecture and design.
- Involved in configuring multi-nodes fully distributed Hadoop cluster.
- Supported Map Reduce Programs those are running on the cluster.
- Responsible to manage data coming from different sources.
- Importing and exporting data into HDFS using Flume.
- Experienced in analyzing data with Hive and Pig.
- Worked with Talend to perform ETL operations.
- Experienced in defining job flows.
- Responsible for operational support of Production system.
- Loading log data directly into HDFS using Flume.
- Experienced in managing and reviewing Hadoop log files.
- Worked on performance tuning.
- Involved in creating Best Practice checklist, data quality and governance check lists and coding standards.
- Analyzing data with Hive, Pig and Hadoop Streaming.
Environment: Hadoop, Linux, HDFS, Hive, Sqoop, Flume, Kafka, Zookeeper, Talend and HBase
Confidential, Stamford, CT
Java/J2ee Developer
Responsibilities:
- Performed requirement analysis, designed, developed the system using iterative software engineering processes and extreme programming practices.
- Designed the application using UML use case diagrams, class diagrams, and sequence diagrams.
- Designed and developed the front-end using JSP’s for the entire application using Struts framework as centralized controller of the application with the help of JSTL and Struts tag libraries.
- Used different spring components like Spring AOP, Spring JDBC, Spring MVC, Spring Tiles etc to manage business layer.
- Developed the application using JSF for front end, spring to manage business layer components and Hibernate for performing database transactions.
- With Data Integration projected user views into single web application
- Developed Web-Services module for integration using SOAP, WSDL, JAX-RPC, SAAJ and JAXP.
- Programmed in PL/SQL to debug queries and stored procedures for business logic and batch jobs.
- Also created Oracle Stored Procedures for processing the carrier’s response and saving the response to database
- Used CVS for Version Control.
- Modified Struts configuration files as per application requirements.
- Developed multiple formatting, validation utilities in Java, JavaScript functions and CSS Style Sheets so that they can be reused across the application.
- Worked on multithread applications
- Developed UI for Customer Service Modules and Reports using JSF, JSP’s and My Faces Components.
- Worked on DB2 thoroughly to do database transactions, which helps in attaining DML and DDL operations
- Performed multiple code reviews, unit & system testing of application to ensure that the QA and application is working as per the business requirement.
- Wrote build files to configure build process using ANT.
- Developed and deployed the application on Web Sphere Application Server.
Environment: Java, JDBC, Servlets, JSF, JSTL, Eclipse, Spring, EJB, JNDI, XML, JavaScript, CSS, JAXP, SOAP, WSDL, WSAD, Clear Quest, Clear Case, Junit, CVS, Multithreading Web Logic, Db2 Windows XP, Unix
Confidential, El Segundo, CA
Java/J2ee Developer
Responsibilities:
- Extensively written JAVA code.
- Developed SOAP Web services using WSDL on SOA Environment.
- Extensively worked on J2EE components and Services.
- Extensively written CORE JAVA code in application
- Developed Java, JSP, Java Script, HTML, and CSS for UI & business logic.
- Implemented Validator framework to implement Client side and server side for Validations.
- Written JDBC statements, prepared statements, and callable statements in Java, JSPs and Servlets.
- Followed Scrum approach for the development process
- Extensively worked on Struts with Spring Integration for communicating with business components and also worked on Spring with Hibernate Integration for ORM mappings.
- Modified and added database functions, procedures and triggers pertaining to business logic of the application.
- Used TOAD to check and verify all the database turnaround times and also tested the connections for response times and query round trip behaviour.
- Used ANT Builder to build the code for production line.
- Used Eclipse IDE for all recoding in Java, Servlets and JSPs.
- Used IBM Clear Case for Versioning and Maintenance.
- Involved in discussions with the business analysts for bug validation and fixing.
- Modified technical design document, functional design document to accommodate change requests.
- Wrote JUnit test cases for system testing, Used Log4j for logging.
- Used JIRA as a bug-reporting tool for updating the bug report.
- Involved in performance tuning where there was a latency or delay in execution of code
Environment: Java, J2EE, Servlets, JSP, Struts, Spring, Hibernate, JDBC, JNDI, JMS, JavaScript, XSLT, DTD, SAX, DOM, XML, UML, TOAD, Jasper Reports, Oracle10g, SQL, Eclipse RCP, IBM Clear Case, WebSphere, Unix/Windows.
Confidential, Columbus, OH
Java/J2ee Developer
Responsibilities:
- Understanding and analyzing the project requirements.
- Analysis and Design with UML and Rational Rose.
- Created Class Diagrams, Sequence diagrams and Collaboration Diagrams
- Used the MVC architecture.
- Worked on Jakarta Struts open framework.
- Developed Servlets in order to deal with requests for account activity,
- Developed Controller Servlets and Action Servlets to handle the requests and responses.
- Developed Servlets and created JSP pages for viewing on a HTML page.
- Developed the front end using JSP.
- Developed various EJB's to handle business logic.
- Designed and developed numerous Session Beans deployed on Web logic Application Server.
- Implemented Database interactions using JDBC with back-end Oracle.
- Worked on Database designing, Stored Procedures, and PL/SQL.
- Created triggers and stored procedures using PL/SQL.
- Written queries to get the data from the Oracle database using SQL.
Environment: J2EE, Servlets, JSP, Struts, Spring, Hibernate, Oracle, TOAD, Weblogic Server
Confidential
Associate JAVA Developer
Responsibilities:
- Involved in the complete SDLC software development life cycle of the application from requirement analysis to testing.
- Developed the modules based on struts MVC Architecture.
- Developed The UI using JavaScript, JSP, HTML, and CSS for interactive cross browser functionality and complex user interface.
- Created Business Logic using Servlets, Session beans and deployed them on Weblogic server.
- Used MVC struts framework for application design.
- Created complex SQL Queries, PL/SQL Stored procedures, Functions for back end.
- Prepared the Functional, Design and Test case specifications.
- Involved in writing Stored Procedures in Oracle to do some database side validations.
- Performed unit testing, system testing and integration testing so as to ensure quality.
- Developed Unit Test Cases. Used JUNIT for unit testing of the application.
- Provided Technical support for production environments resolving the issues, analyzing the defects, providing and implementing the solution defects. Resolved more priority defects as per the schedule.
Environment: Java, JSP, Servlets, Weblogic, Oracle, JUnit, SQL, XML, Toad, Testing, QA