Hadoop Developer Resume
SUMMARY
- 12 years of experience in analysis, architectural design, development, testing and maintenance by using Java, Big Data (Apache Hadoop, MapReduce, HDFS, HBase, Pig, Hive, Impala, Zookeeper and NoSQL HBase), JEE, Web Services, EJB, JMS, Spring, Hibernate, Oracle, Linux and Windows.
- 3 years of experience as a Hadoop Developer using various technologies in Hadoop eco system.
- Over 2 years of experience in large scale enterprise systemsarchitecturaldesign including high and low level design
- Hands on experience in MapReduce, HDFS, Pig, Hive, Oozie, HBase, Zookeeper, Flume, and Sqoop.
- Experience in NoSQL databases like Hbase
- Strong knowledge in cleansing and analyzing data using HiveQL, Pig Latin, and MapReduce programs in Java.
- Experienced in writing custom UDFs and UDFs for extending Hive and Pig core functionalities.
- Ability to develop Pig UDFs to pre - process the data for analysis.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS), Teradata and vice versa.
- Ability to develop, schedule and monitoring the Oozie jobs.
- Hands-on experience with message broker such as Apache Kafka.
- Good experience working with Distributions such as Cloudera, Hortonworks and Data lake.
- Having good knowledge on Hadoop Administration like Cluster configuration, Single Node Configuration, Multi Node Configuration, Data Node Commissioning and Decommissioning, Name Node Backup and Recovery, HBase, HDFS and Hive Configuration, Monitoring clusters, Access control List.
- Experience working with JAVA, JEE, JDBC, JSP, Eclipse, MS SQL Server.
- Hands on experience in enterprise/web application development using Java, RDBMS, and Linux shell scripting
- Strong experience in applying Design Patterns and Development of multi-tier applications.
- Extensive experience with SQL, PL/SQL and database concepts.
- Expertise in debugging and optimizing SQL Queries.
- Strong experience in Implementing multi-threading in Core Java.
- Strong experience in Object Oriented Programming, Agile Methodologies, Test Driven Programming, SDLC, OOAD and UML
- Have strong knowledge static analysis, profiling and performance tuning.
- Experience in all stages of SDLC (Agile, Waterfall), writing Technical Design document, Development, Testing and Implementation of Enterprise level Data mart and Data warehouses.
- Good exposure to software configuration management and continuous integration tools in the process of code review, build automation and deployment. Jenkins, Sonar, Review Board, Git, PMD, Check Style, SVN, VSS, Clear case and Perforce
- Hands on experience in java and JEE design patterns like Singleton, Abstract factory, Business Delegate, Service Locator, Session Facade, DAO, Business Delegates and Front controller
- Strong working experience and knowledge on the MVC architectures implementations like Struts, and Spring Framework.
TECHNICAL SKILLS
Hadoop/Big Data: MapReduce, HDFS, Oozie, Hive, Pig, Sqoop, Flume, Zookeeper, HBaseCloudera, Hortonworks: Java Technologies java, JEE, RESTful and SOAP Web services, JMS, EJB, Spring, Spring: batch, Spring Security, Hibernate, JSP, Servlets
DBMS/Databases: Oracle, MySQL, SQL Server and DB2
Programming Languages: java, C, C++
Operating Systems: Windows, UNIX/LINUX
Methodologies: Agile (Scrum), Iterative and Waterfall
NOSQL Databases: HBase
Version Control Tools: Git, SVN, CVS, Clear case and Perforce
Tools: Maven, Gradle, Ant, Jenkins, Sonar, PMD, Check Style, Drools
PROFESSIONAL EXPERIENCE
Confidential
Hadoop Developer
Environment: Data Lake, MapReduce, HDFS, Hive, Impala, Pig, SQL, Sqoop, Oozie, Shell scripting, Cron Jobs, Apache Kafka, Java and JEE.
Responsibilities:
- Involved in loading and transforming large sets of structured, semi structured and unstructured data from relational databases into HDFS using Sqoop imports.
- Developed Sqoop scripts to import export data from relational sources and handled incremental loading on the customer, transaction data by date.
- End-to-end usecase implementation including architectural design, coding and testing
- Developed simple and complex MapReduce programs in Java for Data Analysis on different data formats.
- Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Worked on partitioning Hive tables and running the scripts in parallel to reduce run-time of the scripts.
- Worked on Data Serialization formats for converting Complex objects into sequence bits by using Avro, Parquet, JSON, CSV formats.
- Responsible for analyzing and cleansing raw data by performing Hive queries and running Pig scripts on data.
- Administration, installing, upgrading and managing distributions of Hadoop, Hive, HBase.
- Involved in performance of troubleshooting and tuning Hadoop clusters.
- Created Hive tables, loaded data and wrote Hive queries that run within the map.
- Implemented business logic by writing Hive UDFs in Java.
- Developed Shell scripts and some of Perl scripts based on the user requirement.
- Used Oozie Operational Services for batch processing and scheduling workflows dynamically.
- Technology stack assessment implementing various proof of concepts (POC) to eventually adopt them to benefit from the Big Data Hadoop initiative.
Confidential, San Jose
Hadoop Developer
Environment: MapReduce, HDFS, Hive, Pig, HBase, PL/SQL, SQL, Sqoop, Flume, Oozie, Zookeeper, Java, Cloudera, Eclipse.
Responsibilities:
- Responsible for Installation and configuration of Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
- Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
- Developed simple and complex MapReduce programs in Java for Data Analysis on different data formats.
- Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Worked on partitioning Hive tables and running the scripts in parallel to reduce run-time of the scripts.
- Use data ingestion tools such as Flume for streaming data and Sqoop for bulk data transfer
- Set up and maintain a Hadoop cluster running HDFS and MapReduce on YARN
- Worked on Data Serialization formats for converting Complex objects into sequence bits by using Avro, Parquet, JSON, CSV formats.
- Good Exposure to the Cloudera distribution
- Responsible for analyzing and cleansing raw data by performing Hive queries and running Pig scripts on data.
- Installing, Upgrading and Managing Hadoop Clusters
- Administration, installing, upgrading and managing distributions of Hadoop, Hive, HBase.
- Advanced knowledge in performance troubleshooting and tuning Hadoop clusters.
- Created Hive tables, loaded data and wrote Hive queries that run within the map.
- Implemented business logic by writing Pig UDFs in Java and used various UDFs from Piggy Bank and other sources.
- Used Oozie Operational Services for batch processing and scheduling workflows dynamically.
- Extensively worked on creating End-End data pipeline orchestration using Oozie.
- Writing SQL, PL/SQL Stored Procedures to analyze the data before ingestion into lake
- Populated HDFS and HBase with huge amounts of data using Apache Kafka.
- Processed the source data to structured data and store in NoSQL database HBase.
- Design and develop Java API (Commerce API) which provides functionality to connect to the HBase through Java services.
- Responsible for continuous monitoring and managing Elastic MapReduce cluster through AWS console.
- Evaluated suitability of Hadoop and its ecosystem to the above project and implementing / validating with various proof of concept (POC) applications to eventually adopt them to benefit from the Big Data Hadoop initiative.
Confidential
Java/JEE Developer
Environment: JAVA, JEE, Web Services (SOAP and REST), WSDL, JMS, spring, Hibernate, JUnit, Sonar, Para soft, Jerkin’s, XML-Schema, XML and LINUX.
Responsibilities:
- Analysis, Design and Development
- Owning the modules and providing continuous supporting implementing new features
- Understanding existing components and converting those components into Web Services.
- Deprecating some of the old components and adding the new features
- Used Java Mail API to send confirmation Email to customers.
- Involved in the design of XML Schema for transferring the trade message from front office to back office using JAXP (DOM) which maps the data into objects for presentation.
- Experience in writing SQL, PL/SQL Stored Procedures.
- Developed system prototype and Presentation to the Vendors.
- Developed the Sequence diagrams, Class diagrams using UML and the Rational Rose tool.
- Responsible for troubleshooting and fixing production defects, contributing to code reviews.
- Participating in review meetings and make sure every method should have Unit test cases and generating code coverage reports by using Cobertura.
- Worked closely with deployment teams during testing and production deployments of the application.
- Taking care of daily build management.
Confidential
Java / J2EE Developer
Environment: Java 1.5, JEE, Web Services (SOAP and REST), WSDL, JMS, spring, Hibernate, TestNG, XML-Schema, Sonar, XML, Eclipse 3.6, LINUX, Zookeeper, JMX, Dynatrace, Oracle and Perforce
Responsibilities:
- Understanding existing components and converting those components into Web Services.
- Deprecating some of the old components and adding the new features
- Participating in review meetings and make sure each method should have Unit test cases and generating code coverage reports by using Cobertura.
- Implemented Enterprise Logging service using JMS and apache CXF.
- Wrote SQL queries, stored procedures, and triggers to perform back-end database operations
- Developed SQL stored procedures and prepared statements for updating and accessing data from database.
- Involved in Unit testing, Integration testing and User Acceptance testing.
Confidential
Java / J2EE Developer
Environment: Web Services, SOAP, WSDL, JAVA 1.5, J2EE, spring, Mule (ESB), Drools, Hibernate, XPath and XML-Schema, XML, Tibco EMS, JPDL, Eclipse 3.3 and AccuRev
Responsibilities:
- Gathered business requirements and wrote functional specifications and detailed design documents
- Extensively used Core Java and XML
- Designed the logical and physical data model, generated DDL scripts, and wrote DML scripts for Oracle 9i database
- Implemented Enterprise Logging service using JMS and apache CXF.
- Developed Unit Test Cases, and used JUNIT for unit testing of the application
- Implemented JMS producer and Consumer using Mule ESB.
- Closely worked with QA, Business and Architect to solve various Defects in quick and fast to meet deadlines
Confidential
Java Developer
Environment: java, Web Services, Java 1.5, J2EE, LWI (Light weight Infrastructure) Apache Ant and JUnit, Eclipse, Lotus Expeditor
Responsibilities:
- Developed the Sequence diagrams, Class diagrams using UML and the Rational Rose tool.
- Coded (EJB) Session Bean and the Entity Beans for the registration process to register a new Vendor.
- Involved in coding the Enterprise Java Beans to implement Business logic and accessing using Java Server Pages (JSP).
- Responsible for developing and deploying the EJB’s in WebSphere Application Server.
- Wrote several Build.xml files to build separate EJBs.
- Responsible for Notification component that would send mail to notify the Vendors if there are any changes in the order or shipment specific notifications. JMS and Java Mail are used for the same.
Confidential
Java Developer
Environment: Struts, JSP, Servlets, HTML, Oracle, Web Services, Hibernate Ant, JUnit, CVS and WebLogic.
Responsibilities:
- Involved in writing the validation rules classes for general server side validations for implementing validation rules as part observer J2EE design pattern.
- Used OR mapping tool Hibernate for the interaction with database. Involved in writing Hibernate queries and Hibernate specific configuration and mapping files.
- Developed EJB tier using Session Facade, Singleton and DAO design patterns, which contains business logic, and database access functions.
- Involved in developing JSP pages and custom tag for presentation layer in Spring framework. Developed tabs with rollover effect using JavaScript plus other effects.
- Developed web services using SOAP and WSDL with Apache Axis 2.