We provide IT Staff Augmentation Services!

Hadoop Developer Resume

2.00/5 (Submit Your Rating)

CA

PROFESSIONAL SUMMARY:

  • 8.6 Years of extensive IT experience in Analysis, Design, Development, Implementation and experience in Big Data using Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Oozie, Flume, Hbase, Cassandra, ZooKeeper, Impala, Spark, Kafka, Java/J2EE and Reporting.
  • High Exposure on Big Data technologies and Hadoop ecosystem, In - depth understanding of Map Reduce and the Hadoop Infrastructure.
  • Capable of processing large sets of structured, semi-structured and unstructured data and supporting systems application architecture.
  • Importing and exporting data into HDFS. Having experience in writing PIG Scripts.
  • Involved in creating HIVE tables, loading with data and writing HIVE queries.
  • Excellent understanding of Hadoop architecture and its components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce programming paradigm.
  • Experience in providing support to data analyst in running Pig and Hive queries.
  • Experience in working with different data sources like xml files, Json files, Sql server, Oracle to load data into Hive and HBasetables.
  • Worked on Performance Tuning of Hadoop jobs by applying techniques such as Map Side Joins, Partitioning and Bucketing.
  • Having good knowledge in NoSQL databases like Hbase, Cassandra.
  • Experience in real-time Big Data solutions using HBase handling billions of records. 
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa. 
  • Experienced in implementing unified data platforms using Kafka producers/ consumers, implement pre-processing using storm topologies. 
  • Experienced in migrating map reduce programs into Spark RDD transformations, actions to improve performance. 
  • Experience utilizing Java tools in Business, Web, and Client-Server environments including Java, J2ee, Hibernate, restful webservices, Jasper Reports, report builder,Jboss Fuse ESB and oracle.
  • Expertise in developing reports by using reporting tools like Report builder 3.0 and BIRT.
  • Expertise in developing application by using JBoss Fuse ESB 6.1.
  • Strong experience in writing database objects like Stored Procedures, Functions, Triggers, PL/SQL packages and Cursors for Oracle, SQL Server, MySQL and TelescopeEnterprise8.4.
  • Experience on Source control repositories like SVN, CVS and GITHUB. 
  • Experienced in build/deploy multi module applications using Maven, Jira, Docker and Jenkins.
  • Worked primarily in the domains of Medical, Manufacturing and telecommunication and main area of experience has been involved in project development of java and j2ee applications.

TECHNICAL SKILLS:

Platforms: Windows (2000/XP), Linux, CentOS and Mac

Big Data Ecosystems: Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Oozie, Flume, Hbase, Flume, impala, CDH4, Spark and Kafka.

Programming Languages: Java and j2ee

Scripting Languages: Jsp, Servlets, HTML, Jquery, Angular JS.

Databases: Oracle, SQL Server, NoSQL.

Frameworks: Struts, Hibernate, Fuse ESB 6.1 and RestfulWeb Services

Tools: My Eclipse, Maven Build Tool, Report builder 3.0, Hortonworks, MapR, Jira, Docker.

Servers: JBoss, Tomcat, BEA Web Logic 8.1, Web Sphere 6.

Methodologies: UML, Design Patterns.

Concepts: JMS, Birt Reports, Jasper Reports

PROFESSIONAL EXPERIENCE:

Confidential, CA

Hadoop Developer

Responsibilities: 

  • Involved in complete SDLC of project includes requirements gathering, design documents, development, testing and production environments.
  • Developed Java Map Reduce programs on log data and analysed to support IOT. 
  • Implemented Hive Generic UDF's to incorporate business logic into Hive Queries. 
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
  • Good knowledge about Cassandra architecture, read, write paths and quering using Cassandra shell.
  • Implemented Hive Generic UDF's to incorporate business logic into Hive Queries. 
  • Configuring Spark Streaming to receive real time data from the Kafka and Store the stream data to HDFS. 
  • Extracted the data from multiple data sources into Hbase,Hive using Sqoop.
  • Monitored workload, job performance and capacity planning using Cloudera Manager. 
  • Involved in Agile methodologies, daily scrum meetings, sprint planning. 
  • Created Hive tables to load large sets of structured, semi-structured and unstructured data coming from Sql Server and a variety of portfolios.
  • Supported code/design analysis, strategy development and project planning.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Produced and REST based web services.
  • Implemented command design pattern to support micro services.
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.

Environment: Hadoop, HDFS , MapReduce , Hive, Java, J2ee, Restful web services, Report Builder, Birt, HBase, Sqoop, CentOS, Cloudera.

Confidential

Hadoop Developer

Responsibilities: 

  • Involved in complete SDLC of project includes requirements gathering, design documents, development, testing and production environments. 
  • Responsible to manage data coming from different sources and involved in HDFS maintenance and loading of structured and unstructured data.
  • Developed Java Map Reduce programs on mainframe data to transform into structured way. 
  • Performed data analysis in Hive by creating tables, loading it with data and writing hive queries which will run internally in a MapReduce way.
  • Created Hive External tables and loaded the data in to tables and query data using HQL.
  • Developed optimal strategies for distributing the mainframe data over the cluster. Importing and exporting the stored mainframe data into HDFS and Hive. 
  • Implemented Hive Generic UDF's to incorporate business logic into Hive Queries. 
  • Implemented Hbase API to store the data into Hbase table from hive tables.
  • Writing Hive queries for joining multiple tables based on business requirement.
  • Monitored workload, job performance and capacity planning using Cloudera Manager. 
  • Implemented JZOS API to convert mainframe data to text file using core java.
  • Involved in build applications using Maven and integrated with CI servers like Jenkins to build jobs. 
  • Involved in Agile methodologies, daily scrum meetings, spring planning. 
  • Involved in Build, Deployment and Integration.

Environment: Hadoop, HDFS , MapReduce,Hbase, Hive, Impala, Cloudera ,Maven , Java, Jzos, Mainframe.

Confidential

Hadoop Developer

Responsibilities: 

  • Experience in developing solutions to analyze large data sets efficiently.
  • Developed Map Reduce application to find out the useful metrics from the data. Did a thorough testing in local mode and distributed mode found bugs with the code and ensured 100% issue free delivery to production.
  • Developed Java Map Reduce programs on log data to transform into structured way. 
  • Developed optimal strategies for distributing the data over the cluster; importing and exporting the stored RDBMS data into HDFS and Hive using Impala. 
  • Implemented Hive Generic UDF's to incorporate business logic into Hive Queries. 
  • Writing Hive queries for joining multiple tables based on business requirement.
  • Integrated impala with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (such as Map-Reduce, Hive, and impala) as well as system specific jobs (such as Java programs). 
  • Created web UI with angular JS to interact with spring application to execute hive quires.
  • Monitored workload, job performance and capacity planning using Cloudera Manager. 
  • Involved in Agile methodologies, daily scrum meetings, spring planning. 
  • Prepared design documents and functional documents. 
  • Involved in Build, Deployment and Integration.

Environment: Hadoop, HDFS ,MapReduce , Hive, Impala, Cloudera , Java, Spring, Maven , Angular JS

Confidential

Hadoop Developer

Responsibilities: 

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Setup and benchmarked Hadoop clusters for internal use.
  • Developed Simple to complex Map/reduce Jobs using Java programming language that are implemented using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Analyzed the data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior. Used UDF’s to implement business logic in Hadoop.
  • Implemented business logic by writing UDFs in Java and used various UDFs from other sources.
  • Experienced on loading and transforming of large sets of structured and semi structured data.
  • Managing and Reviewing Hadoop Log Files, deploy and Maintaining Hadoop Cluster.
  • Involved in implementation of JBoss Fuse ESB 6.1 .
  • Consumed REST based web services.

Environment:Hortonworks, Hadoop-Hdfs, Hive, Impala, Java, J2ee, RestServices, MapReduse, Jboss Fuse ESB 6.1.

Confidential

Hadoop Developer/ Sr. Java Developer

Responsibilities: 

  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
  • Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
  • Created Hive tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Involved in implementation of JBoss Fuse ESB 6.1 and Restful web servies.
  • Supported code/design analysis, strategy development and project planning.
  • Created reports for BI using Sqoop to export data into HDFS and Hive.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Assisted with data capacity planning and node forecasting.
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.

Environment: Hadoop, HDFS, HIVE, PIG, Map Reduce, Sqoop, Java /J2ee,Restservices, JBoss Fuse ESB 6.1,BIRT

Confidential

Sr.Java Developer

Responsibilities: 

  • Involved in end to end development and code review.
  • Developed all reports based Report Builder.
  • Developed object relational mapping model using Hibernate In persistence DAO layer.
  • Developed Web application based on Struts MVC framework.
  • Implemented various Design patterns like Singleton, MVC, DAO.
  • Developed Simple to complex Map/reduce Jobs using Java programming language that are implemented using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Implemented business logic by using Struts Action classes and Struts controller components.
  • Coordinating the team regularly to discuss on the better functionality.
  • Followed good practices while developing code.
  • Analyzing issues and performs impact analysis.
  • Debugging, Defect Fixing and Maintenance.
  • Identifying test cases and Unit testing.
  • Developed stored procedures, functions and triggers in SQL server 2008.

Environment: Struts, Hibernate, Java, JSP, Servlets, WebSphere, XML, TelescopeEnterprise, MS Sql Report Builder 3.0, MSSql Server,Eclipse, MapReduce.

We'd love your feedback!