We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

CA

SUMMARY

  • 8.6 Years of extensive IT experience in Analysis, Design, Development, Implementation and experience in Big Data using Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Oozie, Flume, Hbase, Cassandra, ZooKeeper,Impala, Spark, Kafka.Java/J2EE and Reporting.
  • High Exposure on Big Data technologies andHadoopecosystem, In - depth understanding of Map Reduce and theHadoopInfrastructure.
  • Capable of processing large sets of structured, semi-structured and unstructured data and supporting systems application architecture.
  • Importing and exporting data into HDFS. Having experience in writing PIG Scripts.
  • Involved in creating HIVE tables, loading with data and writing HIVE queries.
  • Excellent understanding of Hadoop architecture and its components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce programming paradigm.
  • Experience in providing support to data analyst in running Pig and Hive queries.
  • Experience in working with different data sources like xml files, Json files, Sql server, Oracle to load data into Hive and HBasetables.
  • Worked on Performance Tuning of Hadoop jobs by applying techniques such as Map Side Joins, Partitioning and Bucketing.
  • Having good knowledge in NoSQL databases like Hbase, Cassandra.
  • Experience in real-time Big Data solutions using HBase handling billions of records.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Experienced in implementing unified data platforms using Kafka producers/ consumers, implement pre-processing using storm topologies.
  • Experienced in migrating map reduce programs into Spark RDD transformations, actions to improve performance.
  • Experience utilizing Java tools in Business, Web, and Client-Server environments including Java, J2ee, Hibernate, restful webservices, Jasper Reports, report builder,Jboss Fuse ESB and oracle.
  • Expertise in developing reports by using reporting tools like Report builder 3.0 and BIRT.
  • Expertise in developing application by using JBoss Fuse ESB 6.1.
  • Strong experience in writing database objects like Stored Procedures, Functions, Triggers, PL/SQL packages and Cursors for Oracle, SQL Server, MySQL and TelescopeEnterprise8.4.
  • Experience on Source control repositories like SVN, CVS and GITHUB.
  • Experienced in build/deploy multi module applications using Maven, Jira, Docker and Jenkins.
  • Worked primarily in the domains of Medical, Manufacturing and telecommunication and main area of experience has been involved in project development of java and j2ee applications.

TECHNICAL SKILLS

Platforms: Windows (2000/XP), Linux, CentOS and Mac

Big Data Ecosystems: Hadoop, MapReduce, HDFS, Hive, Pig, Sqoop, Oozie, Flume, Hbase, Flume, impala, CDH4, Spark and Kafka.

Programming Languages: Java and j2ee

Scripting Languages: Jsp, Servlets, HTML, Jquery, Angular JS.

Databases: Oracle, SQL Server, NoSQL.

Frameworks: Struts, Hibernate, Fuse ESB 6.1 and RestfulWeb Services

Tools: My Eclipse, Maven Build Tool, Report builder 3.0, Hortonworks, MapR, Jira, Docker.

Servers: JBoss, Tomcat, BEA Web Logic 8.1, WebSphere 6.

Methodologies: UML, Design Patterns.

Concepts: JMS, Birt Reports, Jasper Reports

PROFESSIONAL EXPERIENCE

Confidential - CA

Hadoop Developer

Responsibilities:

  • Involved in complete SDLC of project includes requirements gathering, design documents, development, testing and production environments.
  • Developed Java Map Reduce programs on log data and analysed to support IOT.
  • Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
  • Good knowledge about Cassandra architecture, read, write paths and quering using Cassandra shell.
  • Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
  • Configuring Spark Streaming to receive real time data from the Kafka and Store the stream data to HDFS.
  • Extracted the data from multiple data sources into Hbase,Hive using Sqoop.
  • Monitored workload, job performance and capacity planning using Cloudera Manager.
  • Involved in Agile methodologies, daily scrum meetings, sprint planning.
  • Created Hive tables to load large sets of structured, semi-structured and unstructured data coming from Sql Server and a variety of portfolios.
  • Supported code/design analysis, strategy development and project planning.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Produced and REST based web services.
  • Implemented command design pattern to support micro services.
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.

Environment: Hadoop,HDFS, MapReduce, Hive, Java, J2ee, Restful web services, Report Builder, Birt, HBase, Sqoop, CentOS, Cloudera.

Confidential

Hadoop Developer

Responsibilities:

  • Involved in complete SDLC of project includes requirements gathering, design documents, development, testing and production environments.
  • Responsible to manage data coming from different sources and involved in HDFS maintenance and loading of structured and unstructured data.
  • Developed Java Map Reduce programs on mainframe data to transform into structured way.
  • Performed data analysis in Hive by creating tables, loading it with data and writing hive queries which will run internally in a MapReduce way.
  • Created Hive External tables and loaded the data in to tables and query data using HQL.
  • Developed optimal strategies for distributing the mainframe data over the cluster. Importing and exporting the stored mainframe data into HDFS and Hive.
  • Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
  • Implemented Hbase API to store the data into Hbase table from hive tables.
  • Writing Hive queries for joining multiple tables based on business requirement.
  • Monitored workload, job performance and capacity planning using Cloudera Manager.
  • Implemented JZOS API to convert mainframe data to text file using core java.
  • Involved in build applications using Maven and integrated with CI servers like Jenkins to build jobs.
  • Involved in Agile methodologies, daily scrum meetings, spring planning.
  • Involved in Build, Deployment and Integration.

Environment: Hadoop,HDFS, MapReduce,Hbase, Hive, Impala, Cloudera,Maven, Java, Jzos, Mainframe.

Confidential

Hadoop Developer

Responsibilities:

  • Experience in developing solutions to analyze large data sets efficiently.
  • Developed Map Reduce application to find out the useful metrics from the data. Did a thorough testing in local mode and distributed mode found bugs with the code and ensured 100% issue free delivery to production.
  • Developed Java Map Reduce programs on log data to transform into structured way.
  • Developed optimal strategies for distributing the data over the cluster; importing and exporting the stored RDBMS data into HDFS and Hive using Impala.
  • Implemented Hive Generic UDF's to in corporate business logic into Hive Queries.
  • Writing Hive queries for joining multiple tables based on business requirement.
  • Integrated impala with the rest of theHadoopstack supporting several types ofHadoopjobs out of the box (such as Map-Reduce, Hive, and impala) as well as system specific jobs (such as Java programs).
  • Created web UI with angular JS to interact with spring application to execute hive quires.
  • Monitored workload, job performance and capacity planning using Cloudera Manager.
  • Involved in Agile methodologies, daily scrum meetings, spring planning.
  • Prepared design documents and functional documents.
  • Involved in Build, Deployment and Integration.

Environment: Hadoop,HDFS,MapReduce, Hive, Impala, Cloudera, Java, Spring, Maven, Angular JS

Confidential

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop.
  • Installed and configured Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Setup and benchmarked Hadoop clusters for internal use.
  • Developed Simple to complex Map/reduce Jobs using Java programming language that are implemented using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Analyzed the data by performing Hive queries (HiveQL) and running Pig scripts (Pig Latin) to study customer behavior. Used UDF’s to implement business logic in Hadoop.
  • Implemented business logic by writing UDFs in Java and used various UDFs from other sources.
  • Experienced on loading and transforming of large sets of structured and semi structured data.
  • Managing and Reviewing Hadoop Log Files, deploy and Maintaining Hadoop Cluster.
  • Involved in implementation of JBoss Fuse ESB 6.1 .
  • Consumed REST based web services.

Environment: Hortonworks,Hadoop-Hdfs, Hive, Impala, Java, J2ee, RestServices, MapReduse, Jboss Fuse ESB 6.1.

Confidential

Hadoop Developer/ Sr. Java Developer

Responsibilities:

  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
  • Installed and configured MapReduce, HIVE and the HDFS; implemented CDH3 Hadoop cluster on CentOS. Assisted with performance tuning and monitoring.
  • Created Hive tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Involved in implementation of JBoss Fuse ESB 6.1 and Restful web servies.
  • Supported code/design analysis, strategy development and project planning.
  • Created reports for BI using Sqoop to export data into HDFS and Hive.
  • Developed multiple MapReduce jobs in Java for data cleaning and preprocessing.
  • Assisted with data capacity planning and node forecasting.
  • Collaborated with the infrastructure, network, database, application and BI teams to ensure data quality and availability.

Environment: Hadoop, HDFS, HIVE, PIG, Map Reduce, Sqoop, Java /J2ee,Restservices, JBoss Fuse ESB 6.1,BIRT

Confidential

Sr.Java Developer

Responsibilities:

  • Involved in end to end development and code review.
  • Developed all reports based Report Builder.
  • Developed object relational mapping model using Hibernate In persistence DAO layer.
  • Developed Web application based on Struts MVC framework.
  • Implemented various Design patterns like Singleton, MVC, DAO.
  • Developed Simple to complex Map/reduce Jobs using Java programming language that are implemented using Hive and Pig.
  • Optimized Map/Reduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Implemented business logic by using Struts Action classes and Struts controller components.
  • Coordinating the team regularly to discuss on the better functionality.
  • Followed good practices while developing code.
  • Analyzing issues and performs impact analysis.
  • Debugging, Defect Fixing and Maintenance.
  • Identifying test cases and Unit testing.
  • Developed stored procedures, functions and triggers in SQL server 2008.

Environment: Struts, Hibernate, Java, JSP, Servlets, WebSphere, XML, TelescopeEnterprise, MS Sql Report Builder 3.0, MSSql Server,Eclipse, MapReduce.

We'd love your feedback!