We provide IT Staff Augmentation Services!

Hadoop Developer Resume

O Fallon, MissourI

SUMMARY:

  • Around 7 years of experience in IT industry covering Development, Analysis, Design, Testing and System Maintenance.
  • Good Understanding of Hadoop Architecture and various components such as HDFS, Job tracker, Task tracker, NameNode, Data Node and MapReduce programming paradigm.
  • Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Zookeeper and Flume.
  • Good Exposure on Apache Hadoop Map Reduce programming, PIG Scripting and Distribute Application and HDFS.
  • Good knowledge on Hadoop cluster architecture and monitoring the cluster.
  • In - depth understanding of data structure and algorithms.
  • Experience in managing and reviewing hadoop log files.
  • Implemented in setting up standards and processes for Hadoop based application design and implementation.
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
  • Experience in managing Hadoop clusters using Cloudera Manager tool.
  • Developed applications using Core Java, Multithreading, Collections, JDBC, Swing, Networking, Reflections.
  • Java/J2EE Software Developer with experience of Core Java and Web based applications with expertise in reviewing client requirement; prioritize requirements, creating project proposal (scope, estimation) and baseline project plan.
  • Experience in designing Use Cases, Class diagrams, Sequence and Collaboration diagrams for multi-tiered object-oriented system architectures.
  • Extensive experience with design and development of J2EE based applications involving technologies such as Java Server Pages (JSP), Java Messaging Service (JMS), Java Data Base Connectivity (JDBC).
  • Experience on working with Tez for processing Hive queries.
  • Worked on data ingestion methodologies
  • Established connections to ingest data in and from HDFS
  • Excellent knowledge on Apache Spark and its stack

TECHNICAL SKILLS:

Big Data Ecosystem: HDFS, HBase, Hadoop MapReduce, Zookeeper, Hive, Pig, Sqoop, Flume, Oozie

Languages: C, C++, Java, SQL/PLSQL, COBOL, ASSEMBLER

Methodologies: Agile, Waterfall

Database: Oracle 10g, DB2, MySQL, My SQL Server, Teradata, MongoDB

Web Tools: XML, ODBC, JDBCIDE / Testing Tools: Eclipse, Net Beans

Operating System: Windows, UNIX, Linux

Scripts: JavaScript, Shell Script

PROFESSIONAL EXPERIENCE:

Confidential, O’Fallon, Missouri

Hadoop Developer

Responsibilities:

  • Developed multiple mapreduce jobs in java for data cleaning and pre-processing.
  • Developed simple to complex mapreduce jobs using hive and pig.
  • Involved in creating hive tables loading and analyzing data using hive queries.
  • Involved in running hadoop jobs for processing millions of records of text data.
  • Responsible for managing data from multiple sources
  • Implemented best payoff logic using pig scripts.
  • Assisted in exporting analyzed data to relational databases using sqoop.
  • Involved in loading data from UNIX file system to HDFS.
  • Created HBase tables to store different data formats.
  • Installed and configured Hive.
  • Experience in managing and reviewing hadoop log files.
  • Export the analyzed data to the relational databases using sqoop for visualization and to generate reports for the BI team.
  • Analyzed large amounts of datasets to determine optimal way to aggregate and report on it.
  • Supported in setting up QA environment and updating configurations for implementing scripts with pig and sqoop.
  • Established connections to ingest data in and from HDFS
  • Monitored jobs on Informatica monitoring tool.
  • Fetched data from oracle and written into HDFS
  • Used Hive connections to analyze data from Oracle.

Environment: Hadoop, HDFS, MapReduce, Hive, Flume, HBase, Sqoop, PIG, Java (JDK 1.6), Eclipse, MySQL, Zookeeper, Oozie

Confidential, Strongsville, Ohio

Hadoop Developer

Responsibilities:

  • Analyzed large amounts of datasets to determine optimal way to aggregate and report on it.
  • Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Installed and configured Pig and also written Pig Latin scripts.
  • Wrote MapReduce jobs.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Developing scripts and Batch job to schedule various Hadoop programs.
  • Written Hive queries for data analysis to meet the business requirements.
  • Creating Hive tables and working on them using Hive QL.
  • Importing and exporting data into HDFS and Hive using sqoop.

Environment: Hadoop, MapReduce, HDFS, Hive, Java, Hadoop distribution of Cloudera, Pig, HBase, Linux, XML, MySQL, MySQL Workbench, Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Informatica, Spark.

Confidential

Hadoop Developer

Responsibilities:

  • Gathering functional requirements. Analyzed large amounts of datasets to determine optimal way to aggregate and report on it.
  • Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
  • Installed and configured Pig and also written Pig Latin scripts.
  • Wrote MapReduce jobs.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Developing scripts and Batch job to schedule various Hadoop programs.
  • Written Hive queries for data analysis to meet the business requirements.
  • Creating Hive tables and working on them using Hive QL.
  • Importing and exporting data into HDFS and Hive using sqoop.

Environment: Hadoop, MapReduce, HDFS, Hive, Java, Hadoop distribution of Cloudera, Pig, HBase, Linux, XML, MySQL, MySQL Workbench, Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Java,Eclipse,cobol,DB2,Oracle, DB2, PL/SQL, Toad, QMF.

Confidential

Responsibilities:

  • Collect and collate appropriate data for use in databases and conduct related research.
  • Design and development of Java classes and interfaces.
  • Worked in core Java for the client side validations
  • Design and developed JDBC connection objects for the data retrieval and update.
  • Loading the data into DB2 environment.
  • Working on java and mainframes.
  • Implemented JReport for annual report generation with charts.
  • Developed Database applications using SQL and PL/SQL
  • Scheduling of jobs.
  • Responsible for analyzing and developing the data.
  • Coordinating with cross-functional teams in different locations for quality data and analysis.
  • Designing and developing software applications, working following rules of SDLC.
  • Designing and developing software applications, working following rules of SDLC (Software Development Life Cycle)
  • Creating production and analysis report and Handling Production Issues and Releases.
  • Monitor and maintain the quality of database systems and secure access and use.

Hire Now