Hadoop Developer Resume
O Fallon, MissourI
SUMMARY:
- Around 7 years of experience in IT industry covering Development, Analysis, Design, Testing and System Maintenance.
- Good Understanding of Hadoop Architecture and various components such as HDFS, Job tracker, Task tracker, NameNode, Data Node and MapReduce programming paradigm.
- Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop MapReduce, HDFS, HBase, Hive, Sqoop, Pig, Zookeeper and Flume.
- Good Exposure on Apache Hadoop Map Reduce programming, PIG Scripting and Distribute Application and HDFS.
- Good knowledge on Hadoop cluster architecture and monitoring the cluster.
- In - depth understanding of data structure and algorithms.
- Experience in managing and reviewing hadoop log files.
- Implemented in setting up standards and processes for Hadoop based application design and implementation.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems and vice-versa.
- Experience in managing Hadoop clusters using Cloudera Manager tool.
- Developed applications using Core Java, Multithreading, Collections, JDBC, Swing, Networking, Reflections.
- Java/J2EE Software Developer with experience of Core Java and Web based applications with expertise in reviewing client requirement; prioritize requirements, creating project proposal (scope, estimation) and baseline project plan.
- Experience in designing Use Cases, Class diagrams, Sequence and Collaboration diagrams for multi-tiered object-oriented system architectures.
- Extensive experience with design and development of J2EE based applications involving technologies such as Java Server Pages (JSP), Java Messaging Service (JMS), Java Data Base Connectivity (JDBC).
- Experience on working with Tez for processing Hive queries.
- Worked on data ingestion methodologies
- Established connections to ingest data in and from HDFS
- Excellent knowledge on Apache Spark and its stack
TECHNICAL SKILLS:
Big Data Ecosystem: HDFS, HBase, Hadoop MapReduce, Zookeeper, Hive, Pig, Sqoop, Flume, Oozie
Languages: C, C++, Java, SQL/PLSQL, COBOL, ASSEMBLER
Methodologies: Agile, Waterfall
Database: Oracle 10g, DB2, MySQL, My SQL Server, Teradata, MongoDB
Web Tools: XML, ODBC, JDBCIDE / Testing Tools: Eclipse, Net Beans
Operating System: Windows, UNIX, Linux
Scripts: JavaScript, Shell Script
PROFESSIONAL EXPERIENCE:
Confidential, O’Fallon, Missouri
Hadoop Developer
Responsibilities:
- Developed multiple mapreduce jobs in java for data cleaning and pre-processing.
- Developed simple to complex mapreduce jobs using hive and pig.
- Involved in creating hive tables loading and analyzing data using hive queries.
- Involved in running hadoop jobs for processing millions of records of text data.
- Responsible for managing data from multiple sources
- Implemented best payoff logic using pig scripts.
- Assisted in exporting analyzed data to relational databases using sqoop.
- Involved in loading data from UNIX file system to HDFS.
- Created HBase tables to store different data formats.
- Installed and configured Hive.
- Experience in managing and reviewing hadoop log files.
- Export the analyzed data to the relational databases using sqoop for visualization and to generate reports for the BI team.
- Analyzed large amounts of datasets to determine optimal way to aggregate and report on it.
- Supported in setting up QA environment and updating configurations for implementing scripts with pig and sqoop.
- Established connections to ingest data in and from HDFS
- Monitored jobs on Informatica monitoring tool.
- Fetched data from oracle and written into HDFS
- Used Hive connections to analyze data from Oracle.
Environment: Hadoop, HDFS, MapReduce, Hive, Flume, HBase, Sqoop, PIG, Java (JDK 1.6), Eclipse, MySQL, Zookeeper, Oozie
Confidential, Strongsville, Ohio
Hadoop Developer
Responsibilities:
- Analyzed large amounts of datasets to determine optimal way to aggregate and report on it.
- Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
- Installed and configured Pig and also written Pig Latin scripts.
- Wrote MapReduce jobs.
- Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
- Developing scripts and Batch job to schedule various Hadoop programs.
- Written Hive queries for data analysis to meet the business requirements.
- Creating Hive tables and working on them using Hive QL.
- Importing and exporting data into HDFS and Hive using sqoop.
Environment: Hadoop, MapReduce, HDFS, Hive, Java, Hadoop distribution of Cloudera, Pig, HBase, Linux, XML, MySQL, MySQL Workbench, Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Informatica, Spark.
ConfidentialHadoop Developer
Responsibilities:
- Gathering functional requirements. Analyzed large amounts of datasets to determine optimal way to aggregate and report on it.
- Installed and configured Hadoop Mapreduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing.
- Installed and configured Pig and also written Pig Latin scripts.
- Wrote MapReduce jobs.
- Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
- Developing scripts and Batch job to schedule various Hadoop programs.
- Written Hive queries for data analysis to meet the business requirements.
- Creating Hive tables and working on them using Hive QL.
- Importing and exporting data into HDFS and Hive using sqoop.
Environment: Hadoop, MapReduce, HDFS, Hive, Java, Hadoop distribution of Cloudera, Pig, HBase, Linux, XML, MySQL, MySQL Workbench, Java 6, Eclipse, Oracle 10g, PL/SQL, SQL*PLUS, Java,Eclipse,cobol,DB2,Oracle, DB2, PL/SQL, Toad, QMF.
ConfidentialResponsibilities:
- Collect and collate appropriate data for use in databases and conduct related research.
- Design and development of Java classes and interfaces.
- Worked in core Java for the client side validations
- Design and developed JDBC connection objects for the data retrieval and update.
- Loading the data into DB2 environment.
- Working on java and mainframes.
- Implemented JReport for annual report generation with charts.
- Developed Database applications using SQL and PL/SQL
- Scheduling of jobs.
- Responsible for analyzing and developing the data.
- Coordinating with cross-functional teams in different locations for quality data and analysis.
- Designing and developing software applications, working following rules of SDLC.
- Designing and developing software applications, working following rules of SDLC (Software Development Life Cycle)
- Creating production and analysis report and Handling Production Issues and Releases.
- Monitor and maintain the quality of database systems and secure access and use.