We provide IT Staff Augmentation Services!

Hadoop Developer Resume

Irving, TX


  • Highly motivated Hadoop developer and data enthusiast with 7+ years of experience in Information Technology
  • 3+ years of Hadoop system experience translating complex Big Data problems into meaningful solutions
  • 4+ years of experience in database administration and PL/SQL scripting, web services, java and Unix shell scripting
  • Developed databases and projects using R and Python, PL/SQL, Java, NoSQL/MySQL
  • Designed and implemented Big Data solutions using Hadoop, Spark, Hive, Pig, Flume, Sqoop
  • Experienced in Cloudera Distribution including Apache Hadoop (HDFS, YARN, Hive, Pig, Sqoop, Impala, and Flume) using Java and ETL
  • Expertise in writing ETL Jobs for analyzing data using Pig Latin scripting
  • Working experience of HBase and good understanding of NOSQL databases like Cassandra and Mongo DB.
  • Worked with Talend.
  • Experience in importing and exporting data using Sqoop from HDFS to RDBMS and vice - versa
  • Responsible for the field architecture and educating those new to Hadoop on its value to their organization through whiteboard sessions, demos, Technical group presentations, external sessions and reporting, proof of concepts, reference architecture, and training
  • Experience in Java, JSP, Servlets, EJB, WebLogic, WebSphere, Hibernate, Spring, JBoss, JDBC, Java Script, Ajax, JQuery, XML, and HTML
  • Exceptional ability to quickly master new concepts and capable of working in groups as well as independently
  • Excellent interpersonal skills with great teamwork experience
  • Experience in debugging, troubleshooting production systems, profiling and identifying performance bottlenecks


Operating Systems: Linux (Ubuntu, CentOS), Windows, MAC OS

Big Data Frameworks: Hadoop, Spark, HDFS, MapReduce, Hive, Pig, Sqoop, Flume, Ozzie and Zookeeper, Talend

Programming Languages: Java, J2EE (Servlets, Spring,SOAP,REST, JSP, JDBC, Maven), R

Web Technologies: HTML, CSS, XML, JavaScript

Scripting Languages: Python, Bash, PL/SQL, UNIX Shell Scripting

Databases: ORACLE 10g, MySQL, NOSQL HBase


Confidential - Irving, TX

Hadoop Developer


  • Responsible for building scalable distributed data solutions using Hadoop and Spark
  • Implemented complete Big Data solutions including data acquisition, storage, transformation and analysis
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it
  • Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop
  • Worked with different data sources like Avro data files, XML files, Json files,SQL server and Oracle to load data into Hive tables.
  • Worked with Talend for Master Data Management.
  • Coordination with Hadoop Admin’s during deployment to production
  • Developed Pig scripts in the areas where extensive coding needs to be reduced
  • Responsible for creating Hive tables, loading data and writing hive queries
  • Handled importing data from various data sources, performed transformations using Hive and Map Reduce, streamed using Flume and loaded data into HDFS
  • Exported the patterns analyzed by R to and from HDFS using Sqoop
  • Installed Ozzie workflow engine to run multiple MapReduce, Hive, Impala, Zookeeper and Pig jobs which run independently with time and data availability
  • Worked with NoSQL database HBase to create tables and store data
  • Involved in loading data from Linux file system to HDFS.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis
  • Developing Scripts and Batch Job to schedule various Hadoop Program
  • Written Hive queries for data analysis to meet the business requirements

Hadoop Developer



  • Analyzing data using Hadoop components Hive, Impala and Pig
  • Worked hands on with ETL process using Talend.
  • Analyzed large amounts of data sets to determine optimal way to aggregate and report on it
  • Involved in data ingestion into HDFS using Apache Sqoop from a variety of sources using connectors like JDBC and import parameters
  • Used Hive to analyze the partitioned and bucketed data and compute various metrics for reporting
  • Provided quick response to ad hoc internal and external client requests for data
  • Loaded and transformed large sets of structured, semi structured and unstructured data
  • Involved in loading data from UNIX file system to HDFS
  • Responsible for creating Hive tables, loading data and writing hive queries
  • Written Hive queries for data analysis to meet the business requirements
  • Continuously monitoring and managing the Hadoop cluster through Cloudera Manager
  • Developed Hive queries to process the data and generate the data cubes for visualizing
  • Analyzed the application usage in a day-day basis on a sample of machines log data using Spark, Hive and Pig

Environment: Cloudera Hadoop, MapReduce, Hive, Pig, Sqoop,Oozie Scheduler, UNIX, Java 7.0, JSON, Spark, HDFS,YARN, Flume, Ozzie and Zookeeper, Mahout, Talend, Base, MySQL

Confidential, NJ

Oracle Database Developer


  • Involved into updating procedures, functions, triggers, and packages based on the change request
  • Involved in moving the data from flat files to staging area tables using SQL * Loader
  • Implemented Complex business Logic as per the client requirement
  • Applied constraints and writing triggers to apply the business rules
  • Involved in handling errors and transaction processing in the procedures
  • Created database objects like tables, views, synonyms, materialized views, procedures, packages using oracle tools like SQL* plus
  • Extensively used for all and bulk collect to fetch large volumes of data from table
  • Worked closely with Production Support team to Troubleshoot and fix any production issues
  • Reviewed data transformation rules and provide technical suggestions in data transformation logic and pseudo code (PL/SQL) program asynchronously
  • Worked with different types of sub queries and used loops to suit the business logic
  • Worked with performance tuning tools like Oracle optimizer to improve databases performance
  • Developed and maintained ETL process for Extraction from different data source, Transformation of data to meet data warehouse requirements, and Loading of the data
  • Developed batch programs using UNIX shell scripts and created partitions on the tables to improve the performance
  • Developed SQL* loader scripts and PL /SQL code to load data for updating payment terms
  • Created UNIX Shell Scripts for automating the execution process
  • Created indexes on tables and Optimizing Stored Procedure queries

Environment: Oracle 9, SQL*Plus, SQL*Loader, PL/SQL, UNIX, Data Warehouse, Shell script

Confidential, Thousand Oaks, CA

Oracle PL/SQL Developer


  • Designed and developed Oracle objects such as Tables, Views, Indexes, Stored Procedures and Functions in PL/SQL, Packages in PL/SQL, Triggers, Materialized Views, and Dynamic SQL
  • Extensively wrote and developed PL/SQL stored procedures, functions, triggers and packages to meet the business requirements and update the existing objects
  • Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Oracle database
  • Created and maintained database objects like tables, views, materialized views, synonyms, procedures, functions and database triggers to meet the business requirements
  • Loaded 30 tables and checked those data using a validation checking function
  • Used External Tables to query the data from Oracle tables unit testing on reports and queries
  • Used standard Oracle Interfaces for designing and developing conversion programs pertaining to Employee Hire and Terminations, Assignments, Elements and Salaries
  • Worked with the DBA to analyze and improve Java code tuning, Performance tuning, code modification, table space creation etc.
  • Designed, Developed and tested various reports
  • Provided additional technical support, performed test setups and troubleshooting for applications

Environment: Oracle 9i, Oracle Reports 6i, PL/SQL, UNIX, Webservices, Java, Spring MVC


Oracle PL/SQL Developer


  • Developed an online bank system using Java Eclipse and JSP as backend and HTML, CSS, JavaScript and JQuery as frontend.
  • Stored user and passwords, pin numbers and identification registries into Oracle database
  • Developed codes using common concepts of inheritance, collections and polymorphism
  • REST and SOAPWeb services with deployment on Servlet, Spring MVC and Struts
  • Involved in creation of database, modified tables, triggers, and procedures to clean and validate the data entries.

Environment: Java 6, Spring 3.2, 5, WebSphere, JSP, HTML, CSS, Apache Tomcat, Spring batch, JavaScript, Angular JS, JQuery Spring Web services,DB2,J2EE,XML,JSON

Hire Now