We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

Dallas, TX

SUMMARY

  • Over 8+ years of overall IT experience in a variety of industries, which includes hands on experience of 3+years in Big Data technologies.
  • In depth understanding/knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node, and MapReduce concepts and experience in working with MapReduce programs using Apache Hadoop for working with Big Data to analyze large data sets efficiently.
  • Hands on experience in working with Ecosystems like Hive, Pig, Sqoop, Map Reduce, Flume, Oozie.Strong knowledge of Pig and Hive’s analytical functions, extending Hive and Pig core functionality by writing custom UDFs.
  • Experience in importing and exporting terra bytes of data using Sqoop from HDFS to Relational Database Systems and vice - versa.
  • Knowledge of job workflow scheduling and monitoring tools like oozie and Zookeeper, of NoSQL databases such as HBase, Cassandra, and of administrative tasks such as installing Hadoop, Commissioning and decommissioning, and its ecosystem components such as Flume, Oozie, Hive and Pig.
  • Involved in designing and developing non-trivial ETL processes within Hadoop using tools like Pig, Sqoop, Flume, and Oozie
  • Experience in design, development and testing of Distributed, Internet/Intranet/E-Commerce, Client/Server and Database applications mainly using technologies Java, EJB, Servlets, JDBC, JSP, Struts, Hibernate, Spring, JavaScript on WebLogic, Apache Tomcat Web/Application Servers and with Oracle and SQL Server Databases on Unix, windows NT platforms.
  • Extensive work experience in Object Oriented Analysis and Design, Java/J2EE technologies including HTML, XHTML, DHTML, JavaScript, JSTL, CSS, AJAX and Oracle for developing server side applications and user interfaces.
  • Working knowledge of Software Design Patterns, Big Data Technologies (Hadoop, Hortonworks Sandbox) and Cloud Technologies & design.
  • Experience in developing Middle-tier components in distributed transaction management system using Java. Good understanding of XML methodologies (XML,XSL,XSD) including Web Services and SOAP.
  • Extensive experience in working with different databases such as Oracle, IBM DB, RDBMS, SQL Server, MySQL and writing Stored Procedures, Functions, Joins and Triggers for different Data Models.
  • Handled several techno-functional responsibilities including estimates, identifying functional and technical gaps, requirements gathering, designing solutions, development, developing documentation, and production support.
  • An individual with excellent interpersonal and communication skills, strong business acumen, creative problem solving skills, technical competency, team-player spirit, and leadership skills.

TECHNICAL SKILLS

Database: DB2, MySQL, Oracle, MS SQL Server, IMS/DB

Languages: Core Java, PIG Latin, SQL, HiveQL, Shell Scripting and XML

API’s/Tools: Mahout, Eclipse, Log4j, SVN, Maven

Web Technologies: HTML, XML, JavaScript

BigData Ecosystem: HDFS, PIG, MAPREDUCE, HIVE, SQOOP, FLUME, HBase

Operating System: Unix, Linux, Windows XP, IBM Z/OS

BI Tools: Tableau, Talend

PROFESSIONAL EXPERIENCE

Confidential, Dallas TX

Hadoop Developer

Responsibilities:

  • Developed simple and complex MapReduce programs in Java for Data Analysis on different data formats
  • Implemented Daily Cron jobs that automate parallel tasks of loading the data into HDFS and pre-processing with Pig using Oozie co-ordinator jobs.
  • Successfully migrated Legacy application to Big Data application using Hive/Pig/HBase in Production level
  • Transformed date related data into application compatible format by developing apache Pig UDFs
  • Optimized MapReduce jobs to use HDFS efficiently by using various compression mechanisms
  • Creating Hive tables, loading with data and writing Hive queries which will run internally in MapReduce way
  • Worked with NoSQL databases like Hbase in creating Hbase tables to load large sets of semi structured data coming from various sources.
  • Responsible for performing extensive data validation using Hive
  • Used Oozie workflow engine to run multiple Hive and Pig jobs
  • Involved in installing and configuring Hive, Pig, Sqoop, Flume and Oozie on the Hadoop cluster.
  • Experienced in moving data from Hive tables into Cassandra for real time analytics on hive tables.
  • Experience in Upgrading Hadoop cluster Hbase/Zookeeper from CDH3 to CDH4.
  • Involved in designing and developing non-trivial ETL processes within Hadoop using tools like Pig, Sqoop, Flume, and Oozie
  • Used DML statements to perform different operations on Hive Tables
  • Developed Hive queries for creating foundation tables from stage data
  • Developed the Apache Storm, Kafka, and HDFS integration project to do a real time data analyses.
  • Used Pig as ETL tool to do transformations, event joins, filter and some pre-aggregations
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior
  • Implemented business logic by writing Pig UDFs in Java and used various UDFs from Piggybanks and other sources. Involved in joining and data aggregation using Apache Crunch
  • Worked with Sqoop to export analyzed data from HDFS environment into RDBMS for report generation and visualization purpose
  • Involved in writing, testing, and running MapReduce pipelines using Apache Crunch
  • Queried and analyzed data from Datastax Cassandra for quick searching, sorting and grouping
  • Involved in loading data from UNIX file system to HDFS.

Environment: Apache Hadoop, HBase, Hive, Pig, Sqoop, ZooKeeper, NoSQL, HBase, Storm, Map-Reduce, Cloudera, HDFS, Flume, MySQL, JDK1.6, J2ee, JDBC, Servlets, JSP, Struts 2.0, Spring 2.0, Hibernate, WebLogic, SOAP, MongoDB.

Confidential, Peoria IL

Hadoop Developer

Responsibilities:

  • Involved in Requirement gathering, Business Analysis and translated business requirements into Technical design in Hadoop and Big Data
  • Importing and exporting data into HDFS from database and vice versa using Sqoop.
  • Written Hive jobs to parse the logs and structure them in tabular format to facilitate effective querying on the log data
  • Written Map Reduce code to process and parsing the data from various sources and storing parsed data into HBase and Hive using HBase-Hive Integration.
  • Involved in creating hive tables, loading with data and writing hive queries that will run internally in map reduce way.
  • Involved in creating workflow to run multiple hive and Pig Jobs, which run independently with time and data availability
  • Responsible for importing log files from various sources into HDFS using Flume.
  • Involved in developing shell scripts and automated data management from end to end integration work
  • Used Pig as a ETL tool to do Transformations, even joins and some pre-aggregations before storing data into HDFS
  • Developed Map Reduce program for parsing and loading into HDFS information.
  • Built reusable Hive UDF libraries for business requirements which enabled users to use these UDF's in Hive Querying.
  • Configured and optimized the Cassandra cluster and developed real-time java based application to work along with the Cassandra database.
  • Automating and scheduling the Sqoop jobs in a timely manner using Unix Shell Scripts.
  • Used Oozie workflow engine to manage interdependent Hadoop jobs and to automate several types of Hadoop jobs such as Java map-reduce Hive, Pig, and Sqoop.
  • Experienced in using Zookeeper and OOZIE Operational Services for coordinating the cluster and scheduling workflows.
  • Using Hbase to store majority of data which needs to be divided based on region.

Environment: Hadoop, HDFS, HBase, MapReduce, Hortonworks, Storm, Zookeeper, AVRO Files, Hive, Pig, Sqoop, Flume, Kafka, Cassandra, Oozie, Hue, SQL, ETL, Cloudera Manager, MySQL, MongoDB, Java, JDK 1.5, J2EE 1.4, Struts 1.3.

Confidential, Chicago IL

Java/J2EE Developer

Responsibilities:

  • Created design documents and reviewed with team in addition to assisting the business analyst / project manager in explanations to line of business.
  • Responsible for understanding the scope of the project and requirement gathering.
  • Involved in analysis, design, construction and testing of the application
  • Developed the web tier using JSP to show account details and summary.
  • Designed and developed the UI using JSP, HTML, CSS and JavaScript.
  • Utilized JPA for Object/Relational Mapping purposes for transparent persistence onto the SQL Server database.
  • Used Tomcat web server for development purpose.
  • Involved in creation of Test Cases for JUnit Testing.
  • Used Oracle as Database and used Toad for queries execution and also involved in writing SQL scripts, PL/SQL code for procedures and functions.
  • Used CVS for version controlling.
  • Developed application using Eclipse and used build and deploy tool as Maven.
  • Used Log4J to print the logging, debugging, warning, info on the server console.

Environment: Java, J2EE Servlet, JSP, JUnit, AJAX, XML, JavaScript, Log4j, CVS, Maven, Eclipse, Apache Tomcat, and Oracle.

Confidential

Java Developer

Responsibilities:

  • Developed web components using JSP, Servlets and JDBC
  • Designed tables and indexes
  • Designed, Implemented, Tested and Deployed Enterprise Java Beans both Session and Entity using WebLogic as Application Server
  • Developed stored procedures, packages and database triggers to enforce data integrity. Performed data analysis and created crystal reports for user requirements
  • Provided quick turn around and resolving issues within the SLA.
  • Implemented the presentation layer with HTML, XHTML and JavaScript
  • Used EJBs to develop business logic and coded reusable components in Java Beans
  • Development of database interaction code to JDBC API making extensive use of SQL
  • Query Statements and advanced Prepared Statements.
  • Used connection pooling for best optimization using JDBC interface
  • Used EJB entity and session beans to implement business logic and session handling and transactions. Developed user-interface using JSP, Servlets, and JavaScript
  • Wrote complex SQL queries and stored procedures
  • Actively involved in the system testing
  • Prepared the Installation, Customer guide and Configuration document which were delivered to the customer along with the product

Environment: Windows NT 2000/2003, XP, and Windows 7/ 8 C, Java, UNIX, and SQL using TOAD, Microsoft Office Suit, Microsoft project

Confidential

Java Developer

Responsibilities:

  • Application was built on MVC architecture with JSP 1.2 acting as presentation layer, Servlets as controller and developed the application using Jakarta Struts 1.1 Framework: developed action classes, form beans and Used Struts Validation Framework for validating front end forms.
  • Extensively used XML Web Services for transferring/retrieving data between different providers.
  • Developed complete Business tier with Session beans and CMP Entity beans with EJB 2.0 standards using JMS Queue communication in authorization module.
  • Designed and implemented Business Delegate, Session Facade and DTO Design Patterns
  • Involved in implementing the DAO pattern
  • Used JAXB API to bind XML Schema to java classes
  • Used the report generation in the databases written in PL/SQL
  • Used Maven for building the enterprise application modules
  • Used Log4J to monitor the error logs
  • Used JUnit for unit testing
  • Used SVN for Version control
  • Deployed the applications on WebLogic Application Server.

Environment: Struts 1.1, EJB 2.0, Servlets 2.3, JSP 1.2, SQL, XML, XSLT, Web Services, JAXB, SOAP, WSDL, JMS1.1, JavaScript, TDD, JDBC, Oracle 9i, PL/SQL, Log4J, JUnit, WebLogic, Eclipse, Rational XDE, SVN, Linux

We'd love your feedback!