We provide IT Staff Augmentation Services!

Hadoop Developer Resume

4.00/5 (Submit Your Rating)

Houston, TX

SUMMARY

  • Technically accomplished professional with 7 years of total experience in Hadoop (2 years) and Java (5 years) technologies in IT industry.
  • 2 years of experience in Hadoop Development/Administration built on 5 years of experience in Java Application Development.
  • Good understanding of all stages of Software Development Life Cycle (SDLC) and support of systems application architecture.
  • In - depth knowledge of Hadoop architecture and various components such as HDFS, JobTracker, TaskTracker, NameNode, and DataNode.
  • Sound knowledge on Hadoop MRV1 and Hadoop MRV2 (or) YARN Architecture.
  • Developed MapReduce Applications to process large structured and unstructured data.
  • Working experience in MapReduce programming model and Hadoop Distributed File Systems.
  • Proven knowledge of writing Hive Queries to generate reports using Hive Query Language.
  • Working experience in writing Sqoop queries for transferring bulk data between Apache Hadoop and structured data stores.
  • Working knowledge on Oozie, a workflow scheduler system to manage apache Hadoop jobs by Directed Acyclic graph (DAG) of actions with control flows.
  • Working experience in Pig Latin, a Scripting Language for Hadoop Distributed File system.
  • Good Knowledge in Flume, Avro and Zoo Keeper Architecture.
  • Working knowledge in Hadoop HDFS Admin Shell commands.
  • Good working knowledge on NoSQL databases including HBase, Cassandra, and MongoDB.
  • Good knowledge on Hadoop MRV1 and Hadoop MRV2 (or) YARN Architecture.
  • Extending Hive and Pig core functionality by using custom User Defined Function's (UDF), User Defined Table-Generating Functions (UDTF) and User Defined Aggregating Functions (UDAF) for Hive and Pig.
  • Hands on experience in installing, configuring Cloudera Apache Hadoop ecosystem components like Flume, HBase, Zoo Keeper, Oozie, Hive, Sqoop and Pig.
  • Good understanding of Avro and JSON.
  • Hands on development experience with RDBMS, including writing complex SQL queries, PLSQL, views, stored procedure, triggers, etc.
  • Experience in Object Oriented Analysis and Design (OOAD) using UML (Use Cases, Activity, Sequence, Class Diagrams etc.).
  • Expertise in design and development of various web and enterprise applications using Hibernate, JSP, JavaScript, Servlets, JDBC, Web Services, JAX-WS, Axis and JMS
  • Basic Knowledge of UNIX and shell scripting.
  • Working experience on Struts, Spring, and Hibernate.
  • Experience in deploying enterprise applications in Apache-Tomcat, Web Logic, and Web Sphere.
  • Experience in machine learning and data mining.
  • Sound knowledge of algorithms.
  • Sound knowledge in Security concept (SSL, Kerberos autantication).
  • Experience in Business Intelligence and Analytics.
  • Good experience to set up standards and processes for Hadoop based application design and implementation
  • Good Knowledge of using IDE Tools like Eclipse and NetBeans.
  • Extensive knowledge of Front End technologies like HTML, CSS, Java Script.
  • Experience in developing Object Relational Mapping using Hibernate.
  • Extensive experience using Apache Jakarta Struts 1.x and Casper Framework Java Server Faces Framework implementing MVC framework in J2EE architecture.
  • Expertise teh logging teh application using LOG4J.
  • Involved in log file management where teh logs greater than 7 days old were removed from log folder and loaded into HDFS and stored for 3 months.
  • Expertise in debugging teh applications and Unit Testing teh application using JUnit.
  • Experience using Ant, Maven for building and deploying teh working enterprise application.
  • Expertise in working with different databases like Oracle, MS-SQL Server, MySQL, Teradata, and Netezza along with exposure to Hibernate for mapping an object-oriented domain model to a traditional relational database.
  • Expertise in development support activities including installation, configuration and successful deployment of changes across all environments.
  • Involved in Scrum methodologies, Test Driven Development and continuous integration.
  • Experience in production support and application support by fixing bugs.
  • Good communication skills, work ethics and teh ability to work in a team efficiently with good leadership skills.
  • Ability to perform at a high level, meet deadlines, adaptable to ever changing priorities.

TECHNICAL SKILLS

Big Data: Hadoop, MapReduce, HDFS, Hive, Pig, Oozie, Flume, Zookeeper, Avro, HBase, Cassandra, MongoDB, CDH4, Cloudera Manager, MapR

Languages: Java, C, C++, PHP, Python, HTML, XML, SQL

Platforms: Linux (CentOS, RedHat, Ubuntu), Windows, Mac

Databases: Oracle, Netezza, MySQL, Teradata, MS SQL Server

BI Tools: Tableau, Crystal Reports, MicroStrategy, SSIS, SSAS, SSRS

ETL: Informatica, Vertica, Pentaho

Statistical Analysis: R, Stata, SPSS, SAS

PROFESSIONAL EXPERIENCE

Confidential, Houston, TX

Hadoop Developer

Responsibilities:

  • Transferring data between Oracle, Teradata, Netezza and HDFS using Sqoop with connectors.
  • Moving log data periodically into HDFS using Flume. Building multi - hop flows, fan-out flows, and failover mechanism.
  • Developed MapReduce Applications for log processing and analytical batch processing of large structured and unstructured data.
  • Optimizing MapReduce jobs using Combiners and Partitioners to enhance performance.
  • Processing large amount of GPS messages (xml format) using Avro. Defining and compiling schemas, serializing and deserializing data.
  • Developed Pig Latin scripts for data cleansing and preparing.
  • Creating and populating Hive tables and writing Hive queries for data analysis to meet teh business requirements.
  • Migrating data from Oracle database to HBase. Running MapReduce jobs to access HBase data from application using Java Client APIs.
  • Exported teh analyzed data to Oracle using Sqoop for visualization and reporting for teh BI team.
  • Data Warehousing and Decision Support System (Business Intelligence) implemented DR strategies using Hadoop Replication.
  • Automating teh jobs using Oozie. Defining Oozie workflow jobs to chain together Sqoop imports, MapReduce jobs, and Pig scripts (multiple decision, fork and join nodes), and defining Oozie coordinator jobs to execute reoccuring workflow jobs.
  • Running unit tests using JUnit with MRUnit Library and writing MRUnit test classes for MapReduce jobs
  • Used SVN for version control.
  • Involved in solution-driven Agile development methodology and actively participated in daily scrum meetings.
  • Defined custom Counters to count invalid records for problem diagnosis.

Environment: Hadoop, HDFS, MapReduce, Sqoop, Hive, Pig, Oozie, Flume, Avro, HBase, Cassandra 2.0, AWS, SVN, CDH4, Cloudera Manager, MRUnit, Oracle, Teradata, Netezza, Eclipse, Application Lifecycle Management (ALM), CentOS

Confidential, Peoria, IL

Hadoop/Big Data Consultant

Responsibilities:

  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase NoSQL database and Sqoop.
  • Importing and exporting data in HDFS and Hive using Sqoop.
  • Data read and write on teh grid using HCatalog.
  • Data read, write and update in Cassandra.
  • Created indexes in Cassandra to allow efficient querying by column values.
  • Sound knowledge of Cassandra s Architecture.
  • Performed database operations (create, update, find, delete records and indexing) in MongoDB using JavaScript.
  • Involved in creating Hive tables, loading with data and writing hive queries.
  • Designed and implemented log analysis for application tuning.
  • ETL data processing and migration with Informatica and Vertica
  • Analyzed business data to identify patterns and correlations among teh various data points.
  • Provides data analysis and data mining expertise to support various projects.
  • Cluster coordination services through Zookeeper. Configuring ZooKeeper ensemble and setting watches on znodes.
  • Actively participated in software development lifecycle (design, implement, deploy, test), including design and code reviews, test development, test automation.

Environment: Java 6, Eclipse, Hadoop, Hive, HBase, MongoDB, Linux, MapReduce, HDFS, HCatalog, Shell Scripting, MySQL, R, SPSS, Informatica, Vertica

Confidential, Chicago, IL

Java/J2EE Developer

Responsibilities:

  • Involved in Java, J2EE, struts, web services and Hibernate in a fast paced development environment.
  • Rich experiences of database design and hands on experience of large database systems: Oracle 8i and Oracle 9i.
  • Involved in design and implementation of web tier using Servlets and JSP.
  • Used Apache POI for Excel files reading.
  • Written curl scripts, background batch process and on demand process for indexing to Solr using SolrJ API.
  • Written build scripts with Ant for deploying war and ear applications.
  • Developed user and technical documentation.

Environment: Java, J2EE, JDBC, Struts, SQL language. Hibernate, Eclipse, Apache POI, Apache Solr, CSS

Confidential, San Diego, CA

Java/J2EE Developer

Responsibilities:

  • Developed teh user interface using JSP and Java Script to view all online trading transactions.
  • Wrote SQL for DAO access.
  • Coded Java Server Pages for teh Dynamic front end content dat use Servlets and EJBs.
  • Coded HTML pages using CSS for static content generation with JavaScript for validations.
  • Used JDBC API to connect to teh database and carry out database operations.
  • Used JSP and JSTL Tag Libraries for developing User Interface components.
  • Performing Code Reviews.
  • Performed unit testing, system testing and integration testing.
  • Followed agile methodology, interacted directly with teh client provide/take feedback on teh features, suggest/implement optimal solutions, and tailor application to customer needs.

Environment: s: Java, J2EE, Tomcat, Ant, Eclipse, JavaScripts, CSS, Servlets, JSP, XML, HTML, JDBC

Confidential

Java Developer

Responsibilities:

  • Involved in teh analysis, design, implementation, and testing of teh project.
  • Implemented teh presentation layer with HTML, XHTML and JavaScript.
  • Developed web components using JSP, Servlets and JDBC.
  • Designed and developed Data Access Objects (DAO) to access teh database.
  • Used DAO Factory and value object design patterns to organize and integrate teh JAVA Objects
  • Implemented database using SQL Server.
  • Involved in fixing bugs and unit testing with test cases using JUnit.
  • Involved in building and deployment of application in Linux environment.
  • Deploying application in Development and Production servers.

Environment: s: Java/J2EE, JSP, Servlets, Websphere 6.x, JDBC, JavaScript, SQL Server, JUnit, Eclipse IDE

We'd love your feedback!