We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

Seattle, WA

SUMMARY

  • Over 7+ years of total IT experience, with over 2+ years of experience in all phases of Hadoop and HDFS development along with 4+ years of experience in analysis, design, development, testing and deploying various software applications with emphasis on Object Oriented Programming.
  • In depth understanding/knowledge of Hadoop and YARN Architecture.
  • Experienced on major Hadoop ecosystem’s projects such as Pig, Hive, HBase, Impala and Zookeeper.
  • Experience in developing and implementing Map Reduce jobs using java to process and perform various analytics on large datasets.
  • Experience in migrating data to and from RDBMS and unstructured sources into HDFS using Sqoop & Flume.
  • Experience in developing PigLatin and HiveQL scripts for Data Analysis and ETL purposes and also extended the default functionality by writing User Defined Functions (UDFs) for data specific processing.
  • Good knowledge in job scheduling and monitoring through Oozie and ZooKeeper.
  • Knowledge of NoSQL databases such as HBase and Cassandra.
  • Generated dashboards & visualizations from various databases using Tableau.
  • Expertise in writing Shell scripting in UNIX and integrating them with other solutions.
  • Expertise in Object - oriented analysis and design (OOAD) like UML and use of various design patterns.
  • Fluent with the core Java concepts like I/O, Multi-threading, Exceptions, RegEx, Collections, Data-structures and Serialization.
  • Hands-on programming experience in developing enterprise solutions using JAVA, J2EE, JSP, Servlets, JDBC, XML, Struts, Spring, Hibernate, EJB, JSF, MVC, Web Services.
  • Performed unit testing using JUnit Testing Framework and Log4J to monitor the error logs.
  • Have dealt with end users in requirement gathering, user experience and issues.
  • Followed Agile Methodology and SCRUM meetings to track, optimize and tailor features to customer needs.
  • Expertise in developing Maven scripts to build and deploy the application.
  • Hands-on experience in developing SQL queries, Stored Procedures, Triggers, Cursors, Indexes and Packages.
  • Team player with excellent analytical, communication and project documentation skills.
  • Have the motivation to take independent responsibility as well as ability to contribute and be a productive team player.

TECHNICAL SKILLS

Hadoop Ecosystem: HDFS, MapReduce, Sqoop, Flume, Hive, Pig, HBase, Impala, HUE, ZooKeeper, Oozie, Cloudera Manager.

Hadoop Distributions: Apache Hadoop, Cloudera, Hortonworks.

Databases: MySQL, Oracle, SQL Server, DB2.

Programming Languages: Java, C/C++, C#, UNIX Shell Scripting.

Web Technologies: J2EE, JSP, Servlets, HTML, XHTML, CSS, JavaScript, JQuery, AJAX.

Operating Systems: Windows XP/7, LINUX/UNIX.

Tools: Eclipse, Microsoft Office Suite,Tableau, Concurrent Versions System (CVS), ClearCase.

PROFESSIONAL EXPERIENCE

Confidential, Seattle, WA

Hadoop Developer

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Impala, Cassandra, Cloudera Manager, Sqoop, Flume, Oozie,ZooKeeper, Java (jdk 1.6), Java, MySQL, Eclipse, Tableau.

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop.
  • Involved in gathering requirements and participating in the Agile planning meetings in-order to finalize the scope of each development.
  • Developed simple to complex MapReduce programs to analyze the datasets as per the requirement.
  • Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms.
  • Configured periodic incremental imports of data from MySQL into HDFS using Sqoop.
  • Responsible for migrating tables from traditional RDBMS into Hive tables using Sqoop and later generate required visualizations and dashboards using Tableau.
  • Responsible for Load, aggregate and move large amounts of log data using Flume.
  • Involved in loading data from UNIX file system to HDFS.
  • Worked on loading and transformation of large datasets of structured, semi structured and unstructured data into Hadoop ecosystem.
  • Responsible to manage data coming from different data sources.
  • Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior.
  • Implemented business logic by writing custom UDF’s in Java and used various UDF’s from Piggybanks and other sources.
  • Used Hive and created Hive tables and involved in data loading and writing custom HiveUDF’s.
  • Created Partitions, Dynamic Partitions and Buckets for granularity and optimization using HiveQL.
  • Involved in identifying job dependencies to design workflow for Oozie and resource management for YARN.
  • Used Cassandra to store the analyzed and processed data for scalability.
  • Responsible for maintaining and implementing code versions using CVS for the entire project.
  • Coordinated with testing teams to resolve issues during QA testing.

Confidential, Bloomington, IL

Hadoop Developer

Environment: Java, Hadoop, Pig, Hive, Hue, Oozie, Sqoop, Flume, HBase, ZooKeeper, Oracle 10g, Eclipse.

Responsibilities:

  • Gathered the business requirements from the Business Partners and Subject Matter Experts.
  • Involved in writing MapReduce programs and tested using MRUnit.
  • Managing and scheduling Jobs on a Hadoop cluster using Oozie.
  • Responsible for designing and managing the Sqoop jobs that uploaded the data from Oracle to HDFS and Hive.
  • Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
  • Worked on Hue interface for querying the data.
  • Developed Pig scripts for data analysis and extended its functionality by developing custom UDF’s.
  • Created Hive tables to store the processed results in a tabular format.
  • Created HBase tables to store variable data formats.
  • Utilized cluster co-ordination services through ZooKeeper.
  • Implemented test scripts to support test driven development and continuous integration.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Analyzed large amounts of data sets to determine the optimal way to aggregate and report on it.
  • Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.

Confidential, Baltimore, MD

Java/J2EE Developer

Environment: Java, JDK 1.5, Servlets, Hibernate, AJAX, Oracle 10g, Eclipse, Web Services, Apache Axis, Web Logic Server, JavaScript, HTML, CSS, XML.

Responsibilities:

  • Responsible for gathering and analyzing requirements and converting them into technical specifications.
  • Used Rational Rose for creating sequence and class diagrams.
  • Developed presentation layer using JSP, Java, HTML and JavaScript.
  • Used Spring Core Annotations for Dependency Injection.
  • Designed and developed a ‘Convention Based Coding’ utilizing Hibernates persistence framework and O-R mapping capability to enable dynamic fetching and displaying of various table data with JSF tag libraries.
  • Designed and developed Hibernate configuration and session-per-request design pattern for making database connectivity and accessing the session for database transactions respectively.
  • Used HQL and SQL for fetching and storing data in databases.
  • Participated in the design and development of database schema and Entity-Relationship diagrams of the backend Oracle database tables for the application.
  • Implemented Web Services with Apache Axis.
  • Designed and developed Stored Procedures, Triggers in Oracle to cater the needs for the entire application. Developed complex SQL queries for extracting data from the database.
  • Designed and built Web Service interfaces implemented in Java.
  • Used Maven for the build process.
  • Used ClearCase for version control and ClearQuest for bug tracking.

Confidential, Chicago, IL

Java/J2EE Developer

Environment: Java, J2EE, XML, XML Schemas, JSP, HTML, CSS, PL/SQL, JUnit, Log4j, IBM WebSphere Application Server.

Responsibilities:

  • Involved in creation of UML diagrams like Class, Activity and Sequence Diagrams using modeling tools of IBM Rational Rose.
  • Involved in the development of JSP’s and Servlets for different User Interfaces.
  • Used Struts action forms and developed Action Classes, which act as the navigation controller in Struts framework.
  • Implemented the template-based cauterization of presentation content using Struts Tiles. MVC implementation using Struts framework.
  • Involved in Unit Testing of various modules based on the Test cases.
  • Involved in Bug fixing of various modules that were raised by the Testing teams in the application during the Integration testing phase.
  • Involved and participated in code reviews.
  • Used Log4j logging framework for logging messages.
  • Used Rational ClearCase for version control.
  • Used Rational ClearQuest for bug tracking.
  • Involved in deployment of application on IBM WebSphere Application Server.

Confidential

Jr. Java Developer

Environment: Java/J2EE, Eclipse, Web Logic Application Server, Oracle, JSP, HTML, Java Script, JMS, Servlets, UML, XML, Struts, Web Services, WSDL, UDDI.

Responsibilities:

  • Responsible for understanding the business requirement.
  • Worked with Business Analyst and helped representing the business domain details in technical specifications.
  • Was also actively involved in setting coding standards and writing related documentation.
  • Developed Java code using Eclipse as IDE.
  • Developed JSP’s and Servlets to dynamically generate HTML and display the data to the client side.
  • Developed application on Struts,MVC architecture utilizing Action Classes, Action Forms and validations.
  • Tiles were used as an implementation of Composite view pattern.
  • Was responsible in implementing various J2EE Design Patterns like Service Locator, Business Delegate, Session Façade and Factory Pattern.
  • Code Review & Debugging using Eclipse Debugger.
  • Was responsible for developing and deploying EJB’s (Session & MDB).
  • Configured Queues in WebLogic server where the messages using JMS API were published.
  • Consumed Web Services (WSDL, UDDI) from third party for authorizing payments to/from customers.
  • Writing and manipulating the database queries.
  • Build web application using Maven as build tool.
  • Used CVS for version control.
  • Performed unit testing using JUnit Testing Framework and Log4j to monitor the error log.
  • Responsible for understanding the business requirements.

We'd love your feedback!