Hadoop Developer Resume
Seattle, WA
SUMMARY
- Over 7+ years of total IT experience, with over 2+ years of experience in all phases of Hadoop and HDFS development along with 4+ years of experience in analysis, design, development, testing and deploying various software applications with emphasis on Object Oriented Programming.
- In depth understanding/knowledge of Hadoop and YARN Architecture.
- Experienced on major Hadoop ecosystem’s projects such as Pig, Hive, HBase, Impala and Zookeeper.
- Experience in developing and implementing Map Reduce jobs using java to process and perform various analytics on large datasets.
- Experience in migrating data to and from RDBMS and unstructured sources into HDFS using Sqoop & Flume.
- Experience in developing PigLatin and HiveQL scripts for Data Analysis and ETL purposes and also extended the default functionality by writing User Defined Functions (UDFs) for data specific processing.
- Good knowledge in job scheduling and monitoring through Oozie and ZooKeeper.
- Knowledge of NoSQL databases such as HBase and Cassandra.
- Generated dashboards & visualizations from various databases using Tableau.
- Expertise in writing Shell scripting in UNIX and integrating them with other solutions.
- Expertise in Object - oriented analysis and design (OOAD) like UML and use of various design patterns.
- Fluent with the core Java concepts like I/O, Multi-threading, Exceptions, RegEx, Collections, Data-structures and Serialization.
- Hands-on programming experience in developing enterprise solutions using JAVA, J2EE, JSP, Servlets, JDBC, XML, Struts, Spring, Hibernate, EJB, JSF, MVC, Web Services.
- Performed unit testing using JUnit Testing Framework and Log4J to monitor the error logs.
- Have dealt with end users in requirement gathering, user experience and issues.
- Followed Agile Methodology and SCRUM meetings to track, optimize and tailor features to customer needs.
- Expertise in developing Maven scripts to build and deploy the application.
- Hands-on experience in developing SQL queries, Stored Procedures, Triggers, Cursors, Indexes and Packages.
- Team player with excellent analytical, communication and project documentation skills.
- Have the motivation to take independent responsibility as well as ability to contribute and be a productive team player.
TECHNICAL SKILLS
Hadoop Ecosystem: HDFS, MapReduce, Sqoop, Flume, Hive, Pig, HBase, Impala, HUE, ZooKeeper, Oozie, Cloudera Manager.
Hadoop Distributions: Apache Hadoop, Cloudera, Hortonworks.
Databases: MySQL, Oracle, SQL Server, DB2.
Programming Languages: Java, C/C++, C#, UNIX Shell Scripting.
Web Technologies: J2EE, JSP, Servlets, HTML, XHTML, CSS, JavaScript, JQuery, AJAX.
Operating Systems: Windows XP/7, LINUX/UNIX.
Tools: Eclipse, Microsoft Office Suite,Tableau, Concurrent Versions System (CVS), ClearCase.
PROFESSIONAL EXPERIENCE
Confidential, Seattle, WA
Hadoop Developer
Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Impala, Cassandra, Cloudera Manager, Sqoop, Flume, Oozie,ZooKeeper, Java (jdk 1.6), Java, MySQL, Eclipse, Tableau.
Responsibilities:
- Responsible for building scalable distributed data solutions using Hadoop.
- Involved in gathering requirements and participating in the Agile planning meetings in-order to finalize the scope of each development.
- Developed simple to complex MapReduce programs to analyze the datasets as per the requirement.
- Optimized MapReduce Jobs to use HDFS efficiently by using various compression mechanisms.
- Configured periodic incremental imports of data from MySQL into HDFS using Sqoop.
- Responsible for migrating tables from traditional RDBMS into Hive tables using Sqoop and later generate required visualizations and dashboards using Tableau.
- Responsible for Load, aggregate and move large amounts of log data using Flume.
- Involved in loading data from UNIX file system to HDFS.
- Worked on loading and transformation of large datasets of structured, semi structured and unstructured data into Hadoop ecosystem.
- Responsible to manage data coming from different data sources.
- Analyzed the data by performing Hive queries and running Pig scripts to study customer behavior.
- Implemented business logic by writing custom UDF’s in Java and used various UDF’s from Piggybanks and other sources.
- Used Hive and created Hive tables and involved in data loading and writing custom HiveUDF’s.
- Created Partitions, Dynamic Partitions and Buckets for granularity and optimization using HiveQL.
- Involved in identifying job dependencies to design workflow for Oozie and resource management for YARN.
- Used Cassandra to store the analyzed and processed data for scalability.
- Responsible for maintaining and implementing code versions using CVS for the entire project.
- Coordinated with testing teams to resolve issues during QA testing.
Confidential, Bloomington, IL
Hadoop Developer
Environment: Java, Hadoop, Pig, Hive, Hue, Oozie, Sqoop, Flume, HBase, ZooKeeper, Oracle 10g, Eclipse.
Responsibilities:
- Gathered the business requirements from the Business Partners and Subject Matter Experts.
- Involved in writing MapReduce programs and tested using MRUnit.
- Managing and scheduling Jobs on a Hadoop cluster using Oozie.
- Responsible for designing and managing the Sqoop jobs that uploaded the data from Oracle to HDFS and Hive.
- Involved in moving all log files generated from various sources to HDFS for further processing through Flume.
- Worked on Hue interface for querying the data.
- Developed Pig scripts for data analysis and extended its functionality by developing custom UDF’s.
- Created Hive tables to store the processed results in a tabular format.
- Created HBase tables to store variable data formats.
- Utilized cluster co-ordination services through ZooKeeper.
- Implemented test scripts to support test driven development and continuous integration.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Analyzed large amounts of data sets to determine the optimal way to aggregate and report on it.
- Participate in requirement gathering and analysis phase of the project in documenting the business requirements by conducting workshops/meetings with various business users.
Confidential, Baltimore, MD
Java/J2EE Developer
Environment: Java, JDK 1.5, Servlets, Hibernate, AJAX, Oracle 10g, Eclipse, Web Services, Apache Axis, Web Logic Server, JavaScript, HTML, CSS, XML.
Responsibilities:
- Responsible for gathering and analyzing requirements and converting them into technical specifications.
- Used Rational Rose for creating sequence and class diagrams.
- Developed presentation layer using JSP, Java, HTML and JavaScript.
- Used Spring Core Annotations for Dependency Injection.
- Designed and developed a ‘Convention Based Coding’ utilizing Hibernates persistence framework and O-R mapping capability to enable dynamic fetching and displaying of various table data with JSF tag libraries.
- Designed and developed Hibernate configuration and session-per-request design pattern for making database connectivity and accessing the session for database transactions respectively.
- Used HQL and SQL for fetching and storing data in databases.
- Participated in the design and development of database schema and Entity-Relationship diagrams of the backend Oracle database tables for the application.
- Implemented Web Services with Apache Axis.
- Designed and developed Stored Procedures, Triggers in Oracle to cater the needs for the entire application. Developed complex SQL queries for extracting data from the database.
- Designed and built Web Service interfaces implemented in Java.
- Used Maven for the build process.
- Used ClearCase for version control and ClearQuest for bug tracking.
Confidential, Chicago, IL
Java/J2EE Developer
Environment: Java, J2EE, XML, XML Schemas, JSP, HTML, CSS, PL/SQL, JUnit, Log4j, IBM WebSphere Application Server.
Responsibilities:
- Involved in creation of UML diagrams like Class, Activity and Sequence Diagrams using modeling tools of IBM Rational Rose.
- Involved in the development of JSP’s and Servlets for different User Interfaces.
- Used Struts action forms and developed Action Classes, which act as the navigation controller in Struts framework.
- Implemented the template-based cauterization of presentation content using Struts Tiles. MVC implementation using Struts framework.
- Involved in Unit Testing of various modules based on the Test cases.
- Involved in Bug fixing of various modules that were raised by the Testing teams in the application during the Integration testing phase.
- Involved and participated in code reviews.
- Used Log4j logging framework for logging messages.
- Used Rational ClearCase for version control.
- Used Rational ClearQuest for bug tracking.
- Involved in deployment of application on IBM WebSphere Application Server.
Confidential
Jr. Java Developer
Environment: Java/J2EE, Eclipse, Web Logic Application Server, Oracle, JSP, HTML, Java Script, JMS, Servlets, UML, XML, Struts, Web Services, WSDL, UDDI.
Responsibilities:
- Responsible for understanding the business requirement.
- Worked with Business Analyst and helped representing the business domain details in technical specifications.
- Was also actively involved in setting coding standards and writing related documentation.
- Developed Java code using Eclipse as IDE.
- Developed JSP’s and Servlets to dynamically generate HTML and display the data to the client side.
- Developed application on Struts,MVC architecture utilizing Action Classes, Action Forms and validations.
- Tiles were used as an implementation of Composite view pattern.
- Was responsible in implementing various J2EE Design Patterns like Service Locator, Business Delegate, Session Façade and Factory Pattern.
- Code Review & Debugging using Eclipse Debugger.
- Was responsible for developing and deploying EJB’s (Session & MDB).
- Configured Queues in WebLogic server where the messages using JMS API were published.
- Consumed Web Services (WSDL, UDDI) from third party for authorizing payments to/from customers.
- Writing and manipulating the database queries.
- Build web application using Maven as build tool.
- Used CVS for version control.
- Performed unit testing using JUnit Testing Framework and Log4j to monitor the error log.
- Responsible for understanding the business requirements.