Senior Hadoop Developer Resume
Milwaukee, WI
SUMMARY:
- 7+ years of IT experience which includes 4 years of experience with Hadoop, HDFS, MapReduce and Hadoop Ecosystem (MapReduce, PIG, HIVE, Sqoop, Oozie, Flume)
- Good knowledge of Hadoop Architecture and various components such as HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts.
- Hands on experience in writing MapReduce joins like Map - side joins using Distributed Cache API.
- Experience in analyzing data using Hive QL, Pig Latin and custom MapReduce programs in Java.
- Strong experience working with different Hadoop distributions like Cloudera, Horton works, Map R and Apache distributions.
- Hands-on experience in scheduling jobs on Oozie.
- Experience in developing customized UDF's in java to extend Hive and Pig Latin functionality.
- Collecting and aggregating large amount of Log Data using Apache Flume and storing data in HDFS for further analysis.
- Experience in importing and exporting data using Sqoop from Relational Database Systems to HDFS and vice-versa .
- Highly Knowledgeable in streaming data from different data sources like Log files, JMS,applications sources into HDFS.
- Experience in Installation, Configuration, Testing, Backup, Recovery, Customizing and Maintenance of clusters using Apache Hadoop, Cloudera Hadoop.
- Java Programmer with 4 years of Extensive programming experience in developing web based applications and Client-Server technologies.
- Participated in the data migration activities in between Development and production environments.
- Expertise with using Oracle, MySQL,DB2 databases and writing highly complex SQL queries.
- Experienced in using Integrated Development environments like Eclipse, NetBeans, IntelliJ.
- Worked in complete Software Development Life Cycle (analysis, design, development, testing, implementation and support) using Agile Methodologies.
- Extensive knowledge skills and developed UI web-based applications using jQuery UI,jQuery CSS HTML, XHTML and Java Script.
- Excellent technical, communication, analytical and problem solving skills and ability to get on well with people including cross-cultural backgrounds and trouble-shooting capabilities.
TECHNICAL SKILLS:
Bigdata Ecosystem: Hadoop,Mapreduce,hive,Pig,Sqoop,Flume,oozie,Hbase
Operating Systems: Windows XP,Windows 7/8,Linux Distro(Ubuntu,Mint,Fedora)
Languages: Java,C
WebTechnologies: HTML,XML,JavaScript
ScriptingLanguage: Unix Shell Scripting
RDBMS DB: Oracle,MySQL,IBM-DB2
NOSQL DB: Hbase
Servers: Tomcat,JBoss
PROFESSIONAL EXPERIENCE:
Senior Hadoop Developer
Confidential, Milwaukee, WI
Responsibilities:
- Strong experience with big data processing using Hadoop technologies HDFS, MapReduce,Hive and Pig.
- Worked with application teams to install operating system,Hadoop updates, patches, upgrades as required.
- Strong experience in developing, debugging and tuning Map Reduce jobs in Hadoop environment.
- Created and maintained Technical documentation for launching Hadoop Clusters and for Hive queries and Pig Scripts.
- Experience in writing ETL jobs using PIG Latin and HIVE QL. Injected final Results into HP Vertica Database for reporting layers consumption BO and Tableau.
- Good troubleshooting skills on over all Hadoop stack components, ETL services and Hue, which provides GUI for developers/business users for day-to-day activities.
- Setup flume for different sources to bring the log messages from outside to Hadoop hdfs.
- Validated final data sets by comparing RDBMS source systems and writing SQL, Hive and Solr queries.
- Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitored job performance and capacity planning using Ambari.
- Create queues and allocated the clusters resources to provide the priority for jobs.
- Designed and created AVRO schemas and implemented AVRO serialization to serialize object data to and from HBase.
- Working experience on maintaining MySQL databases creation and setting up the users and maintain the backup of cluster metadata databases with cron jobs.
- Create Hive tables and analyzing large flat files in Hive using HiveQL.
- Building, testing, and running MapReduce pipelines using Apache Crunch.
- Developed multi-core CPU pipeline applications to analyze large data-sets.
- Created custom MapReduce programs using Hadoop for big data processing.
- Created monitoring scripts using shell scripting to check the data completenes in hive tables.
Environment: HDFS,Hive,Sqoop,ShellScripting,Crontab,Ubuntu 13.04,Spring MVC, J2EE, Java 6.0.
Senior Hadoop Developer
Confidential, Green Bay, WI
Responsibilities:
- Worked in Multi Clustered Hadoop Echo-System environment
- Analysing new requirements, preparation of feasibility study documents for major enhancements.
- Designed and Developed MapReduce jobs in Java, PIG Defined job flows, managing and reviewing Hadoop log files.
- Working on Big Data Integration and Analytics based on Hadoop, SOLR and web methods technologies.
- Load and transform large sets of unstructured data from UNIX system to HDFS.
- Developed Use cases and Technical prototyping for implementing PIG, HDP, HIVE and HBASE.
- Delivered Working Widget Software using EXTJS4, HTML5, RESTFUL Web services, JSON Store, Linux,HADOOP, ZOOKEEPER, NO SQL databases, JAVA, SPRING Security, JBOSS Application Server for Big Data analytics.
- Experienced in running Hadoop streaming jobs to process terabytes of CSV format.
- Supported Map Reduce Programs those are running on the cluster.
- Led the evaluation of Big Data software like Splunk, Hadoop for augmenting the warehouse, identified use cases and led Big Data Analytics solution development for Customer Insights and Customer Engagement teams.
- Developing data pipeline using Flume, Sqoop, Pig and Java map reduce to ingest customer behavioral data and financial histories into HDFS for analysis.
- Exporting data to Oracle DB using Sqoop.
- Written PIG Scripts to analyse Hadoop logs.
- Created tables, loading with data and writing HIVE queries which will run internally in map.
- Handled Service Request (SRs) and Queries from customers, MUR TAC team and ASR5K Line Managers, including Non-Business and Weekends/Holidays.
- Participated in Project Planning and prioritization between, new requirements, enhancements, Bugs raised from the field and Customer queries and Preparation of Release Scheduling.
- Performed Single Point of Technical Contact for different application teams and DEV, QA, Line Managers.
- Provided non-business and weekend/holidays technical support to various teams to support critical cases from the field
- Coordinating with business and technical manager to gather new requirements and converted them into Functional Specification documents.
Environment: Hadoop 1x, HDFS, Map Reduce, Hive 10.0, Pig, Sqoop, Hbase, Shell Scripting, Oozie, Oracle 10g,Ubuntu 13.04, Spring MVC, J2EE, Java 6.0, JDBC, Apache Tomcat.
Hadoop Developer
Confidential, Melville, NY
Responsibilities:
- Installed, configured and deployed a 50 node MapR Hadoop cluster for development, production.
- Worked with the Data Science team to gather requirements for various data mining projects.
- Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
- Worked on setting up high availability for major production cluster and designed automatic failover.
- Involved in running Hadoop jobs for processing millions of records of text data.
- Configured Hivemetastore, which stores the metadata for Hive tables and partitions in a relational database.
- Configured Flume for efficiently collecting, aggregating and moving largeamounts of log data.
- Worked on configuring security for hadoopcluster(Kerberos, Active Directory)
- Responsible to manage data coming from different sources
- Installed and configured Zookeeper for Hadoop cluster.
- Load and transform large sets of structured, and semi structured data.
- Implemented test scripts to support test driven development and continuous integration.
- Tuning MR Programs those are running on the Hadoop cluster.
- Involved in HDFS maintenance, Upgrading the cluster to latest versions of CDH
- Wrote Map Reduce job using Java API.
- Worked closely with business units to define development estimates according to Agile Methodology.
- Imported/exported data fromRDMS to HDFS using Sqoop.
- Wrote Hive queries for data analysis to meet the business requirements.
- Created Hive tables and working on them using Hive QL.
- Worked on Git hub repository, branching, merging, etc.
Environment: Hadoop,HDFS, MapReduce,Sqoop,Hive,Flume,Oozie,Zookeeper, Kerberos, Cloudera Manager, MySQL, Eclipse.
Hadoop Developer
Confidential, Medford, MA
Responsibilities:
- Involved in the Complete Software development life cycle (SDLC) to develop the application.
- Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, Hbase database and Sqoop.
- Experience in managing and reviewing Hadoop log files.
- Involved in loading data from LINUX file system to HDFS.
- Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
- Identified trends, opportunities, and risks to current forecasts and the next period's plan.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Supported in setting up QA environment and updating configurations for implementing scripts with Pig and Sqoop.
- Installed and configured HadoopMapReduce, HDFS, Developed multiple MapReduce jobs in java for data cleaning and preprocessing
- Created Pig Latin scripts to sort, group, join and filter the enterprise wise data.
- Involved in creating Hive tables, loading with data and writing hive queries that will run internally in mapreduce way.
- Supported MapReduce Programs those are running on the cluster.
- Installed Oozie workflow engine to run multiple Mapreduce jobs.
- Worked with application teams to install operating system, Hadoop updates, patches,version upgrades as required.
Environment: Windows & Linux, Java 6, Eclipse, Hadoop CDH 2, MapReduce, HDFS, Hive, Pig, Shell Scripting, Sqoop, Oracle 10g, Javascripts, Servlets.
Java Developer
Confidential
Responsibilities:
- Responsible and active in the analysis, design, implementation and deployment of full Software Development Lifecycle (SDLC) of the project.
- Designed and developed user interface using JSP, HTML and JavaScript.
- Defined the search criteria and pulled out the record of the customer from the database. Make the required changes and save the updated record back to the database.
- Validated the fields of user registration screen and login screen by writing JavaScript validations.
- Used DAO and JDBC for database access.
- Developed stored procedures and triggers using PL/SQL in order to calculate and update the tables to implement business logic.
- Design and develop XML processing components for dynamic menus on the application.
- Involved in postproduction support and maintenance of the application.
- Involved in the analysis, design, implementation, and testing of the project.
- Implemented the presentation layer with HTML, XHTML and JavaScript.
- Developed web components using JSPand JDBC.
- Implemented database using SQL Server.
- Designed tables and indexes.
- Wrote complex SQL and stored procedures.
- Involved in fixing bugs and unit testing with test cases using JUnit.
- Developed user and technical documentation.
Environment: Java 1.5,Oracle 11g,HTML, XML, SQL, J2EE, JUnit, Tomcat 6.
Java Developer
Confidential
Responsibilities:
- Analyzing and preparing the requirement Analysis Document.
- Deploying the Application to the JBOSS Application Server.
- Requirement gatherings from various parties involved in the project
- Estimate timelines for development tasks.
- Used to J2EE and EJB to handle the business flow and Functionality.
- Interact with Client to get the confirmation on the functionalities.
- Involved in the complete SDLC of the Development with full system dependency.
- Actively coordinated with deployment manager for application production launch.
- Provide Support and update for the period under warranty.
- Produces detailed low-level designs from high level design
- Specifications for components of low level complexity.
- Develops, builds and unit tests components of low level
- Complexity from detailed low-level designs.
- Developed user and technical documentation.
- Monitoring of test cases to verify actual results against expected results.
- Performed Functional, User Interface test and Regression Test
- Carrying out Regression testing to track the problem tracking.
- Develops, builds and unit tests components of low level
- Complexity from detailed low-level designs.
- Developed user and technical documentation.
- Implemented Model View Controller (MVC) architecture at the Web tier level to isolate each layer of the application to avoid the complexity of integration and ease of maintenance along with Validation Framework
Environment: Java, EJB, UNIX, XML, Work Flow, JMS, JIRA, Oracle, JBOSS.