We provide IT Staff Augmentation Services!

Hadoop Developer Resume

0/5 (Submit Your Rating)

Miami, FL

SUMMARY

  • Dedicated and versatile IT Professional with over 6+ years of experience designing, creating, testing software that implements sound technical and business solutions.
  • Worked with all levels of management and peers with the ability to adapt to change and effectively do my part to get the job done right.
  • Experience of 3+ years in Big Data and Hadoop Ecosystem.
  • Worked extensively on Hadoop platform using Pig, Hive, and Java to build highly scalable data processing pipelines
  • Strong Object - Oriented Programming and Design skills.
  • Strong knowledge in using MapReduce programming model for analyzing the data stored in Hadoop.
  • Extensive experience in Data migration from existing data stores to Hadoop.
  • Experience in developing custom applications to run Hadoop Jobs on the cluster usingMap Reduce, Hive, Pig, Sqoop, Flume, Zookeeper,Oozie, NoSQL,Cassandra, HBase for structured, semi structured and unstructured data.
  • Proficient in Hadoop, MapReduce, Hive, Pig, Sqoop, Flume, Zookeeper,Oozie, NoSQL,Cassandra, HBase, Java, JDBC, JDK 1.5, JSP 2.0, Servlets, Struts 1.2, Hibernate 3.2, Spring, JSP, Java Script,Shell Script,
  • Experience in writing (UDF's) for Hiveand Pig. Analyzing data with Hive and Pig.
  • Worked on Oozie to manage data processing jobs for Hadoop.
  • Working knowledge of database such as Oracle 8i/9i/10g, Microsoft SQL Server,MySQL,DB2
  • Strong experience in database design, writing complex SQLQueries and Stored Procedures
  • Experience in using Zookeeper for coordinating the distributed applications.
  • Experience in developing Map-Reduce programs and custom UDF’s for data processing using Java
  • Experience in importing and exporting the data from different relational databases using Sqoop to HDFS and collecting web logs using flume to HDFS.
  • Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS.
  • Experience in IBM specific Methodologies, SDLC by using Waterfall, Agile(SCRUM) and Iterative system development methodologies.
  • Experience in developing technical specifications based on business requirements, developing custom reports for analytics customer expectation management.
  • Good experience in design and execution of test scenarios, test cases and experience in Functionality Testing, Black-Box Testing and Regression Testing.
  • Versatile team member with problem-solving and trouble-shooting capabilities.

TECHNICAL SKILLS

Ecosystems: MapReduce(0.2 YARN), HDFS, Hive, Pig, Oozie, Flume, Zookeeper, HBase, Sqoop

Platforms/Operating Systems: Linux (CentOS, RedHat, Ubuntu), Windows, Mac

Programming Languages: C, C++, Java, Html, PIG LATIN, Shell Scripting

DB and Client/Server Technologies: Oracle, MySQL, MS SQL Server

Framework/ Architecture: Tools/Utilities/ IDE

Struts, Spring, Hibernate: Eclipse, Net beans, MS office, MS access, MS Visio

Application Servers: Tomcat 6.x, Oracle Application Server 10g

PROFESSIONAL EXPERIENCE

Confidential, Miami, FL

Hadoop Developer

Responsibilities:

  • Designed complete end-to-end ETL Infrastructure and created workflows using Oozie framework.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
  • Configured Hive Metastore to use MySQL database to establish multiple user connections to hive tables.
  • Configured the Hadoop MapReduce and HDFS core properties as a part of performance tuning to achieve high computational performance.
  • Performed analysis on the data stored in HDFS using HIVE. This analysis mainly targeted the study of user behavior and patterns.
  • Worked with external tables, Partitions, Joins, Views in Hive.
  • Developed simple to complex MapReduce Jobs using Hive.
  • Developed java UDF’s for hashing data, data encryption and decryption.
  • Used Struts Framework integrated with spring and Hibernate.
  • Automated workflows using shell scripts and Oozie jobs to pull data from various databases into Hadoop
  • Expertise on HIVE performance tuning and setting the appropriate properties.
  • Worked on custom Pig Loaders and Storage classes to work with a variety of data formats such as JSON, Compressed CSV, etc.
  • Expertise in loading the data using pig.

Environment: Hadoop, Hive, Flume, MapReduce, Sqoop, Shell Script, Java, JDK 1.5, JSP 2.0, JSF, Servlets, Struts 1.2, Hibernate 3.2, Spring, Oozie, Zookeeper.

Confidential, Bloomington, IL

Hadoop Developer

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop
  • Developed Simple to complex Map/reduce Jobs using Hive and Pig
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from MySQL into HDFS using Sqoop
  • Used UDF’s to implement business logic in Hadoop
  • Implemented business logic by writing UDFs in Java and used various UDFs from Piggybanks and other sources.
  • Continuous monitoring and managing the Hadoop cluster using Cloudera Manager
  • Worked with application teams to install operating system, Hadoop updates, patches, version upgrades as required
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team

Environment: Hadoop, MapReduce, HDFS, Hive, Pig, Java, SQL, Cloudera Manager, Sqoop, Flume, Oozie, Java (jdk 1.6), Eclipse

Confidential, Glen Allen, VA

Hadoop Engineer

Responsibilities:

  • Build a production, development and testing cluster.
  • Configure the cluster properties to gain the high cluster performance by taking cluster hardware configuration as key criteria.
  • Designed the rack topology script for the production Hadoop cluster.
  • Ran the cluster Benchmarking tool on the all Hadoop cluster to check the performance by changing the size of the data for the benchmarking.
  • Wrote the Map-Reduce Programs to check the cluster performance on the Benchmark data.
  • Implemented the resource-monitoring tool Ganglia on the Hadoop clusters.
  • Deployed the Hadoop cluster using Kerberos to provide secure access to the cluster.
  • Dump the data from MYSQL to HDFS using SQOOP.
  • Loaded the HDFS data into the hive warehouse and generated the data cubes using HIVE and JAVA Map-Reduce as per the business requirement.
  • Deployed Hadoop cluster on Rack Space using apache Whirr.

Environment: Hadoop, JAVA, Hadoop Eco-System, Shell Scripting

Confidential, Bloomington, IL

Hadoop Engineer

Responsibilities:

  • Proactively monitored systems and services, architecture design and implementation of Hadoop deployment, configuration management, backup, and disaster recovery systems and procedures.
  • Involved in analyzing system failures, identifying root causes, and recommended course of actions.
  • Worked with systems engineering team to plan and deploy new Hadoop environments and expand existing Hadoop clusters.
  • Monitored multiple Hadoop clusters environments using Ganglia and Nagios. Monitored workload, job performance and capacity planning using Cloudera Manager.
  • Installed and configured Flume, Hive, Pig, Sqoop and Oozie on the Hadoop cluster.
  • Used Flume to collect, aggregate, and store the web log data from different sources like web servers, mobile and network devices and pushed to HDFS.
  • Analyzed the web log data using the Hiveql to extract number of unique visitors per day, page- views, visit duration, most purchased product on website.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Integrated Oozie with the rest of the Hadoop stack supporting several types of Hadoop jobs out of the box (like Java MapReduce, Pig, Hive, Sqoop) as well as system specific jobs (such as Java programs and shell scripts).
  • Involved in installing and configuring Kerberos for the authentication of users and Hadoop daemons.

Environment: CDH4, Flume, Hive, Sqoop, Pig, Oozie, Cloudera Manager, Java, Linux, CentOS

Confidential

Java/J2EE Developer

Responsibilities:

  • Developed Admission & Census module, which monitors a wide range of detailed information for each resident upon pre-admission or admission to your facility.
  • Involved in comprehensive library of problems, goals and approaches.
  • You have the option of tailoring (adding, deleting, or editing problems, goals and approaches) these libraries and the disciplines you will use for your care plans.
  • Involved in development of General Ledgermodule, which streamlines analysis, reporting and recording of accounting information.
  • General Ledger automatically integrates with a powerful spreadsheet solution for budgeting, comparative analysis and tracking facility information for flexible reporting.
  • Developed UI using HTML, JavaScript, and JSP, and developed Business Logic and Interfacing components using Business Objects, XML, and JDBC.
  • Designed user-interface and checking validations using JavaScript.
  • Managed connectivity using JDBC for querying/inserting & data management including triggers and stored procedures.
  • Developed various EJBs for handling business logic and data manipulations from database.
  • Involved in design of JSP’s and Servlets for navigation among the modules.
  • Designed cascading style sheets and XML part of Order entry Module & Product Search Module and did client side validations with java script.

Environment: J2EE, Java/JDK, JDBC, JSP, Servlets, JavaScript, EJB, JNDI, JavaBeans, XML, XSLT, Oracle 9i, Eclipse, HTML/ DHTML, SVN.

We'd love your feedback!