We provide IT Staff Augmentation Services!

Big Data / Hadoop Developer Resume

4.00/5 (Submit Your Rating)

TX

SUMMARY:

  • Around 4 years of working experience as a Big Data and Hadoop Developer.
  • Thorough understanding of Software Development Life Cycle (SDLC), Software Test Life Cycle (STLC) and processes across multiple environments and platforms.
  • Excellent understanding of Hadoop architecture and various components such as HDFS, YARN, High Availability and MapReduce programming paradigm.
  • Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop, MapReduce, HDFS, HBase, Oozie, Hive, Spark, Storm, Sqoop and Flume.
  • Experience in design, development of web - based applications using HTML, DHTML, CSS, JavaScript, jQuery, JSP and Servlets.
  • Sound Relational Database concepts and extensively worked with ORACLE, DB2. Very good in writing complex SQL queries and PL/SQL procedures.
  • Experience in Performance Tuning, tuned various large critical OLAP and OLTP databases of Terabytes size using various tools like STATSPACK, AWR, SQL-TRACE, TKPROF, ADDM and Explain Plan Methods.
  • Experienced in implementing complex analytical algorithms using Map reduce design patterns.
  • Experience in analysed data using HiveQL, HBase and custom MapReduce programs in Python.
  • Extended Pig and Hive core functionality by writing custom UDFs.
  • Developed real-time read write access to very large datasets via HBase.
  • Used NoSQL Database including HBase, MongoDB, Cassandra.
  • Implemented Sqoop jobs for large sets of structured and semi-structured data migration between HDFS and/or other data storage like Hive or RDBMS.
  • Good expertise using the various Development tools like Eclipse and NetBeans.
  • Presenting data in a visually appealing tool Tableau.
  • Used Agile methodology to work with IT and business to progress efficient system development.
  • Multi-talented and consistently rewarded for success in planning and operational improvements.
  • Ability to work effectively with associates at all levels within the organization.
  • Possess excellent communication, interpersonal and analytical skills along with positive attitude.

TECHNICAL SKILLS:

SDLC, STLC, Agile, Hadoop, HDFS , YARN , MapReduce, HBase, Oozie, Pig, Hive, Spark, Storm, Sqoop, Flume, HTML, DHTML, CSS, JavaScript, jQuery, JSP, Servlets, ORACLE, DB2, SQL, PL/SQL, OLAP, OLTP, STATSPACK , AWR, SQL-TRACE, TKPROF, ADDM, HiveQL, Python, Eclipse, NetBeans, Tableau, Windows, Linux

PROFESSIONAL EXPERIENCE:

Confidential, TX

Big Data / Hadoop Developer

Responsibilities:

  • Gathered the business requirements from the Business Partners and Subject Matter Experts.
  • Involved in installing Hadoop Ecosystem components. Used to manage and review the Hadoop log files.
  • Supported Map Reduce Programs those are running on the cluster.
  • Involved in HDFS maintenance and loading of structured and unstructured data.
  • Wrote MapReduce job using Pig Latin.
  • Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
  • Written Hive queries for data analysis to meet the business requirements.
  • Configured Zookeeper, worked on Hadoop High Availability with Zookeeper failover controller, add support for scalable, fault-tolerant data solution.
  • Used Flume to collect, aggregate, and store dynamic web log data from different sources like web servers, mobile devices and pushed to HDFS.
  • Applied MLlib to build statistical model to classify and predict.
  • Created HDFS Snapshots in order to do data backup, protection against user errors and disaster recovery.
  • Developed SQL queries to fetch complex data from different tables in remote databases using database links.
  • Participated in application planning, design activities by interacting and collecting requirements from the end users.

Confidential

Big Data / Hadoop Developer

Responsibilities:

  • Involved in complete SDLC life cycle of big data project that includes requirement analysis, design, coding, testing and production.
  • Developed map reduce jobs to pre-process data using PIG.
  • Created hive tables defined with appropriate static and dynamic partitions, intended for efficiency and worked on them using HIVE QL
  • Exported the result set from HIVE to MySQL using Shell scripts.
  • Involved in running Hadoop streaming jobs to process terabytes of text data.
  • Worked on various kinds of transformations like Expression, Aggregator, Stored Procedure, Lookup, Filter, Joiner, Rank, Router and Update Strategy.
  • Prepared test Scenarios and Test cases in HP Quality Center and involved in unit testing of mappings, system testing and user acceptance testing.
  • Defect Tracking and reports are done by Rational Clear Quest.
  • Responsible for building scalable distributed data solutions using Hadoop.
  • Extracted the data from the flat files and other RDBMS databases into staging area and populated onto Data warehouse.
  • Worked on different tasks in Workflows like sessions, events raise, event wait, decision, e-mail, command, worklets, Assignment, Timer and scheduling of the workflow.
  • Used Debugger to test the mappings and fixed the bugs.

We'd love your feedback!