Big Data / Hadoop Developer Resume
4.00/5 (Submit Your Rating)
TX
SUMMARY:
- Around 4 years of working experience as a Big Data and Hadoop Developer.
- Thorough understanding of Software Development Life Cycle (SDLC), Software Test Life Cycle (STLC) and processes across multiple environments and platforms.
- Excellent understanding of Hadoop architecture and various components such as HDFS, YARN, High Availability and MapReduce programming paradigm.
- Hands on experience in installing, configuring, and using Hadoop ecosystem components like Hadoop, MapReduce, HDFS, HBase, Oozie, Hive, Spark, Storm, Sqoop and Flume.
- Experience in design, development of web - based applications using HTML, DHTML, CSS, JavaScript, jQuery, JSP and Servlets.
- Sound Relational Database concepts and extensively worked with ORACLE, DB2. Very good in writing complex SQL queries and PL/SQL procedures.
- Experience in Performance Tuning, tuned various large critical OLAP and OLTP databases of Terabytes size using various tools like STATSPACK, AWR, SQL-TRACE, TKPROF, ADDM and Explain Plan Methods.
- Experienced in implementing complex analytical algorithms using Map reduce design patterns.
- Experience in analysed data using HiveQL, HBase and custom MapReduce programs in Python.
- Extended Pig and Hive core functionality by writing custom UDFs.
- Developed real-time read write access to very large datasets via HBase.
- Used NoSQL Database including HBase, MongoDB, Cassandra.
- Implemented Sqoop jobs for large sets of structured and semi-structured data migration between HDFS and/or other data storage like Hive or RDBMS.
- Good expertise using the various Development tools like Eclipse and NetBeans.
- Presenting data in a visually appealing tool Tableau.
- Used Agile methodology to work with IT and business to progress efficient system development.
- Multi-talented and consistently rewarded for success in planning and operational improvements.
- Ability to work effectively with associates at all levels within the organization.
- Possess excellent communication, interpersonal and analytical skills along with positive attitude.
TECHNICAL SKILLS:
SDLC, STLC, Agile, Hadoop, HDFS , YARN , MapReduce, HBase, Oozie, Pig, Hive, Spark, Storm, Sqoop, Flume, HTML, DHTML, CSS, JavaScript, jQuery, JSP, Servlets, ORACLE, DB2, SQL, PL/SQL, OLAP, OLTP, STATSPACK , AWR, SQL-TRACE, TKPROF, ADDM, HiveQL, Python, Eclipse, NetBeans, Tableau, Windows, Linux
PROFESSIONAL EXPERIENCE:
Confidential, TX
Big Data / Hadoop Developer
Responsibilities:
- Gathered the business requirements from the Business Partners and Subject Matter Experts.
- Involved in installing Hadoop Ecosystem components. Used to manage and review the Hadoop log files.
- Supported Map Reduce Programs those are running on the cluster.
- Involved in HDFS maintenance and loading of structured and unstructured data.
- Wrote MapReduce job using Pig Latin.
- Imported data using Sqoop to load data from MySQL to HDFS on regular basis.
- Written Hive queries for data analysis to meet the business requirements.
- Configured Zookeeper, worked on Hadoop High Availability with Zookeeper failover controller, add support for scalable, fault-tolerant data solution.
- Used Flume to collect, aggregate, and store dynamic web log data from different sources like web servers, mobile devices and pushed to HDFS.
- Applied MLlib to build statistical model to classify and predict.
- Created HDFS Snapshots in order to do data backup, protection against user errors and disaster recovery.
- Developed SQL queries to fetch complex data from different tables in remote databases using database links.
- Participated in application planning, design activities by interacting and collecting requirements from the end users.
Confidential
Big Data / Hadoop Developer
Responsibilities:
- Involved in complete SDLC life cycle of big data project that includes requirement analysis, design, coding, testing and production.
- Developed map reduce jobs to pre-process data using PIG.
- Created hive tables defined with appropriate static and dynamic partitions, intended for efficiency and worked on them using HIVE QL
- Exported the result set from HIVE to MySQL using Shell scripts.
- Involved in running Hadoop streaming jobs to process terabytes of text data.
- Worked on various kinds of transformations like Expression, Aggregator, Stored Procedure, Lookup, Filter, Joiner, Rank, Router and Update Strategy.
- Prepared test Scenarios and Test cases in HP Quality Center and involved in unit testing of mappings, system testing and user acceptance testing.
- Defect Tracking and reports are done by Rational Clear Quest.
- Responsible for building scalable distributed data solutions using Hadoop.
- Extracted the data from the flat files and other RDBMS databases into staging area and populated onto Data warehouse.
- Worked on different tasks in Workflows like sessions, events raise, event wait, decision, e-mail, command, worklets, Assignment, Timer and scheduling of the workflow.
- Used Debugger to test the mappings and fixed the bugs.