We provide IT Staff Augmentation Services!

Big Data Hadoop Developer Resume

0/5 (Submit Your Rating)

SUMMARY

  • Having 12+ years of professional experience including 4+ years of extensive experience into Big Data Technologies in design and development of standalone and web applications in multi - tiered environment using Java, Hadoop, Hive, HBase, Pig, Flume and Sqoop.
  • Working experience on designing and implementing complete end-to-end Hadoop Infrastructure including Pig, Hive and Sqoop
  • Expertise in importing and exporting data into HDFS and Hive using Sqoop.
  • Expertise in developing Pig Latin scripts and using Hive Query Language.
  • Experience in developing custom UDF's for Pig and Hive.
  • Collected data from different sources like web servers and social media using flume for storing in HDFS and analyzing the data using other Hadoop technologies.
  • Ability to import and export data between HDFS and Relational Data Management Systems using Scoop
  • In Depth understanding in installing and configuring Pig, Hive, HBase, Flume, Sqoop on the Hadoop Clusters.
  • Experienced in loading and transforming large sets of structured, semi-structured and unstructured data.
  • Worked on NoSQL database like HBase, MongoDB and Cassandra.
  • Expertise in job workflow scheduling and monitoring tools like Oozie and Zookeeper.
  • Expertise in Linux commands and shell scripting.
  • Experience in working with BI team and transform big data requirements into Hadoop centric technologies.
  • In depth knowledge of databases like Oracle 8i/9i/10g, SQL Server 2000, MySQL and extensively worked in writing SQL queries, Stored Procedures, Triggers, Cursors, Functions and Packages
  • In depth knowledge of Informatica and extensively worked on mappings and workflows
  • In depth knowledge of data migration and extensively worked on migrating data from different source systems into Siebel CRM
  • Extensively worked on Siebel Enterprise Integration Manager
  • In depth knowledge of scheduling tools and the process to schedule the data load jobs and extensively worked on Autosys and ControlM
  • A self-motivated professional and natural communicator possessing good technical, initiating, leadership and problem-solving skills and has proven to be a good team player.
  • Ability to quickly ramp up and start producing results on given any tool or technology.
  • Excellent communication, analytical, interpersonal and presentation skills.

TECHNICAL SKILLS

Big Data Ecosystem: Hadoop, MapReduce, HDFS, HBase, Zookeeper, Hive, Pig Utilities Oozie, Sqoop, HBase, Flume.

Languages: Java 1.4/1.5/1.6, PL/SQL

Web Technologies: HTML, DHTML, Java Script, VB Script, XML

RDBMS: Oracle, SQL Server 2000/2005/2008

CRM: Siebel

Operating Systems: Windows XP/Vista/7, UNIX, Linux.

PROFESSIONAL EXPERIENCE

Big Data Hadoop Developer

Confidential

Responsibilities:

  • Gathered the business requirements from the Business Partners and Subject Matter Experts
  • Involved in installing Hadoop ecosystem components
  • Responsible to manage data coming from different sources
  • Supported Map Reduce programs those are running on the cluster
  • Involved in HDFS maintenance and loading of structured and unstructured data
  • Wrote MapReduce jobs using Java API and Pig Latin as well
  • Involved in managing and reviewing Hadoop log files
  • Developed Scripts and Batch Job to schedule various Hadoop Program
  • Written Hive queries for data analysis to meet the business requirements

Environment: Map Reduce, HDFS, Hive, Pig, Linux, Java, Flume, Sqoop, MySQL

Big Data Hadoop Developer

Confidential

Responsibilities:

  • Involved in review of functional and non-functional requirements.
  • Importing and exporting data into HDFS and Hive using Sqoop.
  • Experienced in defining job flows.
  • Experienced in managing and reviewing Hadoop log files.
  • Extracted files from through Sqoop and placed in HDFS and processed.
  • Load and transform large sets of structured, semi structured and unstructured data.
  • Responsible to manage data coming from different sources.
  • Got good experience with NOSQL database.
  • Supported Map Reduce Programs those are running on the cluster.
  • Involved in loading data from UNIX file system to HDFS.
  • Installed and configured Hive and also written Hive UDFs.
  • Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduce way.
  • Setup and benchmarked Hadoop/HBase clusters for internal use
  • Setup Hadoop cluster on Amazon EC2 using whirr for POC.

Environment: Map Reduce, HDFS, Hive, Pig, Linux, Java, Flume, Sqoop, MySQL

Siebel Data Migration Lead

Confidential

Responsibilities:

  • Data migration requirements gathering.
  • Data Mapping between Siebel and different source systems
  • Create/Extend new columns as part of the Data Mapping
  • Prepare Data Migration Design Document and entity wise data mapping spreadsheets.
  • Build Informatica Mapping between Source Systems and Siebel EIM Tables
  • Build Informatica workflows and schedule jobs for the daily data loads
  • Create IFB files for the load of various entities identified.
  • Build and test SQL scripts to insert data for one-time data loads.
  • Build Shell scripts to automate batch jobs

Environment: Siebel CRM, Informatica, Oracle 11g, Linux

Siebel Data Migration Lead

Confidential

Responsibilities:

  • Data migration requirements gathering.
  • Data Mapping between Siebel and different source systems
  • Interaction with the business on different acquisitions
  • Prepare Data Migration Design Document and entity wise data mapping spreadsheets.
  • Build Informatica Mapping between Source Systems and Siebel EIM Tables
  • Build Informatica workflows and schedule jobs for the daily data loads
  • Create/modify DTS packages for the data loads
  • Create IFB files for the load of various entities identified.
  • Build and test SQL scripts to insert data for one-time data loads.
  • Build Autosys jobs for job scheduling

Environment: Siebel CRM, Informatica, MS SQL Server, Windows, Autosys

Siebel Data Migration Consultant

Confidential

Responsibilities:

  • Interaction with the client on the data related developments / issues
  • Develop/Unit Test Siebel EIM Programs
  • Performance tuning and revamping of existing EIM scripts.
  • Worked closely with business analyst / power users during the Unit and System testing phases
  • As a part of the production support, was responsible for planning and executing any application downtime for enhancements and regular maintenance activities.
  • Upon failures/Outages, was responsible for bringing in all the related groups on to a bridge call and analyzing the failure reasons and taking appropriate actions to rectify and also informing the management about the same.
  • Attending weekly Change Control Meetings to discuss the changes planned, in any of the source systems and analyze the effects of the same at report level/Universe level and come out with action plans to mitigate the same.
  • Documenting any Outages, identifying the cause and providing short-term / long-term solutions for the problems occurred.
  • Develop/Unit Test Extract, Transformation and Load Programs using T-SQL
  • Coordinating with Data Warehouse team and other data source providers to ensure update to date data, thus resulting in close to real-time Analytical reports

Environment: Siebel CRM, MS SQL Server, Windows

We'd love your feedback!