Big Data Hadoop Developer Resume
SUMMARY
- Having 12+ years of professional experience including 4+ years of extensive experience into Big Data Technologies in design and development of standalone and web applications in multi - tiered environment using Java, Hadoop, Hive, HBase, Pig, Flume and Sqoop.
- Working experience on designing and implementing complete end-to-end Hadoop Infrastructure including Pig, Hive and Sqoop
- Expertise in importing and exporting data into HDFS and Hive using Sqoop.
- Expertise in developing Pig Latin scripts and using Hive Query Language.
- Experience in developing custom UDF's for Pig and Hive.
- Collected data from different sources like web servers and social media using flume for storing in HDFS and analyzing the data using other Hadoop technologies.
- Ability to import and export data between HDFS and Relational Data Management Systems using Scoop
- In Depth understanding in installing and configuring Pig, Hive, HBase, Flume, Sqoop on the Hadoop Clusters.
- Experienced in loading and transforming large sets of structured, semi-structured and unstructured data.
- Worked on NoSQL database like HBase, MongoDB and Cassandra.
- Expertise in job workflow scheduling and monitoring tools like Oozie and Zookeeper.
- Expertise in Linux commands and shell scripting.
- Experience in working with BI team and transform big data requirements into Hadoop centric technologies.
- In depth knowledge of databases like Oracle 8i/9i/10g, SQL Server 2000, MySQL and extensively worked in writing SQL queries, Stored Procedures, Triggers, Cursors, Functions and Packages
- In depth knowledge of Informatica and extensively worked on mappings and workflows
- In depth knowledge of data migration and extensively worked on migrating data from different source systems into Siebel CRM
- Extensively worked on Siebel Enterprise Integration Manager
- In depth knowledge of scheduling tools and the process to schedule the data load jobs and extensively worked on Autosys and ControlM
- A self-motivated professional and natural communicator possessing good technical, initiating, leadership and problem-solving skills and has proven to be a good team player.
- Ability to quickly ramp up and start producing results on given any tool or technology.
- Excellent communication, analytical, interpersonal and presentation skills.
TECHNICAL SKILLS
Big Data Ecosystem: Hadoop, MapReduce, HDFS, HBase, Zookeeper, Hive, Pig Utilities Oozie, Sqoop, HBase, Flume.
Languages: Java 1.4/1.5/1.6, PL/SQL
Web Technologies: HTML, DHTML, Java Script, VB Script, XML
RDBMS: Oracle, SQL Server 2000/2005/2008
CRM: Siebel
Operating Systems: Windows XP/Vista/7, UNIX, Linux.
PROFESSIONAL EXPERIENCE
Big Data Hadoop Developer
Confidential
Responsibilities:
- Gathered the business requirements from the Business Partners and Subject Matter Experts
- Involved in installing Hadoop ecosystem components
- Responsible to manage data coming from different sources
- Supported Map Reduce programs those are running on the cluster
- Involved in HDFS maintenance and loading of structured and unstructured data
- Wrote MapReduce jobs using Java API and Pig Latin as well
- Involved in managing and reviewing Hadoop log files
- Developed Scripts and Batch Job to schedule various Hadoop Program
- Written Hive queries for data analysis to meet the business requirements
Environment: Map Reduce, HDFS, Hive, Pig, Linux, Java, Flume, Sqoop, MySQL
Big Data Hadoop Developer
Confidential
Responsibilities:
- Involved in review of functional and non-functional requirements.
- Importing and exporting data into HDFS and Hive using Sqoop.
- Experienced in defining job flows.
- Experienced in managing and reviewing Hadoop log files.
- Extracted files from through Sqoop and placed in HDFS and processed.
- Load and transform large sets of structured, semi structured and unstructured data.
- Responsible to manage data coming from different sources.
- Got good experience with NOSQL database.
- Supported Map Reduce Programs those are running on the cluster.
- Involved in loading data from UNIX file system to HDFS.
- Installed and configured Hive and also written Hive UDFs.
- Involved in creating Hive tables, loading with data and writing hive queries which will run internally in MapReduce way.
- Setup and benchmarked Hadoop/HBase clusters for internal use
- Setup Hadoop cluster on Amazon EC2 using whirr for POC.
Environment: Map Reduce, HDFS, Hive, Pig, Linux, Java, Flume, Sqoop, MySQL
Siebel Data Migration Lead
Confidential
Responsibilities:
- Data migration requirements gathering.
- Data Mapping between Siebel and different source systems
- Create/Extend new columns as part of the Data Mapping
- Prepare Data Migration Design Document and entity wise data mapping spreadsheets.
- Build Informatica Mapping between Source Systems and Siebel EIM Tables
- Build Informatica workflows and schedule jobs for the daily data loads
- Create IFB files for the load of various entities identified.
- Build and test SQL scripts to insert data for one-time data loads.
- Build Shell scripts to automate batch jobs
Environment: Siebel CRM, Informatica, Oracle 11g, Linux
Siebel Data Migration Lead
Confidential
Responsibilities:
- Data migration requirements gathering.
- Data Mapping between Siebel and different source systems
- Interaction with the business on different acquisitions
- Prepare Data Migration Design Document and entity wise data mapping spreadsheets.
- Build Informatica Mapping between Source Systems and Siebel EIM Tables
- Build Informatica workflows and schedule jobs for the daily data loads
- Create/modify DTS packages for the data loads
- Create IFB files for the load of various entities identified.
- Build and test SQL scripts to insert data for one-time data loads.
- Build Autosys jobs for job scheduling
Environment: Siebel CRM, Informatica, MS SQL Server, Windows, Autosys
Siebel Data Migration Consultant
Confidential
Responsibilities:
- Interaction with the client on the data related developments / issues
- Develop/Unit Test Siebel EIM Programs
- Performance tuning and revamping of existing EIM scripts.
- Worked closely with business analyst / power users during the Unit and System testing phases
- As a part of the production support, was responsible for planning and executing any application downtime for enhancements and regular maintenance activities.
- Upon failures/Outages, was responsible for bringing in all the related groups on to a bridge call and analyzing the failure reasons and taking appropriate actions to rectify and also informing the management about the same.
- Attending weekly Change Control Meetings to discuss the changes planned, in any of the source systems and analyze the effects of the same at report level/Universe level and come out with action plans to mitigate the same.
- Documenting any Outages, identifying the cause and providing short-term / long-term solutions for the problems occurred.
- Develop/Unit Test Extract, Transformation and Load Programs using T-SQL
- Coordinating with Data Warehouse team and other data source providers to ensure update to date data, thus resulting in close to real-time Analytical reports
Environment: Siebel CRM, MS SQL Server, Windows