Hadoop Big Data Consultant Resume
Columbus, OH
SUMMARY:
- Overall 10+ years of wide Experience in various technologies like Big Data Hadoop, Data Warehouse, ETL, Java and Mainframe.
- 3 + Years of experience in Big Data Hadoop.
- 2 + Years of Java, Struts, JQuery, Oracle JDBC.
- 2+ years of experience in Informatica, ETL, and Data Warehouse/Data Mart design.
- 3+ years experienced on COBOL, DB2, JCL, and IDMS Technology in Banking, Insurance and Health Care domains.
- Big Data Hadoop trained and certified Professional.
- Extensive knowledge on Programming Language Java and Python in which have coded lot of Map Reduce Programs.
- Experienced in HIVE, PIG Query handling and implemented lot of scrubbing rules in HIVE.
- Implemented queries written in BigSQL and HIVE.
- Experience in handling the Big Data Projects with the hands on expertise in designing the Hadoop Eco System for the Project.
- Well versed in using HBase and is involved in the HBase database design including Row Key and Column Family designs.
- In Depth understanding of the Data Structure, Data flow, loading the data to HDFS, Map Reducing framework and query handling using PIG, HIVE, BigSQL.
- Have extensively worked in SQOOP to load the data from Teradata to HDFS.
- Have strong knowledge in UNIX scripting, shell commands.
- Involved in the design of HIVE and BigSQL queries where many optimization techniques are used.
- Hand on Experience on Production Support model where the data injection happens for every hour into HDFS and processing of data using Java, Python, HIVE and BigSQL.
- Able to assess business rules, collaborate with stakeholders and perform source - to-target data mapping, design and review.
- Designed and implemented application using JSP, JavaScript, JQuery, Struts, Spring MVC, SQL, Maven, Oracle and Web Services.
- Responsible for development, production support, bug fixing and enhancements of multiple applications.
- Implemented MVC-Model architecture using Struts and other JEE design patterns for the application development.
- Involved in installing and configuring Maven for application builds and deployment.
- Was involved in Query handling to address the Customers adhoc requests.
- Strong Data Warehousing ETL experience of using Informatica 9.0.1/9.5.1/9.6.1 Power Center Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor.
- Strong experience in Extraction, Transformation and Loading (ETL) data from various sources into Data Warehouses and Data Marts using Informatica Power Center.
- Exposure to the enterprise wide best practices of Data Modelling and having strong knowledge on Dimensional modelling.
- Extensively worked on Informatica Power Center Transformations such as Source Qualifier, Lookup, Filter, Expression, Router, Joiner, Update Strategy, Rank, Aggregator, Stored Procedure, Sorter, Sequence Generator, Normalizer and Union.
- Extensive experience in all the Mainframe configuration Management, debugging, File Management and Version Control Tools.
- Proficient in analyzing and translating business requirements to technical requirements and architecture.
- Strong database skills with extensive expertise on DB2 and IDMS Databases.
- Worked on development, Maintenance and Support related Projects and worked as SME and application Owner of major Projects.
- Experienced in Leading the team and handling multiple projects simultaneously.
- A well-organized, Creative, goal-oriented, highly motivated effective team member with excellent analytical, troubleshooting, and problem solving Skills.
- Extensively worked both on Batch and real time projects which are based on Waterfall, Agile and Scrum Methodology.
- SCRUM Master certified and worked as Product Owner and SCRUM Master.
TECHNICAL SKILLS:
Domain: Insurance, Banking, Health Care
Big Data: Apache Hadoop, Big Insight, Map Reduce, HDFS, HIVE, BigSQL, SQOOP, PIG, HBase, Python, OOZIE, UNIX, Shell Scripting
Java: Java/JEE, JSP, Strut, XML, SQL, Eclipse, SVN
Informatica: Power Center 9.0.1/9.5.1/9.6.1 , Oracle, UNIX, Perl, ESP Scheduler, Toad
Mainframe: COBOL, JCL, VSAM, DB2, IDMS, CICS Changeman, Endevour, File AID, File Manager, Xpediter, CTM, ESP
PROFESSIONAL EXPERIENCE:
Confidential, Columbus, OH
Hadoop Big Data Consultant
Responsibilities:
- Involved in the design and architecture of end to end transformation/data flow from source feeds to SRE Web Application.
- Involved in the architectural design and building of Hadoop Eco System.
- Have worked on implementation of scripts to load the data to HDFS using SQOOP and Web HDFS commands.
- Involved in defining the Job flow, scheduling the Hadoop Jobs in Scheduler, managing and reviewing the Job logs.
- Have implemented various performance optimization techniques in HIVE, PIG, BigSQL and HBase.
- Handled importing of data from various data sources, loading of the data to HDFS, performing the transformations using Hive, MapReduce, finally loading the data to HBase.
- Standardize the scripts to load the data from HDFS to the different HIVE tables.
- Developing scripts to extract the data from HDFS and load to HBase.
- Was involved in the design of HBase Row Key and Column Family.
- Have written UDFs in Java and Python and implemented in HIVE.
- Performed data validation on the data ingested using MapReduce by building a custom model to filter all the invalid data and cleanse the data.
- Developed and implemented the queries in HIVE and BIGSQL.
- Performed complex Joins on the tables in Hive.
Environment: Hadoop, HDFS, Map Reduce, HIVE, PIG, HBase, Python, BigInsight
Confidential
Project Lead
Responsibilities:
- Involved in solving the quick requests from customers with the help of Adoc one time code.
- Worked on the enhancement and operational activities.
- Involved in all the maintenance related activities by keeping the System UP and running without causing SLA delays.
- Actively participated in the daily SCRUM meetings to produce quality deliverables within time.
- Worked on Java, JavaScript, XML and MVC.
- Monitoring the Batch Jobs scheduled in ESP.
Environment: Java, JSP, Strut, Oracle, ESP, Corba
Confidential
Sr. Analyst Programmer
Responsibilities:
- Involved both in Build and RUN team where worked on development, Enhancement and Production Support model.
- Worked on an application to perform ETL from Mainframe System which had 35 applications to Life ODS.
- Involved in building the ETL architecture and Source to Target mapping to load data into Data warehouse.
- Created mapping documents to outline data flow from sources to targets.
- Extracted the data from the flat files and other RDBMS databases into staging area and populated onto Data warehouse.
- Used various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings in the Informatica Designer.
- Used existing ETL standards to develop these mappings.
- Worked on the enhancement and operational activities.
- Involved in all the maintenance related activities by keeping the System UP and running without causing SLA delays.
- Created Reusable Transformations and Mapplets to use in Multiple Mappings.
- Supported daily loads and worked with business users to handle rejected data.
Environment: Informatica, Power Center, UNIX, ESP, Oracle, DB2
Analyst Programmer
Confidential, Ga
Responsibilities:
- Worked as Onsite Coordinator in Columbus, GA and was involved in client interaction, gathering requirement, and Implementation activities.
- Involved in all the phases of major Projects from Design till Implementation and Production Support.
- Review the project requirements and interacting with product management.
- Involved in preparation of Estimates, LOE’s (Level of Effort) and Design documents.
Environment: COBOL, JCL, VSAM, CTM, Changeman, CTM
Confidential
Software Engineer
Responsibilities:
- Analyzing the User Requirements specifications so that requirements are technically feasible.
- Preparing design document from business requirement.
- Actively involved in batch and online programming for the development projects.
- Work with QA teams to create test scripts and scenarios for enhancements.
- Performing Unit Testing, System Integration Testing, Regression Testing.
Environment: COBOL, JCL, VSAM, DB2, Changeman, Xpeditor CICS