We provide IT Staff Augmentation Services!

Hadoop Big Data Consultant Resume

5.00/5 (Submit Your Rating)

Columbus, OH

SUMMARY:

  • Overall 10+ years of wide Experience in various technologies like Big Data Hadoop, Data Warehouse, ETL, Java and Mainframe.
  • 3 + Years of experience in Big Data Hadoop.
  • 2 + Years of Java, Struts, JQuery, Oracle JDBC.
  • 2+ years of experience in Informatica, ETL, and Data Warehouse/Data Mart design.
  • 3+ years experienced on COBOL, DB2, JCL, and IDMS Technology in Banking, Insurance and Health Care domains.
  • Big Data Hadoop trained and certified Professional.
  • Extensive knowledge on Programming Language Java and Python in which have coded lot of Map Reduce Programs.
  • Experienced in HIVE, PIG Query handling and implemented lot of scrubbing rules in HIVE.
  • Implemented queries written in BigSQL and HIVE.
  • Experience in handling the Big Data Projects with the hands on expertise in designing the Hadoop Eco System for the Project.
  • Well versed in using HBase and is involved in the HBase database design including Row Key and Column Family designs.
  • In Depth understanding of the Data Structure, Data flow, loading the data to HDFS, Map Reducing framework and query handling using PIG, HIVE, BigSQL.
  • Have extensively worked in SQOOP to load the data from Teradata to HDFS.
  • Have strong knowledge in UNIX scripting, shell commands.
  • Involved in the design of HIVE and BigSQL queries where many optimization techniques are used.
  • Hand on Experience on Production Support model where the data injection happens for every hour into HDFS and processing of data using Java, Python, HIVE and BigSQL.
  • Able to assess business rules, collaborate with stakeholders and perform source - to-target data mapping, design and review.
  • Designed and implemented application using JSP, JavaScript, JQuery, Struts, Spring MVC, SQL, Maven, Oracle and Web Services.
  • Responsible for development, production support, bug fixing and enhancements of multiple applications.
  • Implemented MVC-Model architecture using Struts and other JEE design patterns for the application development.
  • Involved in installing and configuring Maven for application builds and deployment.
  • Was involved in Query handling to address the Customers adhoc requests.
  • Strong Data Warehousing ETL experience of using Informatica 9.0.1/9.5.1/9.6.1 Power Center Client tools - Mapping Designer, Repository manager, Workflow Manager/Monitor.
  • Strong experience in Extraction, Transformation and Loading (ETL) data from various sources into Data Warehouses and Data Marts using Informatica Power Center.
  • Exposure to the enterprise wide best practices of Data Modelling and having strong knowledge on Dimensional modelling.
  • Extensively worked on Informatica Power Center Transformations such as Source Qualifier, Lookup, Filter, Expression, Router, Joiner, Update Strategy, Rank, Aggregator, Stored Procedure, Sorter, Sequence Generator, Normalizer and Union.
  • Extensive experience in all the Mainframe configuration Management, debugging, File Management and Version Control Tools.
  • Proficient in analyzing and translating business requirements to technical requirements and architecture.
  • Strong database skills with extensive expertise on DB2 and IDMS Databases.
  • Worked on development, Maintenance and Support related Projects and worked as SME and application Owner of major Projects.
  • Experienced in Leading the team and handling multiple projects simultaneously.
  • A well-organized, Creative, goal-oriented, highly motivated effective team member with excellent analytical, troubleshooting, and problem solving Skills.
  • Extensively worked both on Batch and real time projects which are based on Waterfall, Agile and Scrum Methodology.
  • SCRUM Master certified and worked as Product Owner and SCRUM Master.

TECHNICAL SKILLS:

Domain: Insurance, Banking, Health Care

Big Data: Apache Hadoop, Big Insight, Map Reduce, HDFS, HIVE, BigSQL, SQOOP, PIG, HBase, Python, OOZIE, UNIX, Shell Scripting

Java: Java/JEE, JSP, Strut, XML, SQL, Eclipse, SVN

Informatica: Power Center 9.0.1/9.5.1/9.6.1 , Oracle, UNIX, Perl, ESP Scheduler, Toad

Mainframe: COBOL, JCL, VSAM, DB2, IDMS, CICS Changeman, Endevour, File AID, File Manager, Xpediter, CTM, ESP

PROFESSIONAL EXPERIENCE:

Confidential, Columbus, OH

Hadoop Big Data Consultant

Responsibilities:

  • Involved in the design and architecture of end to end transformation/data flow from source feeds to SRE Web Application.
  • Involved in the architectural design and building of Hadoop Eco System.
  • Have worked on implementation of scripts to load the data to HDFS using SQOOP and Web HDFS commands.
  • Involved in defining the Job flow, scheduling the Hadoop Jobs in Scheduler, managing and reviewing the Job logs.
  • Have implemented various performance optimization techniques in HIVE, PIG, BigSQL and HBase.
  • Handled importing of data from various data sources, loading of the data to HDFS, performing the transformations using Hive, MapReduce, finally loading the data to HBase.
  • Standardize the scripts to load the data from HDFS to the different HIVE tables.
  • Developing scripts to extract the data from HDFS and load to HBase.
  • Was involved in the design of HBase Row Key and Column Family.
  • Have written UDFs in Java and Python and implemented in HIVE.
  • Performed data validation on the data ingested using MapReduce by building a custom model to filter all the invalid data and cleanse the data.
  • Developed and implemented the queries in HIVE and BIGSQL.
  • Performed complex Joins on the tables in Hive.

Environment: Hadoop, HDFS, Map Reduce, HIVE, PIG, HBase, Python, BigInsight

Confidential

Project Lead

Responsibilities:

  • Involved in solving the quick requests from customers with the help of Adoc one time code.
  • Worked on the enhancement and operational activities.
  • Involved in all the maintenance related activities by keeping the System UP and running without causing SLA delays.
  • Actively participated in the daily SCRUM meetings to produce quality deliverables within time.
  • Worked on Java, JavaScript, XML and MVC.
  • Monitoring the Batch Jobs scheduled in ESP.

Environment: Java, JSP, Strut, Oracle, ESP, Corba

Confidential

Sr. Analyst Programmer

Responsibilities:

  • Involved both in Build and RUN team where worked on development, Enhancement and Production Support model.
  • Worked on an application to perform ETL from Mainframe System which had 35 applications to Life ODS.
  • Involved in building the ETL architecture and Source to Target mapping to load data into Data warehouse.
  • Created mapping documents to outline data flow from sources to targets.
  • Extracted the data from the flat files and other RDBMS databases into staging area and populated onto Data warehouse.
  • Used various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings in the Informatica Designer.
  • Used existing ETL standards to develop these mappings.
  • Worked on the enhancement and operational activities.
  • Involved in all the maintenance related activities by keeping the System UP and running without causing SLA delays.
  • Created Reusable Transformations and Mapplets to use in Multiple Mappings.
  • Supported daily loads and worked with business users to handle rejected data.

Environment: Informatica, Power Center, UNIX, ESP, Oracle, DB2

Analyst Programmer

Confidential, Ga

Responsibilities:

  • Worked as Onsite Coordinator in Columbus, GA and was involved in client interaction, gathering requirement, and Implementation activities.
  • Involved in all the phases of major Projects from Design till Implementation and Production Support.
  • Review the project requirements and interacting with product management.
  • Involved in preparation of Estimates, LOE’s (Level of Effort) and Design documents.

Environment: COBOL, JCL, VSAM, CTM, Changeman, CTM

Confidential

Software Engineer

Responsibilities:

  • Analyzing the User Requirements specifications so that requirements are technically feasible.
  • Preparing design document from business requirement.
  • Actively involved in batch and online programming for the development projects.
  • Work with QA teams to create test scripts and scenarios for enhancements.
  • Performing Unit Testing, System Integration Testing, Regression Testing.

Environment: COBOL, JCL, VSAM, DB2, Changeman, Xpeditor CICS

We'd love your feedback!