We provide IT Staff Augmentation Services!

Data Warehousing Resume

SUMMARY

  • Over 8 years of experience in the field of Information Technology Data warehousing mainly emphasizing in Extraction, Transformation and Loading (ETL) mechanism using Ab Initio Maintenance & Development projects
  • Analyze functional and business requirements and prepare high level and low - level design document.
  • Integrate multiple systems by building ETL procedures using Ab Initio/Pentaho PDI.
  • Define and design data transformations using different Ab Initio Components like Input, Output (Table/File), Filter by expression, Rollup, Scan, Normalize, Sort within groups, De-dup sort, Reformat, Split, Read Json and Join.
  • Good Knowledge on generic graphs build and Parameter sets.
  • Good Knowledge on conduct >it plans.
  • Sound Knowledge on reading and parsing JSON data using components.
  • Good Knowledge on implementing both Batch and Continuous flows.
  • Knowledge on Big Data technologies and cloud technology.
  • Implemented the Parallel application by replicating the components and processing modules into number of partitions.
  • Strong working experience in Scrum/ Agile framework and Waterfall project methodologies.
  • Good understanding of. GIT repository for the code migrations.
  • Good Experience working with various Heterogeneous Source Systems: Oracle, Teradata and DB2, and in various Scheduling tools: tws master and Control Center
  • Good Knowledge on Technical repository management console.
  • Tuned the Graphs by removing unnecessary Sorts, used Lookup files, to enhance the performance.
  • Good knowledge of ETL and Data warehousing concepts.
  • Ability to handle process related tasks like code migration, tag creation, access request.
  • Excellent knowledge and experience in SQL and UNIX shell scripting.
  • Good knowledge on writing complex Teradata recursion SQL’s.
  • Have good understanding on Teradata utilities (Fast Load, Fast Export, PT Load and Multiload)
  • Hands on experience in NO SQL data bases like MongoDB and AWS DynamoDB.
  • Handled all aspects of software development lifecycle - Developing, Production Support and QA.
  • Expertise in Bug Tracking Process, Familiar with Bug Reporting and Tracking using Bug tracking tools like Mantis.

TECHNICAL SKILLS

ETL Tool: Ab Initio GDE 3X/ 3.1.5/3.2.5 Co>operating system 3X/ 3.1.5/3.2.5 , Pentaho PDI, Apache NIFI, Alteryx

Database: Oracle 10g/11g, Teradata, Redshift, MongoDB

Operating System: UNIX, Windows XP Professional .

Scheduler : IBM TWS master

Language: Unix Shell scripting and Python

PROFESSIONAL EXPERIENCE

Confidential

Data Warehousing

Responsibilities:

  • Created mapping document as per the functional requirements coordinating with team.
  • Attending the business meetings to discuss about the complex problems and gathering requirements.
  • Attending scrum meetings like stand ups and planning sessions.
  • Extracting, Transforming and Loading the data from Source to Staging and Staging to Target according to the Business requirements.
  • Involved in writing Shell Scripts to transfer files from FTP servers to ETL servers with some validations.
  • Extensive knowledge on parsing JOSN messages and converting data into JSONs messages using Abinitio.
  • Extensively used Multifile System across the project.
  • Implemented ICFF’s for history data storage.
  • Developed the generic procedure to compare the database DDL’s in Abinitio.
  • Understanding the code issues and providing the break fixes in short time.
  • Good knowledge on creating the tags for migrations to higher environments.
  • Good Knowledge in technical repository management console.
  • Involved in code migrations to higher environments.
  • Writing Technical design document and ETL design document as per the business requirements.
  • Good at understanding and implementing the recursive loop logics in Teradata.
  • Implemented the CDC logic’s in Teradata.
  • Writing Tivoli Jobs/Job streams to automate the graph runs.
  • Wrote Teradata SQL queries and shell scripts as per functional requirements.
  • Build Complex SQL Queries to Join Different Input Teradata Tables, ensure that there are Indexes applied, if there is not, then coordinate with the Business Architect/Modelers to get them added.

Confidential

Data Warehousing

Responsibilities:

  • Design and implemented lot of functionalities on Data migration from 711 to RCX.
  • Co-ordinate and execute the loads to get the data from 711 to RCX.
  • Identify and address the gaps between the client system and RCX provided the solutions for the gaps.
  • Extensively worked on SQLite3 to convert csv file into JSON object during the migration.
  • Conducted the user testing through interactive sessions for the UAT AND CRP phases.
  • Requirement analysis and prepared Design Document.
  • Perform Delta and full refresh load.
  • Extensively used git repository for the code migrations.
  • Written UNIX shell scripts to enable the checkpoint features in the Pentaho ETL tool.
  • Written SQL queries to load data from stage to target tables in redshift DB.
  • Writing complex SQL statements to solve business problems.
  • Migrate DB objects and ETL code from Development to till PROD.
  • Prepare design documents and Visio that describe input, output, logical operations and convert them into deliverable code

Confidential

Data Warehousing

Responsibilities:

  • Created mapping document as per the functional requirements coordinating with onsite.
  • Created generic graphs and parameter sets.
  • Extracting, Transforming and Loading the data from Source to Staging and Staging to Target according to the Business requirements using Ab-initio.
  • Good knowledge on using transform components like sort, De-dup sort, filter by expression, rollup, scan, normalize etc.
  • Extensively using partition and Departition components.
  • Actively participated in migration tasks from DEV to PROD.
  • Involved in writing Shell Scripts as per the business requirement.
  • Involved in creating Technical design document and ETL design document.
  • Involved in creating DOU document for EPS team.
  • Involved in performance tuning for the Graphs and SQL queries.
  • Created tags and migrated code from one environment to another.
  • Created and scheduled Jobs using TWS master.
  • Build Complex SQL Queries to Join Different Input Teradata Tables, ensure that there are Indexes applied, if there is not, then coordinate with the Business Architect/Modelers to get them added.
  • Writing complex SQL statements to solve business problems.

Hire Now