We provide IT Staff Augmentation Services!

Lead Datastage/qualitystage Developer Resume

5.00/5 (Submit Your Rating)

Bentonville, AR

SUMMARY:

  • 10+ years of experience in Data warehousing experience with special emphasis on Development, Analysis, Design of ETL methodologies in all the phases of the Data Warehousing life cycle with knowledge of DSS and OLAP technology.
  • Extensively worked on Requirement Gathering, Designing, Developing, and Implementing ETL Process using IBM Infosphere DataStage and Quality Stage 11.5/9.1/8.5/8.1/8.0 both enterprise and standard editions.
  • Strong working knowledge of SDLC and Agile methodologies.
  • Experience in creating complex jobs using various transformations, and developing strategies for Extraction, Transformation and Loading (ETL) mechanism DataStage Designer and DataStage Director on Server Engines for sectors such as Health Care, Finance, and Insurance.
  • Experience in creating Hive load scripts and Oozie workflow jobs.
  • Strong understanding of the Big Data concepts.
  • Experience in source systems analysis and data extraction from various sources like Flat files, MVS, Oracle9i/10g/11g, SQL Server, DB2, INFORMIX, and Teradata.
  • Expert in Data Warehousing techniques for Data Analysis, Cleansing, Transforming, Testing, Slowly Changing Dimensions, Change Data Capture and Data Loading across source systems. E.g. OLTPs. Vs. DW Structures, Normalization, Fact and Dimensional Tables, Star Schema Modeling, Snowflake Schema Modeling.
  • Migrating jobs from Datastage 8.0 to Datastage 9.1 and Datastage 9.1 to Datastage 11.5
  • Used QualityStage to develop address standardization and data profiling jobs
  • Strong skills in performing Column Analysis, Column Analysis Review, Table Analysis, Table Analysis Review, Primary Key Analysis, Primary Key Review, Cross Table Analysis, Relationship Analysis, and Review, Removing redundant domains, Accept Relations, Normalization, Generating specification and DDL generation using Profile Stage.
  • Strong working experience with SAP BW packs and XML parsers .
  • Experience in using SAP BW packs to work with data from SAP BI.
  • Experience in designing the QA testing strategies and test cases for a thorough testing of developed Data warehousing applications.
  • Strong experience in effort estimation, resource capacity planning, prioritization, and allocation.
  • Extensive experience in developing UNIX and Linux Shell Scripts for automating Extraction, Transformation and Loading process for various feeds and enabling the execution of jobs in a production environment using a scheduler like CA7 and Crontab.
  • Strong working experience in designing DataStage job scheduling outside the DataStage tool and also within the tool as required by Client/Customer company standards
  • Strong knowledge of database concepts such as indexing, views, schemas and other database objects in both Oracle and DB2 databases and Object - Oriented technologies.
  • Strong experience in coding using SQL, SQL*Plus, PL/SQL Procedures/Functions, Triggers, and Packages.
  • Expertise in Logical, Physical, Conceptual Data Modeling and performance tuning of Oracle, DB2 database, SQL, ETL processes.
  • A Clear understanding of Star-Schema Modelling, Snowflakes Modelling, Fact and Dimension Tables.
  • Successfully lead the development teams across multiple locations as the onsite coordinator and in the process mentored the junior resources in the team to bring them up to speed.
  • Technical expertise in Software Development Life Cycle entailing requirement analysis, development, maintenance and support.
  • Hands-on experience in data conversion methods and tools.
  • Excellent communication and interpersonal skills with proven abilities in resolving the complex software issues.
  • A self-motivated, creative and innovative individual willing to work hard.

TECHNICAL SKILLS:

  • IBM Infosphere 11.5/9.1/8.5/8.1 (DataStage
  • IBM Information Analyzer
  • Quality Stage
  • UNIX (AIX
  • SOLARIS
  • ABAP
  • BAPI
  • IDOC Extracts)
  • Hadoop
  • MapReduce
  • HDFS
  • Hive
  • Oozie
  • LINUX
  • CA7
  • BW Pack
  • Erwin
  • Oracle 11g/10g/9i/8i
  • DB2
  • Toad
  • SQL Developer
  • Teradata SQL Assistant
  • SQL Server 2005/2008
  • SQL
  • PL/SQL
  • SQL* Loader
  • MS Visio
  • HTML
  • C
  • C++
  • Teradata V2R5.x
  • XML
  • TDCH
  • Jira
  • GIT.

PROFESSIONAL EXPERIENCE:

Confidential, Bentonville, AR

Lead DataStage/QualityStage Developer

Responsibilities:

  • The primary responsibility included creating QualityStage jobs to run quality rules on the received files from the third party feeds.
  • Identified and documented data sources and transformation rules and also developed the DB column mapping documents.
  • Designed jobs using different parallel job stages such as investigate, data rules, join, Merge, Lookup, Remove Duplicates, Filter, Dataset, Change Data Capture, Slowly Changing Dimensions and Aggregator.
  • Created sequences and implemented error logging for rejected rows and business rule failure. The rejected files had a mechanism to trigger emails to support and business teams explaining the reason for rejects and the count of rejected records.
  • Worked with the mainframe team to set up CA7 batch jobs to trigger the DataStage flows.
  • Developed LINUX shell scripts for handling the SFTP between DataStage server and third-party servers, wrapper scripts to trigger DataStage jobs when parameters are passed from CA7.
  • Developed test cases for Unit Testing and Integration Testing and coordinated with DataStage Administrators to move the developed code to production.
  • Used Jira to log and track the defects during the testing process.
  • Used GIT to maintain the version of the code .

Environment: IBM Infosphere DataStage/QualityStage 11.5/9.1, Oracle 11g DB2, SQL Server, Linux Shell scripts, SQL Developer, Teradata SQL Assistant, Windows 7, Teradata, Informix, Jira, GIT

Confidential, Bentonville, AR

Lead DataStage and Big Data Developer

Responsibilities:

  • The primary responsibility included, creating DataStage jobs to extract source data from true source systems like DB2, Informix, and Oracle and create load ready files.
  • Identified and documented data sources and transformation rules and also developed the DB column mapping documents.
  • Designed jobs using different parallel job stages such as join, Merge, Lookup, Remove Duplicates, Filter, Dataset, Change Data Capture, Slowly Changing Dimensions and Aggregator.
  • Create shell scripts to perform SFTP, Hive table load and historical (TDCH) data loads.
  • Create HQL scripts to facilitate the Hive load from the load ready files.
  • Create Oozie workflow to facilitate batch loads to Hive tables.
  • Worked on performance tuning of the Hive queries.
  • Worked with the mainframe team to set up CA7 batch jobs to trigger the DataStage flows.
  • Developed test cases for Unit Testing and Integration Testing and coordinated with DataStage Administrators to move the developed code to production.
  • Used Jira to log and track the defects during the testing process.
  • Used GIT to maintain the version of the code .

Environment: IBM Infosphere DataStage 11.5/9.1, Hive, Oozie, MapReduce, TDCH, complex flat files, Oracle 11g DB2, SQL Server, Linux Shell scripts, SQL Developer, Teradata SQL Assistant, Windows 7, Teradata, Informix, Jira, GIT

Confidential, Bentonville, AR

Lead DataStage Developer

Responsibilities:

  • The primary responsibility included, creating DataStage jobs to extract Location Master Data from the MDM DB2 Database, markdowns, GL and Retail balances and inventory tracking data from a legacy system in the form of complex flat files, Compliance data from Oracle, Teradata and Informix tables.
  • Analyzed the various sources and designed and developed parallel jobs for extracting data from different databases such as Teradata, Informix, DB2, Oracle and sequential files, flat files.
  • Identified and documented data sources and transformation rules and also developed the column mapping documents.
  • Designed jobs using different parallel job stages such as join, Merge, Lookup, Remove Duplicates, Filter, Dataset, Change Data Capture, Slowly Changing Dimensions and Aggregator.
  • Created sequences and implemented error logging for rejected rows and business rule failure. The rejected files had a mechanism to trigger emails to support and business teams explaining the reason for rejects and the count of rejected records.
  • Worked with the mainframe team to set up CA7 batch jobs to trigger the DataStage flows.
  • Developed LINUX shell scripts for handling the SFTP between DataStage server and third-party servers, wrapper scripts to trigger DataStage jobs when parameters are passed from CA7.
  • Developed test cases for Unit Testing and Integration Testing and coordinated with DataStage Administrators to move the developed code to production.
  • Used Quality Centre to log and track the defects during the testing process.
  • Used before and after job subroutines in parallel jobs to handle the archival process of the source and target files.

Environment: IBM Infosphere DataStage 9.1, complex flat files, Oracle 11g DB2, PL/SQL, Linux Shell scripts, SQL Developer, Teradata SQL Assistant, Windows 7, Teradata, Informix, HP Quality Center

Confidential, Bentonville, AR

Senior DataStage Developer

Responsibilities:

  • The primary responsibility included, extracting Vendor data from legacy flat files, contract data from OSA DB2 tables and anti-corruption data from third party flat files.
  • Developed DataStage jobs to extract the data from the source systems and load it as is to landing tables in the Informix landing database.
  • Developed complex DataStage jobs to cleanse the data extracted from landing database, apply the business rules and load the data to the Informix staging database.
  • Used quality stage to develop address standardization jobs to convert the addresses coming from the source systems into a standard format before loading to the staging database. Used CASS stage to use the files sent by USPS as a reference for addresses.
  • Designed jobs using different parallel job stages such as join, Merge, Lookup, Remove Duplicates, Sequential, XML, Filter, Dataset, Change Data Capture, Slowly Changing Dimensions and Aggregator.
  • Created sequences and implemented error logging for rejected rows and business rule failure. The rejected files had a mechanism to trigger emails to support and business teams explaining the reason for rejects and the count of rejected records.
  • Worked with the mainframe team to set up CA7 batch jobs to trigger the DataStage flows.
  • Developed UNIX shell scripts for handling the SFTP between DataStage server and third-party servers, wrapper scripts to trigger DataStage jobs when parameters are passed from CA7 and also automate file manipulation and data loading procedures.
  • Developed test cases for Unit Testing and Integration Testing and coordinated with DataStage Administrators to move the developed code to production.
  • Used Quality stages built in wizards for parsing and removing duplicate records.
  • Used Quality Centre to log and track the defects during the testing process.

Environment: IBM Infosphere DataStage 8.5, Quality Stage, CASS files, XML Files and Stages, complex flat files, Oracle 10g/11g, DB2, PL/SQL, UNIX Shell scripts, SQL Developer, Teradata SQL Assistant, Windows 7, Informix, HP Quality Center

We'd love your feedback!