Etl/sr. Datastage Developer Resume
Kent, WA
SUMMARY
- Over 8 years of data warehousing experience in the areas of ETL design and development. Involved in complete Software Development life - cycle (SDLC) of various projects, including requirement gathering, system designing, data modeling, application development, enhancement development, maintenance and production support.
- Extensive knowledge in Development, Analysis and Design of ETL methodologies in all the phases of Agile and Waterfall Data Warehousing life cycle.
- Extensive ETL tool experience using IBM Infosphere DataStage 11.3.1, IBM Infosphere DataStage 9.1, DataStage 8.5, IBM Infosphere DataStage 8.1, Ascential DataStage 7.5.
- Extensively used DataStage Client components like DataStage Administrator, Manager, Designer and Director in Data Warehouse development.
- Strong understanding of the principals of DW using fact tables, dimension tables and Star and snowflake schema modeling.
- Worked extensively with Dimensional modeling, Data migration, Data cleansing, ETL Processes for data warehouses.
- Experience as a Lead Developer and worked with Business and stakeholders on projects to ensure requirements are met.
- Created Functional, Design, Technical, and Mapping specifications and documentation.
- Used Enterprise Edition/Parallel Edition stages like Join, Filter stage, Modify, Copy, CDC, Join, Row Generator and many other stages in accomplishing the ETL Coding
- Extensively worked on Change Data Capture process and Slowly Changing Dimensions techniques to maintain the historical details of the data.
- Hands on experience in Netezza, SQL Server, DB2, Oracle 11g/10g, Teradata, databases.
- Extensively used Sequential files, data sets, XML, Web services, MQ etc., for data extraction and target loads.
- Experience in troubleshooting of jobs and addressing production issues like performance tuning and enhancement.
- Created various parameters and param sets for multi-instance jobs and environmental variables to be used across the project.
- Implemented various Shared container for multiple jobs and Local containers for same job as per requirements.
- Experienced in scheduling server and parallel jobs using scheduler tools like TIDAL, BMC Control M, Autosys.
- Experience in version control tools like TortoiseSVN and Apache Subversion.
- Experience in using UNIX and IBM AIX commands and writing UNIX shell scripts
- Knowledge in using PL/SQL to write stored procedures
- Excellent communication, interpersonal, analytical skills and strong ability to manage and motivate the team.
- Handling changing priorities, tight schedules, and correction of critical changes outside normal working hours.
TECHNICAL SKILLS
ETL Tools: IBM Infosphere DataStage 11.3.1 (Designer, Director, Administrator), IBM Infosphere DataStage 9.1, 8.5, IBM Infosphere DataStage 8.1, IBM Websphere Information Server 8.0, DataStage 7.5.X /6.0
Data Modeling: MS Visio Diagram, IBM Rational Rose, ERwin, embarcadero
Databases: Netezza, SQL Server 2003/2005/2008, IBM DB2/UDB, Teradata, Oracle 11g/10g/9i/8i, Sybase
Tools: Aginity, embarcadero, SQL Server Management Studio, Win SQL, IBM DB2 Control Center, SQL*Loader, TOAD, BMC Control M, TIDAL, TortoiseSVN, Apache Subversion
Operating system: IBM AIX 5.2/4.x, Linux, Windows 2000/2003/XP/2007.
Languages: SQL, PL/SQL, Unix Shell Scripting, XML.
PROFESSIONAL EXPERIENCE
Confidential, Kent, WA
ETL/Sr. DataStage Developer
Responsibilities:
- Analyzing, designing, developing, implementing and maintaining ETL jobs using IBM Infosphere Datastage and Netezza.
- Worked on various enterprise projects such as Network fulfillment, Store Inventory Fulfillment which involves new datamart development.
- Worked as Lead Developer for KBM Demographics Project for Customer Data Enrichment and collaborated with the Marketing team and stakeholders on KBM Project to ensure expectations are managed and ensure requirements are met.
- Created functional and technical design documents, the mapping documents for business requirements
- Worked closely with the CDW project team members and stakeholders for CDW Stabilization project to improve the quality of customer data for business.
- Worked on Change Data Capture process to replicate transactional data from Sterling IBM DB2 database to CDW Netezza database.
- Involved in Datastage version upgrade activities for export and import of all the ETL jobs, setting up the environmental variables, testing jobs, modifying the scheduled batch jobs.
- Prepared the Implementation plan for the code migration to QA/Production and performed end-to-end testing of the flow
- Created ELT jobs using UNIX scripts and standard re-usable pearl scripts.
- Co-ordinated and lead the offshore team. Assisted team developers in debugging and researching the code issues in QA environment.
- Unit testing of all deliverables while coding. Assisting QA team to prepare the test plan, execute and coordinate with team.
- Worked on enhancements, maintenance and support of major BI projects.
- Created jobs to extract ECC XML messages using various stages MQ connector, XML transformation stage to load into Netezza tables.
- Scheduling batch jobs using TIDAL, testing and debugging its components, and monitoring the resulting executable versions
- Ensuring reusability of existing code and documentation. Contribution to new development standards and best practices.
- Regular interaction with client for discussing the updates and work related issues.
- Assisted team members in defining interface requirements as well as analytical reporting requirements SAP BOBJ reporting module.
Environment: IBM Infosphere DataStage 11.3, DataStage 9.1, DataStage 8.5, Netezza, Oracle 11g, SQL Server,XML, Flat files, Aginity, Toad, TIDAL Scheduler, WinSQL, Unix
Confidential, Milwaukee, WI
ETL Analyst
Responsibilities:
- Working with the business users in requirements gathering and solution options development.
- Worked closely with Subject Matter Experts on the Requirement analysis, Source/Target data analysis, functional design documents.
- Project participation, including ETL implementations for major eCommerce - BI project.
- Worked on migration of SAP MDM to IBM MDM project converting data from flat files to XML files.
- Analyzed, designed, developed, implemented and maintained server jobs and Parallel jobs using Enterprise Edition of Datastage.
- Worked on Solution design and ETL development for the proposed functional requirement
- Analyzing the existing system and documenting the current state features.
- Designed the mapping documents between source databases and target databases.
- Created and used Shared container and Local containers for multiple jobs as per requirements.
- Developing parallel jobs and server jobs using various processing stages (Aggregator, Transformer, Lookup, Join, Filter, Pivot & Merge, Funnel, Remove Duplicate Stage) and Development/debug stages (Peek stage, Head & Tail Stage, Row, Column generator stage)
- Used Data Stage Manager to import metadata, new job categories and creating new data elements as well as to export job categories.
- Created several jobs on data conversion from flat files to XML output using XML Output stage.
- Created table definitions using the XSD scheme files.
- Created shell scripts for pre/post processing of the files.
- Used the ETL Data Stage Director to schedule running the solution, testing and debugging its components & monitor the resulting executable versions.
- Debug, test and fix the transformation logic applied in the parallel jobs.
- Orientation toward solution design and structured system development methodologies.
- Monitoring ETL jobs developed using BMC Control M scheduling tool and updating reports.
- Created views in DataMart’s on SQL Server to support the reporting requirements.
- Assisted in creating, implementing, modifying the business simple to complex reports using Business objects reporting module.
Environment: IBM Infosphere DataStage 8.1 (PX & Server), IBM Infosphere DataStage 8.5, Ascential DataStage 7.5, SQL Server 2008, DB2, BMC Control M Batch Scheduler, MS Visio
Confidential, Atlanta, GA
ETL Datastage Developer
Responsibilities:
- Involved in Designing the ETL process to Extract translates and load data from OLTP Oracle database system to Teradata data warehouse.
- Gatheird information from different data warehouse systems and loaded into One Sprint Financial Information System Consolidated model using Fast Load, Fast Export, Multi Load, BTEQ and UNIX Shell Scripts.
- Involved in ETL Production Support, maintenance, troubleshooting, problem fixing and ongoing enhancements to the Data mart.
- Conducted researches and created RCA (Root Cause Analysis) documents on production issues.
- Knowledgeable in the areas of relational database logical design, physical design, and performance tuning of the RDBMS.
- Worked extensively on different types of stages like Sequential file, ODBC, Hashed File, Aggregator, Transformer, Change Capture Stage, Join, Lookup, Merge, Sort and Containers (Shared and Local) for developing job.
- TEMPEffectively used DataStage Manager to Import/Export projects from development server to production server. Parameterized jobs for changing environments.
- Responsible for trouble shooting, identifying and resolving data problems, Worked with analysts to determine data requirements and identify data sources, provide estimates for task duration.
- Optimized performance of Mappings by identifying bottlenecks and eliminating them.
- Created FastLoad, FastExport, MultiLoad, TPump, BTEQ scripts for Financial Information System.
- Used the DataStage Director and its run-time engine to schedule running the solution, testing and debugging its components, and monitoring the resulting executable versions.
Environment: Ascential DataStage 7.5.2, Infosphere Datastage 8.1, 8.5 (Administrator, Manager, Designer, Director, Parallel Extender), Teradata Tools & Utilities (BTEQ, Fast Export, Multi Load, Fast load, SQL Assistant), Oracle 10g, AutoSys, ERwin 3.5, Microsoft SQL Server 2005
Confidential, Southfield, MI
ETL Developer
Responsibilities:
- Worked with the Business analysts and the DBAs for requirements gathering, analysis, testing, and metrics and project coordination.
- Involved in the Dimensional modeling of the Data warehouse.
- Developed documents like Source to Target mapping for developing the ETL jobs.
- Worked with DataStage server stages like OCI, ODBC, Transformer, Hash file, Sequential file, Aggregator, Sort, Merge, Link practitioner, Link collector and other stages.
- Involved in designing various jobs using PX and Parallel jobs using Parallel stages like: Merge, Join, Lookup, Transformer (Parallel), Teradata Enterprise Stage, Funnel, Dataset, etc
- Used Remove Duplicates stage in PX (EE) to remove the duplicates in the data.
- Involved in the migration of DataStage jobs from Development to Production environment.
- Designed and implemented several wrappers to execute the DataStage jobs, create job reports out of the DataStage job execution results from shell scripts.
- Worked on SQL joins, views, aggregate conditions, parsing of objects and hierarchies.
- Tuned SQL queries for better performance for processing business logic in the database.
Environment: DataStage 8.1 (Designer, Manager, Director, Administrator), Oracle 9i, TOAD, SQL/PLSQL, Teradata, Erwin 4.0, UNIX (AIX).
Confidential, Santa Clara, CA
DataStage Developer
Responsibilities:
- Interacted with Business analyst on day-to-day basis. Involved in analyzing different modules of facets system and EDI interfaces to understand the source system and source data.
- Worked on extracting data related to various health programs like Medicare, Medicaid/medi-cal, and other mediclaims.
- Used the DataStage Designer to develop processes for extracting, cleansing, transforming, integrating and loading into Oracle Data Warehouse.
- Worked on Datastage Designer, Manager, Administrator and Director.
- Created re-usable components using shared containers for local use or shared use. Exported the universe to the Repository to make resources available to the users.
- Loaded data into load, staging and lookup tables. Staging area was implemented using flat files.
- Created jobs in DataStage to import data from heterogeneous data sources like Sybase, Oracle 9i, Text files and SQL Server.
- Used stored procedures, function to access data from the database.
- Tuned transformations and jobs for Performance Enhancement.
- Utilized Parallel Edition PX, to optimize performance in this very large database environment.
- Executed Pre and Post session commands on Source and Target database using Shell scripting.
Environment: Datastage 8.1, Parallel Edition PX, SQL Server, Oracle 9i, Sybase, Teradata, Text Files, ERwin, UNIX.
Confidential
Developer/Trainee
Responsibilities:
- Worked together with managers and lead developers in coming up with alternative solutions.
- Assisted in building the ETL source to Target specification documents by understanding the business requirements
- Involved in extracting the data from different data sources like Oracle 8i and flat files.
- Creating ETL Job flow design, creating and maintaining Sequencer and Batch jobs using Datastage ETL tool and performing Unit Testing and Integration
- Used Oracle to write SQL queries dat create/alter/delete tables and to extract the necessary data
- Created Post UNIX scripts to perform operations like gunzip, remove and touch files. Used designer and director to schedules and monitor resulting executable versions.
- Contributed and followed ETL standards and best practices.
Environment: Ascential DataStage 6.0 (Designer, Manager, Director), Oracle 8i, Flat files, UNIX, etc.