We provide IT Staff Augmentation Services!

Sr. Etl Datastage Developer Resume

3.00/5 (Submit Your Rating)

Chicago, IL

PROFESSIONAL SKILLS

  • Over 5+ Years of experience in IT industry with 4+ years of experience in designing, developing, testing, enhancement, production support and maintenance of Data Warehouse applications using IBM WebSphere /Infosphere Data Stage and Information analyzer.
  • Good Knowledge about the principles of DW like Data marts, OLTP, OLAP, Dimensional Modeling, fact tables, dimension tables and star/snowflake schema modeling.
  • Excellent in using highly scalable parallel processing infrastructure using parallel jobs with multi - node configuration files.
  • Participated in discussions with Project Manager, Business Analysts and Team Members on any technical and/or Business Requirement issues.
  • Knowledge of Data Warehouse Architecture and Designing Star Schema, Snow flake Schema, Fact and Dimensional Tables.
  • Knowledge on Job Sequences to control the execution of the job flow using various Activities & Triggers (Conditional and Unconditional) like Job Activity, Email Notification, Sequencer, Routine activity and Exec Command Activities.
  • Designed Mapping documents ETL Architecture documents and specifications.
  • Experienced in scheduling sequence, parallel and server jobs using DataStage Director, UNIX scripts and scheduling tools.
  • Designed and developed parallel job and sequence jobs using DataStage Designer.
  • Experience in writing UNIX Shell scripts for various purposes like file validation, automation of ETL process and job scheduling using CA Work Load Automation (Autosys), Automic and Cronttab.
  • Extensively made use of Parallel stages like Sequential, Aggregator, Head, Tail, Sort, Lookup, Merge, Join, Change Capture, SCD stage, Peek stages, Dataset, Filter, Enterprise database stages in Parallel Extender job.
  • Worked and extracted data from various data sources such as GreenPlum, Sybase IQ, Sybase ASE, Oracle, MS-SQL Server, Teradata, DB2 and Flat files.
  • Good working Knowledge with GreenPlum, SybaseIQ, SybaseASE, Teradata V2R5/V2R6/12.0and Sound knowledge ofOracle 10g/11g, MS SQL Server 2000, DB2 7.0.
  • Knowledge in using PostGreSql, PL/SQL to write stored procedures, functions, and triggers.
  • Extensive database experience using Teradata SQL and fine tuning.
  • Extensive experience in using BTEQ, FLOAD, MLOAD, FEXPORT utilities.
  • Extensive experience in Unit Testing, Functional Testing, System Testing, Integration Testing, Regression Testing, User Acceptance Testing and Performance Testing.
  • Created local and shared containers to facilitate ease and reuse of jobs.
  • Experience in Technical documentation (Source-Target Mapping, ETL Design, Impact Analysis, Production Support Handover)
  • Extensive experience in UNIX Korn shell scripting.
  • Used the Data Stage Director and its run-time engine to schedule running the solution, testing and debugging its components, and monitoring the resulting executable versions (on an ad hoc or scheduled basis).
  • Extensive experience in loading high volume data and performance tuning.
  • Quick learner and adaptive to new and challenging technological environments.

TECHNICAL SKILLS:

IBM Software: IBM DataStage 8.7/8.1/8.0/7.5.3/7.5.2/7.5.1 (Designer, Director, manager, Administrator), Parallel extender, Server Edition Business Glossary, Fast

Track and Information Analyzer: Operating Systems

Languages: Shell Scripting, Teradata BTEQ Scripts, SQL, PL/SQL

Databases: GreenPlum, SybaaseIQ, SybaseASE, Oracle10g/9i/8i/8.0/7.0,DB2UDB7.2/8.1/9.0 and MS SQL Server 2005, 2008, Teradata 13,14

Data Modeling Tools: Erwin 7.0, Microsoft Visio

Applications and Tools: HP Quality Centre, Autosys, Toad, SQL Developer, Putty,PGADMIN

Data Warehousing: Data Marts, OLTP, Normalization, Dimensional Modeling, Star Schema, Snowflake

PROFESSIONAL EXPERIENCE

Confidential, Chicago, IL

Sr. ETL DataStage Developer

Responsibilities:

  • Analyzing the Business requirements and System specifications to understand the Application.
  • Designed the ETL jobs using IBM Infosphere DataStage9.1to Extract, Transform and load the data into staging,ODS and EDW.
  • Designed and developed the ETL jobs usingParallel Editionwhich distributed the incoming data concurrently across all the processors, to achieve the best performance.
  • Designed parallel jobs using stages such asJoin, Merge, Lookup, Remove Duplicates, Copy, Filter, Funnel, Dataset, Lookup, Pivot, and Sort, Surrogate key Generator, Change Data Capture (CDC), Modify, Row Generator and Aggregator.
  • Importing data from source files like flat files using Teradata load utilities likeFastLoad,Multiload, andTPump.
  • Creating adhoc reports by usingFastExportand BTEQ.
  • Designed sequence jobs using the activities such asJob Activity, Nested Condition, Notification Activity, Sequencer Activity, Terminator Activity and Execute Command.
  • Performed theIntegrationandSystem testingon the ETL jobs.
  • Responsible for Tuning Report Queries and ADHOC Queries.
  • Wrote transformations for data conversions into required form based on theclient requirementusing Teradata ETL processes.
  • ExperiencedinTuning SQL Statementsand Procedures for enhancing theload performancein various schemas across databases.Tuning the queriesto improve the performance of the report refresh time.
  • DevelopedMLoad scriptsand shell scripts to move data from source systems to staging and from staging to Data warehouse in batch processing mode.
  • Exported data from Teradata database using TeradataFast Export.
  • AutomatedUnix shell scriptsto verify thecount of records added everyday due to incremental data load for few of the base tables in order to check for the consistency
  • Making modifications as required for reporting process by understanding the existing data model and involved in retrieving data from relational databases.
  • Assisted operation support team for transactional data loads in developingSQL&UNIX scripts.
  • Imported therequired Metadata from heterogeneous sources at the process level.
  • CreatedJob ParametersandEnvironment variablesto run the same job for differentsources and targets.
  • UsedMulti-job-compilerduring deployment of projects.
  • Created Batches (DS job controls) and Sequences to control set of jobs.
  • Used theDirectorto schedule running the job, testing and debugging its components, and monitoring.
  • Created Shared Containers for Re-using the Business functionality.
  • Involved in working with SSA requestor responsibilities which will be assigned for both project and support requests.
  • Managing queries by creating, deleting, modifying, and viewing, enabling and disabling rules.
  • Loading the data into the warehouse from different flat files.
  • Database testing by writing and executingSQLqueries to ensure that data entered has been uploaded correctly into the database.
  • Transfer files over various platforms usingsecure FTPprotocol.
  • Involved in creating Unit test plans for and testing the data for various applications

Environment:IBM WebSphere DataStage 9.x (Administrator, Designer, Director), IBM Information Analyzer 8.1.1a, Teradata 13, Oracle 11g, Microsoft SQL 2005/2008, IBM DB2 9.1, AIX6.0, Oracle 11g, Toad 9.5, shell scripts, PUTTY, WinSCP, ERwin R8.1,Mercury Quality Center,IBM Rational Rose, Autosys.

Confidential, Plano, Texas

Sr. Data Stage Developer

Responsibilities:

  • Worked with the TeradataSQL Assistant to interface with the Teradata.
  • Built tables, views, UPI, NUPI, USI and NUSI.
  • Worked on loading of data from several flat files sources to Staging using TeradataMLOAD (MultiLoad), FLOAD(FastLoad).
  • Developed processes using utilities such as TPT, BTEQ
  • Worked on exporting data to flat files using TeradataFEXPORT.
  • Query optimization (explain plans, collect statistics, Primary and Secondary indexes)
  • Transferred data between heterogeneous sources.
  • Designed and developed Parallel jobs to extract data, clean, transform, and to load the target tables using the DataStage Designer.
  • Designed developed job sequencer to run multiple jobs with dependency and email notifications.
  • Used DataStage Designer for importing the source and target database schemas, importing and exporting jobs/projects, creating new job categories and table definitions.
  • Developed parallel jobs using various Development/debug stages (Peek stage, Row generator stage, Column generator stage, Sample Stage) and processing stages (Aggregator, Change Capture, Change Apply, Filter, Sort & Merge, Funnel, Remove Duplicate Stage).
  • Developed BTEQ scripts for Teradata.
  • Responsible for tuning the performances of DataStage mappings and TeradataBTEQ scripts.
  • Designed and implemented slowly changing dimensions (SCD's) methodologies.
  • Worked with DBAs to tune the performance of the applications and Backups.
  • Involved in developing Unit Test cases for the developed mappings.
  • Extensively worked in the performance tuning of transformations, Sources, Sessions, Mappings and Targets.
  • Used ETL methodology for supporting data extractions, transformation and loading process.
  • Responsible for preparing ETL strategies for extracting data from different data sources like Oracle, Flat file and XML.
  • Involvedin analyzing existing logical and physical data modeling with STAR schema and SNOW FLAKE schema techniques.
  • Involved in the performance tuning of the application through creation of necessary indexes.
  • Prepared Fast Export scripts to generate flat files from Teradata.
  • Worked with UNIX Shell scripts.
  • Performed Unit testing, Integration testing and generated various Test Cases.
  • Performed Data analysis and Data validations.

Confidential, Cincinnati, OH

DataStage Developer

Environment: DataStage 7.5.1 Enterprise Edition/Parallel Extender, Quality Stage, UNIX Shell Scripting, SQL Server 2005, Oracle 10g, Erwin 7.0, MS Access, TOAD, AutoSys.

Responsibilities:

  • Involved in meetings with Team leader, Group Members and Technical Manager regarding any Technical and Business Requirement issues.
  • Used DataStage Designer to develop processes for extracting, cleansing, transforming, integrating, and loading data from Oracle, MS Access and mapped the data into a target business warehouse.
  • Extensive experience in creating and loading Data Warehouse tables like dimensional, fact and aggregate tables using DataStage
  • Developed logical and physical data models, design data access strategies and wrote SQL code to access and manipulate data stored in OLAP databases.
  • Constructed data volumes mostly adhering to Star Schema using strong knowledge of Data Warehousing, Star Schema, Fact and Dimension Tables.
  • Extensively used Pivot stage to pivot the source data to achieve required table structures like converting data from rows into a column.
  • Involved in Designing, Source to Target mappings between sources to operational staging targets. Implemented logic for Slowly Changing Dimensions.
  • Followed the concepts of Naming conventions, parameterizing the variables from Global Environment to stage level.
  • Developed jobs in Parallel Extender using different stages like Transformer, Aggregator, Lookup, Source Dataset, external filter, Row generator and Column Generator.
  • Imported Metadata from Oracle database. Imported metadata definitions into the repository. Exported and imported DataStage components using DataStage Manager.
  • Used Parallel Extender for parallel processing of data extraction and transformation.
  • Extensively worked with various stages of Parallel Extender like Sequential file, Dataset, lookup, peek, transformer, Merge, Aggregator, row generator, surrogate key generator and many more to design jobs and load the data in to Fact and Dimension tables.
  • Worked on DataStage Oracle stages like Oracle Bulk Loader and Oracle OCI with DataStage jobs in loading the data.
  • Monitoring the DataStage jobs on Daily/Weekly/Monthly basis by running the AutoSys Script and made a force start whenever job fails
  • Developed job sequence by identifying independent and dependent flows with proper rollback strategies incorporated in the sequences

We'd love your feedback!