Sr. Etl / Datastage Developer Resume
Chicago, IL
SUMMARY
- 8 years of experience in building and integrating Large Enterprise Data Warehouse and Data Processing Pipelines.
- Experience in Data Warehousing concepts and Principals of Ralph Kimball
- Experience in System Analysis, design, development and implementation of Relational Database and Data Warehousing Systems using IBM Data Stage 11.5/8.1
- Experienced in Database programming for Data Warehouses (Schemas), proficient in Dimensional Modeling (Star Schema modeling, and Snowflake modeling)
- Experience in both Structured and Object - Oriented Programming Languages such as UNIX Shell Scripting, PL/SQL, Java and Python
- Experience in handling multiple Relational Database Management Databases (RDBMS) Oracle, SQL Server, My SQL, Teradata
- Excellent Experience in Designing, Developing, Documenting, Testing of ETL jobs and mappings in Server and Parallel jobs using Data Stage to populate tables in Data Warehouse and Data marts
- Expert in designing Parallel jobs using various stages like Join, Merge, Lookup, Remove duplicates, Filter, Dataset, Lookup file set, Complex flat file (CFF), Modify, Aggregator, XML
- Expert in designing Server jobs using various types of stages like Sequential file, ODBC, Hashed file, Aggregator, Transformer, Sort, Link Partitioner and Link Collector
- Familiar in using highly scalable parallel processing infrastructure using parallel jobs and multiple node configuration files (APT)
- Experienced in integration of various data sources (DB2-UDB, SQL Server, Oracle, Teradata, Netezza, XML and MS-Access, SAS, HDFS and JSON) into data staging area
- Experience in Plug-in Teradata stages like TeraData Load and TeraData Multiload
- Expert in Data Warehousing techniques for Data Cleansing, Slowly Changing Dimension Phenomenon(SCD),Surrogate Key Assignment and CDC(Change Data Capture)
- Experience in implementing Datastage Partition Hash, Modulus, Random, Range, Round Robin and Same
- Expert in working wif Data Stage Manager, Designer, Administrator, and Director
- Experience in analyzing the data generated by the business process, defining the granularity, source to target mapping of the data elements
- Proven track record in troubleshooting of Data Stage jobs and addressing production issues like performance tuning and enhancement
- Experience in writing complex Oracle PL/SQL Packages, Stored Procedures, Functions, Triggers and Collections
- Experience in building complex Analytical SQLs and performed data analysis
- Experience IBM High Performance Data Warehouse appliance “Netezza”, NZSQL, NZLOAD, NZMIGRATE, Zone Map, Distribution Key Handling, and Optimizing Query Plan
- Experience in batch job management in multiple batch job schedulers such as Autosys, Control-M and TIBCO Ops Console
- Experience in Agile methodology and all Agile Ceremonies including SOS, Daily Standup, Sprint Planning, Retrospective
- Worked wif cross functional teams like Architecture, Development, Production and Operation
- Experience in working wif offshore and onshore team model
- Ability to grasp any new technology and languages very quickly
TECHNICAL SKILLS
Languages: JAVA and Python
Big Data Eco Systems: Hadoop, HDFS, Spark, PIG, HBase, Hive, Sqoop, Zoo Keeper, Oozie, Kafka
ETL Tools: IBM Infosphere Datastage 11.5, 8.1, ODI, OBIEE
Relational Databases: Oracle, My SQL, SQL Server, Netezza, Teradata, DB2, MS Access
No SQL Databases: HBase
Scripting Languages: UNIX Shell Scripting, SQL, NZSQL, PL/SQL, SAS
Front End tools: XML, HTML, Ajax, Jquery and JSON
Tools: Eclipse, IDLE, SQL Developer, DB Visuablizer, TOAD
Operating Systems: UNIX, Linux, Windows XP and Windows 7
Data Modeling Tools: Erwin and SQL Developer Data Modeler
PROFESSIONAL EXPERIENCE
Confidential, Chicago, IL
Sr. ETL / Datastage Developer
Responsibilities:
- Responsible for gathering business requirements and preparing architecture design documents
- Prepared Data Mapping Documents and Design the ETL jobs based on the DMD wif required Tables
- Involved in design of dimensional data model - Star schema and Snow Flake Schema using Erwin Modeler
- Generating DB scripts from Data modeling tool and Creation of physical tables in DB
- Used Parameter Sets, Environment Variables, Stage Variables and Routines for developing Parameter Driven Jobs
- Developed parallel jobs using various Development/debug stages (Peek stage, Head & Tail Stage, Row generator stage, Column generator stage, Sample Stage) and processing stages (Aggregator, Change Capture, Change Apply, Filter, Sort & Merge, Funnel, Remove Duplicate Stage)
- Used DataStage as an ETL tool to extract data from sources systems, loaded the data into the ORACLE and Netezza database
- Worked SCDs to populate Type I and Type II slowly changing dimension tables from several operational source files
- Extensively worked wif Join, Look up (Normal and Sparse) and Merge stages
- Extensively worked wif sequential file, dataset, file set and look up file set stages
- Extensively used Parallel Stages like Row Generator, Column Generator, Head, and Peek for development and de-bugging purposes
- Extensively worked on Job Sequences to Control the Execution of the job flow using various Activities & Triggers (Conditional and Unconditional) like Job Activity, Wait for file, Email Notification, Sequencer, Exception handler activity and Execute Command
- Collaborated in developing Java Custom Objects to derive the data using Java API
- Implemented multi-node declaration using configuration files (APT Config file) for performance enhancement
- Deployed different partitioning methods like Hash by column, Round Robin, Entire, Modulus, and Range for bulk data loading and for performance improvement
- Involved in creating UNIX shell scripts for database connectivity and executing queries in parallel job execution
- Used the ETL Data Stage Director to schedule and running the jobs, testing and debugging its components & monitoring performance statistics
- Experienced in using SQL *Loader and import utility in TOAD to populate tables in the data warehouse
- Developed complex PL/SQL Stored Procedures using input/output parameters, Cursors, Views, Triggers and complex queries using Temp tables and Joins
- Documented ETL test plans, test cases, test scripts, and validations based on design specifications for unit testing, system testing, functional testing, prepared test data for testing, error handling and analysis
Environment: Datastage 11.5, 8.1 (Designer, Director and Manager), Oracle, Netezza, SQL, SAS, PL/SQL, UNIX Shell Scripting, Sqlplus, Sqlldr, Data Modeling, NZ SQL, NZ Load, NZ Migrate, Zone Map, Putty, WinScp, SQL Developer, DB Visualizer, TOAD, Java, Web Services, HTML 5, CSS, XML & JASON files, Agile Methodology
Confidential, Chicago, IL
ETL/Datastage Developer
Responsibilities:
- Experience in preparing Technical Design Documents for the ETL development
- Analyzed and Designed ETL processes for loading the data in to Data Warehouse
- Mapped the source and target databases by studying the specifications and analyzing the required transforms
- Developed Data stage jobs using different stages like Transformer, Aggregator, Sort, Join, Merge, Lookup, Data Set, Funnel, Remove Duplicates, Copy, Modify, Filter, Change Data Capture, Change Apply, Sample, Surrogate Key, Column Generator, Row Generator
- Implemented Slowly Changing Dimension Phenomenon(SCD),Surrogate key assignment and CDC(Change Data Capture)
- Worked on Complex data coming from Mainframes (EBCIDIC files) and noledge of Job Control Language (JCL)
- Implemented Shared container for multiple jobs and Local containers for same job as per requirements.
- Successfully implemented pipeline and partitioning parallelism techniques and ensured load balancing of data
- Used Parallel Extender for Parallel Processing for improving performance when extracting the data from the sources
- Created hash tables wif referential integrity for faster table look-up and for transforming the data representing valid information
- Converted complex job designs to different job segments and executed through job sequencer for better performance and easy maintenance
- Mapped the source and target databases by studying the specifications and analyzing the required transforms
- Used the Data Stage Director and its run-time engine to schedule running the solution, testing and debugging its components, and monitoring the resulting executable versions on ad hoc or scheduled basis
- Created shell script to run data stage jobs from UNIX and tan schedule this script to run data stage jobs through scheduling tool
Environment: Datastage 8.1 (Designer, Director and Manager), Oracle, Netezza, UNIX Shell Scripting, SQL, PL/SQL, NZ SQL, NZ Load, NZ Migrate, SQL Loader, Putty, WinScp, SQL Developer, DB Visualizer, Java, XML & JASON files, Agile Methodology
Confidential, Chicago, IL
Sr Database Developer
Responsibilities:
- Developing application components in NZSQL, Oracle PL/SQL to support change request and code fix.
- Building Datastage jobs to migrate data from Oracle to Netezza databases.
- Performing complex Data Analysis wif Data Science and providing solutions.
- Building redundant and fail-over application components to enable disaster recovery management.
- Assessing the entire system and performing FMEA to find potential failures.
- Driving System Improvement Plans and bring stability in the platform.
- Handling Incident Management and recovering high severity production issues wifin SLA.
- Providing Root Cause Analysis using standard 5-WHY and Fish-bone Analysis.
- Providing permanent solutions for the high severity incidents through Problem Management.
- Managing project timelines and keeping say-do ratio up.
- Collaborating wif client team and planning low level tasks required for deployment.
Confidential, Chicago, IL
Database Developer
Responsibilities:
- Gathering system requirements to build a connecting system between legacy and enhanced version of application.
- Discussing all system requirements wif architects and users and freeze design approach.
- Developing a connecting system components using Oracle SQL, PL/SQL Stored Procedures and Functions, Oracle Data Integrator (ODI), OBIEE, Shell Scripting.
- Rewriting Store Facts Modeling (SFM) algorithm in SAS SQL dat calculates Store ACV, Selling Sqft, Number of Employee and other metrics for a particular store irrespective of country/region.
- Perform Unit Testing and documenting test cases and test results.
- Stabilizing system performance by looking at Oracle SQL query plan and rewriting SQL query.
- Providing assistance in SQA and UAT testing.
- Planning the cut-over approach to transition into enhanced application from legacy application.
- Supporting post production issues wif high quality recovery and resolution.
- Documenting high level design, low level design, UML diagrams, and unit test results artifacts.