Datawarehouse Engineer Resume
Denver, CO
OBJECTIVE:
To obtain a challenging position in enterprise application integration environment where technology integrates with business functionalities to enhance business process; I look forward to the opportunities where I can use my analytical skills in combination with technical expertise to improve all phases of enterprise development, implementation, maintenance and performance tuning life cycle.
SUMMARY:
- Highly skilled IT professional with over 4 years of diversified experience in Software analysis, Design, Implementation and Application development using Ab Initio, Teradata, DB2, Java, Oracle, Netezza, Windows NT and Unix environments with emphasis in Data Warehousing and Business Intelligence
- Over 3 years of experience in developing strategies for ETL (Extraction, Transformation and Loading) mechanism using Ab Initio tool in complex, high volume Data Warehousing projects.
- Strong development skills including the ability to work through the entire development life (SDLC) cycle from gathering requirements through implementation, development, production, support and documentation of the complete project.
- Worked extensively in the Confidential (Graphical Development Environment) configuring, connecting, and testing the components to produce executable flow graphs on UNIX environment, proficient with various Ab Initio Data Cleansing, Parallelism, EME(Enterprise Meta Environment) and Multi File System techniques. .
- Involved in providing production support to various Ab - Initio ETL jobs and Teradata Database programs.
- Involved with Teradata Tuning experts, and provided technical advice and guidance to other application developers on my projects on systems and database programming techniques.
- Experience with Ab Initio Co>Operating System, application tuning, debugging strategies and proficient with various Ab Initio Parallelism and Multi File System technique.
- Implemented dependency analysis using the EME web interface to conduct the project analyses of the dependencies within and between graphs
- Implemented various parallelism techniques (4 way, 8 way, 16 way multi file system) with Partition by Key, Partition by Expression and Round Robin techniques on flat files
- Ability to demonstrate command and use of industry accepted Ab Initio best practices and has completed terabyte level solutions preferably at multiple separate environments.
- Prepared and implemented data verification and testing methods for the Data Warehouse as well as to design and implement data staging methods and stress testing of ETL routines to make sure that they don’t break on heavy loads
- Used Teradata SQL for creating aggregate tables, controls tables, staging tables and base tables in Teradata.
- 24x7 Production Support for ETL jobs for daily, monthly and weekly schedule and also worked with scheduling tools to schedule the ETL batch process.
- Excellent analytical, functional, design and development skills with excellent communication skills, both written and verbal.
TECHNICAL SKILLS:
ETL Tools: Ab-Initio ( Confidential 1.15,Co>Operating System 2.15, EME ), Informatica
BI Tools: Hyperion Intelligence(Brio) 8.2, Microstrategy 8.0, Business Objects 5.1
Database: Teradata V2R6, MS Access, Oracle 10g/9i/8i, DB2 9.0, Netezza 3.0 and SQL Server
Operating Systems: MS-DOS, UNIX (Sun Solaris 8.0, IBM AIX), Linux, Windows XP/2000/NT
Languages: Java, C, C++, Perl, SQL, Unix Shell scripting, HTML, XML, Java script
Scheduling tools: Maestro (Tivoli Work Load scheduler 8.3), Control-M, Autosys
EMPLOYMENT HISTORY:
Confidential, DENVER, CO
Datawarehouse Engineer
Responsibilities:
- Closely worked with data modelers to assist them in framing the business rules and help them in preparing the mapping documents for ETL process
- Responsible to configure the db configuration file (dbc files) to connect to the Oracle and Netezza database to load into data warehouse and as well as various data marts.
- Developed Netezza best practices in loading data into dimensions and fact tables on a daily basis.
- Assisted in SQL tuning for user based ad hoc queries.
- Large scale parallel processing using Ab Initio and Oracle PL/SQL
- Implemented various parallelism techniques (4 way, 8 way, 16 way multi file system) using Ab Initio in ETL environment as well as Oracle parallelism techniques to achieve best performance.
- 24x7 Production Support for ETL jobs for daily, monthly and weekly schedules and also worked with Maestro scheduling tools to schedule the ETL batch process.
Environment: Ab-Initio CO>OS 2.15, Confidential 1.15, AIX Unix, Oracle 10g, Netezza 3.0, Clear Case, Clear Quest, Maestro Scheduler, Microstrategy 8, Windows NT/2000.
Confidential, COLUMBUS, OH
ETL Developer
Responsibilities:
- Involved in analyzing the business requirements mentioned in the business requirements document (BRD) with the Business Analysts and create a detailed design in my technical requirements document (TRD) and develop complex abinitio graphs.
- Closely worked with the business analysts to assist them in framing the business rules and help them in preparing the mapping documents.
- Implemented various parallelism techniques (4 way, 8 way, 16 way multi file system) with Partition by Key, Partition by Expression and Round Robin techniques on flat files in ETL environment.
- Involved in coding complex Ab Initio transformation rules to derive new fields and modify logic for existing fields in various data marts
- Worked heavily on the Confidential build common components for building Ab Initio graphs during the data integration process
- Performed the tuning of ETL process by means of maximizing the use of MFS, using lookup file instead of join if possible, removing unnecessary phasing, removing unnecessary components, making use of the indexes on the tables, etc. and thus to achieve an exceptional reduction of the total process time.
- Developed various Ab Initio graphs for data cleansing using Ab Initio function like is valid, is defined, is error, is defined, sting substring, string concat and other string * functions .
- Developed applications to migrate, maintain, and monitor data between existing applications and the data warehouse.
- 24x7 Production Support for ETL jobs for daily, monthly and weekly schedules and also worked with Maestro scheduling tools to schedule the ETL batch process.
- Involved in unit testing and integration testing of the developed work,analyzed production support issues and diagnose problems
- Extensively worked under the Unix Environment using Korn Shell Scripts
- Heavily worked on the brio reporting tool to query, pivot, and chart and create dashboards for data exploration and decision making in Credit data mart and Credit Origination data mart on a monthly and daily basis
- Provide technical support and assistance to other ETL developers in Confidential credit team and business user communities in various line of businesses of JP Morgan Chase
Environment: Ab-Initio CO>OS 2.13, Confidential 1.13, Informatica, AIX Unix, DB2 9.0, Polytron Version Control System (PVCS), Concurrent Version System (CVS), Maestro Scheduler (TWS 8.3), Hyperion Intelligence(Brio) 8.2, Windows NT/2000.
Confidential, charlotte, NC
ETL Developer
Responsibilities:
- Involved in analyzing the business requirements mentioned in the BRD (Business Requirements Document) with the Business Analysts and create a detailed design in my TRD (Technical Requirements Document) and develop abinitio graphs.
- Instrumental in developing and incorporating framework for Ab Initio deployment that allows specification, validation, translation and transformation of data/documents.
- Implemented an 8 way multifile system that is composed of individual files on different nodes that are partitioned and stored in distributed directories (using Multidirectories).
- Responsible for creating Common components in Ab Initio which are useful for the plugging into the code in development environment.
- Implemented Data Parallelism through graphs, which deals with data, divided into segments and operates on each segment simultaneously through the Ab-Initio partition components to segment data.
- Improved the overall performance of response time by utilizing Ab Initio data parallelism and modifying component parameters.
- Implemented the parallel application by replicating the components-data sets and processing modules into number of partitions.
- Involved in all phases and created various technical and functional design documents for Unit testing, Integration testing, Component level testing and detail design.
- Involved in writing wrapper scripts responsible for the automation of Ab Initio graphs using korn shell scripts created Ab Initio Data flow processes for the loading data in Teradata database from various source systems.
- Performance tuning for Data warehouse Database (Teradata) and Data warehouse creating ETL Scripts & procedures extract data and populate the data warehouse using Ab Initio tool.
- Responsible for creating parameter sets for running the load jobs for clean loading of the both the staging and the base tables in Teradata database.
- Implemented dependency analysis using the EME web interface to conduct the project analyses of the dependencies within and between graphs
- Heavily used the Ab Initio m command and air commands to interact with co-operating system and EME.
- Responsible to create different Generic graphs using the Ab-Initio components for development environment.
- Interacting with end users, to get the business rule modifications in design, development and testing phases.
- Deployed and ran Ab initio graphs as executable korn shell scripts in the development, testing, pre-production and production envirornments.
Environment: Ab-Initio CO>OS 2.14, Confidential 1.14, EME, AIX Unix, Sun Solaris, Teradata V2R6, Oracle 11i/10g, DB2, Informix, Sun Solaris, Test Director, Quality Center, Autosys, CA Harvest 5.2.1,Windows NT/2000.
Confidential, FOSTER CITY, CA
ETL Developer
Responsibilities:
- Involved in understanding the Requirements of the End Users/Business Analysts and Developed Strategies for ETL processes
- Used the following components of Ab Initio in creating graphs. Dataset components (Input file, output file, lookup file, and intermediate file), Database components (Input table, output table, RunSql, Truncated Table), Transform Components (Aggregate, Dedup Sorted, Filter by Expression, Join, Normalize, Reformat, Rollup and Scan Components), Partitioning Components (Broad Cast, partition by expression, partition by key, partition by round robin), Gather logs, Redefine format, Replicate, Run program components
- Extensively used the Ab Initio tools feature of Component, Data and Pipeline parallelism
- Configured the source and target database connections using .dbc files
- Heavily used Ab Initio components such as Read XML, Write XML and XML Reformat to operate on XML data in graphs.
- Created checkpoints, phases to avoid dead locks and tested the graphs with some sample data then committed the graphs and related files into Repository from sandbox environment.
- Worked heavily with various built-in transform components to solve the slowly changing dimensional problems and creating process flow graphs using Ab Initio Confidential and Co>Operating System
- Improved the overall performance of response time by utilizing data parallelism and modifying component parameters.
- Extensively worked under the Unix Environment using Shell Scripts
Environment: Ab Initio ( Confidential 1.13, Co>Operating System 2.13, EME), UNIX (Sun Solaris), Control-M, Micro strategy 7.0, Windows NT, Oracle 9i
Confidential, Cincinatti, OH
ETL Developer
Responsibilities:
- Responsible to create different Generic Graphs using the Ab-Initio components for ODS and DSS environment.
- Implemented Ab-Initio configuration and well tuned setting up environment for Production and development environment.
- Responsible for designing Parallel Partition ( Confidential bound) Ab-Initio Graph for high volume data warehouse for 24-month warehouse.
- Database Query Optimization and I/O tuning techniques have been used for performance enhancements.
- Implemented Data Parallelism through graphs, which deals with data, divided into segments and operates on each segment simultaneously through the Ab-Initio partition components to segment data.
- Implemented a 8 way multifile system that is composed of individual files on different nodes that are partitioned and stored in distributed directories
- Responsible to configure the db configuration file to connect to the ODS and DSS to load the data.
- Responsible for deploying Ab-Initio graphs and running them through the Cooperating systems mp shell command language.
- Responsible for creating test plans for validating analysis results.
- Involved in gather business requirements from end user and translating them into technical documents
- Working with the testing team to help them in running some complex queries against the Data warehouse.
- Involved in 24x7 production support for ETL jobs for daily, monthly and weekly schedule using Maestro Scheduling tool.
Environment: Ab-Initio CO>OS 2.1x, Confidential 1.1x, EME, AIX UNIX, PL/SQL, SQL server 2000,Oracle 8i/9i, DB2, MVS, Erwin 3.5.2, Sun Solaris, Windows NT/2000.