We provide IT Staff Augmentation Services!

Data Warehouse Architect/etl Engineer Resume

4.00/5 (Submit Your Rating)

New York, NY

SUMMARY:

  • Over 13 yrs IT experience starting with Mainframes, Oracle DBA, ETL, Reporting and Data warehousing. Extensive experience in Architecting and Developing Data Warehousing projects with proof for ROI outcomes. Data
  • Requirement Gathering in terms of Desired outcomes/performance objectives
  • Bridging gap between Business owners/drivers and technical teams
  • Identifying bottlenecks in processes, tools and technologies
  • Setting up BI Systems sizing, installation and configuration
  • Defining metadata driven BI integration
  • Requirements - test cases - sample data/outcomes - source data model -
  • staging data model - target data model - ETL dependencies -
  • reporting dependencies - ETL optimization - BI optimization -
  • BI feedback and reduction - ETL feedback and reduction - Redefining staging and target data models - Redefining requirements and providing proof for business outcomes
  • Optimization of resource utilization wrt Operating System, ETL, Databases, Reporting & BI and Batch window availability.
  • Metadata Analysis apart from Master Data Management.
  • Data model optimization
  • Integration of multiple ETL tools with a variety of Reporting tools.
  • Optimizing ER & Kimball Models for specific domain requirements.
  • Specialization in open source and enterprise ETL tools wrt dynamic processing and scheduling.
  • Performance Testing, Grid Optimization and Performance Benchmarking.
  • Dynamic SQL scripting and data generation, Data partitioning and archival.

TECHNICAL SKILLS

Oracle Database: I worked as an Oracle DBA during the initial phase of my IT career. This has provided me a strong understanding of data, databases and data models.

Microsoft Excel: Simplicity and power in one go. I have used excel to solve Sudoku, magic squares and mockups for applications. Macros, VB Scripts, Pivot tables & charts, Formulas and references.

Qlikview: In-memory reporting tool, great for mockups of Data warehouses since it combines Data modeling, ETL, Reporting and Analysis in the fastest possible way.

Tableau: In Memory reporting similar to Qlikview but more closer to enterprise reporting tools like BO, Cognos.

Pentaho BI Suite: Data Integration and Analysis. Being open source its metadata is readily available and can be harnessed for dynamic development. Used it extensively to create product data warehouses which can respond dynamically to changing business requirements.

PL/SQL: Expert in writing queries and processing data. Most ETL components are extensions PL/SQL and Reporting tools are extensions of Spool.

ETL Tools: Ab-Initio Confidential Datastage Confidential Infosphere Streams Informatica Kettle Oracle Data integrator Talend Microsoft SSIS

Reporting/BI Tools: Business Objects Google Analytics Confidential Cognos Jasper Pentaho BI Suite Qlikview Tableau Microsoft SSRS Crystal Reports

Databases: Oracle DB2/ IDMS Sqlserver

Programming: Cobol JCL CICS Shell scripting PL/SQL Perl Python VB

PROFESSIONAL EXPERIENCE:

Confidential, New York, NY

Data Warehouse Architect/ETL Engineer

Responsibilities:

  • Requirement Gathering from the Business Users in Advertizing and Reporting
  • Data warehouse Modeling using DBWrench to create source data model, and Star schema for Data warehouse.
  • Defining Source-target mappings for ETL, dependency matrix for ETLs and Reporting.
  • Creating partitions/indexes for optimized reporting access. Defining strategies for scheduling Pentaho Mondrian cubes.
  • Defining Data lineage and reprocessing logic.
  • Creating master metadata for ETL and reporting.
  • Creating Master ETL scripts Jobs/transformations in Pentaho and informatica.
  • Defining and creating logging and alerts.
  • Optimizing scheduling dependencies by identifying critical path and dynamically triggering parallel threads based on system resource utilization.
  • Incorporating reconciliation using Google Analytics and source target data profiling.
  • Dynamic ETL using Sugar CRM metadata to keep pace with the numerous changes on the CRM Analytics and ensuing that the ETL development is no longer a bottleneck for business users.
  • Creating dashboards in Pentaho BI for business health check.
  • Incorporating data from Redshift and web logs in hadoop.
  • Production support.

Environment: Pentaho BI Suite, Informatica, DBWrench, Unix, Oracle

Confidential, Houston, TX

Product Technical Architect

Responsibilities:

  • Source System Analysis The Source systems was mainly oracle data bases along with Flat-Files or XML feeds from Non-Finacle Systems. The systems were analyzed to bring all data sources to a de-normalized format to be processed in a generic format that can be loaded to target Dimensions and Facts using a Pre-staging and Staging Database. Source-Target Mappings for 200 targets.
  • Target Data Warehouse Design The Data warehouse was designed using Ralph Kimball Star Schema with minor snow-flaking and Global Dimensions. The Tables were optimized for downstream reporting.
  • Designing ETL Architecture Implementing the Source-Target Mapping using Transformations created in Datastage and Pentaho. Implementing Dependencies, SCD, Partitioning using ETL and Oracle, Scheduling Optimization, generating Hashkeys for data lineage, etc
  • Reporting Design Designing 50 off the shelf reports as a product. These reports were created in Qlikview and Tableau format for easier deployment.
  • Shell Scripting to control ETL Steps and Reporting Refresh, monitoring and feedback to support teams.
  • Performance benchmarking Using bulk generation of data all sources were populated and product tested for performance using Confidential Labs and analyzing Oracle AWR and nmon reports
  • A key feature of this product was Dynamic ETL using Product metadata to create ETL and all objects dynamically using semantic layers.
  • Presenting Solutions to Clients and incorporating new features.

Environment: Confidential Infosphere Information Server, Pentaho BI Suite, Informatica, Erwin, Unix, Oracle, Qlikview, Jaspersoft, Rational Clearcase

Confidential, San Jose, CA

Data Warehouse Architect

Responsibilities:

  • Source System Analysis The Source systems consisted of many relational databases, equipment generated log files, XML files. Typically any equipment connected to the network creates at least 2-5 types of data having different metadata. The systems were analyzed to bring all data sources to a pre-staging database.
  • Target Data Warehouse Design The Data warehouse was designed using Ralph Kimball Star Schema with major snow-flaking and Global Dimensions. The Tables were optimized for downstream reporting.
  • Designing ETL Architecture Implementing the Source-Target Mapping using Transformations created in Datastage, Pentaho, PERL & Python. Implementing Dependencies, SCD, Partitioning using ETL and Oracle
  • Reporting Design Designing 100 off the shelf reports as a product in Cognos. These reports were later created in Qlikview and Tableau format for easier deployment.
  • Shell Scripting to control ETL Steps and Reporting Refresh, monitoring and feedback to support teams.
  • KPI incorporation The telecom domain has 1000s of Key Performance Indicators and all of them cannot be shipped as a part of the product. An Application interface was created to provide drag and drop framework which enabled KPIs to be directly created as ETL components. This enabled in-memory Reporting from QlikView and Tableau to directly connect to Datawarehouse bypassing the Cognos Framework.
  • A key feature of this product was Excel Source-Target Mapping being directly used to create ETL Source Target Mappings using semantic layers.
  • Presenting Product Demo to Clients and incorporating new features.

Environment: Confidential Infosphere Information Server, Confidential Cognos, Oracle, Erwin, Unix, Tableau, Rational Clearcase

Confidential

Data Warehouse Architect

Responsibilities:

  • Datamart Design A couple of datamarts were designed as Star Schema and optimized for ETL and Reporting
  • Dynamic SQLs generation process used Metadata directly from ETL requirements. This was optimized to reduce the development efforts considerably.

Environment: Microsoft SSIS/SSRS Unix, Sqlserver, Erwin

Confidential, Austin, Tx

Data Warehouse Architect

Responsibilities:

  • Requirement Analysis Analyze site analytics and performance data. Defining site performance metrics. Comparing multiple ETL and reporting tools.
  • Data Warehouse Design Creating Data Warehouse Star Schema Model and OLAP Cubes for Reporting
  • ETL Design Creating POC ETL components in Pentaho Kettle
  • OLAP Cubes Design Creating OLAP cubes and Refresh Strategy
  • Reports and Dashboard Design In-Memory Reporting tools Qlikview and Tableau were used to Design 20+ Reports and 5 Dashboard including Reports Bursting and Email delivery
  • Requirement Analysis and Database design Reconciliation requirements were translated into Database design using Data from Flat files and ERCOT Oracle databases Most of the Reporting Requirements were re-conciliation based data from Retail usage and ERCOT.
  • Two datamarts were created with 40-50 tables and around 50 PL/SQL packages.
  • Packages were tuned re-cursively when data volumes increased to more than 10 million records/day.
  • Datastage ETL was used to connect to flat files and secondary Data-sources

Environment: Erwin, Confidential information server, Oracle, unix

Confidential

Oracle DBA

Responsibilities:

  • Database Installation and Configuration for Manugistics
  • Database Support Post installation, monitoring and single point reconciliation of all databases, troubleshooting issues using Enterprise Manager, Statspack, AWR reports
  • Optimizing Databases Processes like Indexing, Partitioning, Defragmentation, SQLs tuning, DB tuning.
  • Critical Production Support
  • Requirement Analysis Data Sources in Flat-File, SAP-IDOC, EDIFACTs, SWIFT, XML, etc. Message broker service included configuring protocols like FTP, SFTP, MQ Series, etc. This data was loaded into Sybase Db and finally delivered via other broker services.
  • ETL Development Runtime Maps were created using AMTrix EAI tool
  • Programming/Testing Developing and Maintaining COBOL Programs and JCLs
  • GUI Development Developing and Maintaining IDMS-ADSO and DB2-CICS Applications

Environment: Confidential Mainframes, CICS, ADSO, JCL, IDMS, DB2

We'd love your feedback!