We provide IT Staff Augmentation Services!

Lead Data Warehouse Consultant Resume

5.00/5 (Submit Your Rating)

Jersey City, NJ

SUMMARY:

  • Over 7.3 years of strong experience focusing on ETL, Data Quality, Data warehousing, Data Quality, Data Management, Data Governance, Data modelling, Data integration, Data Migration, Data analysis and Business Intelligence.
  • Complete understanding of Software Development Lifecycle and Core area of experience in to validate end - to-end business scenarios of B2B (Business to Business) applications and specifically Supply Chain
  • Expert in Designing IBM Infosphere datastage parallel jobs using stages like Join, Merge, Lookup, Remove Duplicates, Funnel, Filter, Dataset, Lookup file set, Complex Flat File, change data capture, Modify, Aggregator, XML, FTP, Modify Stage, database connectors, modifying IBM Infosphere datastage configuration file.
  • Expert in designing all type of ETL loads including type-1 and type-2 SCD, CDC and facts.
  • Extensively created several types of ETL load using Informatica and SSIS.
  • Using IBM data quality tools and stage to read data from heterogeneous source system for data profiling, integrated rules analysis and data quality rules for greater validation which reduced the risk of proliferating incorrect information.
  • Implementation of ETL concepts using IBM Infosphere datastage like Reconciliation, Audit Frameworks, Alerts, Email Notifications, data quality, error handling, Performance Tuning of IBM Infosphere datastage jobs and automation of processes using ETL to support integration, critical and high-volume data management with fast turnaround timing.
  • Evaluated vendors and guided recommendation for master data management and commercial data warehouse solutions, ensuring compliance and integrating Value Centric aggregated hub and sales data, sales and marketing activities - including product sales.
  • Extensive experience with T-SQL in constructing Triggers, Tables, implementing stored Procedures, Functions, Views, User and Data Integrity
  • Involved in various projects related to Data Modeling (Relational, dimensional, Star and Snowflake Schema), System/Data Analysis, Design and Development for both OLTP and Data warehousing environments using Erwin tool.
  • Exposure on the Hadoop cluster, Yarn configuration and export/import data from RDBMS, Netezza db. to Hadoop.
  • Extensively used IBM Cognos Framework Manager, Query Studio, Analysis Studio, ad-hoc reporting and advanced report authoring.
  • Involved in data profiling at various levels -providing recommendations on rule development, configuring and code data quality validation rules in designated tools using SQL, IBM Quality stage and IBM Info Analyzer.
  • Extensively worked in Netezza for Data Cleanup procedures, transformations, Scripts, Stored Procedures and execution of test plans for loading the data successfully into the target
  • Experienced with invoking SOAP and REST web services through IBM Infosphere datastage.
  • Experienced in IBM Infosphere datastage Administrator to create/modify/delete projects and cleanup projects and Implemented security among IBM Infosphere datastage users and projects
  • Experienced in loading/reading data from Oracle 10.2 / 9i / 8i, SQL Server 05/08/12/14, PL/SQL, SQL*Plus, different file sources (. xml, csv.txt), Netezza and applications.
  • Expert in data profiling /data analysis for new Source Systems, Business Requirements, Identifying Business Rules for Decision Support Systems and Data Feeds and Solution Scoping
  • Identify, analyze, and interpret trends or patterns in complex data sets
  • Management skills include Team Management, Project Planning, Monitoring, and self-starter with effective communication skills and ability to work independently and as part of a team.
  • Possess strong domain knowledge on Oil and Gas Domain - Supply and Trading and Retail and Finance

TECHNICAL SKILLS:

Technical: ETL, Reporting, Data Mining, Data Warehousing, Data Modeling, Predictive Analysis & Forecasting, Trend Analysis, Data Analysis

ETL Tools: IBM Infosphere datastage 7.5/8.7/9.1/11.3, Informatica 9.1, SSIS

Reporting Tools: IBM Cognos

Data Profiling\ Modeling Tools\Scheduling\Versioning: Information Analyzer, Information Metadata Workbench, Information Server Operations Console Erwin, UC4, Crontab, TFS, SVN, D-Series

Databases: Sql Server 2005/2008/2012 Oracle 9i, 11g,12g, Netezza, Microsoft Access

Languages: Unix, SQL, PL-SQL

Tool: and Utilities: Toad, Putty, Microsoft Visio, Winscp, Db Visualizer, SVN

Operating System: Unix, Windows 7x, AIX

PROFESSIONAL EXPERIENCE:

Confidential, Jersey City, NJ

Lead Data Warehouse Consultant

Responsibilities:

  • BI Lead for the end to end project implementation for building the client repository and trade and asset data into the data warehouse.
  • Creating the ETL mapping documents, report templates and extensively building the complex ETL mappings to load data into data warehouse.
  • Interacting with Business Analyst \Users to understand Business Requirement Documents BRD's and converting Functional Requirements into Technical Specifications
  • Created data quality reports for business to show quality and quantitative trends to define key metrics
  • Identifying load or bottleneck issues, and resolve them from the root cause and direct/guide offshore development team to perform required steps.
  • Performed Source System Data analysis as per the Business Requirement.
  • Implemented controls and audit to track the ETL loads.
  • Using IBM data quality tools to read data from heterogeneous source system for data profiling, integrated rules analysis and data quality rules for greater validation which reduced the risk of proliferating incorrect information.
  • Organizing daily technical discussions with the Onsite team and offshore workstream leads and set expectations for Offshore delivery.
  • As part of data governance and management team, provided and built technical solution using distinct set of BI tools and technologies for complex data-related challenges across a globally diverse organization .
  • Coordination of system/Integration/UAT testing with other teams involved in project and review of test strategy.
  • Filter and “clean” data, and review reports and performance indicators to locate and correct code problems.
  • Worked on IBM Big Data Integrate POC to move the existing datastage system to IBM big Insights Hadoop cluster leveraging the Yarn configuration

Environment: IBM Infosphere datastage 9.1/11.5 Designer, Director, Oracle, Crontab, DB2, SQL Server 14, Oracle 11g, Erwin, Winscp, UC4, DB visualizer, Information Analyzer, Unix - shell scripting, IBM Big Insights

Confidential, TX

Business Intelligence Lead \ Sr. ETL Developer

Responsibilities:

  • Creating high volume (terabytes of data) and complex ETL datastage mappings for data integration of retail and store inventory management, customer data management, feedback management and maintaining employee database into Netezza database.
  • Created high volume complex ETL development using Sql and CTAS, temp table techniques in the Netezza environment for fast loading of data than any other ETL tool.
  • Creating IBM Infosphere datastage reusable shared components, parameter sets to reduce the IBM Infosphere datastage build time and reduce the maintenance cost.
  • Working closely with the business users to understand the requirements and converting them into project level technical capabilities
  • Created all type of ETL jobs loading tables to staging, ODS, Fact and Type -1, 2 SCD’s.
  • E xtensively worked on Job Sequences to Control the Execution of the job flow using various Activities & Triggers (Conditional and Unconditional) like Job Activity, wait for file, Email Notification, Sequencer, Exception handler activity and Execute Command.
  • Used IBM datastage quality stages, Information Analyzer and Information Metadata Management tool to perform standardization ensure consistency, removing data anomalies and spelling errors of the source information on Customer Analytics Data.
  • Performing trend and data analysis on the existing model and jobs to improve the efficiency of the current high-volume loads and bottlenecks.
  • Used IBM Netezza fluid query to offload the data from Netezza to HDFS.
  • Exposure on Hadoop Cluster architecture and monitoring the cluster
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports for the BI team.
  • Created different kind of data quality reports for data validation, data lineage for executives, BA’s to show quality and quantitative trends to define key metrics.
  • Improving the performance of high volume IBM Infosphere datastage jobs by splitting the jobs implementing intermediate datasets, best use of partitioning techniques for each IBM Infosphere datastage stage, maintain load balancing between IBM Infosphere datastage and database and designing the configuration file based on the type of jobs.
  • Reverse engineered the Cognos reports and identified the Data Elements (in the source systems), Dimensions, Facts and Measures required for new enhancements of reports.
  • Conduct Design discussions and meetings to come out with the appropriate Data Mart at the lowest level of grain for each of the Dimensions involved.
  • Designed a STAR schema for the detailed data marts and Plan data marts involving confirmed dimensions.
  • Scheduling IBM Infosphere datastage jobs through automatic scheduler UC4, crontab utility and IBM Infosphere datastage director.

Environment: IBM Infosphere datastage 9.1/11.3 Designer, Director, Oracle, Crontab, DB2, SQL Server 14, Erwin, Aix, Netezza, Winscp, UC4, DB visualizer, Information Analyzer, Informatica, Unix - shell scripting, IBM Big Integrate, Hadoop

Confidential, Houston TX

Sr. Data warehouse Developer

Responsibilities:

  • Extensively created different type of SSIS 2014 jobs and IBM Infosphere datastage jobs to load the data into SQL 2014 data warehouse.
  • Extensively worked on IBM Infosphere datastage jobs for splitting bulk data into subsets and to dynamically distribute to all available processors to achieve best job performance
  • Created ETL jobs in loading/reading data from Oracle 10.2 / 9i / 8i, SQL Server 05/08/12/14, PL/SQL, SQL*Plus, different file sources(.xml,.csv,.txt), Netezza and applications
  • Used DataStage stages namely Hash file, Sequential file, Transformer, Aggregate, Sort, Datasets, Join, Lookup, Change Capture, Funnel, Peek, Row Generator stages in accomplishing the ETL Coding.
  • Implemented multi-node declaration using configuration files (APT Config file) for performance enhancement.
  • Created SCDs to populate Type I and Type II slowly changing dimension tables from several operational sources.
  • Used Quality Stages to coordinate the delivery, consistency, removing data anomalies and spelling errors of the source information
  • Built shell scripts using IBM Infosphere ds jobs commands to capture the job logs, details to perform data lineage and audit.
  • Created new IBM Infosphere datastage jobs replacing high volume old jobs with improved logics and new stages of the tool reducing the cycle load time by 50% and thus users could see the data more frequently and saving the customer more than $500K annually.
  • Collaborated with EDW team in High level design documents for extract, transform, validate and load ETL process data dictionaries, Metadata descriptions, file layouts and flow diagrams
  • Worked on Netezza table and database maintenance and basic administration by running groom and generating statistics commands on table through scripts, nasal and unload utilities to export\import the data from file or external system and to checking long running queries, resource estimation, and priority taken.
  • Used Informatica Power for extraction, transformation and load (ETL) of data in the data warehouse.
  • Created shell scripts to feed data from various sources to ETL JOBS.
  • Documented ETL test plans, test cases, test scripts, and validations based on design specifications for unit testing, system testing, functional testing, prepared test data for testing, error handling and analysis.
  • Experienced in troubleshooting of IBM Infosphere datastage jobs and addressing production issues like performance tuning and fixing the data issues
  • Performing IBM Infosphere datastage administration such as killing jobs or process, creating new project importing exporting IBM Infosphere datastage projects troubleshooting IBM Infosphere datastage logs issues.
  • Interaction with end-users and business analysts to identify and develop business requirements and transform it into technical requirements and ultimately responsible for delivering the solution.

Environment: IBM Infosphere datastage 8.7/9.1, Designer, Director, Oracle 9i,10g, Crontab, SQL Server 14, SSIS, IBM Information analyzer 8.7, IBM Metadata Workbench 8.7, Erwin, IBM Cognos10.1, Unix, Informatica.

Confidential

ETL Lead

Responsibilities:

  • Delivered an Infra Upgrade project - Aimed to IBM Infosphere datastage upgrade from 7.5 to 8.7 and SQL SERVER database from 2005 to 2008R2.
  • Owned customer engagement, product roadmap, requirements specification, development schedule, and release schedule.
  • Created the prototype migration of jobs from IBM Infosphere datastage7.5 to IBM Infosphere datastage 8.7
  • Created the POC from migration of jobs from UNIX to windows in turn added the high value for customer benefit.
  • Verified new deployments and contributed to evolution of various tools to improve data quality
  • Involved in design of dimensional data model - Star schema and Snow Flake Schema
  • Successfully implemented pipeline and partitioning parallelism techniques and ensured load balancing of data in IBM Infosphere datastage jobs
  • Used Administrator to administer the locks on the jobs and other Administration activities for IBM Infosphere datastage Server.

Environment: IBM Infosphere datastage 8.7, Designer, Director, Oracle 9i, 10g, Confab, SQL Server 12, SSIS, UNIX

Confidential

Sr. ETL Developer\Data Analyst

Responsibilities:

  • Created Several IBM Infosphere datastage jobs to populate the data into dimensions and fact tables.
  • Extensively worked on Job Sequences to Control the Execution of the job flow using various Activities & Triggers (Conditional and Unconditional) like Job Activity, Wait for file, Email Notification, Sequencer, Exception handler activity and Execute Command.
  • Designed and created complex source to target mappings using various transformations inclusive of but not limited to Aggregator, Look Up, Joiner, Source Qualifier, Expression, Sequence Generator, and Router Transformations.
  • Extracted data from flat files and then transformed according to the requirement and Loaded into target tables using various stages like sequential file, Look up, Aggregator, Transformer, Join, Remove Duplicates, Change capture data, Sort, Column generators, Funnel and Oracle Enterprise.
  • Acted as customer relations liaison for internal business applications, defining and prioritizing customer needs, developing requirements and design specifications, conducting industry research, and performing usability testing.
  • Built efficient SSIS packages for processing fact and dimension tables with complex Transforms and type 1 and type 2 slowly changing dimensions.
  • Used DTS/SSIS and T-SQL stored procedures to transfer data from OLTP databases to staging area and finally transfer into data marts.
  • Extensive Knowledge on Deploying the RTI jobs as Web service using the RTI console.
  • Manually modified the SQL in Report Studio to tune and/or to write complicated reports. Used union/Join objects in Report Studio
  • Involved in testing of jobs and creating test cases to ensure proper functionality of production interfaces.
  • Involved in change of existing IBM Infosphere datastage jobs to improve performance in production environment and ensure data integrity.
  • Fine Tuning PL/SQL queries, views and stored procedures and Creating the Backup and lookup tables to handle Historical data and Id problems
  • Necessary changes in the DWH structure or Unix shell Scripts or PLSQL codes to satisfy the requirement
  • Involved in database modeling to improve the performance of database.
  • Fabricated all aspects of project documentation, including functional, technical specifications, testing templates, ETL mapping, report layout
  • Helped in creating a script to capture the logs of IBM Infosphere datastage jobs in table since logs get purged and helped in lot of time saving for researching issues and idea was very well appreciated by customer.
  • Repartitioned job flow by determining IBM Infosphere datastage PX best available resource consumption

Environment: IBM Infosphere datastage 7.5,8.1,8.7, Designer, Director, Oracle 9i, Crontab, SQL Server 2005/08/12, SSIS, IBM Information analyzer 8.7, IBM Metadata Workbench 8.7, Erwin, IBM Cognos10.1, Netezza, Informatica 9.1, Unix

We'd love your feedback!