Lead Data Warehouse Consultant / Data Architect /big Data Consultant Resume
Austin, TX
PROFESSIONAL SUMMARY:
- I.T Professional with around 10 years of experience in application, designing, development and implementation of ETL/Data warehouse and RDBMS projects for Oil & Gas, Retail, IT services industries.
- Expertise in data warehousing, ETL architecture, Data Profiling using Informatica PowerCenter/ Informatica Data Quality 9.6/91/8.6/8.5/8.1/7.1 - Client and Server tools and building Design and building Enterprise Data warehouse/Data Marts.
- Adept at understanding business processes / requirements and implementing them through mappings and transformations.
- Experience with Cloud Data bases (Azure SQL Data warehouse, Azure SQL Database) and integrating it Hadoop HDI cluster.
- Experience in Database design, entity relationship modeling and dimensional modeling using Star and Snowflake schemas.
- Extensively worked with mappings using different transformations like Filter, Joiner, Router, Source Qualifier, Expression, Union, Update Strategy, Unconnected / Connected Lookup, Aggregator and SCD Type-2.
- Experience in tuning Mappings and Sessions for better Performance.
- Experience in loading various data sources like Oracle, SQL Server, Teradata, DB2 and Flat Files into Datamarts.
- Experience with various ERP Systems SAP s/4 hana & Epicor.
- Good working knowledge on Informatica Big Data Edition, loading data from various sources to Hadoop File Systems(HDFS) (Environment: - Hortonworks/Cloudera)
- Worked in Production support team for maintaining the mappings, sessions and workflows to load the data in Data warehouse.
- Good knowledge on Data stage. Working knowledge of Designer, director & Manager.
- Working knowledge of creating data stage jobs, scheduling & monitoring.
- Used complex SQL to validate/reconcile the data including the business scenarios.
- Experience in Performance Tuning and Debugging of existing ETL processes.
- Experience in reverse engineering the existing ETL (Informatica)/PL/SQL code for documentation.
- Experience in preparing and scheduling and/or running Sessions/tasks, Work Flows and Batch processes using Workflow Manager or PMCMD command
- Experience in Oracle 10g/9i/8i.
- Experience in writing Triggers, Stored Procedures, Functions, and Packages etc using PL/SQL.
- Experience in UNIX Shell Scripting.
TECHNICAL SKILLS:
ETL Tools: Informatica PowerCenter/PowerMart/Power Designer/Real Time 9.6/9.5/9.1.1/8.6.1/7.1 , ILM
Databases: Oracle 11g/10g/9i/8i/7.3, MS SQL Server 2000/2005; TeraData V12/V2R4, DB2, MySql, MS Access, Office, Netezza
Modelling Tools: Erwin
Programming Skills: SQL,T-SQL, PL/SQL, UNIX shell scripting
Tools/Utilities: TOAD 7.x/8.x/9.x, Win SQL, SQL Developer, Winscp, Business Objects 5.x/ 6.x,Tidal, PuTTY, WinSCP.
Scripts: Perl, Unix Shell Scripting, Java Scripts .
PROFESSIONAL EXPERIENCE:
Lead Data warehouse Consultant / Data Architect /Big Data Consultant
Confidential, Austin, TX
Responsibilities:
- Working with Finance team on the reporting requirements and enhancements to the data warehouse.
- Design the ETL architecture for the conversion of source from legacy to new ERP (SAP).
- Working with SAP team to build Confidential in Azure SQL data warehouse.
- Build & modify the Data warehouse with the new source system (SAP s/4 hana) data structures.
- Data analysis of legacy ERP (Epicor) and new ERP (SAP) for efficient building of the data warehouse.
- Created procedures to load the data from staging (external tables) to the data mart.
- Provide production support for the existing Data lake jobs and Data warehouse jobs.
- Built Sqoop scripts to extract data from the SAP.
- Created hive external and internal tables in the data lake.
- Reconciled the data between data lake and data warehouse.
- Design, modelled & built data mart for sales & order report.
- Responsible for data integration (Epicor and SAP) and data management.
- Scheduled jobs using Data factory to move data from Data lake to Azure SQL DW(MPP).
- Created stored procs, functions to build Dimensions & FACTs (T-SQL).
- Wrote SQL queries for the data validation between Data Lake and DW Staging layer.
- Created hive external and internal tables in the data lake.
- Created DDL for Staging and Data mart tables in SQL.
- Created partitioned snapshots fact tables in Azure SQL DW for tracking the order walk.
- Created and maintained databases, tables, views, users, logins & indexes using T-SQL.
- Direct repots to VP of IT.
- Used SQL analytical and window functions for grouping based on business logic.
- Built Indexes and partitions for better query performance on Azure SQL DW.
- Design and implemented the logic for cut-over strategy in data warehouse (for migration day).
- Created partitioned snapshots fact tables in Azure SQL DW for tracking the order walk.
- Created and maintained databases, tables, views, users, logins & indexes using T-SQL.
- Validated DW data with standard SAP reports (VA05, VL060, ZSHIP PENDING..)
- Managed Azure portal subscriptions.
- Created Azure data factory jobs for scheduling and migrating data.
- Manage a team of 4.
- Worked with Tableau team to get the dashboards done and customized as per business needs (Custom SQL, multiple data sources).
Environment:- Azure SQL Data warehouse(Cloud), Hadoop HDI 3.5, Hive, Sqoop, SQL Server, SAP, Microsoft Azure, Data warehouse, Azure Data factory, Tableau.
Lead ETL Developer/ Data Modeler/ Sr. Data Analyst
Confidential, Houston, TX
Responsibilities:
- Created mappings to move the historical data from ODS to EDW, and generated the purge mapping to purge the historical data from ODS (non-current records) to improve performance and cleanup the ODS.
- Informatica object metadata search in the Informatica repository (using metadata queries) before decommissioning or to find out the dependencies in the DWH.
- Fine-tuned the complex SQL package (Deal Audit entity) to improve performance, time and cost of the process.
- Created reconciliation reports to reconcile data from Endure to ODS load for price value entity.
- Involved in cost analysis (explain plan) of complex queries, gathering stats and copying baseline plan from one to other DB environments.
- Reverse engineering existing data model and made changes as per the user needs using ERwin.
- Reverse engineering complex ETL mappings in-order to document and map the existing production logic to the new systems.
- Generated First of the month reports for (Power, GAS & NGL) using SQL and Informatica.
- Involved in mentoring the new joiner in the team.
- Fine-tuned existing (ETL logic/SQL code) of load from ODS to EDW. Simplified the process, improved the performance by eliminating the source/target bottlenecks.
- Develop the reconciliation reports to reconcile the data movement from ODS to EDW.
- Notified the downstream stake holders for any changes to ODS/EDW.
- Implemented various Performance Tuning techniques on Sources, Targets, Mappings, and Workflows.
- Used deployment groups to migrate the Informatica code from Dev to INT, Test and Prod. Also created labels and queries.
- Involved in resolving issues in FRR/LTR reports (Federal regulatory reports for trading) using SQL.
- Worked with the DA team to improve the performance of the existing ETL processes.
- Worked closely with the business analyst’s team in order to solve the Problem Tickets, Service Requests. Helped the 24/7 Production Support team.
Environment: Informatica PowerCenter 9.5, SQL Server, Endur, Oracle 11g, SQL, TOAD, Tableau, PL/SQL, STAR Schema, Remedy, Teradata v2r12
Confidential, San Antonio, TX
Data Migration specialist/Data Conversation Specialist / Data Quality Consultant/ Lead Informatica/Data Architect
Responsibilities:
- Interacted with business/users and gathered requirements based on changing needs. Incorporated identified factors into Informatica mappings to build the Data warehouse/Data Mart.
- Managed a team of 5 (onsite-offshore).
- Created the data model for data migration Source to Landing to Staging to Target XMLs.
- Developed a standard ETL framework to enable the re-usability of similar logic across the board. Involved in System Documentation of Dataflow and methodology.
- Standardized the US/UK Addresses with Informatica Address doctor,
- Data profiling of the Product data.
- Documented Source to Target mapping with transformation rules for Customer, Invoices & Payments entities.
- Build logic for migration of open bills, unbilled usage, billed usage.
- Worked with Informatica Data Quality 9.5 (IDQ) toolkit to perform Analysis, data cleansing, data matching, exception handling, and reporting and monitoring capabilities.
- Created stored procedures/functions to load the data from landing to staging tables.
- Created IDQ Data profiles rules on the source system to measure the quality and pattern of the dataset.
- Worked on Informatica DataQuality (IDQ) and PowerCenter to build Data Quality Sanity Check and Business rules to capture the bad records as per the business requirements.
- Done POC’s on various business scenarios to facilitate business decisions.
- Used Informatica Designer to create complex mappings using different transformations like Filter, Router, Connected & Unconnected lookups, Stored Procedure, Joiner, Update Strategy, Expressions and Aggregator transformations to pipeline data to Data Mart.
- Written shell scripts to process make changes in XMLs generated.
- Developed complex business reconciliation reports/data validations reports using SQL.
- Conducted Database testing to check Constrains, field size, Indexes, Stored Procedures, etc.
- Used complex SQL queries to generate reconciliation reports, to reconcile between legacy source system and target system.
- Use ServiceNow for defect tracking.
- Build test cases and business use cases and assisted in created test cases for ETL.
- Implemented various Performance Tuning techniques on Sources, Targets, Mappings, and Workflows.
- Validation of Informatica mappings for source compatibility due to version changes at the source level.
- Used Source Analyzer and Warehouse designer to import the source and target database schemas, and the Mapping Designer to map the sources to the target.
- Implemented Address Cleansing of all the Customer Addresses, migration from the legacy Systems in Oracle BRM for proper taxation.
- Create and publish reports using Qlikview.
- Developed Test Scripts for validating the ETL code and the data (User acceptance testing)
- Built SSIS Packages to load data various ODS systems.
- Setup and maintain Informatica folders, user access of the Informatica repository.
- Create and maintain connections (relational, application, etc.) and system username/passwords.
Environment: - Informatica PowerCenter 9.6/9.5/9.1/8.6 , Oracle 11g, SQL Server 2012, SAP, PostgreSQL UNIX, Qlikview, Informatica Data Quality(IDQ) 9.5/8.6, SSIS, Informatica Data Profiling,QuickBooks, ServiceNow, Oracle BRM, XML’s.
Sr. Informatica Developer
Confidential, Phoenix, AZ
Responsibilities:
- Gathering the requirement changes from the functional team and incorporates them in Informatica and Business Objects.
- Interaction with direct Business Users and Data Architect for changes to data warehouse design in on-going basis.
- Designed the ETL processes using Informatica tool to load data from Oracle, flat files into the target Oracle Database.
- Followed & Implemented Standards for BI & DW at various levels of the SDLC.
- Built facts and dimension to build the DataMart.
- Developed complex mappings in Informatica to load the data from various sources using transformations like Source Qualifier, Expression, Lookup (connected and unconnected), Aggregator, Update Strategy, Filter, Router, Transaction Control etc.
- Have used Teradata BTEQ, FLOAD, MLOAD Teradata utilities to export and load data to/from Flat files.
- Used Informatica workflow Manager to create, schedule, monitor and send the messages in case of process failures.
- Designed SQL queries with multiple joins to pull relative data during import state.
- Designed and modified PL/SQL Stored Procedures to modify data flow.
- Used triggers in order to enforce business rules.
- Loading of Data from Siebel CRM to Siebel Analytics Database.
- Performed Data Analysis and Data matching between Siebel CRM and Siebel OBIEE Database.
- Developed FTP scripts to send the data extracts to various downstream applications using Informatica.
- Providing support for user BO report issues and Informatica loading issues.
- Monitoring Daily ETL data load in DAC.
- Implemented Confidential using PowerExchange.
- Tuning and performance improvement of the jobs in Informatica. Translated business requirements to Informatica Mappings. Involved in Unit testing of mappings.
- Delegating and Tracking the change requests in Informatica.
- Scheuled jobs using control-M.
- Created the transformation routines to transform and load the data. Developed processes for automation of loading data using parameter driven sessions for batch schedule processes, verification and re-conciliation of data stored in several different source systems.
- Worked with analysts and data source systems experts to map requirements to ETL code.
Environment: Informatica Power centre PowerExchange, 9.1/8.6, TOAD, PL/SQL Developer, Siebel CRM 7.8, DAC 7.7.1, Data Mining, Oracle, DB2, Teradata 12, Erwin 4.0, Windows 2000, XML, SQL, PL/SQL and Unix/Perl/Shell Scripting, Contol-M
Database Developer/Analyst
Confidential, Denton, TX
Responsibilities:
- Contributed in the development of system requirements and design specifications
- Participated in the design and development of Dimensional modelling.
- Developed complex mappings in Informatica to load the data from various sources using various transformations like Source Qualifier, Look up (connected and unconnected), Expression, Aggregate, Update Strategy, Joiner, Filter and Router
- Developed Mapplets to implement business rules using complex logic
- Extensively used PL/SQL for creating packages, procedures and functions.
- Converted the PL/SQL Procedures and SQL*Loader scripts to Informatica mappings
- Tuned the Sessions for better performance by eliminating various performance bottlenecks
- Created and scheduled Sessions and Batches through the Informatica Server Manager Wrote UNIX shell scripts to automate the data transfer (FTP) process to and from the Source systems, to schedule weekly and monthly loads/jobs
Environment: Informatica PowerCenter 8.5, SQL Server 2005/2000, Microsoft Visual Studio 2005,.NET Frame Work 2.0, Oracle 9i, SQL, PL/SQL, IBM AIX, UNIX Shell Scripts, Actuate Report tool, SSRS, Erwin, STAR team, Remedy.
ETL Developer
Confidential
Responsibilities:
- Contributed in the development of system requirements and design specifications
- Participated in the design and development of Dimensional modelling
- Developed complex mappings in Informatica to load the data from various sources using various transformations like Source Qualifier, Look up (connected and unconnected), Expression, Aggregate, Update Strategy, Joiner, Filter and Router
- Developed Mapplets to implement business rules using complex logic
- Converted the PL/SQL procedures and SQL*Loader scripts to Informatica mappings
- Tuned the Sessions for better performance by eliminating various performance bottlenecks
- Created and scheduled Sessions and Batches through the Informatica Server Manager
- Wrote UNIX shell scripts to automate the data transfer (FTP) process to and from the Source systems, to schedule weekly and monthly loads/jobs
Environment: Informatica PowerCenter 6.2, Business Objects, Oracle Applications 11i, Oracle 9i, SQL Server, SQL* Loader, HP UNIX, ERwin 4.0, WinRunner.