Sr. Data Warehouse Etl Consultant Resume
New York, NY
SUMMARY:
- Over fourteen years of IT consulting experience with design, development, implementation and maintenance of Data Warehouse / DataMarts, Data Extraction, Transformation and Loading (ETL).
- Worked with whole spectrum of software development life cycle, including requirement gathering, system designing, data modeling, development, testing, deployment, enhancement, maintenance and support.
- Assisted the management in managing the expectations (schedule, usability, expandability) for different modules in the project.
- Good understanding of Data warehouse design methodologies (Kimball, Inmon etc).
- Strong knowledge of ETL toolset (Informatica Power Center, PowerExchange), Pentaho Data Integrator and experience with Oracle Data Integrator, IBM Datastage, Microsoft DTS and SSIS.
- Experienced in a variety of databases like Oracle, DB2, GreenPlum, Teradata, SQLServer, PostGreSQL, MySQL etc.
- Working knowledge in various reporting tools like Crystal Reports, Business Objects, Qlikview, Pentaho Report Designer etc.
- Experienced in designing Logical and Physical Database models (Star and Snow Flake Schema) using ERWIN, ER Studio etc.
- Experienced in designing ETL architecture for data migration, data warehouse loads.
- Experienced in mentoring the staff on Data warehouse/ETL technologies.
- Created/Executed test plans for unit and integrated testing processes for different modules in a data warehousing projects.
- Resolved data issues by following the data standards and business metadata.
- Experienced in the migration of ETL objects from development to production and maintenance of ETL environments.
- Worked with UNIX scripts for data cleansing, automation of ETL jobs etc.
- Can understand and evaluate different source system file structures and data formats.
- Ability to quickly understand business goals, objectives, and adapt to existing architectures.
- Excellent verbal, written communication and mentoring skills.
- Hard working and can work very well in a team as well as independently under minimum supervision.
TECHNICAL SKILLS:
ETL Tools: Informatica Power Center 9.6.1/9.1.4/ 8.5.1/7.1/6.2/5.1 , Pentaho Data Integrator 7.1/4.0/3.2.4 (Kettle, Spoon, Carte), Oracle Data Integrator 11g, SQL Server DTS, SSIS
Databases: GreenPlum 4.2.7.2, Teradata V2R12/V2R6/V2R5/V2R4, Oracle 12c/11i/11g/10.2 g/9i/8i, MS SQL Server 2016/2012/2008/2005 , DB2 UDB 9.8/8.2/7.0, Adabas, PostGreSQL 2.0, AS 400
DB Tools & Utilities: PgAdmin III, Quest Central 4.0, Golden Gate, DB Visualizer 4.0.2, SQL*Plus, SQL* Loader, PLSQL Developer 7.0, Query Analyzer 8.0, Enterprise Manager 8.0, TOAD 6.3.3.1, BTEQ, Queryman (SQL Assistant), TPump, Mload, Fload, Fast Export, MS DB2 Import 1.5.0.1, EMS PostGreSQL Manager 2.8.0.3
Data Modeling: Erwin 7.0/4.1.2/4.0 , ER Studio, Star and Snowflake Schema Modeling
Languages: SQL, DB2 SQL Dialect, Confidential: SQL, PL/SQL, XML, HTML, UNIX Shell Scripting
Reporting Tools: Espress Reports, Actuate, Crystal Reports, Business Objects, Cognos, Qlikview
OLAP Tools: Business Objects 5.1, Info view, WEBI 2.7/2.6/2.5
Scheduling Tools: Crontab, Autosys, Tidal
Platforms: DOS, Windows 95/98/NT/2000/XP, UNIX, Linux 3.0, AIX 5L
Version Control: Microsoft Visual SourceSafe 6.0, Subversion
SFTP/SSH Clients: WinSCP 4.0.2, PuTTY 0.54
Big Data: Hadoop, Hive
WORK EXPERIENCE:
Confidential, New York, NY
Sr. Data Warehouse ETL Consultant
Responsibilities:
- Designing the ETL Architecture for bringing the Rave data into Confidential DB.
- Designing Metadata driven approach for loading the tables thus reducing the ETL code base by 90%.
- Developing and testing the ETL code using Pentaho Data Integrator 7.1
- Designing Control and Metadata tables for aiding the data loads, alerting mechanisms.
- Writing Sumo Logic reports for Logs and Metrics management.
- Writing Unix scripts to run ETL jobs.
- Deploying the code to various environments like Sandbox, Distro, Validation in Medistrano.
- Assisting in Production deployments and Production Support.
Environment: Pentaho Data integrator 7.1 (Kettle, Spoon, Carte), Oracle 12c, MS SQL Server 2016, Sumo Logic, SourceTree, GitHub, Docker, SQL Developer 4.1, DB Visualizer 9.5.6
Confidential, New York, NY
Sr. Data warehouse Consultant
Responsibilities:
- Designing ETL strategy for Global Data Initiative, Data Masking, Confidential Mobile, ECD, Risk projects.
- Analyzing reporting needs, gaps in current systems and designing solutions.
- Performing Source Data Analysis including flat files, databases and Target Data requirements.
- Aid in designing of Confidential Mobile, GDIS Datamarts, ECD datawarehouse and its supporting staging areas in Greenplum.
- Developing, testing the ETL code (Informatica PowerCenter), GreenPlum database functions and Unix scripts.
- Identifying the keys on the tables for optimal distribution of data on various segment servers or nodes of GreenPlum database.
- Analyzing, Vacuuming, Full vacuuming the tables for keeping the data dictionary of the GreenPlum database up - to-date.
- Utilizing GreenPlum writer instead of an odbc in combination with Informatica to improve the data load speeds.
- Optimizing the SQL queries executing on GreenPlum to avoid any spilling on the disks.
- Identifying sensitive data related to employees, vendors and terminal.
- Masking sensitive data with minimal impact to the existing processes.
- Establishing controlled access of the sensitive data to the users.
- Writing and executing unit and system test cases and ensuring that the code meets the requirements.
- Writing Unix scripts to validate the source files, run etl jobs, raise alerts for success and data validation failures and load failures.
- Deploying the code to Production and doing Support for the Production related Issues.
- Writing and modifying Qlikview and QlikSense reports per business requirement.
- Maintaining Source Inventory List for all the sources feeding DLDW, Mobile, GDIS Datamarts and ECD Datawarehouse.
- Creating databases, tables, and data pipeline in Hadoop cluster.
- Writing ETL jobs and hive scripts to migrate data between HDFS cluster and Greenplum database.
Environment: GreenPlum 4.2.7.2, Informatica PowerCenter 9.5.1, Oracle 11i, SQL Server 2008, PgAdmin III, Aginity Workbench, SQL Management Studio, Qlikview, QlikSense, SQL Tools 1.4.2, Hive
Confidential, Mt Laurel, NJ
Sr. Data Integration Consultant
Responsibilities:
- Design, develop, document, implement and validate database solutions and data integration methods.
- Working with business groups to understand the user requirements, develop business and functional requirement documents.
- Understanding business context and analyzing data from a business and technical perspective.
- Identifying sources, producers and consumers of data, collect and analyze metadata
- Developing System Requirement Specifications, Detailed Design Specifications.
- Perform data analysis and data profiling on large sets of financial data, comprehend and rationalize business data, develop SQL queries.
- Creating ETL specification documents, Writing and Executing Test Cases.
- Assisting in the implementation and support of various solutions to meet the end user requirements.
- Ensuring compliance Confidential different levels of project development.
- Coordinating between business users and the technology teams.
- Leading the data movement efforts from requirement gathering phase to implementation and support phase.
Environment: Oracle 11i, Quality Center 10.0, Source Off Site, Informatica PowerCenter 9.1.4, MS Word, MS Excel.
Confidential, Atlanta, GA
Sr. DW Consultant/ETL Architect
Responsibilities:
- Interacting with business users and business analysts to gather requirements.
- Developing high level architectural documents for the business and detailed architectural documents for ETL development.
- Data model changes/enhancements to source, staging and target environments to accommodate changing business requests.
- Analysis of ETL tools in the market to cater the current and future needs of the company.
- Implementation and maintenance of Pentaho 3.2 and 4.8 Business Suite and its components.
- Managing Users and Folders on various repositories.
- Hands on ETL Coding (transformations and jobs) using Pentaho Data Integrator (Kettle, Spoon).
- Writing and executing SSIS packages.
- Developing shell scripts for automation of ETL jobs, data archiving, data cleansing.
- Creating exception, error reports and alerts for the users.
- Working with a variety of data sources and targets including Flat Files, XML Files and Databases.
- Fine tuning and improving the performance of jobs and transformations.
- Writing the ETL Process Flow Documents for reference and production support.
- Writing and Tuning SQL scripts.
- Validating the transformation logic according to the pre-defined ETL standards.
- Analysis/Proposal of creating/altering of Indexes on tables for faster execution of queries.
- Fixing Production issues and ensuring prompt completion of etl loads.
- Code Review sessions with ETL / Reporting team members.
- Deploying Golden Gate data replication software for real time or near real time data loads into the Operation Data Store area, and then using the Pentaho Data Integrator to load the Datawarehouse on a more frequent schedule.
Environment: MS SQL Server 2012/2008, Pentaho Data Integrator 4.2.2/3.2.4 , Pentaho BI Integration Suite 4.8/3.5.2, SSIS, Golden Gate, Linux, Toad 5.0, Putty 0.60, Win SCP 4.2
Confidential, Atlanta, GA
Sr. ETL Architect
Responsibilities:
- Developing, testing, tuning Informatica Mappings, Sessions, Workflows etc.
- Analyzing data load speeds and introducing/making changes to partitioning Confidential the database level and ETL level.
- Developing, testing Teradata BTEQ, Fast Export, Mload scripts.
- Performing skew analysis on database tables.
- Developing Unix scripts to automate data analysis, generate audits on data and send alerts to business and technical users.
- Analyze any data issues and design sweeps to fix incorrect data.
- Using Change Data Capture to capture the data changes on the source systems on more real time basis, and load them into a data warehouse, so that end users will have access to more up-to-date data.
- Data model change recommendations to accommodate new business requests.
- Writing and analyzing BO reports.
- Using INFA Data Quality to ensure data consistencies and Data Masking to control, access of sensitive information to users.
- Developing the AID (Application Interface Design), DTD (Detailed Technical Design) and Technical specifications documents needed for various projects.
- Using Oracle streams for propagating data using user applied filters.
- Automation of jobs using crontab as well as Autosys.
- Troubleshooting production failures.
- Working on sustainment projects in resolving issues raised by business.
Environment: Teradata V12R12, Oracle 10g/9i, Informatica Power Center 9.0/8.5.1, Informatica MDM/DQ/Data Masking, DB2 UDB 9.8, Teradata SQL Assistant, GoldenGate, TOAD 7.6.0.11, XML, Unix, AIX, Autosys, Erwin Data Modeler 7.0, Putty
Confidential, Atlanta, GA
Sr. Informatica ETL Developer
Responsibilities:
- Developing mappings, sessions and workflows using Informatica Power Center 8.1.1 to populate the FDW staging area and the data warehouse.
- Following the pre defined ETL standards while developing the mappings.
- Developing the ETL technical design documents.
- Identifying the potential sources and analyzing the source data for data migration.
- Debugging the mappings to fix the bugs and improve the performance of mappings.
- Interacting with business analysts, data warehouse architects to better understand the business requirements and transform it into technical requirements.
- Data analysis to understand the issues with the data and designing solutions to fix them.
Environment: Informatica PowerCenter 8.1.1, Oracle 10g, PLSQL Developer 7, XML, Unix
Confidential, Alpharetta, GA
ETL Architect/Analyst
Responsibilities:
- Involved in redesigning the ETL architecture to facilitate the Hotlining process of the device, in case of device being lost/stolen.
- Developing the Technical documents needed for building the Informatica maps.
- Involved in writing the use cases for the business for the Hotlining process.
- Developing the Informatica mappings, mapplets, sessions, worklets and workflows and automating jobs using Autosys scheduler and managing daily ETL schedules.
- Extracting data from source systems including relational databases, Flat files, XML files
- Performance tuning the maps to reduce the lag between the occurred events on the device and updates in the system.
- Using the Data Quality product of Informatica to identify the data inconsistencies, data entry errors etc.
Environment: Informatica Power Center 7.1.1, Oracle 9i, Teradata V2R6, Autosys, TOAD 7.6.0.11, Sun Solaris Unix, XML, Erwin Data Modeler 7.0
Confidential, Atlanta, GA
Data Migration Lead/Manager
Responsibilities:
- Liaison between business users, business analysts, data modelers, development and testing teams.
- Managing the expectations (schedule, usability, expandability) of the Management Team for different modules in the project.
- Resource allocation and managing the timelines among various components in the project.
- Building the ETL architecture for migrating data from legacy systems.
- Developing and maintaining the technical documents needed for data migration.
- Mentoring the staff on design methodologies and development/testing process.
- Working with the data owners to resolve any issues with data from several sources and doing Data Cleansing, Data scrubbing.
- Writing necessary test plans to ensure the successful execution of the data conversion.
- Designing staging tables for assistance in the data conversion.
- Designing, Developing, Testing and maintaining several ETL loads like General Ledger, Bank, Location, Reporting Groups, Black Book, Repo, Auction etc.
- Developing mappings, mapplets, transformations, sessions, worklets and workflows in Informatica PowerCenter.
- Using CDC Change Data Capture features to reduce the amount of data being extracted for ETL loads and speed up the data load process.
- Automating the ETL loads through shell scripts, Production support for data migration and other ETL feeds.
- Extracting data from different source systems like Flat Files, SQL Server, XML, AS 400, Dataflex, DB2 using Power Exchange and Power Center products of Informatica.
- Handling alerting mechanisms and performance statistics of the loads.
- Checking the integrity of data and validating the numbers after the data loads.
- Tuning the performance of the ETL loads.
- Creating users and maintaining folders using the Repository Manager.
Environment: Informatica Power Center 7.1.2/6.2, DB2 UDB 8.1/7.0, SQL Server 7.0, DTS, Dataflex, PostGreSQL, Quest Central for DB2, DB Visualizer, EMS DB2 Import, EMS PostGreSQL Manager, AS 400, Espress Reports, Actuate, Windows XP, Linux, AIX.