Managing Consultant - Api/data Governance Consultant Resume
NJ
SUMMARY:
- 8+ years of focused experience in Information Technology on Development as well as with a strong background in Data Warehousing, Database development projects.
- Strong background in Data Integration ETL process, working as Data Integration Specialist - Greenplum, Netezza and Datastage.
- Hands on working experience in Netezza stored procedures and functions, Zone Map CBT, Groom, NZLoad and NZLUA package.
- Work Experience in Hadoop, Hive, Pig, Scoop, Spark, Cassandra.
- Hands on work Experience in Various GreenPlum Postgresql and pl/sql, External Tables, Distribution, Vacuuming and Table Partition etc.
- Used various stages on Quality Stage 8.7 like Investigate stage, Standardize stage, Match Frequency stage, De-duplicate Match stage, Reference Match stage, Survive stage.
- Working Experience in Data Governance and InfoSphere Governance Catalog, Data profiling using InfoSphere Information Analyzer.
- Strong Hands on working experience in Data Stage 8.7, Parallel 8.1/7.5.1/7.5 x2 PX and Server Jobs.
- Strong experience in Data Warehousing and ETL using Informatica Power Center 9.1.
- Hands on working experience in Data Quality using Quality Stage 8.7.
- Used various stages on Data stage 8.7 Parallel jobs; like Sequential File, Transformer, Dataset stage, Aggregator, Merger, File set, Range lookup, Joiner, Copy, Modify, filter, funnel, Oracle Connector Job Sequence, peak stage, look up stage, Joiner stage, Remove duplicate etc
- Used various partition Techniques like Round robin, Same, Entire, Modulus etc.
- Sound Knowledge on 7.5 and 8.7 Data stage Architecture and internal process of Datastage
- Used various stages on Data stage 7.5 PX and server jobs; like Sequential File, Transformer, Aggregator, Hashed File, ODBC, Link Partitioned, Link Collector, IPC, Merger, Oracle OCI and Job Sequence, peak stage, look up stage, joiner stage, Remove duplicate, Stored Procedure Stage etc.
- Hands on Performance Tuning in ETL Jobs and Sessions and SQL Tunings.
- Writing of PL/SQL stored procedures and functions, Packages, Database Triggers, oracle Collections and Cursor, Dynamic sql that enhances the functionality of an application.
- Configure Data Stage Director and scheduling data integration jobs and Sequencer.
- Used Data stage Director to run, monitor and schedule the jobs.
- Sound knowledge on Data profiling.
- Strong Work experience on Zena Scheduler.
- Sound Knowledge in Multiload, BTEQ, FastLoad, MultiLoad, FastExport, TPump and Teradata Parallel Transporter.
- Created Data stage job sequences to handle the load dependency.
- Good exposure in UNIX commands and Shell Scripting to execute data stage jobs.
- Also involve in preparing various documents like Application Design Document, Unit Test case, Deployment plan.
- Good client interfacing skills and demonstrated the same by interacting with client on onshore Offsite model.
TECHNICAL SKILLS:
ETL: Quality Stage 11.3/8.1, Data stage 11.5/11.3/8.7/8.0/7.5/7.1 Parallel/Server job, Ab Initio 4.0
Case Tool: SCME
Script: Shell Script, Sql Script
Data Modeling: Erwin 7.2/4.1.4, Microsoft Visio, PowerDesigner 12.5
RDBMS: Oracle 9i/10g PL/SQL, Teradata, Greenplum and Netezza
Operating System: AIX5.3, Solaris, Unix-IBM AIX, Red Hat Linux, Windows98, 2000Server, Win-XP, Win-7
Big Data: Hadoop, Hive, Pig, Scoop, Spark, Cassandra.
Job Scheduler: Cron Tab, Zena.
Tools: Sql* Loader, Toad VII, Aginity Workbench
Remedy Tools: AOTS ( Confidential Proprietary tool), Star Team and Jtracker.
PROFESSIONAL EXPERIENCE:
Confidential, NJ
Tools: Info sphere Information Suite 11.5(Business Glossary, IGC, IMAM, Information Analyzer, Quality Stage, Datastage), Oracle Exa-data 11g, Unix Shell Scripts.
Managing Consultant - API/Data Governance Consultant
Responsibilities:
- Establish and govern an enterprise data governance implementation strategic priorities for development of information-based capabilities
- Roll out an enterprise wide data governance framework, with a focus on improvement of data quality and the protection of sensitive data through modifications to organization behavior policies and standards, principles, governance metrics, processes, related tools.
- Define roles and responsibilities related to data governance and ensure clear accountability for stewardship of the company’s principal information assets
- Facilitate the development and implementation of data quality standards, data protection standards and adoption requirements across the enterprise
- Define indicators of performance and quality metrics and ensure compliance with data related policies, standards, roles and responsibilities, and adoption requirements
- Align data steward by domain and educate GPT Product teams and business community
- Create collaboration forums to prioritize and approve API designs
- Generate policies and guidelines for API design and alignment enterprise logical model
- Process model the go-forward engagement for API governance (review, prioritize and approvals)
- Transition responsibilities to ADP’s GPT Product team resources
Confidential, CA
Tools: Info sphere Information Suite 11.5(Business Glossary, Metadata Workbench Analyzer, Quality Stage, Datastage), Oracle Exa-data 11g, Unix Shell Scripts.
Managing Consultant - ETL - Data Integration Architect/Data Analyst
Responsibilities:
- Worked as Data Integration Lead/Data Analyst in Confidential Project.
- Worked with Solution Architect to create Macro and Micro Design Document for Data Integration.
- Worked Business Analysts as a Data Analysts to analyze the source system based on business requirement and come up with Source to Target Mapping Document.
- Worked with Data Modeler to Finalized the Logical and Physical Target Data Model.
- Work With the DBA / Data Architect in Physicalized the target data model.
- Define the Data Integration Strategy and Data Loading Plan for target subject area load.
- Performing Data Analysis Source and Target information and creating Mapping Documents.
- Select the systems and capabilities to meet business data requirements, evaluate new techniques for improved functionality.
- Performing data analysis and data profiling to understand how to translate data into EDW model and validating quality of data
- Member of EPD(Enterprise Data Platform) working group for setting up standards and best practices
- Delivering documents like high level and detailed technical design, source to target mapping document, unit test cases, implementation plans, rollback strategies
- Delivering data flow diagrams for complex integration between various systems
- Leading efforts for data mapping between various layers of information architecture
- Designing data interface layer
- Producing POC for complex integration by designing ETL pipes using ETL tool
- Designing reusable ETL components
- Successfully performance tuned various ETL processes in order to achieve the execution time within the given SLA
- Performing SQL tuning using various SQL HINTS Delivering documents like high level and detailed technical design, source to target mapping document, unit test cases, implementation plans, rollback strategies
- Delivering data flow diagrams for complex integration between various systems
- Leading efforts for data mapping between various layers of information architecture
- Designing data interface layer
- Producing POC for complex integration by designing ETL pipes using ETL tool
- Designing reusable ETL components
- Successfully performance tuned various ETL processes in order to achieve the execution time within the given SLA
- Performing SQL tuning using various SQL HINTS
- Defining different database objects like Tables, Views, Indexes, Partitions, Packages, Collections, Bulk processing, Analytical Functions, Procedures, Sequences, Synonyms, Database links, Contexts, VPD policies
- Providing SIT and UAT support and resolving defects and delivering solution with quick turnaround time
Confidential, IL
Tools: Info sphere Information Suite 11.3(Business Glossary, Metadata Workbench Analyzer, Quality Stage, Datastage), Oracle Exadata 11g, Unix Shell Scripts.
Senior Consultant
Responsibilities:
- Analyze Source system architecture and help establish Source to Target Mapping
- Perform Data Analysis on the underlying table and document the Schematics of the Table
- Create Technical Specifications out of the Business Requirement document adapting agile software development methodology - maintaining version control for every iteration
- Design, Develop and Unit Test ETL DataStage Jobs using IBM Information Server 8.1/9.1 utilizing various component stages
- Read data from Disparate source systems including flat files, Oracle and DB2 database Tables and load aggregated and line level detail records into data mart tables
- Design Sequencers that calls the jobs on adhoc basis and need basis.
- Create Exception Report and suggest DataStage Best practices across the LCR Project
- Set up Environment Variables to be used across the Projects.
- Develop Unix Shell Scripts and debug existing scripts
- Create Audit Jobs that captures Record Statistics for multiple work areas and design the process the trigger based on validation
- Provide estimates of ETL job development for self and counterparts offshore and perform peer reviews and suggest alternative approach/solution to enhance performance of DataStage Jobs
- Point of contact for Production Job Migration and successful implementation of Job runs in Development, UAT, SIT environments and perform data validation post ETL job implementation run
Environment: IBM Infosphere Data Stage 8.1, Oracle 11i, DB2, Delimited Flat Files, Comma Separated Files, Advanced Query Tool (AQT), WinScp, MS Visio, MS SharePoint, HP Quality Center, Windows 7, AIX, Unix, MS Office Components.
Confidential
Tools: Informatica Power center 9.1, Greenplum Database 4.2, PostgreSQL 8.2, Ms SQL Server, Unix Shell Scripts
Datawarehouse Analyst
Responsibilities:
- Created source-target mapping spreadsheet using MS Excel to facilitate better understanding of the mapping and transformations.
- Experience working with Greenplum/Pivotal Databases/Developer.
- Convert Oracle Procedures and functions into Greenplum.
- Proficient with Performance and Tuning of DBMS configurations, SQL, Capacity Planning.
- Proven expertise with Backup and Disaster recovery (multiple data centers).
- Designed and developed the ETL jobs using Informatica Mappings which distributed the incoming data concurrently across all the processors, to achieve the best performance.
- Used Informatica Power Center 8.6 for extraction, transformation and load (ETL) of data in the data warehouse.
- Extensively used Transformations like Router, Aggregator, Normalizer, Joiner, Expression and Lookup, Update strategy and Sequence generator and Stored Procedure.
- Designed mappings and command task to Involved in the extraction, transformation and loading of the source data to the target database.
- Used various transformations like Source Qualifier, Expression, Aggregator, Joiner, Filter, Lookup, Update Strategy Designing and optimizing the Mapping.
- Implemented performance tuning logic on targets, sources, mappings, sessions to provide maximum efficiency and performance.
- Parameterized the mappings and increased the re-usability.
- Used Informatica Power Center Workflow manager to create sessions, workflows and batches to run with the logic embedded in the mappings.
- Extensively used Informatica debugger to figure out the problems in mapping. Also involved in troubleshooting existing ETL bugs.
- Performed import and export mappings and various components.
- Review all Project Specific Documents before taking it forward.
- Leading the team and initiate review meetings and get regular updates on the tasks to be completed.
- Designed and written mappings, performed Unit tests, set up the System Test environments, generate test cases, Regression test where needed and support User Acceptance Testing.
- Implemented procedures for development of detailed specifications along with interfaces and platforms.
- Integrated programs with production systems and applications by designing of system testing requirements.
- Debugged programs for development of code and custom systems as per IT guidelines.
- Coordinated with other client teams for organizing and conducting systems and integration testing services.
Confidential
Tools: Unix Shell Scripts, Data stage 8.7 Parallel Jobs, Greenplum Database, Ms SQL Server, Netezza.
Datastage Consultant
Responsibilities:
- Created source-target mapping spreadsheet using MS Excel to facilitate better understanding of the mapping and transformations.
- Designed and developed the ETL jobs using Parallel Extender which distributed the incoming data concurrently across all the processors, to achieve the best performance
- Designed parallel jobs using stages such as Join, Merge, Lookup, Remove Duplicates, Copy, Filter, Funnel, Dataset, Lookup, Sort, Surrogate key Generator, Change Data Capture and Aggregator.
- Created Master controlling sequencer jobs using Data Stage Job Sequencer.
- Involved in the extraction, transformation and loading of the source data to the target database.
- Performed import and export data stage jobs and various components.
- Review all Project Specific Documents before taking it forward.
- Leading the team and initiate review meetings and get regular updates on the tasks to be completed.
- Designed and written mappings, performed Unit tests, set up the System Test environments, generate test cases, Regression test where needed and support User Acceptance Testing.
Confidential
Tools: Oracle, Unix Shell Scripts, Data stage 8.5 Parallel Jobs, Data stage 8.0 Parallel Jobs, Data stage 7.1 & 7.5 server job.
Datastage ETL Developer
Responsibilities:
- Developing the jobs for loading data from ICORE to our Target EM multipleFeed files.
- Creation and debugging of Jobs which will use Joins, database levels, Transformers, Sequential files, look-up and other utilities.
- Table creations and altering, materialized views, Views, sequences, stored procedures, functions, complex queries and performance tuning.
- Owning Data stage configuration and deployment.
- Owning all kinds deployment related ETL track on all the Environments Setting up Data stage Parameters.
Confidential
Tools: Oracle PL/SQL, Unix Shell Scripts, Data stage 8.5 Parallel Jobs
Datastage ETL Developer
Responsibilities:
- Developing the jobs for loading data from GPS to our Target (PTY3) schema.
- Creation and debugging of Jobs which will use Joins, database levels, Transformers, Sequential files, look-up and other utilities.
- Table creations and altering, materialized views, Views, sequences, stored procedures, functions, complex queries and performance tuning.
- Owning Data stage configuration and deployment.
- Owning all kinds deployment related activities on all the Environments.
Confidential
Tools: Oracle PL/SQL, Unix Shell Scripts, Data stage 8.7 Parallel Jobs
Datastage ETL Developer
Responsibilities:
- Developing the jobs for loading data from NICE CORE to our NICECORE STAGING schemas and from there generating the feed files.
- Developing the jobs for 4 interfaces(MCDB TO GCP,REMEDY TO GCP,OVPI TO GCP,BEDW TO GCP)
- GCP staging tables to support AERS SLA reporting
- Creation and debugging of Jobs which will use Joins, database levels, Transformers, Sequential files, look-up and other utilities.
- Table creations and altering, materialized views, Views, sequences, stored procedures, functions, complex queries and performance tuning.
- Owning Data stage configuration and deployment.
Confidential
Tools: /Platforms: Oracle PL/SQL, Unix Shell Scripts, Data stage 8.5 Parallel Jobs
Datastage ETL Developer
Responsibilities:
- Developing the jobs for loading data from Source system to Calnet Vain
- Analyzing the requirement and transfer it to technical jobs.