Lead Data Integration Consultant Resume
EXPERIENCE SUMMARY:
- Over 12 years of IT experience in analyzing, developing, migrating and maintaining ETL applications through complete project life cycle (Waterfall and Agile)
- IBM Certified Professional with Over 10 years of experience using IBM InfoSphere Suite of products(DataStage 11.3/9.1/8.7/8.5/8.0.1) QualityStage, Information Analyzer, Metadata workbench, Business Glossary), Ascential Datastage 7.5.3
- Experience in installation, configuration and administration of IBM InfoSphere Information server suite
- Sound knowledge in IBM Information Server Suite Architecture and familiarity with various Installation configurations
- Proficient in ETL techniques to integrate and synchronize data from disparate source systems to target systems, Data Warehousing concepts, Data Governance, Metadata Management, Slowly Changing Dimension phenomenon, surrogate key assignment and change data capture
- Proficient in ETL Architectural Design, Capacity Planning and tools assessments
- Expertise in various RDBMS like Oracle, SQL SERVER, DB2 UDB, Teradata, Sybase
- Specialist in optimizing high volume environments for optimal performance, Fine tuning, Performance analysis, Bench marking and Trouble Shooting of ETL Processes. Handled data volumes close to 1 Billion
- Experience in Shell Scripting for Process Automation, Reconciliation and Database specific procedural SQL programs (PL/SQL, T - SQL)
- Strong analytical experience in understanding Product specifications, functional requirements and Technical Requirements
- Conceptual knowledge of BigData Technologies- Data Ingestion to HDFS, NoSql Databases (Cassandra, HBase), Query Engines (Impala), Resource management (YARN), MapReduce, WorkFlow Scheduler (Oozie), HiveQL, Pig Latin
- Highly self - motivated and able to set effective priorities to achieve immediate and long - term goals
TECHNICAL SKILLS:
ETL Tools: IBM InfoShpere Information Server 9.1/8.7/8.5/8.0.1, Ascential DataStage 7.5.3, Talend Open Studio 4.2, Informatica PowerCenter
Programming: UNIX Shell Scripting, Awk, PL/SQL, T-SQL, C, C++, Python
Databases: Oracle 10gR2/Oracle 10g/Oracle 9i, SQL Server 2008R2/2005/2003, DB2 UDB 8.1, Teradata 14.0, Sybase
Environment: Win 2008/2003/NT/XP, AIX 6.0/5.0, RHEL 6.0
Data modeling: Erwin 4.x/3.x, Dimension modeling (Star/Snowflake Schema Design)
Scheduling Tools: ESP, CRONTAB, AUTOSYS, TWS
PROFESSIONAL EXPERIENCE:
Confidential
Lead Data Integration Consultant
Responsibilities:
- Meet with Business regularly to articulate Technical requirements in an Agile development Environment
- Analyze partner specific formats(NDB, CMS and EPDL) and create Source to Enterprise Canonical model(ECM) mapping
- Create Canonical to MDM data model Mappings
- Create MDM consumable XML transaction files for HealthCare Organizations(HCO), HealthCare Professionals(HCP) and Solo Practioners
- Implement address verification/Certification using USPS postal address data base (CASS)
- Deploy Address verification jobs as a web service for multiple consumers
- Build processes to write to MQ, Pub/Sub model for downstream applications
- Build processes to interact with ODM for JRules using iLog Connector using Java XOM as well as Dynamic XOM with Classic Rule Engine (CRE) as well as DE (Decision Rule Engine Mode)
- Involved in Building framework for scheduling jobs on Tivoli Work load Scheduler (TWS)
- Automate reference files update through a service call to RDM
- Consume NPPES file and extract NPI, Demographic, License and Taxonomy information and write to MDM
Environment: IBM InfoSphere Datastage 9.1/8.7/8.5, Oracle 11g/10gR2/10g,, RHEL 6.0, Sybase, Altova XMLSpy, Tivoli Workload Scheduler (TWS), IBM MDM, RDM, ODM, DB Visualizer, DB2
Confidential, TN
Data Integration Developer
Responsibilities:
- Translate Business needs to Technical requirements and prepare documentation
- Analyze partner specific formats and create source to target mappings
- Develop ETL processes to parse data from multiple sources, validate and load data to different RDBMS (Oracle, SqlServer) and DW Environment
- Create MEDai consumable Relational Extracts, test for relational integrity, feed them through predictive analytics Engine, Monitor and push the outcome to Repository.
- Process high volumes of data leveraging concepts of parallelism in DataStage and Database
- Create job sequences with restartability, exception handling and job dependencies
- Utilize Web Services in the jobs to retrieve the data from external applications
- Build Static and Dynamic models for resource estimation by enabling record performance data
- Generate resource utilization report (CPU, Memory and Disk utilization) using the feature performance analysis
- Automate ETL Batch processes using Scheduling tools
- Fine tune resource intensive jobs for optimal performance, debug Jobs for performance bottlenecks and propose solutions
- Install and Upgrade DS Environments through multiple versions of Ascential/InfoSphere DataStage/QualityStage from 7.5.3 to 8.0.1/8.5/8.7/9.1
- Install native clients for database access (Oracle Native Client), configure TNS for Oracle RAC
- Configure ODBC data source connections (.odbc.ini, uvodbc.config) to utilize DataDirect 5.0/6.0 wire protocol drivers
- Post install configuration tasks including adding library locations to dsenv, Configuring Parallel Engine, Configuring Information Analyzer, Creating Configuration files
- Apply patches, perform health checks on Datastage servers periodically
- Execute batch scripts for regular DataStage Projects Export and Archiving
- Perform server maintenance tasks - clearing resource and scratch pools, &PH& directory, purging core dumps and temporary Datasets in DS Environments
- Create QualityStage jobs for implementing standardization rules using various quality stages like Investigate, Match Frequency, Standardize and Reference Match.
- Maintain user roles and Metadata Management for Information Analyzer projects
- Data Synchronization between UAT and Prod environments for development and reporting team needs
- Document end to end implementation including Mock - Ups, Business requirements, Technical requirements, Test cases, System and Integration Test Results
- Conduct Scrums, Brainstorming sessions, code walk thru and peer review
Environment: IBM InfoSphere Datastage 9.1/8.7/8.5/8.0.1, Ascential DataStage 7.5.3 Enterprise Edition, Oracle 11g/10gR2/10g, Unix AIX 5.3/6.0, RHEL 6.0, SQL Developer, TOAD, SSMS,SQL Server 2008R2/2005/2003, Teradata 14.0, BTEQ
Confidential, OR
DataStage Developer
Responsibilities:
- Developed DataStage jobs for Extracting data from heterogeneous databases and Loading of the data into Data Warehouse
- Built DataStage jobs involving Slowly Changing Dimensions, surrogate key generation and for Incremental data loading
- Created and used DataStage Shared Containers and Local Containers for reusability
- Extensively Worked using Various Stages like Dataset, Fileset, Lookup File Set, Join, Lookup, Sort, Change Capture, Filter, Remove Duplicates, Surrogate key generator
- Extensively used stages like Job Sequencer, Wait for file Activity, Exception Handler, Execute Command in Job sequences
- Worked on Importing/Exporting Data stage jobs from Development to production
- Debugged the mappings to fix the bugs and involved in Performance tuning of the DataStage jobs
- Used DataStage Manager for importing table definitions, jobs and exporting objects.
- Used the DataStage Director extensively to run, monitor, debug and test the jobs in development, and to obtain the performance statistics
- Developed stored procedures and SQL queries for production reports with attention on query performance
- Used PVCS Version Manager in order to maintain, manage different versions and for secure Import and Export of DataStage Jobs
- Documented test cases, conducted Unit tests, Integration tests and migrated jobs from development to testing and eventually to production environment.
- Participated in discussions with Team leader, Group Members and Technical Manager for technical and Business Requirements
Environment: Ascential DataStage 7.5.2 Enterprise Edition (Designer, Director, Manager), Oracle 9i, UNIX AIX 4.2, UNIX Shell Scripts, SQL Developer, Windows xp, PVCS, Syncsort 3.9.0
Confidential
Database Developer/Production support
Responsibilities:
- Worked in production support team for PL/SQL applications and Unix Scripts
- Developed Stored Procedures, Created Triggers and PL/SQL packages
- Involved in Performance Tuning, Analyzing query plans and query tuning
- Worked with logical and physical database modeling along with design and implementation
- Involved in analyzing problems, root cause analysis and problem resolution related to application support tasks
- Provided timely updates on risks, issues and escalations along with an action plan to the upper management
- Interacted with users for report requirements
Environment: Oracle 8i, Erwin, TOAD, PL/SQL, HP-UX, Windows Server
