Etl/ Informatica Developer Resume
St Louis, MO
SUMMARY
- Experience in Data management using Informatica Master Data Management (MDM), Informatica Data Quality (IDQ), Informatica Power Center, Python Scipt, Shell script, PL SQL and SQL with focus on analysis, design, development and maintenance of various data warehouse applications in an effective way to deliver leading edge software solutions.
- Involved in all phases of Software Development Life Cycle from gathering requirements, analysis, design, development, and support.
- Proficient in Master Data Management concepts, Methodologies and have ability to apply this knowledge in building Master Data Management Solutions.
- Expertise in installation and configuration of Informatica MDM Hub components like MDM Hub Console, Hub Store, Hub server, Cleanse Match Server, Resource Kit, and Address Doctor.
- Designed and Configured various components of MDM Hub such as Landing tables, Staging tables, Base Objects, Hierarchies, foreign - key relationships and lookups.
- Expert at creating Mapping, Trust and Validation rules, Match Path, Match Column, Match rules, Merge Properties.
- Configured custom queries and advance queries, Packages both using Hierarchies and packages.
- Designed and developed cleansing and standardization scenarios using Address Doctor.
- Expertise in configuring IDD, creating subject area groups, subject areas, subject area child and packages.
- Implemented Hierarchies using entity objects, relationship type objects, relationships, packages and profiles for hierarchy management in MDM Hub implementation.
- Skilled in utilizing Informatica Power Center to import the data into Informatica MDM.
- Expert in architecting and developing Data reconciliation designs between various systems to generate reports.
- Worked with Repository manager, Mapping designer, Transformations, Workflow manager, Task Developer, IDQ, Informatica Powercenter for ETL, and Autosys,& Tidal scheduling tools. Worked with heterogeneous source systems like flat files, Oracle, sql server, Netteza, Teradata, Mango DB and parameterizing variables.
- Created IDQ mappings to standardize and validate the data. Cleansed, labeled and fixed data gaps with IDQ.
- Expertise in python programming for connecting to multiple databases using various python modules for data reconciliation and analysis.
- Expertise in writing shell scripts to connect to Teradata via BTEQ script for data analysis and creating reports.
- Team player with good problem-solving skills and excellent communication skills.
TECHNICAL SKILLS
EIM Tools: Informatica MDM Multi-Domain, IDD, Informatica Data Quality (IDQ)
Cleanse Adapter: Address Doctor, Trillium
Application Servers: JBoss
Data Integration tool: Informatica Power Center
Databases: Oracle 11g/10g/9i/8i, SQL Server 2005, MySQL, MS-Access, Netezza,, Mongo DB, Teradata
Development Tools: SQL Plus, Oracle SQL Developer, Microsoft SQL server management Studio, Advance Query Tool
Configuration Tool: Erwin Data Modeler
Operating Systems: Windows 7, Windows XP, Windows NT/2000, Windows Server 2008 R2, Redhat Linux
Programming Skills: C, C++, java, SQL, PL/SQL, XML, Unix Shell Scripting, Python
Package: MS Office (Access, Excel, Word, PowerPoint, Word)
Tools: and Utilities: SSA Name3, SoapUI, ServiceNow, Jeera, HP Quality Center
PROFESSIONAL EXPERIENCE
Confidential, St. Louis MO
ETL/ Informatica Developer
Responsibilities:
- Gathered requirements from Business users and Business analysts for various Data Quality issues in the Centene’s Enterprise Data warehouse
- Refined the stories and analyzed all the requirements and designed architecture for development of stories.
- Worked on Informatica Data Quality to Profile the data sources before starting the development to identify the data anomalies and reported them to source systems for correction.
- Designed Informatica Data Cleaning mappings to standardize and cleanse the data before moving to source.
- Created Data Reconciliation programs to compare data across the organization for ACA member data. The report generated out of the data reconciliation process provides information to business users about discrepancies between various systems across organization and for solving the data discrepancies.
- Created DIFF and MISMATCH reports using Shell script to connect to Teradata via BTEQ scripts for showing business exact records which are having discrepancies across UMV, AMISYS, TRUCARE, EDW, RX and SOFTHOEN Source systems.
- Created shell scripts for triggering Informatica jobs via Tidal scheduler after doing inbound File check.
- Developed mappings as per the business requirements using transformations such as source qualifier, aggregator, expression, lookup, filter, sequence generator, router, union, update strategy for loading the data from UMV, AMISY, TRUCARE and RX source systems into Teradata for Member data reconciliation process.
- Fine tuning mappings and sessions to improve performance.
- Analyzed issues reported by Business team for Member data reconciliation process and fixed those issues to increase the matching percentage based on various business rules by modifying Teradata PLSQL Procedure
- Created python script to reconcile data between UMV which is on Mongo DB and UMV STAGE OWN database which is on Teradata database and created reports (for Missing in Source, Stale in Target and Missing in Target) using python modules TeradataDS.py and MongoDS.py. Raised Service Now incident ticket to the source team to get the data fixed on downstream.
- Created python script to reconcile data between Trucare which is on Mongo DB and UMV STAGE OWN database which is on Teradata database and created reports (for Missing in Source, Stale in Target and Missing in Target) using python Modules TeradataDS.py and PYPYODBC.py. Raised Service Now incident ticket to the source team to get the data fixed on downstream.
- Created Automated Balance Control systems to validate the data coming from Syndicated source systems into Stage and load area at Enterprise Data warehouse for Authorization data using Shell script and Tidal Automation.
- Worked on creating dynamic BDM mappings to move the encounters data from Hive to EDW HEALTHNET STAGE OWN.
- Created BTEQ, Hive Query scripts and Shell scripts for encounters Jiras.
Environment: Informatica Power center 10.2.0, Informatica BDM 10.2.1, Informatica Data Quality, Tidal, Teradata, Oracle 13g, SQL Developer, Ambari, putty.
Confidential, Schaumburg, IL
Informatica MDM Consultant
Responsibilities:
- Gathered requirements from Business users and Business analysts for Canada UWD desktop implementation, developed Design document and got Sign off from Application Owner.
- Profiled the new Global RCT data to understand the new data and identify the data issues that need to be corrected and standardized before sending into MDM.
- Created IDQ mappings to replace existing MDM Cleanings mappings between landing tables and stage tables for the MDM 9.1 to 10.1 upgrade project and implemented new business rules specified by RCT Business analysts.
- Developed mappings as per the business requirements using transformations such as source qualifier, aggregator, expression, lookup, filter, sequence generator, router, union, update strategy etc.
- Fine tuning mappings and sessions to improve performance.
- Used address doctor validator in IDQ mappings to standardize the customer and distributor address data coming from EDB, DMS, ZST, SFDC source systems.
- Created WSDL connections in MDM utilize IDQ Mapplets in MDM cleanse functions.
- Added Canada Population Set and made related changes at MDM repos tables and MDM base objects to accommodate new match rules.
- Created ETL mappings for Canada Historical Data Cleansing and SIP POP column update for historic data.
- Worked with SFDC team and ZSOA team to accommodate SFDC subscription changes. Created Message Queues to send data to SFDC.
- Created MDM components to Synchronize SFDC’s APEX ID into MDM’s pkey src object column. As part of this, created Landing tables, Staging tables lookup for load by rowid object, Mappings to accommodate Load by rowed object.
- Worked on pl-sql procedures and ETL mappings to Soft delete the SFDC records which do not have APEX ID.
- Worked on fixing Strata Plan issue by creating a historic and on the go fix. Eyeballed all the records that need to be unmerged via Execute Batch Unmerge. Unmerged the identified records and updated the Siperian Process Code to 93 for avoiding those records to be merged again for historic fix. Modified the cleanse functions for each source system to automatically update the Siperian Process Code to 93 for new records entering into MDM with full nm having Strata Plan.
- Worked on SIP POP issues by creating Historic and on the go fix. Created ETL mappings to pick all the records with ISO2 standard country code in SIP POP and converting them to ISO3 Standard. Created MDM components, Landing tables, Staging tables, lookups and Mappings for load by rowid object.
- Involved in RCT project requirement analysis, design, development and test support, which includes, adding new source system RCT-WF to MDM, configuring new MDM components to accommodate Global data. Configured IM process for Global data to get data from Global DNB.
- Analyzed the issues occurred in the downstream by analyzing the data flow via several layers from upstream, to MDM, MDM ODS, ZEA and ZNAW till BOE reporting by analyzing the data flow via MDM and ETL.
- Identified the root cause for issues caused and provided solutions to business. Created fixes to several issues to correct the data.
- Identified the formation of huge clusters of data in Prod after an implementation and worked with Informatica to unmerge those clusters. Eyeballed thousands of records to manually unmerge them using SIF call ExecuteBatchUnmerge call.
- Identified the formation of Orhpan records in Production data, i.e. Party record with no PostalAddress record. Identified the root cause for the issue to a newly created fix. Provided the solution to business to fix the issue. Created ETLs to read the Address records from file and load them into respective Address landing tables. Executed batch run adhocly to fix the Orphan records.
- Analyzed the issue reported in BOE reporting layer, where MDM EIDs are not flowing downstream. Analyzed the data in MDM and identified that formation of duplicates is the cause for the issue. The duplicate formation was due to invalid SIP POP values in the Base object records as the match rules are configured with SIP POP exact rules causing duplicates. Provided solution to Business to change the IDQ rules to permanently fix the issue.
- Worked on issue with missing EIDs reported at BOE Layer, analyzed the ETLs, MDM and Stored Procedures and identified the issue. Missing hierarchy data in MDM and downstream has prevented the existing MDM data to not go downstream. Modified the MDM Monthly DNB Brick Load ETLs which are preventing the Canadian data which are part of hierarchy and resolved the issue.
- Closely worked with testing team to fix any issues that are identified during testing phase.
- Worked on defects and incidents that are raised by business with in the SLA and closed them with business team’s approval.
- Worked closely with Informatica to resolve the issues identified in the Informatica MDM Tool.
- Ensured all the information is passed to offshore and resolved their queries by clarifying with the client.
- Verified and validated the developed application by end to end unit testing in DEV before migrating changes to QA.
- Created detailed implementation Plan, Post implementation validation plan for the deployments and created change request for production deployment. Coordinated with other application areas during the deployment and got their sign off after post deployment resolving issues.
- Managed daily, weekly, monthly, quarterly and Adhoc batch jobs, solved data issues and other technical issues that occurred during the batch runs.
- Used Autosys scheduler to schedule the jobs, to check the status of jobs, or to send events such as to change status of job, to mark it success, to place it on ice or on hold etc.
- Managed issues that occurred in MDM Hub, IDD, Batch jobs and Real time web service calls.
- Monitored Informatica jobs that flow from source systems to MDM and from MDM to MDM ODS and to ZEA reporting layer. Debugged Informatica failures by checking the session logs and work flow logs.
Environment: Informatica Multi Domain MDM 10.1, Informatica Data Director (IDD), Informatica Power center 9.6.1, Oracle 13g, JBOSS 6.4, Red hat Linux, SQL Developer, Address Doctor, Trillium, Informatica Data Quality, Service Now, HP Quality Center, Element Tool, Autosys, putty, WinSCP, Aginity Work bench, AQT, Netezza, DB2.
Confidential, Frazer, PA
Informatica MDM Consultant
Responsibilities:
- Installed Informatica MDM application on JBoss application server and installed Hub and Cleanse Server.
- Created Landing, Base, Stage tables based on the Data Model and the Source Systems.
- Defined relationships among the base objects and defined lookups for the Stage tables.
- Enabled Trust for the columns in the base tables and defined trust scores for the source systems as per the requirements.
- Developed the Mappings for the Stage Process by using various Cleanse Functions in the MDM Hub.
- Created Hierarchies using the Hierarchy Manager in the Hub as per the business need.
- Executed the Stage process to get the data loaded into the Staging tables and the load process to load the data into the base objects.
- Defined Match and Merge rules and ran the Auto Match and Merge process.
- Analyzed the Match rules by capturing the data from MTCH table.
- Fine-tuned the Match rules after analyzing the data from the match table and reduced the number of rules.
- Created the IDD application for basic search and extended search.
- Configured extended search capabilities and found duplicates for new and existing records.
Environment: Informatica Multi Domain MDM 9.7.1, Informatica Data Director (IDD), Informatica Power center 9.6, Oracle 11g, TOAD, Jboss 6.2.0, Address Doctor 5.
Confidential, Washington D.C
Informatica MDM Consultant
Responsibilities:
- Involved in complete life cycle implementation of Informatica MDM 9.7 implementation for Customer Master.
- Gathered requirements from the business users and developed functional and technical documents.
- Involved in building the data model and did the necessary changes in the documents.
- Installed Informatica MDM platform on JBoss Application Server and did the configuration.
- Created prototypes for Data steward workflows, tasks, roles & Privileges.
- Created landing tables, base object tables and staging tables based on the data model and source systems.
- Involved in discussions with ETL team to get the data loaded into the Landing tables.
- Defined foreign-key relationships between base objects, lookups on staging tables, query groups and queries.
- Developed the mappings and used the address doctor for the stage process.
- Created Match Path Components, Match Columns and Match Rule Sets for Match and Merge Process.
- Executed Stage Process, Load Process, Match and Merge Processes using the Batch Viewer in development phase.
- Developed IDD application as per the business needs and created Subject Area Groups, Subject Areas and Involved with data Steward Team for designing, documenting and configuring IDD.
- Defined Manual rules for doing the Manual Merge using the IDD application by Data Stewards.
- Created all project related documents that will be helpful for project team members and support to continue further.
Environment: Informatica Multi domain MDM 9.7, Power Center 9.1, JBoss 6.2.0, Oracle 11g, SQL Developer, Address Doctor 5, IDD, TOAD, Windows 7.
Confidential
Informatica MDM Developer
Responsibilities:
- Involved in landing the data from Seven Different source systems by running external batch process to load business data for creating customer master data. External batch process used was ETL process.
- Defined the Landing tables, Base objects, Staging tables, foreign key relationships, lookup definitions, queries, packages and query groups.
- Analyzed the source systems for erroneous, duplicative, and integrity issues related to the data which resided in various source systems such as ERP, LEGACY, CRM, DNB, SALES system, and defined those all source systems in systems and trust model workbench.
- Developed and managed Relationships between data domains for the data models.
- Created mappings to perform the tasks such as cleansing the data and populate that into staging tables.
- Analyzed the Business data and defined the Match and Merge rules and Trust Scores for MDM Hub.
Environment: Informatica MDM 9.5, Informatica Power Center 8.6, SIF, Oracle 10g, Toad, JBoss 5.1.0, Windows XP, SSA Name3, Windows server 2003