We provide IT Staff Augmentation Services!

Sr. Informatica Developer / Analyst Resume

0/5 (Submit Your Rating)

West Des Moines, IA

SUMMARY

  • 8+ years of IT experience in design, analysis, development, documentation, coding, and implementation including Databases, Data Warehouse, ETL Design, Oracle, PL/SQL, SQL Server databases, SSIS, Informatica Power Center 9.x/8.x/7.x, Informatica Data Quality, etc
  • Expertise in Master Data Management concepts, Methodologies, and ability to apply this knowledge in building MDM solutions
  • Experience in installation and configuration of core Informatica MDM Hub components such as Hub Console, Hub Store, Hub Server, Cleanse Match Server, and Cleanse Adapter in Windows.
  • Expertise in the ETL Tool Informatica and have extensive experience in Power Center Client tools including Designer, Repository Manager, Workflow Manager/ Workflow Monitor
  • Extensively worked with complex mappings using various transformations like Filter, Joiner, Router, Source Qualifier, Expression, Union, Unconnected / Connected Lookup, Aggregator, Stored Procedure, XML Parser, Normalize, Sequence Generator, Update Strategy, Reusable Transformations, User Defined Functions, etc.
  • Extensively worked on Relational Databases Systems like Oracle11g/10g/9i/8i, MS SQL Server, Teradata and Source files like flat files, XML files and COBOL files
  • Excellent background in implementation of business applications and in using RDBMS and OOPS concepts.
  • Worked with Informatica Data Quality 9.6.1 (IDQ) toolkit, Analysis, data cleansing, data matching, data conversion, exception handling, and reporting and monitoring capabilities using IDQ 9.6.1.
  • Expertise in Data warehousing concepts like OLTP/OLAP System Study, Analysis, and E - R modeling, developing database Schemas like Star schema and Snowflake schema used in relational, dimensional, and multidimensional data modeling.
  • Experienced in mapping techniques for Type 1, Type 2, and Type 3 Slowly Changing Dimensions
  • Hands-on experience in Informatica upgrade from 8.6 to 9.1
  • Extensive experience in debugging mappings, identifying bottlenecks/bugs in existing mappings/workflows by analyzing the data flow and evaluating transformations
  • Solid experience in implementing business requirements, error handling, job control & job auditing using Informatica Power Center tools
  • Strong experience in SQL, PL/SQL, Tables, Database Links, Materialized Views, Synonyms, Sequences, Stored Procedures, Functions, Packages, Triggers, Joins, Unions, Cursors, Collections, and Indexes in Oracle
  • Sound knowledge of Linux/UNIX, Shell scripting, experience in command line utilities like pmcmd to execute workflows in non-windows environments
  • Proficiency in working with Teradata utilities like (BTEQ, FASTLOAD, FASTEXPORT, MULTILOAD, Teradata Administrator, SQL Assistant, PMON, Visual Explain).
  • Implemented change data capture (CDC) using Informatica power exchange to load data from clarity DB to TERADATA warehouse.
  • Experience in integration of various data sources with Multiple Relational Databases like DB2, Oracle, SQL Server, MS Access, Teradata, Flat Files, XML files, CSV, and other sources like Salesforce, etc.
  • Experience in Migrating Data from Legacy systems to Oracle database using SQL*Loader
  • Proficient in Oracle Tools and Utilities such as TOAD and SQL*Loader.
  • Experienced in scheduling Sequence and parallel jobs using DataStage Director, UNIX scripts, Linux Scripts and scheduling tool (Control-M v7/v8), Autoflow, CA WA Workstation (ESP)
  • Expert in analyzing Business & Functional Specifications, creating Technical Design Document and Unit Test cases
  • Experience in Performance Tuning of targets, sources, mapping, workflow, system.
  • Identified and fixed bottlenecks and tuned the complex Informatica mappings for better Performance.
  • Involved in SDLC- software development life cycle (Water, Scrum/Agile) of building a Data Warehouse on windows and Unix Platforms.
  • Well versed with onsite/offshore project delivery model and experience in working with offshore teams
  • Designed Applications according to the customer requirements and specifications.
  • Excellent Interpersonal and Communication skills, coupled with strong technical and problem-solving capabilities.
  • Excellent analytical, problem solving, technical, project management, training, and presentation skills.

TECHNICAL SKILLS

Operating System: UNIX, Linux, Windows

Programming and Scripting: C, C++, Java, .Net, Perl Scripting, Shell Scripting, XSLT, PL/SQL, T-SQL.

Specialist Applications & Software: Informatica Power Center 9.5/9.1/8.6/8.1/7.1, Informatica Power Exchange, Metadata Manager, Informatica Data Quality (IDQ), Informatica Data Explorer (IDE), MDM, SSIS, Salesforce, DataStage, etc.

Data Modeling (working knowledge): Relational Modeling, Dimensional Modeling (Star Schema, Snow-Flake, FACT, Dimensions), Physical, Logical Data Modeling, and ER Diagrams.

Databases tools: SQL Server Management Studio (2008), Oracle SQL Developer (3.0), Toad 11.6 (Oracle), Teradata, AQT v9 (Advanced query tool) (Oracle/Netezza), DB Artisan 9.0 (Sybase), SQL Browser (Oracle Sybase), Visio, ERWIN

Scheduling tools: Informatica Scheduler, CA Scheduler(Autosys), ESP, Maestro, Control-M.

Conversion/Transformation tools: Informatica Data Transformation, Oxygen XML Editor (ver.9, ver.10)

Software Development Methodology: Agile, Water fall.

Domain Expertise: Publishing, Insurance/Finance, HealthCare

Others (working knowledge on some): OBIEE RPD creation, OBIEE, ECM, Informatica Data TransformationXMAP, DAC, Rational Clear Case, WS-FTP Pro, DTD, XML

RDBMS: SQL, SQL*Plus, SQL*Loader, Oracle11g/10g/9i/8i, Teradata, IBM DB2, UDB 8.1/7.0, Sybase 12.5, Netezza v9, MS SQL Server 2000/2005/2008, MS Access 7.0/2000.

PROFESSIONAL EXPERIENCE

Confidential, Houston, TX

Sr. Informatica/IDQ Developer

Responsibilities:

  • Involved in gathering and analysis business requirement and writing requirement specification documents and identified data sources and targets.
  • Communicated with business customers to discuss the issues and requirements.
  • Design, document and configure the Informatica MDM Hub to support loading, cleansing of data.
  • Involved in implementing the Land Process of loading the customer/product Data Set into Informatica MDM from various source systems.
  • Worked on data cleansing and standardization using the cleanse functions in Informatica MDM.
  • Developed ETL programs using Informatica to implement the business requirements.
  • Profiled the data usingInformatica Data Explorer(IDE) and performed Proof of Concept forInformatica Data Quality(IDQ)
  • Used Informatica Power Center to load data from different data sources like xml, flat files, CSV, MS Acess and Oracle, Teradata, Salesforce.
  • Imported the IDQ address standardized mappings into Informatica Designer as a Mapplets.
  • Used relational SQL wherever possible to minimize the data transfer over the network.
  • Effectively used Informatica parameter files for defining mapping variables, workflow variables, FTP connections and relational connections.
  • Involved in maintenance and enhancements activities of the data warehouse including tuning, modifying of stored procedures for code enhancements.
  • Effectively worked in Informatica version based environment and used deployment groups to migrate the objects.
  • Used debugger in identifying bugs in existing mappings by analyzing data flow, evaluating transformations.
  • Worked on CDC (Change Data Capture) to implement SCD (Slowly Changing Dimensions) Type 1 and Type 2
  • Involved in creating UNIX shell scripts for Datastage job and Informatica workflow execution.
  • Designed workflows with many sessions with decision, assignment task, event wait, and event raise tasks, used informatica scheduler to schedule jobs.
  • Used the Teradata fast load utilities to load data into tables
  • Used SQL tools like TOAD to run SQL queries and validate the data.
  • Wrote complex SQL override scripts at source qualifier level to avoidInformaticajoiners and Look-ups to improve the performance, as the volume of the data was heavy.
  • Converted all the jobs scheduled in Maestro to Autosys scheduler as the per requirements
  • Worked on maintaining the master data using Informatica MDM
  • Wrote UNIX Shell Scripts for Informatica Pre-Session, Post-Session and Autosys scripts for scheduling the jobs (work flows)
  • Used Erwin for Logical and Physical database modeling of the warehouse, responsible for database schema creation based on the logical models
  • Performed tuning of queries, targets, sources, mappings, and sessions.
  • Used Unix scripts and necessary Test Plans to ensure the successful execution of the data loading process
  • Worked with the Quality Assurance team to build the test cases to perform unit, Integration, functional and performance Testing.
  • Developed UNIX Korn shell scripts for archiving, zipping and cleanup bad files data from logs
  • Reviewed and analyzed functional requirements, mapping documents, problem solving and trouble shooting.
  • Identified problems in existing production data and developed one time scripts to correct them.
  • Provided Knowledge Transfer to the end users and created extensive documentation on the design, development, implementation, daily loads, and process flow of the mappings
  • Fixed the invalid mappings and troubleshoot the technical problems of the database.
  • Provided 24x7 production supports for business users and documented problems and solutions for running the workflows.

Environment: Informatica Power Center 9.6.1, UNIX, SQL, IDQ, IDE, CDC, MDM, Linux, Perl, WINSCP, Shell, PL/SQL, Netezza, Maestro, Teradata, Erwin, MS Access, MS Excel, Salesforce, Microsoft SQL Server 2008, and Microsoft Visual studio

Confidential, West Des Moines, IA

Sr. Informatica Developer / Analyst

Responsibilities:

  • Involved in all phases of SDLC from requirement gathering, design, development, testing, Production, and support for production environment.
  • Strong expertise in installing an d configuring the core Informatica MDM components (Informatica MDM Hub Server, Hub Cleanse, Resource Kit, and Cleanse Adaptors like Address Doctor)
  • Used Informatica Power Center for (ETL) extraction, transformation and loading data from heterogeneous source systems into target database.
  • Created mappings using Designer and extracted data from various sources, transformed data according to the requirement.
  • Involved in extracting the data from the Flat Files and Relational databases into staging area.
  • Created Stored Procedures for data transformation purpose.
  • Involved in the Dimensional Data Modeling and populating the business rules using mappings into the Repository for Data management
  • Worked on Informatica Power Center 9x tool - Source Analyzer, Data warehousing designer, Mapping, Designer, Mapplet & Transformations.
  • Various kinds of the transformations were used to implement simple and complex business logic.
  • Creatednumerous Mappings and Mapplets using Transformations like Filters, Aggregator, Lookups, Expression, Sequence generator, Sorter, Joiner, and Update Strategy.
  • Created and configured workflows, worklets & Sessions to transport the data to target warehouse Oracle tables using Informatica Workflow Manager.
  • Worked in building Data Integration and Workflow Solutions and Extract, Transform, and Load (ETL) solutions for data warehousing using SQL Server Integration Service (SSIS).
  • Mapplets and Reusable Transformations were used to prevent redundancy of transformation usage and maintainability.
  • Involved in analyzing existing logical and physical data modeling with STAR and SNOWFLAKE schema techniques using Erwin in Data warehouse
  • Used Teradata Utilities (BTEQ, Multi-Load, and Fast-Load) to maintain the database.
  • Build a re-usable staging area in Teradata for loading data from multiple source systems using template tables for profiling and cleansing in IDQ
  • Actively involved in Data validations and unit testing to make sure data is clean and standardized before loading in to MDM landing tables.
  • Troubleshoot problems by checking sessions and error logs.
  • Configured sessions using server manager to have multiple partitions Source data & to improve performance.
  • Actively involved in Exceptional handling process in IDQ Exceptional transformation after loading the data in MDM and notify the Data Stewards with all exceptions.
  • Worked on Autoflow as job scheduler and used to run the created application and respective workflow in this job scheduler in selected recursive timings.
  • Wrote SQL override scripts at SQ level to avoid Informatica joiners and Look-ups to improve the performance
  • Generated PL/SQL and Perl scripts for scheduling periodic load processes.
  • Extensively worked on the Triggers, Functions, and Database Constraints.
  • Tuned Informatica Mappings and Sessions for optimum performance.
  • Designed and developed Linux scripts for creating, dropping tables which are used for scheduling the jobs.
  • Invoked Informatica using "pmcmd" utility from the UNIX script.
  • Wrote pre-session shell scripts to check session mode (enable/disable) before running/ schedule batches.
  • Participated in problem solving and troubleshooting for the applications implemented with Informatica.
  • Perform process analysis to provide detailed documentation and recommendations to the Load forecasting team for future improvements
  • Involved in supporting 24*7 rotation system and strong grip in using scheduling tool Tivoli.
  • Involved in Production support activities like batch monitoring process in UNIX
  • Involved in Production Support by performing Normal Loads, Bulk Loads, Initial Loads, Incremental Loads, Daily loads and Monthly loads.
  • Transition support of the Load forecasting applications/tools to an IT support organization and implement standard IT support processes and procedures.
  • Prepared Unit test case documents
  • Have performed Peer Reviews within the project.

Environment: Informatica Power Center 9.6.1, UNIX, Oracle, Linux, Perl, Shell, MS Access, MS Excel, MDM, IDQ, IDS, PL/SQL, Tivoli, Oracle 11g/10g, Teradata 14.0, Maestro.

Confidential, Charlotte, NC

Sr. ETL Developer / Analyst

Responsibilities:

  • Worked with Business Analysts (BA) to analyze the Data Quality issue and find the root cause for the problem with the proper solution to fix the issue...
  • Document the process that resolves the issue which involves analysis, design, construction and testing for Data quality issues
  • Involved in doing the Data model changes and other changes in the Transformation logic in the existing Mappings according to the Business requirements for the Incremental Fixes
  • Worked extensively with Informatica tools like Source Analyzer, Warehouse Designer, Transformation Developer, and Mapping Designer.
  • Extracted data from various heterogeneous sources like DB2, Sales force, Mainframes, Teradata, and Flat Files using Informatica Power center and loaded data in target database.
  • Created and Configured Landing Tables, Staging Tables, Base Objects, Foreign key relationships, Queries, Query Groups etc. in MDM.
  • Defined the Match and Merge Rules in MDM Hub by creating Match Strategy, Match columns and rules for maintaining Data Quality
  • Extensively Implemented SCD TYPE 2 Mappings for CDC (Change data capture) in EDW.
  • Code walkthrough and Review of documents which are prepared by other team members.
  • Involved in doing Unit Testing, Integration Testing, and Data Validation.
  • Migration of Informatica maps extensively involved in migration from Dev to SIT and UAT to Production environment.
  • Developed Unix script to sftp, archive, cleanse and process many flat files
  • Created and ran Pre-existing and debug sessions in the Debugger to monitor and test the sessions prior to their normal run in the Workflow Manager
  • Extensively worked in migrating the mappings, worklets and workflows within the repository from one folder to another folder as well as among the different repositories.
  • Created Mapping parameters and Variables and written parameter files.
  • Implemented various Performance Tuning techniques by finding the bottle necks at source, target, mapping, and session and optimizing them.
  • Used SQL queries and database programming using PL/SQL (writing Packages, Stored Procedures/Functions, and Database Triggers).
  • Worked on Scheduling Jobs and monitoring them through Control M and CA scheduler tool(Autosys).
  • Used SQL tools like TOAD to run SQL queries and validate the data in warehouse.
  • Worked with the SCM code management tool to move the code to Production
  • Extensively worked with session logs and workflow logs in doing Error Handling and trouble shooting.

Environment: Informatica Power Center 9.0.1, Erwin, Teradata, Tidal, MS Access, MS Excel, SQL Assistance, DB2, XML, Oracle 9i/10g/11g, MQ Series,perl, OBIEE 10.1.3.2, IDQ, MDM Toad and UNIX Shell Scripts.

Confidential, New York

Sr. Informatica Developer

Responsibilities:

  • Designing the dimensional model and data load process using SCD Type 2 for the quarterly membership reporting purposes.
  • Derived the dimensions and facts for the given data and loaded them on a regular interval as per the business requirement.
  • Generating the data feeds from analytical warehouse using required ETL logic to handle data transformations and business constraints while loading from source to target layout.
  • Worked on Master Data Management (MDM), Hub Development, extract, transform, cleansing, loading the data onto the staging and base object tables
  • Extracted data from multiple sources such as Oracle, XML, and Flat Files and loaded the transformed data into targets in Oracle, Flat Files.
  • Wrote Shell Scripts for Data loading and DDL Scripts.
  • Designing and coding the automated balancing process for the feeds that goes out from data warehouse.
  • Implement the automated balancing and control process which will enable the control on the audit and balance and control for the ETL code.
  • Improving the database access performance by tuning the DB access methods like creating partitions, using SQL hints, and using proper indexes.
  • All the jobs are integrated using complex Mappings including Mapplets and Workflows using Informatica power center designer and workflow manager.
  • Worked on data cleansing and standardization using the cleanse functions in Informatica MDM.
  • Created ETL mappings, sessions, workflows by extracting data from MDM system, FLASH DC & REM source systems.
  • Use debugger in identifying bugs in existing mappings by analyzing data flow, evaluating transformations, and create mapplets that provides reusability in mappings.
  • Analyzing the impact and required changes to in corporate the standards in the existing data warehousing design.
  • Following the PDLC process to move the code across the environments though proper approvals and source control environments.

Environment: Informatica Power Center 9.0.1, Erwin 7.2/4.5, Business Objects XI, Unix Shell Scripting, XML, Oracle 11g/10g, DB2 8.0, IDQ, MDM, TOAD, MS Excel, Flat Files, SQL Server 2008/2005, PL/SQL, Windows NT 4.0, Sun Solaris 2.6.

Confidential

Informatica Developer

Responsibilities:

  • Involved in analysis, design, development, test data preparation, unit and integration testing, Preparation of Test cases and Test Results
  • Coordinating with client, Business and ETL team on development
  • Developed Batch jobs using extraction programs using COBOL, JCL, VSAM, Datasets, FTP to Load Informatica tables
  • Based on the business requirements created Functional design documents and Technical design specification documents for ETL Process.
  • Designing and developing ETL solutions in Informatica Power Center 8.1
  • Designing ETL process and creation of ETL design and system design documents.
  • Developing code to extract, transform, and load (ETL) data from inbound flat files and various databases into various outbound files using complex business logic.
  • Most of the transformations were used like Source Qualifier, Aggregator, Filter, Expression, and Unconnected and connected Lookups, Update Strategy.
  • Created automated shell scripts to transfer files among servers using FTP, SFTP protocols and download files.
  • Developed Informatica mappings, enabling the ETL process for large volumes of data into target tables.
  • Designed and developed process to handle high volumes of data and high volumes of data loading in each SLA.
  • Created Workflow, Worklets and Tasks to schedule the loads at required frequency using Informatica scheduling tool.
  • Expertise in creating control files to define job dependencies and for scheduling using Informatica.
  • Provided effective support in delivering process and product change improvement solutions.
  • Ensured acceptable performance of the data warehouse processes by monitoring, researching and identifying the root causes of bottlenecks.
  • Acted as a liaison between the application testing team and development team in order to make timely fixes.
  • Involved in jobs scheduling, monitoring and production support in a 24/7 environment.

Environment: Informatica Power Center 8.1, ETL, Business Objects, Oracle 10g/9i/8i, HP - Unix, PL/SQL.

We'd love your feedback!