Sr.etl Informatica Developer /data Analyst Resume
Cary, NC
SUMMARY
- Data analyst/ETL Informatica expert with around 6+ years of Total IT Experience in analysis, design, development, testing, deployment, and production support of Enterprise Data warehouse applications using Informatica across Healthcare, Manufacturing and Retail domains under all the phases of SDLC.
- 6+ years of vast experience in developing ETL interfaces using Informatica 10.0, 9.x, 8.x and IDQ 9.1 for Enterprise Data warehouse and standalone data marts.
- Expertise in developing end - to-end integration solutions using DataStage and multiple databases like ORACLE, DB2, Teradata, Netezza, MySQL, PostGreSQL and SQL Server.
- Superior SQL skills and ability to write and interpret complex SQL statements and skillful in SQL optimization and ETL debugging and performance tuning
- Experience in developing of on-line transactional processing (OLTP), operational data store (ODS) and decision support system (DSS) (e.g., Data Warehouse) databases.
- Worked with Teradata utilities like Fast Load and Multi Load and Tpump and Teradata Parallel transporter and highly experienced in Teradata SQL Programming.
- Experience in Teradata Parallel Transporter (TPT). Used full PDO on Teradata and worked with different Teradata load operators.
- Experienced in writing SQL, PL/SQL programming, Stored Procedures, Package, Functions, Triggers, Views, Materialized Views.
- Experience in Inmon and Kimball data warehouse design and implementation methodologies
- Hands-on knowledge of enterprise repository tools, data modeling tools, data mapping tools, data profiling tools and data and information system life cycle methodologies.
- Experience with dimensional modeling and architecture experience implementing proper data structures for analytical reporting from an enterprise data warehouse.
- Implemented Change Data Capture (CDC) with Informatica Power Exchange.
- Used Informatica Power Exchange to access VSAM files also worked on Flat files, nested JSON and XML files
- Well versed with data quality features like Analyst, IDQ & transformation like Key Generator, Standardizer, Case Converter, Match Consolidation etc.
- Applied Address transformation for Address Validation and Standardization.
- Strong in implementation of data profiling, documenting Data Quality metrics like Accuracy, completeness, duplication, validity, consistency.
- Have good skills on understanding and development of business rules for its Standardization, Cleanse and Validation of data in various formats.
- Very strong knowledge of Informatica Data Quality transformations like Address validator, Parser, Labeler, Match, Exception, Association, Standardizer and other significant transformations.
- Experience in Data modeling; Dimensional modeling and E-R Modeling and OLTP and OLAP in Data Analysis. Very familiar with SCD1 and SCD2 in snowflake schema and star schema.
- Proficient on Data Warehouse Concepts like Dimension Tables, Fact Tables, Slowly Changing Dimensions, DataMart’s and Dimensional modeling schemas.
- Strong Experience in developing Sessions/Tasks, Worklets and Workflows using Workflow Manager tools -Task Developer, Workflow & Worklet Designer.
- Experience in performance tuning of Informatica mappings and sessions to improve performance for the large volume projects.
- Experience in Extraction, Transformation and Loading (ETL) data from various data sources into Data Marts and Data Warehouse using Informatica power center components (Repository Manager, Designer, Workflow Manager, Workflow Monitor and Informatica Administration Console).
- Extensively worked on Informatica Power center transformations as well like Expression, Joiner, Sorter, Filter, Router, Normalizer, rank, Lookups, Stored procedure, Update strategy, Source Qualifier, Union, CDC and other transformations as required.
- Experience in debugging mappings, identified bugs in existing mappings by analyzing the data flow and evaluating transformations.
- Experience in writing UNIX shell scripts, SQL scripts for development, automation of ETL process, error handling and auditing purposes.
- Well versed with waterfall and Agile methodologies in Software Development Life Cycle (SDLC)
- Good knowledge of AWS (Amazon Web Services), S3 Bucket and Redshift (AWS Relational Database).
- Strong Knowledge of BigData / Hadoop Ecosystem like Hive, Impala, HDFS.
- Have got good problem-solving skills. Team player with excellent analytical, communication and multi-tasking skills.
TECHNICAL SKILLS
ETL: Informatica PowerCenter 10.0, 9.5.1, 9.0, 8.1.1
Data Profiling Tools: Informatica IDQ 10.0, 9.5.1, 8.6.1
ETL Scheduling Tools: Control M, ESP.
RDBMS: DB2, Oracle 11g/12c, Teradata 13/15, SQL Server 2008/2012, MySQL, PostgreSQL 9.2
Data Modeling: ER (OPLTP) and Dimensional (Star, Snowflake Schema);
Data Modeling Tools: Erwin 9.3/7.5
UNIX: UNIX, Shell scripting
Reporting Tools: Tableau 9, Cognos 9/10
Defect Tracking Tools: Quality Center, Bugzilla
Operating Systems: Windows XP/2000/9x/NT, UNIX
Source Management: BitBucket, Visual SourceSafe
Cloud Computing: Amazon Services (AWS), S3, Redshift
Programming Languages: C, C++, PL/SQL
Other Tools: Notepad++, Toad, SQL Navigator, Teradata SQL Assistant, Teradata view point, JIRA, Rally
PROFESSIONAL EXPERIENCE
Confidential, Cary, NC
Sr.ETL Informatica Developer / Data Analyst
Responsibilities:
- Involved in all phases of SDLC from requirement, design, development, testing and support for production environment.
- Extensively used Informatica Client tools like Informatica Repository Manager, Informatica Designer, Informatica Workflow Manager and Informatica Workflow Monitor.
- Used Teradata utilities Fast Load, Multi Load, tpump to load data.
- Created Sources, Targets in shared folder and developed re-usable transformations, mapplets and user defined function (UDF) to re-use these objects in mappings to save the development time.
- Developed mappings using Mapping Designer and worked with Aggregator, Lookup (connected and unconnected), Filter, Router, Joiner, Source Qualifier, Expression, Stored Procedure, Sorter and Sequence Generator transformations.
- Created mappings which involved Slowly Changing Dimensions Type 1 and Type 2 to implement business logic and capturing the deleted records in the source systems.
- Involved in analyzing the site usage files data coming from external vendors using sql after loading them in to work database.
- Involved in migration projects to migrate data from data warehouses on Oracle/DB2 and migrated those to Teradata.
- Experience with high volume data sets from various sources like Oracle, Text Files and Teradata Tables
- Used debugger extensively to identify the bottlenecks in the mappings.
- Modified PL/SQL stored procedures for Informatica mappings.
- Created Sessions and Workflows to load data from the SQL server, flat file and Oracle sources that exist on servers located at various locations all over the country.
- Configured the session properties i.e. high value for commit intervals to increase the performance.
- Involved in unit testing, Integration testing and User acceptance testing of the mappings.
- Involved in Migrating the Informatica objects using Unix SVN from Dev to QA Repository.
- Worked on developing workflows and sessions and monitoring them to ensure data is properly loaded on to the target tables.
- Responsible for scheduling workflows, error checking, production support, maintenance and testing of ETL procedures using Informatica session logs.
- Performance tuning on sources, targets mappings and SQL (Optimization) tuning.
Environment: Informatica Power Center 9.5.1/9.6.1, Flat Files, Oracle 11g, Teradata 14, SQL, PL/SQL, TOAD, SQL Assistant, Windows XP, Unix, Maestro, SVN.
Confidential, Detroit, MI
ETL Informatica developer / Data Analyst
Responsibilities:- Involved in all phases of SDLC from requirement gathering, design, development, testing, Production, user training and support for production environment.
- Actively involved in interacting with business users to record user requirements and Business Analysis.
- Involved in Analysis, profiling and cleansing of source data and understanding the business process.
- Translated requirements into business rules & made recommendations for innovative IT solutions.
- Outlined the complete process flow and documented the data conversion, integration and load mechanisms to verify specifications for this data migration project.
- Involved in documentation of Data Mapping & ETL specifications for development from source to target.
- Also implemented the change Data Capture (CDC) while integrating the enterprise data sources.
- Parsing high-level design spec to simple ETL coding and mapping standards.
- Worked with PowerCenter Designer tools in developing mappings and Mapplets to extract and load the data from flat files and Oracle database.
- Maintained warehouse metadata, naming standards and warehouse standards for future application development.
- Created the design and technical specifications for the ETL process of the project.
- Used Informatica as an ETL tool to create source/target definitions, mappings and sessions to extract, transform and load data into staging tables from various sources.
- Responsible for mapping and transforming existing feeds into the new data structures and standards utilizing Router, Lookups Using Connected, Unconnected, Expression, Aggregator, Update strategy & stored procedure transformation.
- Worked on Informatica PowerCenter tool - Source Analyzer, Data Warehousing Designer, Mapping Designer & Mapplets, and Transformations.
- Maintained Development, Test and Production Mappings, migration using Repository Manager. Involved in enhancements and Maintenance activities of the data warehouse.
- Performance tuning of the process at the mapping level, session level, source level, and the target level.
- Implemented various new components like increasing the DTM Buffer Size, Database Estimation, Incremental Loading, Incremental aggregation, Validation Techniques, and load efficiency.
- Strong on Exception Handling Mappings for Data Quality, Data Cleansing and Data Validation.
- Worked with SQL*Loader to load data from flat files obtained from various facilities.
- Created Workflows containing command, email, session, decision and a wide variety of tasks.
- Tuning the mappings based on criteria, creating partitions in case of performance issues.
- Tested End to End to verify the failures in the mappings using scripts.
- Performed data validation after the successful End to End tests and appropriate error handling in ETL processes.
- Resolving the tickets based on the priority levels raised by QA team.
- Facilitated in developing testing procedures, test cases and User Acceptance Testing (UAT)
- Developed Parameter files for passing values to the mappings for each type of client
- Scheduled batch and sessions within Informatica using Informatica scheduler and also wrote shell scripts for job scheduling.
Environment: Informatica Power Center 10.2, Oracle 12c, Cognos 10.2, DB2, Flat files, SQL, putty, UltraEdit-32, shell scripting, Toad, Quest Central, UNIX scripting, Windows NT
Confidential,Houston, TX
ETL Informatica developer
Responsibilities:- Actively involved in interacting with business users to record user requirements and Business Analysis.
- Involved in Analysis, profiling and cleansing of source data and understanding the business process.
- Translated requirements into business rules & made recommendations for innovative IT solutions.
- Outlined the complete process flow and documented the data conversion, integration and load mechanisms to verify specifications for this data migration project.
- Involved in documentation of Data Mapping & ETL specifications for development from source to target.
- Also implemented the change Data Capture (CDC) while integrating the enterprise data sources.
- Parsing high-level design spec to simple ETL coding and mapping standards.
- Worked with PowerCenter Designer tools in developing mappings and Mapplets to extract and load the data from flat files and Oracle database.
- Maintained warehouse metadata, naming standards and warehouse standards for future application development.
- Created the design and technical specifications for the ETL process of the project.
- Used Informatica as an ETL tool to create source/target definitions, mappings and sessions to extract, transform and load data into staging tables from various sources.
- Responsible for mapping and transforming existing feeds into the new data structures and standards utilizing Router, Lookups Using Connected, Unconnected, Expression, Aggregator, Update strategy & stored procedure transformation.
- Worked on Informatica PowerCenter tool - Source Analyzer, Data Warehousing Designer, Mapping Designer & Mapplets, and Transformations.
- Worked with slowly changing dimension Type1, Type2, and Type3.
- Maintained Development, Test and Production Mappings, migration using Repository Manager. Involved in enhancements and Maintenance activities of the data warehouse.
- Performance tuning of the process at the mapping level, session level, source level, and the target level.
- Implemented various new components like increasing the DTM Buffer Size, Database Estimation, Incremental Loading, Incremental aggregation, Validation Techniques, and load efficiency.
- Strong on Exception Handling Mappings for Data Quality, Data Cleansing and Data Validation.
- Worked with SQL*Loader to load data from flat files obtained from various facilities.
- Created Workflows containing command, email, session, decision and a wide variety of tasks.
- Tuning the mappings based on criteria, creating partitions in case of performance issues.
- Tested End to End to verify the failures in the mappings using scripts.
- Performed data validation after the successful End to End tests and appropriate error handling in ETL processes.
- Resolving the tickets based on the priority levels raised by QA team.
- Facilitated in developing testing procedures, test cases and User Acceptance Testing (UAT)
- Developed Parameter files for passing values to the mappings for each type of client
- Scheduled batch and sessions within Informatica using Informatica scheduler and wrote shell scripts for job scheduling.
Environment: Informatica PowerCenter 9,1.0, Informatica PowerExchange 9.1.0, Informatica DataQuality 9,1.0, Cognos 9.0, Linux, SQL, PL/SQL, Oracle 11g, TOAD, Snowflake cloud data warehouse, SQL Server 2012, HDFS, Hive, SparkSQL, Control M, Shell Scripting, XML, SQL Loader, Putty, WinSCP
Confidential, Hartford, CT
Sr. Informatica Developer
Responsibilities:- Worked on Informatica Power Center 9.1 tool - Source Analyzer, Data Warehousing designer, Mapping &Mapplet Designer and Transformation Designer. Developed Informatica mappings and also in tuning of mappings for better performance.
- Participated in deployment planning and in deployment of the system to production.
- Facilitated business user smoke testing of the production system by setting up test data.
- Involved in production support duties including monitoring of nightly batches.
- Responsible for updating business stakeholders and OLTP/OLAP application support teams about the status of various ETL sessions and the impact of failed sessions on data availability.
- Responsible for debugging failed ETL sessions and taking steps to mitigate impact to business users.
- Extensively used informatica Informatica Data Quality tool (IDQ Developer) to create rule-based data validations for profiling.
- Utilized Informatica IDQ to complete the initial data profiling and matching/removing duplicate data for the process of data migration from the legacy systems to the target Oracle Database.
- Created dictionary tables using IDQ analyst tool for data validations.
- Implemented B2B informatica parsers to convert PDF documents to XMLs through Informatica.
- Involved in code review sessions to bring the team’s attention to best practices and to identify areas of improvement.
- Responsible for implementing action items identified during code review sessions focused towards improving data quality and performance.
- Involved in performance tuning of sessions that work with large sets of data by tweaking block size, data cache size, sequence buffer length and target-based commit intervals.
- Developed sessions and batches to move data at specific intervals and on demand using workflow manager.
- Worked on different modules of the project for successful execution.
- Have created / developed diverse types of profiles like Column level profiling, Summary profiles, drill down profiles, Score cards, reports etc. using IDE.
- Have created Match, and Merge rules, developed address validations etc. and developed address validations to countries like US, and developed reusable error handling rules using IDQ.
- Utilized Informatica IDQ to complete the initial data profiling and matching/removing duplicate data for the process of data migration from the legacy systems to the target Oracle Database.
Environment: Informatica power center 9.1, IDQ Developer 9.1, Oracle 10g, PL/SQl, Toad, HP-UX, Informatica B2B, Windows, SQL.
Confidential, Minneapolis, MN
Informatica Developer
Responsibilities:- Involved in gathering business requirements, logical modeling, physical database design, data sourcing and data transformation, data loading, SQL and performance tuning.
- Participated in data analysis, data Profiling, data dictionary and metadata management. Used SQL to do the Data Profiling.
- Collaborated with Business users to collect requirements and prepared ETL technical specifications.
- Developed, supported and maintained the ETL processes for exporting data from application into reporting data mart using Informatica Power center 8.6.
- Designed, built and maintained mappings, sessions and workflows for the data load process using Informatica, PL/SQL and UNIX.
- Implemented customer history data capture for catalogue tables using SCD Type 2.
- Designed mappings for Slowly Changing Dimensions (Type 1 and Type 2), used Lookup (connected and unconnected), Update strategy and filter transformations for loading historical data.
- Experience in debugging execution errors using Data Services logs (trace, statistics and error) and by examining the target data.
- Tuned performance for large data files by increasing block size, cache size and implemented performance tuning logic on sources, workflows, data flow's and SAP ECC 6.0 target system in order to provide maximum efficiency and performance.
- Worked extensively with Informatica tools such as Source Analyzer, Warehouse Builder and Workflow Manager.
- Extensively used transformations like router, aggregator, lookup, source qualifier, joiner, expression and sequence generator transformations in extracting data in compliance with the business logic developed.
- Wrote SQL overrides in source qualifier to filter data according to business requirements.
- Wrote Unix shell scripts for scheduling Informatica pre/post session operations.
- Created different parameter files and started sessions using these parameter files using pmcmd command to change session parameters, mapping parameters, and variables at runtime.
- Tuned the mappings by removing the Source/Target bottlenecks and Expressions to improve the throughput of the data loads.
Environment: Informatica PowerCenter 8.1, Cognos 8, SQL Server 2008, Oracle 11g, PL/SQL, TOAD, Putty, Autosys Scheduler, UNIX, Teradata V2R6, Oracle SQL Developer, ESP, WinSCP, UltraEdit-32, shell Programming, Toad, SQL Developer, UNIX scripting, Windows NT.