We provide IT Staff Augmentation Services!

Informatica Etl Design/ Dev Lead Resume

New York, NY


  • Around 7+ years of experience in design and development of data warehouse and database solutions.
  • Good understanding of Data warehousing and ETL concepts and business application, business dataflow and data relation.
  • Good understanding on Data Quality Management, Master Data Management, Data Integration techniques and Data Migration across Healthcare, Life Sciences, Telecom and Retail industry domains.
  • Thorough knowledge in various tools and technologies like different versions of Informatica Data Quality and Data Analyst, Informatica Power Center (9.X,10) and Informatica Cloud.
  • Experience in designing and developing mappings from varied transformation logic like Unconnected and Connected lookups, Router, Filter, Expression, Aggregator, Joiner, Update Strategy etc.
  • Experience in Performance Tuning the mappings
  • Experience in developing Test Plans, Test Strategies and Test Cases for Data Warehousing projects ensuring the data meets the business requirements, also Good Work experience in SQL, PL/SQL, UNIX
  • Involved in knowledge sharing and knowledge transfer to the new members of the project & Preparation of knowledge transfer documents.
  • Having Good Communication skills, Interpersonal relations, hardworking and result - oriented as an Individual and in Team.
  • Have extensive experience in requirement analysis, database design, development and Testing.
  • Excellent problem-solving skills with a strong technical background and result oriented team player with excellent communication and interpersonal skills
  • Extensively used ETL methodology for supporting Data Extraction, Transformation and Loading processing, in a corporate-wide-ETL Solution using Informatica.
  • Extensively used Teradata BTEQ Scripts for Data Extraction, Transformation and Loading into tables.
  • Extensively used Teradata utilities Fast Load, Multi Load for to load data from flat files to tables.
  • Performed Teradata code tuning on a regular basis.
  • Good understanding and experience in unstructured data management and modeling concepts (Big data belonging to different domain like Social media, Health care, web usage data and relational data).
  • Have Data Analysis experience in analytical projects for the client in Healthcare domain.
  • Knowledge on Big Data, Predictive Modeling, Data Mining and data analysis using R.


ETL Tool: Informatica 9X,10, BDE

Operating Systems: Windows, Unix

Application Tools: ControlM, Confidential Remedy, Putty, Toad, WinScp, Edit Plus, Teradata SQL Assistant, TIDAL scheduler, Cloudera, R

Database: Teradata, Oracle 10g, Netezza 7.1


Confidential, New York, NY

Informatica ETL Design/ Dev Lead


  • Involved in the Designing the ETL Solution.
  • Involved in Analysis and discussion with business regarding the legacy BE HIST/Ariel systems for the assessment of the efforts.
  • Involved in preparation of technical design documents, low level functional design documents and Velocity documents.
  • Developing the PL/SQL and UNIX Shell scripts.
  • Developing the Informatica Mappings/Mapplets/Worklets and workflows for loading the data to MySQL data marts.
  • Designed various UDF’s (User Defined Functions) in Informatica designer to be used in various mappings.
  • Data coming in as source definitions are from different data sources types like Flat files, XML, Sales force, Oracle, SQL Server and Hadoop files (HDFS).
  • Hands on experience in middleware integration components (iPaaS) platforms like such as Informatica Cloud Real Time (ICRT), Informatica Cloud.
  • Proficiency implementing integrations with third party CRM data(Salesforce)
  • Created various shell scripts for pre and post processing of flat files.
  • Optimizing and doing performance tuning of mappings/sessions/workflows to achieve higher response times.
  • Created Cron/Control M jobs to schedule the Informatica data loads.
  • Created test cases and test scenarios to test the incremental loads of data in unit testing.
  • Creating workflow dependency diagram for the code automation.
  • Coordinating all releases successfully according to all compliance rules

Environment: Informatica PowerCenter 10.1, Informatica Analyst, Informatica IDQ/Developer, MySQL 5.7, SQL Server 2005, UNIX Shell Scripting, Crontab, Autosys, Git Version control system, SVN, HP ALM and Agile Methodology


ETL Specialist/ Onsite coordinator


  • Validated data to ensure the quality, validity and accuracy of content.
  • Part of reconciliation activity which ensures the Source and Vision data is in sync
  • Day to Day reporting to the client - Send data verification report (comparing data from multiple systems) before Users login into their systems.
  • Monitored workflows and collected performance data to maximize the session performance.
  • Involved in development of Informatica mappings and tuned for better performance
  • Incident resolution using ALM system, and production support, handling production failures and fix them within SLA.
  • Creating/modifying Informatica Workflows and Mappings (power center and Cloud) also involved in unit testing, internal quality analysis procedures and reviews.
  • Validated and fine-tuned the ETL logic coded into existing Power Center Mappings, leading to improved performance.
  • Part of Informatica cloud integration with amazon redshift
  • Managing Informatica Cloud based tools including Address Dr.
  • Integrated data from various source systems like SAP (ABAP), SQL Server (DTS), Ariba, Oracle (ODBC) so on.
  • Part of SDLC (Software Development Life Cycle) Requirements, Analysis, Design, Testing, Deployment of Informatica Power Center.
  • Very familiar with SCD1 and SCD2 in snowflake schema and star schema.
  • Identifying and tracking the slowly changing dimensions (SCD), used CDC logic (Change Data Capture) for the SCD tables loading in Oracle.
  • Good Experience in writing UNIX shell scripts, SQL scripts for development, automation of ETL process, error handling and auditing purposes.
  • Created Stored Procedures, Packages, Triggers, Cursors and backup-recovery for the various tables
  • Leveraged Explain Plan to improve the performance of SQL queries and PL/SQL Stored procedures.
  • Working and understanding on the end to end features of Hadoop-hive
  • Creating the Sqoop job to store the data in database.
  • Analyzing the FLUME job which will place format and the data to HDFS.
  • Analyzing and modifying PIG scripts
  • Experience is RCA, preparing Business Specification Document and Technical Designs Document.
  • Provided in-depth Root Cause Analysis reports in relation to production load failures and critical change requests.
  • Involved in scheduling the jobs by using TIDAL scheduling tool (ETL workflows invoked using pmcmd, used EXEC to invoke Teradata Macros and run file for BTEQ)
  • Prepared Unit Test Case and Unit Test Results for change requests.
  • Active participation in WSR (Weekly Status Report) meeting to Communicate Requirements, Development and Data related issues to concerned teams.
  • Basic reporting knowledge on Tableau.
  • EDA to understand the behavior of the product and customer at various levels
  • Time based cross validation data preparation
  • Dimension reduction for best possible predictors
  • Built RFM (Regency, Frequency and Monetary) analysis is followed by Cluster Analysis where customers with similar RFM characteristics are grouped together to arrive at segments
  • Used the top ranked customers who has high propensity to buy the products for focused campaign.
  • Calculating the incremental revenue metric. Incremental Revenue is the difference between the actual revenue from the Target Group and the expected revenue from the same group. It is calculated at a campaign level.
  • Understood the data by exploratory data analysis and classified the category into four subgroups: Highly Value Customer, Active Customer, Moderately Active customer and Inactive Customers.
  • Pre-modeling process like missing value imputing, outlier detection, data transformation etc.
  • Variable reduction to get the best predictors
  • Developed a Logistic regression model to get the rejection probability at transaction level
  • Model was tested on the testing data set
  • Developed a scoring mechanism to predict churn of Rejection probability at customer level.

Environment: Informatica 9x, Informatica Cloud, Teradata, TIDAL scheduler, IRIS, Data Science, R, Cloudera


ETL Developer/ Data Analyst


  • Migration of code.
  • Writing basic UNIX shell scripts and PL/SQL packages and procedures.
  • Involved in performance tuning of the mappings, session and SQL queries.
  • Creating/modifying Informatica Workflows and Mappings.
  • Working with various transformations such as Expression, Aggregator, Update Strategy, Look Up, Filter, Router, Joiner and Sequence generator etc. in Informatica 8.6.1 for new requirement.
  • Created reusable Tasks, Sessions, reusable Worklets and workflows in Workflow manager.
  • Sound knowledge with Netezza SQL
  • Managed Netezza queries by performance tuning and techniques such as CBT and Collocation.
  • Identifying and tracking the slowly changing dimensions (SCD), used CDC logic (Change Data Capture) for the SCD tables loading in Oracle.
  • Communicate Requirements, Development and Data related issues to concerned teams.
  • Interacting with Business Users for Business logic Clarifications
  • Conduct data quality analysis on various source data systems
  • Presentation of Issues and findings to subject matter experts and providing in depth root cause analysis.

Environment: Informatica 8.6/9.1, Oracle 9i, Netezza 7.1


Informatica/ ETL Developer


  • Extracting and loading of data from flat files and Oracle sources to Oracle (target warehouse) using transformations in the mappings.
  • Developed a series of mappings for handling different cases of input data.
  • Involved in development of Informatica mappings and tuned for better performance.
  • Worked extensively with mappings using Expressions, Aggregators, Filters, Lookup and Update Strategy to develop and feed Data Warehouse.
  • Monitored workflows and collected performance data to maximize the session performance.
  • Involved in the maintenance of Data Warehouse.
  • Worked on extensively for monitoring session for conformity of Data in Data Warehouse.
  • Testing of mapping, Workflow and Tuning of mapping.
  • Data quality validation includes Creating test cases for Testing and helping for UAT.
  • Proper documentation for all changes to meet the standards of the organization.

Environment: Informatica 8.6, Oracle 9i


Informatica Developer


  • Analyzing the New changes/enhancements.
  • Informatica Designer tools were used to design the source definition, target definition and transformations to build mappings.
  • Designed the ETL processes using Informatica tool to load data from flat files and Oracle into the target Oracle 9i database.
  • Worked extensively with transformations like Expressions, Aggregators, Filters, Look Up, sequence generator and Update Strategy etc.
  • Cleansing and processing the data from stage area to Warehouse using Informatica transformations.
  • Developed various kinds of mappings, Dimensions, Facts, Workflows, worklets and sessions.
  • Created shell scripts and PL/SQL procedures.
  • Performed tasks of validating and executing the sessions in Informatica.
  • Created and Monitored Work Flows using Workflow Manager and Work Flow Monitor.
  • Reviewing Change Requests of User Maintenance.

Environment: Informatica Power Center 7.x, Oracle 9i, Unix

Hire Now