Informatica Etl Design/ Dev Lead Resume
New York, NY
SUMMARY:
- Around 7+ years of experience in design and development of data warehouse and database solutions.
- Good understanding of Data warehousing and ETL concepts and business application, business dataflow and data relation.
- Good understanding on Data Quality Management, Master Data Management, Data Integration techniques and Data Migration across Healthcare, Life Sciences, Telecom and Retail industry domains.
- Thorough knowledge in various tools and technologies like different versions of Informatica Data Quality and Data Analyst, Informatica Power Center (9.X,10) and Informatica Cloud.
- Experience in designing and developing mappings from varied transformation logic like Unconnected and Connected lookups, Router, Filter, Expression, Aggregator, Joiner, Update Strategy etc.
- Experience in Performance Tuning the mappings
- Experience in developing Test Plans, Test Strategies and Test Cases for Data Warehousing projects ensuring the data meets the business requirements, also Good Work experience in SQL, PL/SQL, UNIX
- Involved in knowledge sharing and knowledge transfer to the new members of the project & Preparation of knowledge transfer documents.
- Having Good Communication skills, Interpersonal relations, hardworking and result - oriented as an Individual and in Team.
- Have extensive experience in requirement analysis, database design, development and Testing.
- Excellent problem-solving skills with a strong technical background and result oriented team player with excellent communication and interpersonal skills
- Extensively used ETL methodology for supporting Data Extraction, Transformation and Loading processing, in a corporate-wide-ETL Solution using Informatica.
- Extensively used Teradata BTEQ Scripts for Data Extraction, Transformation and Loading into tables.
- Extensively used Teradata utilities Fast Load, Multi Load for to load data from flat files to tables.
- Performed Teradata code tuning on a regular basis.
- Good understanding and experience in unstructured data management and modeling concepts (Big data belonging to different domain like Social media, Health care, web usage data and relational data).
- Have Data Analysis experience in analytical projects for the client in Healthcare domain.
- Knowledge on Big Data, Predictive Modeling, Data Mining and data analysis using R.
TECHNICAL SKILLS:
ETL Tool: Informatica 9X,10, BDE
Operating Systems: Windows, Unix
Application Tools: ControlM, Confidential Remedy, Putty, Toad, WinScp, Edit Plus, Teradata SQL Assistant, TIDAL scheduler, Cloudera, R
Database: Teradata, Oracle 10g, Netezza 7.1
PROFESSIONAL EXPERIENCE:
Confidential, New York, NY
Informatica ETL Design/ Dev Lead
Responsibilities:
- Involved in the Designing the ETL Solution.
- Involved in Analysis and discussion with business regarding the legacy BE HIST/Ariel systems for the assessment of the efforts.
- Involved in preparation of technical design documents, low level functional design documents and Velocity documents.
- Developing the PL/SQL and UNIX Shell scripts.
- Developing the Informatica Mappings/Mapplets/Worklets and workflows for loading the data to MySQL data marts.
- Designed various UDF’s (User Defined Functions) in Informatica designer to be used in various mappings.
- Data coming in as source definitions are from different data sources types like Flat files, XML, Sales force, Oracle, SQL Server and Hadoop files (HDFS).
- Hands on experience in middleware integration components (iPaaS) platforms like such as Informatica Cloud Real Time (ICRT), Informatica Cloud.
- Proficiency implementing integrations with third party CRM data(Salesforce)
- Created various shell scripts for pre and post processing of flat files.
- Optimizing and doing performance tuning of mappings/sessions/workflows to achieve higher response times.
- Created Cron/Control M jobs to schedule the Informatica data loads.
- Created test cases and test scenarios to test the incremental loads of data in unit testing.
- Creating workflow dependency diagram for the code automation.
- Coordinating all releases successfully according to all compliance rules
Environment: Informatica PowerCenter 10.1, Informatica Analyst, Informatica IDQ/Developer, MySQL 5.7, SQL Server 2005, UNIX Shell Scripting, Crontab, Autosys, Git Version control system, SVN, HP ALM and Agile Methodology
Confidential
ETL Specialist/ Onsite coordinator
Responsibilities:
- Validated data to ensure the quality, validity and accuracy of content.
- Part of reconciliation activity which ensures the Source and Vision data is in sync
- Day to Day reporting to the client - Send data verification report (comparing data from multiple systems) before Users login into their systems.
- Monitored workflows and collected performance data to maximize the session performance.
- Involved in development of Informatica mappings and tuned for better performance
- Incident resolution using ALM system, and production support, handling production failures and fix them within SLA.
- Creating/modifying Informatica Workflows and Mappings (power center and Cloud) also involved in unit testing, internal quality analysis procedures and reviews.
- Validated and fine-tuned the ETL logic coded into existing Power Center Mappings, leading to improved performance.
- Part of Informatica cloud integration with amazon redshift
- Managing Informatica Cloud based tools including Address Dr.
- Integrated data from various source systems like SAP (ABAP), SQL Server (DTS), Ariba, Oracle (ODBC) so on.
- Part of SDLC (Software Development Life Cycle) Requirements, Analysis, Design, Testing, Deployment of Informatica Power Center.
- Very familiar with SCD1 and SCD2 in snowflake schema and star schema.
- Identifying and tracking the slowly changing dimensions (SCD), used CDC logic (Change Data Capture) for the SCD tables loading in Oracle.
- Good Experience in writing UNIX shell scripts, SQL scripts for development, automation of ETL process, error handling and auditing purposes.
- Created Stored Procedures, Packages, Triggers, Cursors and backup-recovery for the various tables
- Leveraged Explain Plan to improve the performance of SQL queries and PL/SQL Stored procedures.
- Working and understanding on the end to end features of Hadoop-hive
- Creating the Sqoop job to store the data in database.
- Analyzing the FLUME job which will place format and the data to HDFS.
- Analyzing and modifying PIG scripts
- Experience is RCA, preparing Business Specification Document and Technical Designs Document.
- Provided in-depth Root Cause Analysis reports in relation to production load failures and critical change requests.
- Involved in scheduling the jobs by using TIDAL scheduling tool (ETL workflows invoked using pmcmd, used EXEC to invoke Teradata Macros and run file for BTEQ)
- Prepared Unit Test Case and Unit Test Results for change requests.
- Active participation in WSR (Weekly Status Report) meeting to Communicate Requirements, Development and Data related issues to concerned teams.
- Basic reporting knowledge on Tableau.
- EDA to understand the behavior of the product and customer at various levels
- Time based cross validation data preparation
- Dimension reduction for best possible predictors
- Built RFM (Regency, Frequency and Monetary) analysis is followed by Cluster Analysis where customers with similar RFM characteristics are grouped together to arrive at segments
- Used the top ranked customers who has high propensity to buy the products for focused campaign.
- Calculating the incremental revenue metric. Incremental Revenue is the difference between the actual revenue from the Target Group and the expected revenue from the same group. It is calculated at a campaign level.
- Understood the data by exploratory data analysis and classified the category into four subgroups: Highly Value Customer, Active Customer, Moderately Active customer and Inactive Customers.
- Pre-modeling process like missing value imputing, outlier detection, data transformation etc.
- Variable reduction to get the best predictors
- Developed a Logistic regression model to get the rejection probability at transaction level
- Model was tested on the testing data set
- Developed a scoring mechanism to predict churn of Rejection probability at customer level.
Environment: Informatica 9x, Informatica Cloud, Teradata, TIDAL scheduler, IRIS, Data Science, R, Cloudera
Confidential
ETL Developer/ Data Analyst
Responsibilities:
- Migration of code.
- Writing basic UNIX shell scripts and PL/SQL packages and procedures.
- Involved in performance tuning of the mappings, session and SQL queries.
- Creating/modifying Informatica Workflows and Mappings.
- Working with various transformations such as Expression, Aggregator, Update Strategy, Look Up, Filter, Router, Joiner and Sequence generator etc. in Informatica 8.6.1 for new requirement.
- Created reusable Tasks, Sessions, reusable Worklets and workflows in Workflow manager.
- Sound knowledge with Netezza SQL
- Managed Netezza queries by performance tuning and techniques such as CBT and Collocation.
- Identifying and tracking the slowly changing dimensions (SCD), used CDC logic (Change Data Capture) for the SCD tables loading in Oracle.
- Communicate Requirements, Development and Data related issues to concerned teams.
- Interacting with Business Users for Business logic Clarifications
- Conduct data quality analysis on various source data systems
- Presentation of Issues and findings to subject matter experts and providing in depth root cause analysis.
Environment: Informatica 8.6/9.1, Oracle 9i, Netezza 7.1
Confidential
Informatica/ ETL Developer
Responsibilities:
- Extracting and loading of data from flat files and Oracle sources to Oracle (target warehouse) using transformations in the mappings.
- Developed a series of mappings for handling different cases of input data.
- Involved in development of Informatica mappings and tuned for better performance.
- Worked extensively with mappings using Expressions, Aggregators, Filters, Lookup and Update Strategy to develop and feed Data Warehouse.
- Monitored workflows and collected performance data to maximize the session performance.
- Involved in the maintenance of Data Warehouse.
- Worked on extensively for monitoring session for conformity of Data in Data Warehouse.
- Testing of mapping, Workflow and Tuning of mapping.
- Data quality validation includes Creating test cases for Testing and helping for UAT.
- Proper documentation for all changes to meet the standards of the organization.
Environment: Informatica 8.6, Oracle 9i
Confidential
Informatica Developer
Responsibilities:
- Analyzing the New changes/enhancements.
- Informatica Designer tools were used to design the source definition, target definition and transformations to build mappings.
- Designed the ETL processes using Informatica tool to load data from flat files and Oracle into the target Oracle 9i database.
- Worked extensively with transformations like Expressions, Aggregators, Filters, Look Up, sequence generator and Update Strategy etc.
- Cleansing and processing the data from stage area to Warehouse using Informatica transformations.
- Developed various kinds of mappings, Dimensions, Facts, Workflows, worklets and sessions.
- Created shell scripts and PL/SQL procedures.
- Performed tasks of validating and executing the sessions in Informatica.
- Created and Monitored Work Flows using Workflow Manager and Work Flow Monitor.
- Reviewing Change Requests of User Maintenance.
Environment: Informatica Power Center 7.x, Oracle 9i, Unix