Etl Developer Resume
Los Angeles, CA
PROFESSIONAL SUMMARY
- Over 8+ years of Software/IT experience in Design, Development, Testing, Implementation and support of Data Warehousing applications.
- Expertise in Data Warehousing, Data Migration, Data Integration using Business Intelligence (BI) tools such as Informatica Power Center, Power Exchange (Power Connect), Metadata Manager, Informatica Data Quality, Informatica Data Explorer, Data Transformation, MDM etc
- Hands on experience with Informatica Data Quality (IDQ) tools for Data Analysis / Data Profiling / IDQ Developer.
- Worked on IDQ Analyst for Profiling, Creating rules on Profiling and Scorecards. Designed IDQ mappings which is used as Maplets in Power center Worked extensively on Informatica Data Quality Versions 9.x
- Worked with Informatica Data Quality toolkit, Analysis, data cleansing, data matching, data conversion, exception handling, and reporting and monitoring capabilities of IDQ 9.6.1
- Experience in designing, development and implementing Extraction, Transformation, and Loading (ETL) techniques on multiple database platforms and operating system environments using Informatica Power Center 9.x/8.x, Power Exchange, SSIS, SQL, PL/SQL, External Loaders, and Shell/batch Scripting etc.
- Experience in major sectors such as Health care, Telecommunications, Banking and Insurance.
- Proficient in interacting with Business Analysts, Users and Business executives to understand the requirements and come up with innovative and efficient solutions by working with the data modelers and Architects.
- Proficient in all phases of the Software development lifecycle (SDLC) using Agile, Waterfall and RUP methodologies.
- Extensive Experience with databases such as Oracle, Sales force, DB2 UDB, Microsoft SQL Server, Sybase, Teradata and database clients such as SQL server Management Studio and SQL Developer, SQL navigator, Toad etc.
- Proficient in Data Modeling, E - R diagrams, Logical, Physical design, Star/Snow flake Schemas using Erwin.
- Developed business intelligence dashboards to provide visualization techniques for the business units
- Experience in Trouble shooting and implementing Performance tuning at various levels such as Source, Target, and Mapping, Session, Database and System level in ETL processes.
- Strong Database experience and back-end procedure development in Oracle 11g/10g/9i, MS SQL Server 2005/2008, Sybase, Teradata, DB2, Netezza, etc.
- Expertise in the development of integration mappings such as Type I, Type II, and Type III Slowly Changing Dimensions and implementing incremental loads, and change data capture.
- Proficient in scheduling and Automation of ETL processes with scheduling tools such as Control-M, Autosys and Tidal.
- Thorough understanding of configuration, administration and upgrading of Informatica Power Center.
- Working knowledge of other integration tools such as SSIS, Data stage, Abinitio, etc.
- Expertise in OLAP and Reporting tools like Qlik view, SSRS, OBIEE, Micro Strategy, and Cognos, etc.
- Hard working individual and a quick learner with strong analytical and communication skills. Experienced in Client interaction and understanding Business applications.
- Gather data necessary to accomplish analysis from a number of sources, accumulate it together in prearranged format and enter the data in several data analysis software
TECHNICAL SKILLS:
ETL Tools: Informatica Power Center, Power Exchange 9.1, 8.6.1, 8.1., Power Exchange (Power Connect), Metadata Manager, Informatica Data Quality, Informatica Data Explorer, SSIS, Data stage, Abinitio, Pentaho Data Integration.
Databases: Oracle 11g/10g/9i, MS SQL Server 2008/2005/2000, MS Access, IBM DB2, Tera Data V12, Siebel CRM, PeopleSoft, Teradata, Sybase, Netezza, DB2, etc.
Data Modeling: Dimensional Data Modeling, Star SchemaModeling, Snow Flake Modeling, FACT and Dimension Tables, Physical and Logical Data Modeling.
Environments: UNIX, Windows 2000/XP/7, Windows server 2000/2003, HP-Unix, Linux
Languages: C, C++, Java, SQL, PL/SQL, T-SQL, HTML, DHTML, XML, UNIX, Shell Scripting, Visual basic, ASP, JSP, XML, Macromedia Software, JCL.
Tools: Toad, Visio, SQL developer, ERWIN, MS Office Suite
Design Skills: Object Oriented Analysis Design using UML.
Scripting Languages: Java Script,UNIX Shell Scripting, Batch scripting.
PROFESSIONAL EXPERIENCE
Confidential, Los Angeles, CA
ETL Developer
Responsibilities:
- Interacted with business users to understand the business requirements
- Designed complete IDQ application from Source extraction through mapping design and Master workflows
- Designed complex application in duplicate/discrepancy data handling by exception and human task and projecting bad data to analyst for data stewards to work on it and manager to review the work of the data stewards
- Designed mapplets in the developer which is used as reusable rules at the profiling level.
- Integrated analyst and developer through creating LDO’s(Logical Data Objects) and created profiling and score cards on top of LDO’s
- Created Accounts for all the data stewards, Data Steward Groups and created connections in the Admin Console.
- Provided various INFACMD commands for automation and scheduling the IDQ jobs
- Designed various Mapping and Mapplets using different transformations such askey generator,match, labeler, case converter, standardizer.
- Use Exception transformation to handle the duplicate and bad data and invoke mapping using workflow objects mapping task, human task, exclusive gateway and notification task.
- Provided Golden Set of Records using Grouping, Matching, Consolidation to the Data team
- Exported theIDQMappings and Mapplets to power center and automated the scheduling process.
- Identified and eliminated duplicates in datasets thorough IDQ 8.6.1 components of Edit Distance, Jaro Distance and Mixed Field matcher.
- Coordinated with offshore team for development and support activities.
- Created Informatica mappings and Workflows to extract data from different source systems /flat files/parameter files to load the staging tables and the master tables.
- Worked extensively with front end team to meet the business requirements.
- Developed various shell scripts for ETL process, to truncate the staging tables, to ftp the files from local server to remote, file validation scripts, file archive scripts, mailing scripts.
- Created stored procedures, packages, triggers, tables, views, synonyms, and test data in Oracle.
- Creating Data Maps, Extracting (incremental) CDC data from Main frame sources, exporting Data Maps to Repository and Importing Mainframe source files on to shared folders.
- Extensively worked in the performance tuning of the programs, ETL Procedures and processes.
- Developed PL/SQL procedures for processing business logic in the database.
- Used SQL tools like TOAD to run SQL queries and validate the data in warehouse.
- Worked closely with Production control to make sure the jobs run in success.
- Associated with 24/7 Production support.
Environment: Informatica Power Centre 9.1, Oracle,SQLServer,Pentaho 5.0,UNIX scripting.IDQ 8.6.1
Confidential,
SR Informatica Developer
Responsibilities:
- Involved in understanding the business requirements and translate them to technical solutions.
- Worked withInformaticaData Quality 9.6.1 (IDQ) toolkit, Analysis, data cleansing, data matching, data conversion, exception handling, and reporting and monitoring capabilities of IDQ 9.6.1.
- Perform System Analysis and Requirement Analysis, design and write technical documents and test plans.
- Created a Hybrid process in IDQ by combining both IDQ Developer and analyst version through LDO(Logical Design Objects)
- Worked on IDQ Analyst for Profiling, Creating rules on Profiling and Scorecards.
- Worked with Management for creating the requirements and estimates on the Project.
- Assisted Business Analyst with drafting the requirements, implementing design and development of various components of ETL for various applications.
- Coordinated with DBA in creating and managing tables, indexes, table spaces, auditing and data quality checks.
- Designed IDQ mappings which is used as Maplets in Power center
- Developed numerous mappings using the various transformations including Address Validator, Association, Case Converter, Classifier, Comparison, Consolidation, Match, Merge, Parser etc.
- Setup Procedures to Extract, Transform and Load (ETL) data from Operational Systems to the Data Warehouse by Using Informatica Designer.
- Build a re-usable staging area in Oracle for loading data from multiple source
- systems using template tables for profiling and cleansing in IDQ or QualityStage.
- Designed and developed ETL and Data Quality mappings to load and transform data from source to ODS using Powercenter 9.1.0 and IDQ 8.6.1
- Designed and developed Informatica mappings for data loads that included Source Qualifier, Aggregator, Unconnected Lookup, Connected Lookup, Filter, Router, Update Strategy, Expression and Sequence Generator transformations.
- Implemented performance tuning in sources and mappings.
- Used the debugger in debugging for some critical mappings to check the flow of data from instance to instance.
- Created workflows using workflow manager and used various tasks like Email, Command and Session.
- Made some enhancements to the existing mapping as per the business requirement.
- Involved in documenting the existing SQL Codes ( simple and complex) and uploading them into SharePoint.
- Executed workflows and sessions using Workflow Manager.
- Scheduled sessions to update the target data using Workflow Manager of Informatica.
- Involved in the error checking and testing of ETL Procedures using Informatica Session log and workflow logs.
- Using pushdown option and increase the system performance.
- Using with CDC, data extraction takes place at the same time the insert, update, or delete operations occur in the source tables, and the change data is stored inside the database in change tables. The change data, thus captured, is then made available to the target systems in a controlled manner.
- Used Metadata Manager to extract metadata from application, business intelligence, data integration, data modeling and relational metadata sources.
- Developed all the mappings according to the design document and mapping specs provided and performed unit testing.
- Prepared ETL mapping documents for every mapping and data migration document for smooth transfer of project from development to testing environment and then to production environment.Unit testing of mappings and testing of various conditions
Environment:Informatica Power Center 9.1.0 (Informatica Designer, Repository Manager, Workflow Manager, Workflow Monitor), Oracle 11g, Toad for Oracle 11.5, UNIX scripting. IDQ 8.6.1
Confidential, Baltimore, MD
Sr. Informatica&Pentaho Developer
Responsibilities:
- Used Pentaho Data Integration (ETL) various transformations and bulk loading data from heterogeneous source systems into target database.
- Created Transformations using Spoon and extracted data from various sources, transformed data according to the requirement.involved in extracting the data from the Flat Files and Relational databases into staging area.
- Developed Pentaho Jobs and Reusable Transformations to facilitate timely Loading of Data of a star schema.
- Created Jobs and extracted data from various sources, transformed data according to the requirement and loading into data warehouse.
- Used various steps like merge Join, Look up, Transform steps, input steps, output steps, UtilityScripting, Validation, Data mining to create robust Transformations in the Pentaho 5.
- Imported various heterogeneous files using Pentaho Input Steps.
- Developed several reusable transformations that were used in other Jobs.
- Prepared Technical Design documents and Test cases.
- Involved in Unit Testing and Resolution of various Bottlenecks came across.
- Implemented various Performance Tuning techniques.
Environment:Pentaho 5,Mysql,UNIX scripting.
Confidential, New Jersey
Sr. ETL Developer
Responsibilities:
- Participated in the Design Team and user requirement gathering meetings.
- Performed business analysis and requirements per high end users and managers and performed requirement gathering.
- Responsible for passing the requirements to offshore team and co-ordinate with them in completing the tasks and then presenting them to the client.
- Used Ralph Kimball approach for designing of database for the data warehouse
- Developed a standard ETL framework to enable the reusability of similar logic across the board.
- Used Teradata utilities such as Multiload, FastLoad, Tpump and TPT to export and load data into Teradata Data Warehouse.
- Extracted source data from Oracle, Flat files, XML files using Informatica, and loaded into target Database.
- Extracted data from Mainframe using Power Exchange and COBOL files and created Recon processes to test the feed from main frames with the feed generated at the Distributed end.
- Developed complex mapping also involving SCD Type-I, Type-II,Type-III mapping in Informatica to load the data from various sources
- Engaged in designing and developing test cases, error control routines, data validation, audit and log controls using Informatica inbuilt error logging techniques and PL/SQL.
- Extensively transformed the existing PL/SQL scripts into stored procedures to be used by Informatica Mappings with the help of Stored Procedure Transformations.
- Involved in extensive Performance Tuning by determining bottlenecks in sources, mappings and sessions.
- Created Models based on the dimensions, levels and measures required for the analysis.
- Validate the data in warehouse and data marts after loading process balancing with source data.
- Worked closely with the business analyst’s team in order to solve the Problem Tickets, Service Requests. Helped the 24/7 Production Support team
Environment:Informatica Power Center 8.6.1, Teradata, Oracle 11g/10g, MS SQL Server, Teradata SQL Assistant, Windows, Unix, Micro Strategy 9.1, ERWin 3.5, Mainframe, TOAD.
Confidential, Des Moines, IA
Sr. ETL developer
Responsibilities:
- Interacted with Business Analysts to understand business and functional specifications.
- Involved in dimensional modeling(Star Schema) of the Data warehouse and used Erwin 3.5 to design the business process, grain, dimensions and measured facts.
- Participated in the Design Meetings, and prepared technical and mapping documentation.
- Extensively migrated data from different source systems such as flat files, relational, and XML sources to ODS, Data marts and Data warehouse and worked on Data partitioning.
- Used Type 1 and Type 2 SCD mappings to update slowly Changing Dimension Tables using Informatica changed data capture (CDC).
- Used shortcuts to reuse objects without creating multiple objects in the repository and inherit changes made to the source automatically.
- Used Meta Data Manager to created lineage report for source/target objects and connected business terms with actual implementation and prepared a final report.
- Performance tuned the Mappings, Tasks, Sessions, SQL and Scripts to optimize Data-Load performance.
- Used Multiload, FastLoad and Tpump utilities to load to Teradata database.
- Read from SAP source using Power Exchange and also loaded data into SAP targets.
- Worked on IDQ tool to identify data duplication, and understand the quality of data. Used IDE tool for profiling the data prior to Data migration.
- Created Financial, Forecast reports for the top management/business users using Micro Strategy 8.1.
- Worked with testers and participated in Unit, Integration, system, and performance testing of ETL applications.
Environment: Informatica Power Center 8.6.1, Informatica Power Exchange(CDC), IDQ, Oracle 11g, SQL, MicroStrategy 8.1, SQL Server 2005, Teradata, Windows NT/2003, UNIX, Autosys.
Confidential, Jackson, MS
Informatica Developer
Responsibilities:
- Gathered requirements from Business Analysts, report writers and Business users and created a Low level technical design and mapping document using high level design document.
- Created PL/SQL scripts for extracting, transforming and loading data to/from a number of both Oracle and non-Oracle sources
- Reverse Engineering Schema’s using ERWIN, and re-wrote PL/SQL routines using Netezzanzsql and nzload utilities
- Used Informatica Power Center to extract, transform and load data from different sources like flat files, COBOL files, XML files and Oracle 10g/11g database.
- Troubleshooted PL/SQL and UNIX shell scripts, Tuned SQL statements; Partitioning/Indexing/compressing ¼ terabyte tables
- Performance tuned existing ETL code, SQL, mappings, sessions, database, etc.
- Created and executed PL/SQL procedures, pre-session and post-session scripts (in Informatica) to ensure timely, accurate processing and ensure balancing of job runs.
- Created various challenging reports such as Support Ticket Summary Report, Open Ticket Report, using Oracle Analytics and Cognos.
- Worked with production support in finalizing scheduling of workflows and database scripts using Control M tool.
- Created UNIX shell scripts for Informatica ETL tool to automate sessions.
- Completed transfer of ETL processes and finished documentation in a period of three months to help start a new Data Management team.
Environment:InformaticaPower Center 8.6.1, PL/SQL, Oracle 10g, Netezza, TOAD, UNIX, Microstrategy..
Confidential, Sandy Springs, GA
Informatica/ ETL Developer
Responsibilities:
- Analyzed existing process flows and business functionalities and prepared ETL process flow document for Predictive analysis.
- Involved in designing Conceptual, Logical and Physical model for Staging and Target data bases.
- Created complex mappings which involved Slowly Changing Dimensions, implementation of Business Logic and capturing the deleted records in the source systems.
- Extensively involved in Recovery process for capturing the incremental changes in the source systems for updating in the staging area and data warehouse respectively.
- Extensively worked on tuning of Informatica mappings for better performance.
- Developed PL/SQL stored procedures for database updates and to create the necessary indexes in the target tables.
- Was involved in basic Informatica administration such as creating users and privileges, folders, optimizing server settings, deployment groups etc.
Environment: Informatica Power Center 8.1, Power Exchange CDC, Oracle 10g, SQL Server 2005, Windows XP, PL/SQL, UNIX.