Etl Developer Resume Profile
Summary:
- Over 8 years of experience in Designing, Analysis, Development, Maintenance, Implementation of Relational Database OLTP and Data Warehousing Systems OLAP using Oracle, Teradata, Netezza, Informatica, Talend Big Data.
- Analyzed, designed and developed Extraction, Transformation and Load ETL processes for Data Warehousing and Data Migration projects using Informatica Power Center including Designer, Repository Manager, Workflow Manager, Workflow Monitor and Administration Console.
- Extracted data from various source systems like Oracle and flat files as per the requirements.
- Extensive experience in writing and executing scripts for validation and testing of the sessions, data integrity between source and target database and for report generation.
- Developed Complex mappings from varied transformation logics like Unconnected/Connected lookups, Router, Filter, Expression, Aggregator, Joiner, Union, Update Strategy and more. Used debugger to test and fix mapping.
- Solid experience in designing ETL Jobs using Talend Open Studio TOS and Talend Integration suite.
- Solid understanding knowledge of Ralph Kimball Bill Inmon Methodologies
- Experience in Data Analysis, ODS, Dimensional Data Modeling, Data Extraction Transformation Loading ETL , Building Cubes, Creating Meta layer, Data Mining, End User Reporting, Web based Reporting Implementation, Creating Test scripts and user signoff documents.
- Experienced in the Integration of various data sources like Oracle, XML, DB2, COBOL files and Flat Files into the staging area, ODS, Data Warehouse and Data Mart.
- Experienced in Dimensional Data Modeling using Erwin, Slowly Changing Dimensions Type I II, Star Schema /Snowflake Schema modeling, Fact Dimensions tables, Physical Logical data modeling.
- Extensively used Informatica Repository Manager and Workflow Monitor.
- Experienced in debugging mappings. Identified bugs in existing mappings by analyzing the data flow and evaluating transformations.
- Wrote complex SQL scripts to avoid Informatica Joiners, Unions and Look-ups to improve the performance as the volume of the data was heavy.
- Used Informatica Power Center Server manager to create sessions, batches to run Post SQL and Per SQL with the logic embedded in the mappings.
- Expertise in creating mappings, mapplets and reusable transformations using Informatica. Good knowledge in Scheduling ETL jobs like U, Tidal, Autosys tools.
- Responsible for implementing all of the technical Data Warehousing solution to the client, in accordance with an agreed technical design.
- Expert in the field of design, development, and implementation of processes for Data Warehousing and Data Integration projects using Informatica PowerCenter and PowerExchange.
- Expertise in performance tuning of SQL and ETL code and writing efficient and complex queries on huge volumes of data.
- Proficient in understanding business processes / requirements and translating them into technical requirements.
- Expert in designing and developing Unix Shell Scripts for automation of ETL Processes, Error handling and Auditing.
- Excellent implementation of data warehouse/data marts using various data warehouse concepts and principles like Change Data Capture, Slowly Changing Dimensions, Normalization/De-normalization Star and snow-flake Schema Modeling .
- Excellent experience working on Slowly Changing Dimensions SCD Type 1, 2 and 3 to keep track of historical data.
- Solid experience in loading and maintaining Data Warehouses in Teradata Netezza.
- Extensively used tools are MLoad, BTeq, FastExport and FastLoad to design and develop dataflow paths for loading transforming and maintaining data warehouse.
- Heavily used nzsql and nzloadNetezza utilities to load data from source to Netezza
- Good programming/debugging/troubleshooting skills in software development.
- Extensively used various Performance Tuning techniques to improve the session performance.
- Designed ETL architecture, created data flow diagrams, process flow diagrams, and high level ETL design.
- Designed and developed ETL job orchestration framework, CDC framework and Data Quality framework.
- Worked on POCs to load data from Oracle to Hadoop to Hive.
- Loaded data into Hadoop Hive using sqoop. Created tables into Hadoop Hive out of a flat file.
- Solid experience in support and knowledge transfer to production team.
Technical Skills:
Databases: Oracle 7/8/8i/9i, SQL Server 2008, MS Access Development Tools: SQL Navigators, Toad, Stylus Studio XML , XMLDB, Trans-SQL ETL Tools: Informatica Power Center 9.1/8.6.1,IBM Data Stage 8.1 Other Tools: XML Spy, Stylus studio, Visual Source Safe. Developer 6i/9i, Oracle Express, Workflow2.5/2.6.Informatica 8.6.1 Languages: SQL, PL/SQL, WSDL, eScript, Siebel VB, Java Script, XML XSD. Operating System: UNIX Sun Solaris , Windows 95/98/2000/NT/XP/7.
Confidential
Sr. ETL Developer
Responsibilities
- Worked closely with Business analysts and Data architects to understand and analyze the user requirements.
- Used Teradata utilities TPT, BTEQ to load data from source to target table.
- Created various kinds of indexes for performance enhancement.
- Performance tuned and optimized various complex SQL queries.
- Worked on Informatica Power Center - Source Analyzer, Data warehouse designer, Mapping Designer Mapplet, Transformations, Work Flow Manager Task Developer, Worklets, and Work Flow Designer and Work Flow Monitor.
- Imported Source/Target Tables from the respective databases and created reusable transformations Joiner, Routers, Lookups, Rank, Filter, Expression, Update, and Aggregator in a Mapplet and created new mappings using Designer module of Informatica.
- Responsible for developing, testing, debugging, support and maintenance for the ETL processes using Informatica Power Center.
- Transform data to various sources using SQL Server Integration Service and Talend Open Studio.
- Scheduled Pentaho ETL jobs and deployed in Pentaho BI Server in production environments.
- Created mappings using the transformations like Source Qualifier, XML Source Qualifier, Aggregator, Expression, Lookup, Router, Normalizer, Filter, Update strategy and Joiner transformations.
- Experience integrating the components of the following DB2, Data Stage ETL .
- Developed and scheduled Workflows using task developer, worklet designer, and workflow designer in Workflow manager and monitored the results in Workflow monitor.
- Created complex SCD type 1 type 2 mappings using dynamic lookup, Joiner, Router, Union, Expression and Update Strategy Transformations.
- Utilized SDLC and Agile methodologies such as SCRUM.
- Extensively used Change data capture concept in Informatica using mapping variables to capture the changes to the datawarehouse.
- Worked on Dimension as well as Fact tables, developed mappings and loaded data on to the relational database.
- Responsible for Performance Tuning at Talend level and SQL queries level.
- Loading the data from different sources into relational tables with Talend ETL.
- I Created reports in pentaho report designer using complex SQL queries.
- Developed complex Talend jobs mappings to load the data from various sources using different components.
- Wrote SQL overrides in source Qualifier in order to filter the data more effectively at the source level.
- Wrote Complex OLAP SQL's in Teradata.
- Successfully migrated the mappings to the testing and production environment.
- Responsible for creating complete test cases, test plans, test data, and reporting status ensuring accurate coverage of requirements and business processes.
Environment: Informatica Power Center 9.5, Erwin, Teradata, Oracle 11g, Visio, TOAD, SQL Plus, XML, Windows, Winscp, UNIX, Putty, Control-M, Business Objects Xi3.1, Teradata SQL Assistant
Confidential
Sr. ETL Developer
Responsibilities
- Published cubes and reports onto Pentaho repository and refreshed Pentaho BI repository after uploading each object to be available for central use.
- Worked on Teradata SQL Assistant, Teradata administrator, Teradata view point and BTEQ scripts.
- Extracted data from various sources like Oracle, Netezza and flat files and loaded into the target Netezza database.
- Wrote complex SQL queries on Netezza and used them in lookup SQL overrides and Source Qualifier overrides.
- Solid experience in performance tuning on TeradataSQL Queries and Informatica mappings.
- Loaded data into Hadoop Hive using sqoop. Created tables into Hadoop Hive out of a flat file.
- Extensively worked on developing Informatica Mappings, Mapplets, Sessions, Workflows and Worklets for data loads from various sources such as Oracle, ASCII delimited Flat Files, EBCDIC files, XML, COBOL, DB2, and SQL Server.
- Used Teradata SQL Assistant, Teradata Administrator, PMON and data load/export utilities like BTEQ, FastLoad, Multi Load, Fast Export, Tpump on UNIX/Windows environments and running the batch process for Teradata.
- Extensively worked on ETL Informatica transformations effectivelyincluding Source Qualifier, Connected - Unconnected Lookup, Filter, Expression, Router, Union, Normalizer, Joiner, Update, Rank, Aggregator, Stored Procedure, Sorterand Sequence Generator and created complex mappings.
- Developed ETL deployment plan for the production environment provided run books to production support team.
- Developed, tested and deployed full code package for EDW in SVN Visual Source Safe.
- Participated in the SIT testing and supported the System Acceptance and User Acceptance testing.
- Worked on writing various stored procedures using PL/SQL on Oracle database to achieve complex functionality.
- Responsible for Requirement Gathering Analysis and End user Meetings
- Worked as Being a part of a POC effort to help build new Hadoop clusters.
- Created Operation Manual documents and developed and updated other technical documentations.
- Removing duplicates, standardizing data mass maintaining , and incorporating rules to eliminate incorrect data from entering the system in order to create an authoritative source of master data Using MDM.
- Extensively used various active and passive transformations like Filter Transformation, Router Transformation, Expression Transformation, Source Qualifier Transformation, Joiner Transformation, and Look up Transformation, Update Strategy Transformation, Sequence Generator Transformation, Rank Transformation, and Aggregator Transformation.
- Responsible for best practices like naming conventions, Performance tuning, and Error Handling
- Responsible for Performance Tuning at the Source level, Target level, Mapping Level and Session Level
- Solid Expertise in using both Connected and Un connected Lookup transformations
- Extensively worked with various lookup caches like Static Cache, Dynamic Cache, and Persistent Cache.
- Worked on writing various stored procedures using PL/SQL on Oracle database to achieve complex functionality.
- Worked on creating various mapplets in Informatica to provide re-usability.
- Worked effectively with Business Analysts and Project managers for assigned tasks and deliverable timelines.
- Responsible for best practices like naming conventions, Performance tuning, and Error Handling.
- Involved in defining the overall strategy for design and standards by creating many checklists for smooth deployments.
- Developed Slowly Changing Dimension Mappings for Type 1 SCD and Type 2 SCD
- Worked on extracting real time data and loading to data warehouse.
- Scheduling and performance tuning of Informatica mappings.
- Worked closely with the project team to formulate and implement a flexible system design that meets functional requirements.
- Created data models diagrams using Visio and Erwin for EDW and datamarts.
- Collaborated with data modellers, ETL developers in the creating the Data Functional Design documents.
- Created the Staging mappings for delta detection using the different Transformations Expression, Lookup, Joiner, Filter, Router, Stored Procedure etc.
- Free hand SQL, reports with calculation context and drill down features.
- Requirements analysis, Design, Code and Design reviews.
Environment: Informatica 9.1 on UNIX, Teradata R13, Netezaa , Oracle 11g, PL/SQL, TOAD, UNIX, FASTLOAD, MULTILOAD, TPUMP, TPT