Sr. Etl Developer Resume
Mason, -oH
SUMMARY:
- Over 11 years of experience in Informatica Power Center 9.x/8.x/7.x/6.x/5.x, Work flow Designer, Workflow Manager, Workflow Monitor and Mapping Designer, Teradata.
- Proficient in all phases of the Software Development Life Cycle (SDLC) In Data Warehousing
- Requirements Gathering, Data Scrubbing, Design, Development, Performance Tuning and Unit Testing.
- Excellent in coding SQL, PL/SQL Procedures, Functions, Triggers and Packages.
- Extensive experience with ETL tools in the areas of analysis, design, development and administration with focus on Talend Data Integration, Talend Big - data Real Time, Informatica power center and Apache Ni-Fi with different databases.
- Worked with heterogeneous data like Oracle, Greenplum, HAWQ, Casandra, Mongo-DB, Flat files and different Hadoop concepts in Bigdata.
- Have built Talend 5.x, 6.x and Informatica environments for high availability for QA and production environments including Installation, upgrade, migration, User management, back/restore, code promotion and OS Profiling in AWS, RHEL and Windows instances
- Experience in ETL Design, Entity-Relationship Modeling and Dimensional Modeling.
- Developed Slowly Changing Dimension Mappings of Type I, and II.
- Extensively involved in creating Complex Mappings and reusable components like Reusable Transformations, Mapplets, Worklets and control tasks to implement reusable business logic.
- Profile source data and determine all source data and metadata characteristics; Design and execute a Data Quality Audit/Assessment; Design and execute the data quality mappings that will cleanse, de-duplicate, and otherwise prepare the project data
- Implement data quality processes including transliteration, parsing, analysis, standardization and enrichment at point of entry and batch modes; Deploy mappings that will run in a scheduled, batch, or real-time environment.
- Document all mappings, mapplets, and rules in detail and hand over documentation to the customer
- Collaborate with various business and technical teams to gather requirements around data quality rules and propose the optimization of these rules if applicable, then design and develop these rules with IDQ
- Designed mappings using Source qualifier, Joiner, Aggregator, Expression, Lookup, Router, Filter, and Update Strategy transformations and Mapplets to load data into the target involving slowly changing dimensions.
- Worked with the Salesforce source system application (Veeva Data)
- Proficient in SQL Tuning to ensure best query performance.
- Involved in Unit testing, System testing to check whether the data loads into target are accurate, which was extracted from different source systems according to the user requirements
- Expertise in using ERwin, MS Visio, Oracle Designer to design process flow diagrams.
- Extensive experience in Administration/maintenance of Informatica Power Center including Installation, upgrading and patching.
- Assisted in Data modelling by creating Star schemas using MS Visio and Erwin tools.
- Developed business rules for cleansing/validating/standardization of data.
- Defined System Trust and Validation rules for the base object columns.
- Excellent experience in creating UNIX scripts.
- Extensively worked with Teradata utilities like BTEQ, Fast Export, Fast Load, Multi, TPump, TPT to Export and Load data to/from different source systems including flat files.
- Written several Teradata BTEQ scripts to implement the business logic.
- Worked on TSQL queries.
- Involved in the admin activities of informatica
- Worked with informatica Power exchange and Informatica cloud to integrate salesforce and load the data from salesforce to oracle Db
- Experience in Informatica Cloud data integration (ICDI/ICRT)
- Interoperability of integration, efficient provisioning and monitoring using Informatica Cloud.
TECHNICAL SKILLS:
ETL: Informatica Power Center 9.5/9.1/8.6/8.5/8.1/7.1
Data Quality Tools: Informatica Data Quality 9.5.1/9.6.1
Operating Systems: Unix, Windows NT/2000/XP
Reporting tools: Qlilkview 11
DBMS: Oracle11g/10g/9i, SQL Server 2005/2008, Teradata 11/12/13
Data modeling tool: Erwin, Oracle Designer 10g, MS Visio 2010
Data Base Tools: SQL* Loader, TOAD, PL/SQL Developer, SQL Developer, Hadoop, Greenplum and HAWQ
Languages: PL/SQL, UNIX Shell Scripting, Spark, Hadoop, Python
ITIL TOOLS: Service now
Languages: PERL 5.8, Unix Shell script
PROFESSIONAL EXPERIENCE:
Confidential - Mason -OH
Sr. ETL Developer
Responsibilities:
- Extensive experience in Informatica (9.x) applications. Designed and developed the Workflows, Worklets, Mappings, Mapplets, Sessions, Tasks, Transformations and schedule the Workflows and Sessions.
- Design and implement the ETL Data model and create staging, source and Target tables in Oracle and Hadoop database.
- Configured SVN, GitHub, Nexus and Jenkins for the Talend code to be managed.
- Implemented Informatica/Talend automations to reduce time and man power.
- Design and Developed interactive Dashboards using Business Insight.
- Managed Hadoop and Oracle databased by using SQL developer and Hive tools.
- Used Oozie and Talend schedulers to schedule the workflows and jobs.
- Interacted with end-users and functional analysts to identify and develop Business Specification Documents (BSD) and transform it into technical requirements.
- Strong experience with Informatica tools - Source Analyzer, Warehouse designer, Mapping designer, Mapplet Designer, Transformations Developer, Informatica Repository.
- Designed and developed complex mappings to move data from multiple sources into a common target area such as Data Marts and Data Warehouse using lookups, Source Qualifier, Router, Filter, Expression, Aggregator, Joiner, Normalizer, Sequence Generator, Update Strategy from varied transformation logics in Informatica.
- Having strong hands on experience in extraction of the data from various source systems ranging from Mainframes like DB2, Flat Files, VSAM files, etc. to RDBMS like Oracle, SQL Server, Teradata etc.
- Implemented Stored Procedures, Functions, views, Triggers, Packages in PL/SQL.
- Extensively used Slowly Changing Dimension (SCD) technique in insurance application. Expertise in OLTP/OLAP System Study, Analysis, E-R modeling, developing Dimensional Models using Star schema and Snowflake schema techniques used in relational, dimensional and multidimensional modeling.
- Worked on optimizing the mappings by creating re-usable transformations and Mapplets. Created debugging and performance tuning of sources, targets, mappings, transformations and sessions.
- Interoperability of integration, efficient provisioning and monitoring using Informatica Cloud.
Environment: Informatica Power Center 9.6, Qlik view XI, Salesforce, Informatica Cloud Teradata 13.0, PL/SQL, SQL*Loader, UNIX, IDQ, Hadoop, Greenplum and HAWQ
Confidential - Columbus -OH
Sr. ETL Developer
Responsibilities:
- Exposure in overall SDLC including requirement gathering, data modeling, development, testing, debugging, deployment, documentation, production support.
- Developed Task Automation using UNIX Scripts, Job scheduling and Communicating with Server using pmcmd. Extensively used Autosys for Job monitoring and scheduling.
- Excellent analytical, problem solving and communication skills with ability to interact with individuals at all levels.
- Worked with informatica Power exchange and Informatica cloud to integrate salesforce and load the data from salesforce to oracle Db
- Import data from MySQL to HDFS, using Sqoop to load data
- Developed and designed a 10-node Hadoop cluster for sample data analysis
- Regularly tune performance of Hive and Pig queries to improve data processing and retrieving
- Run Hadoop streaming jobs to process terabytes of XML data
- Experience in Informatica Cloud data integration (ICDI/ICRT)
- Developed data quality specific ETL jobs (address standardization and validation, email cleanups, name cleanup, parsing, etc.) utilizing IDQ and other ETL tools
- Serve as the primary resource to team members and data stewards for training, problem resolution, data profiling etc.
- Analyze and provide data metrics to management in order to help prioritize areas for data quality improvement
- Proof of Concepts (POC): Jenkins with Talend, Hortonworks Data Platform (HDP) for Apache-NiFi and Talend streeming environment for AWS servers.
- Experience in Task Automation using UNIX Scripts, Job scheduling and Communicating with Server using pmcmd. Extensively used Autosys for Job monitoring and scheduling.
- Excellent analytical, problem solving and communication skills with ability to interact with individuals at all levels.
- Worked with informatica Power exchange and Informatica cloud to integrate salesforce and load the data from salesforce to oracle Db
Environment: Informatica Power Center 9.6, Qlik view XI, Salesforce, Informatica Cloud Teradata 13.0, PL/SQL, SQL*Loader, UNIX, IDQ, Informatica Data Quality 9.5.1/9.6.1
Confidential
Sr. ETL Developer
Responsibilities:
- Extraction and Transformation of data from various sources such as Oracle and Flat files and loading them into the Oracle target database using Informatic Power Centre.
- And extensively worked on IDQ admin tasks and worked as both IDQ Admin and IDQ developer.
- Performed many multiple tasks effectively and involved in troubleshooting the issues.
- Excellent in coding SQL, PL/SQL Procedures, Functions, Triggers and Packages.
- Worked on Informatic Power Centre client tools like Source Analyzer, Warehouse Designer, Mapping Designer, Mapplets and Transformations Developer.
- Developed different types of transformations like Source qualifier, Expression, Filter, Aggregator, Lookup, Stored procedure and update strategies.
- Created Mapplets for reusable business rules.
- Ran workflows and sessions during production support and monitor workflow and session logs for error.
- Worked on ETL strategy to store Data validation rules, Error handling methods to handle both expected and non-expected errors and documented it carefully.
- Develop "matching" plans, help determine best matching algorithm, configure identity matching and analyze duplicates
- Build complex profiles and scripts to execute and test mappings and workflows to implement data stewardship and exception processing
- Worked with informatica Power exchange and Informatica cloud to integrate salesforce and load the data from salesforce to oracle Db
- Experience in Informatica Cloud data integration (ICDI/ICRT)
- Interoperability of integration, efficient provisioning and monitoring using Informatica Cloud.
Environment: Informatica Power Center 9.6, Qlik view XI, Teradata 13.0, PL/SQL, SQL*Loader, UNIX, IDQ, Informatica Data Quality 9.5.1/9.6.1 , Sales force, Talend Big Data 5.x/6.x, Talend Real-Time Big Data 6.x
Confidential - Philadelphia, PA
Sr. ETL Developer
Responsibilities:
- Extensively used Informatica Power Center to create data mappings for extracting the data from various Relational systems, applying appropriate Transformations and Loading.
- Extensively used Informatica client Tools Source Analyzer, Warehouse designer, Mapping designer, Mapplets Designer, Transformation Developer.
- Developed the mapping as per the business rules.
- Creation of customized Mload scripts on UNIX platform for Teradata loads
- Implemented various integrity constraints for data integrity like Referential Integrity, using Primary key and foreign keys relationships.
- Implemented the design of mappings for Implementing SCD TYPE 1, SCD TYPE 2.
- Developed numerous Complex Informatica Mapplets and Reusable Transformations as needed.
- Designed and created complex source to target mapping using various transformations inclusive of but not limited to Lookup, Aggregator, Joiner, Filter, Source Qualifier, Expression and Router Transformations.
- Expertise in using different tasks (Session, Assignment, Command, Decision, Email, Event-Raise, Event- Wait, Control).
- Developed business rules for cleansing/validating/standardization of data
- Performed Data cleansing using external tools like Name Parser and Dataflow.
- Optimized Query Performance, Mapping Performance, Session Performance and Reliability.
- Extracting the Data from the sales force source systems and loading In to respective targets as per the business rules.
- Configured the mappings to handle the updates to preserve the existing records using Update Strategy Transformation (Slowly Changing Dimensions SCD Type-2).
- Implemented Source Pre-Load, Source Post-Load, Target Pre-Load and Target Post-Load functionalities.
- Extensive Performance Tuning of Sources, Targets, Mappings and Sessions.
- Used Debugger and breakpoints to view transformations output and debug mappings.
- Involved in the admin activities of Informatica.
- Worked on TSQL queries.
- Implemented Pipeline Partitioning to improve performance.
- Created very useful UNIX shell scripts while writing cron jobs for batch processing.
- Provided Production Support at the end of every release.
- Documented Technical specifications, business requirements and functional specifications for the all Informatica Extraction, Transformation and Loading (ETL) mappings.
- Expertise in working in Teradata systems and used utilities like Multiload, Fastload, Fastexport, BTEQ, TPump, Teradata SQL
- Implemented Stored Procedures, Functions, views, Triggers, Packages in PL/SQL.
Environment: Informatica Power Center 9.6, Informatica cloud, Qlik view XI, Teradata 13.0, PL/SQL, SQL*Loader, UNIX, sales force, Veeva, Erwin, IDQ. Sales force(Veeva Data Model), Informatica Data Quality 9.5.1/9.6.1
Confidential - Atlanta, GA
ETL Developer
Responsibilities:
- Extraction and Transformation of data from various sources such as Oracle and Flat files and loading them into the Oracle target database using Informatica Power Center.
- Worked on Power Center client tools like Source Analyzer, Warehouse Designer, Mapping Designer, Mapplets and Transformations Developer.
- Developed different types of transformations like Source qualifier, Expression, Filter, Aggregator, Lookup, Stored procedure and update strategies.
- Extensively worked on Informatica IDE/IDQ. Created Mapplets for reusable business rules.
- Involved in massive data profiling using IDQ (Analyst Tool) prior to data staging.
- Created business rules in Informatica Developer and imported them to Informatica power center to load the standardized and good format of data to staging tables.
- Have very good knowledge on all the data quality transformation which will be used throughout the development.
- Develop the pre-processor and post-processor scripts in PERL for the Mercer Funds files to create the index files for the incoming zip files and to update corresponding DB2 tables during the post ingestion process
- Ran workflows and sessions during production support and monitor workflow and session logs for error.
- Worked on ETL strategy to store Data validation rules, Error handling methods to handle both expected and non-expected errors and documented it carefully.
- Used Update Strategies for cleansing, updating and adding to the data in the warehouse.
- Extensively involved in unit and integration testing. Worked closely with QA team during the Testing phase and fixed bugs that were reported.
- Optimized and performance Tuning in mappings to achieve higher response times
- Carried out unit and Integration testing for Informatica mappings, sessions and workflows.
- Expertise in working in Teradata systems and used utilities like Multiload, Fastload, Fastexport, BTEQ, TPump, Teradata SQL, TPT
Environment: Informatica Power Center 9.1, Oracle10g, Teradata 12.0, PL/SQL, SQL*Loader, UNIX, Erwin, IDQ. Informatica Data Quality 9.5.1/9.6.1 , Perl
Confidential - Phoenix, AZ
ETL Developer
Responsibilities:
- Created Informatica Mappings, sessions including Command tasks like Event Wait, Event Raise, and Timer and assignment workflows on business requirements
- Created Technical Specification documents based on high level requirement documents
- Reviewed Technical specification documents with the functional owners.
- Involved in design, development and implementation of the Enterprise Data Warehouse (EDW) and Data Mart.
- Used external tools like Address for cleansing the data in the source systems.
- Designed mappings using Source qualifier, Joiner, Aggregator, Expression, Lookup, Router, Filter, and Update Strategy transformations and Mapplets to load data into the target involving slowly changing dimensions. Used Workflow Manager for creating and maintaining the Sessions and Workflow Monitor to monitor workflows.
- Enhanced existing UNIX shell scripts as part of the ETL process to schedule tasks/sessions.
- Coordinated with end users and reporting teams to correlate Business requirements
- Extraction, transformation and loading of data were carried out from different sources like Flat files, SQL Server, Sales force. Expertise in working in Teradata systems and used utilities like Multiload, Fastload, Fastexport, Bteq, TPump, TPT.
- Writing Teradata SQL queries to join or any modifications in the table.
- Used Debugger and breakpoints to view transformations output and debug mappings.
Environment: Informatica Power Center 8.1, Oracle10g, Teradata12.0, PL/SQL, SQL*Loader, UNIX, Erwin.
Confidential - Chicago, Illinois
ETL Developer
Responsibilities:
- Accomplished data movement process that load data from DB2 into Teradata.
- Extraction, Transformation and Loading of data were carried out from different sources like Flat files, SQL Server Involved in creating and designing mappings and mapplets using Expression, Filter, Router, Joiner, Lookup, Update Strategy, Stored Procedure, Union and other transformations. Reworking if any discrepancies in the flat file extracts.
- Worked on XML transformations to send input to Permits as per specifications.
- Moving the code between development to testing environments
- Fixing the Bugs rose during System and Integration testing.
- Done Audit and Reconciliation for the data during SIT. Worked in Teradata Utilities Bteq, Fastload, Fastexport, Multiload, and improved the design of Bteq, Multiload.
Environment: Informatica Power Center 8.1, Oracle 9i, Teradata 11.0, DB2 SQL, SQL*Loader, UNIX, Erwin.
Confidential
ETL Developer
Responsibilities:
- Extensively worked in the performance tuning of programs, ETL procedures and processes.
- Extraction, Transformation and Loading of data were carried out from different sources like Flat files, SQL Server Involved in creating and designing mappings and mapplets using Expression, Filter, Router, Joiner, Lookup, Update Strategy, Stored Procedure, Union and other transformations. Worked in Teradata Utilities, Fastload, Bteq, multiload
- Error checking & testing of the ETL procedures & programs using Informatica session log.
- Performance Tuned Informatica Targets, Sources, mappings & sessions for large data files by Increasing data cache size, sequence buffer length and target-based commit interval.
- Reviewed Technical specification documents with the functional owners.
- Developed parallel jobs using technical specification documents.
Environment: Informatica Power Center 7.6., Oracle 9i, Teradata 10, PL/SQL, SQL*Loader, UNIX, Erwin.
Confidential - Atlanta, GA
ETL Consultant
Responsibilities:
- Involved different phases of SDLC from requirement, design, development, testing.
- Identified performance issues in existing sources, targets and mappings by analyzing the data flow, evaluating transformations and tuned accordingly for better performance.
- Extraction, Transformation and Loading of data were carried out from different sources like Flat files, SQL Server Involved in creating and designing mappings and mapplets using Expression, Filter, Router, Joiner, Lookup, Update Strategy, Stored Procedure, Union and other transformations. Developed the mappings as per the ETL specifications.
Environment: Informatica Power Center 7.1 Oracle 9i, Teradata 10, SQL, Teradata SQL Assistant, Unix, Erwin.