Etl Architect/lead Resume
South, CarolinA
SUMMARY:
- 14+ years IT experience in Software development, Analysis, design, development, enhancement and maintenance of Banking Financial services (BFS), Retail, and Healthcare applications.
- 4 years’ IT experience in Data conversion projects for existing clients by using Pentaho 5.2 ETL tool like kettle, spoon. Spoon is a GUI for Extraction, Transformation and Loading (ETL) engine that uses a metadata - driven approach.
- Extensive working knowledge in Software Development Life Cycle (SDLC) using Agile/ Waterfall methodologies with a thorough understanding of various phases such as requirement Analysis/design, LLD, Development, Testing and support.
- Experience in working in an onsite-offshore structure and effectively coordinated tasks between onsite and offshore teams.
- Excellent interpersonal and communication skills, and is experienced in working with senior level managers, business people and developers across multiple disciplines.
- Quick learner, open to suggestions and recommendations and easily adapt to changes
- Excellent Data Analysis skills and ability to translate business logic into transformation logic for ETL processes.
- Solid understanding of Data Warehousing concepts and Dimensional Modeling using Star and Snowflake schemas.
- Extensively involved in Optimization and Tuning of transformations, SQL’s and sessions in Informatica by identifying, eliminating bottlenecks and memory management.
- Experienced with Informatica Advanced Techniques - Dynamic Caching, Incremental Aggregation, Parallel Processing and Pushdown optimization to increase Performance.
- Extensive experience in automating ETL processes (Workflow run, import and export XMLs) using UNIX shell scripting and Autosys scheduler.
- Experience in the reports using Cognos 8 Report Studio, Query Studio, Analysis Studio reports.
- Experience in Informatica repository trouble shooting, creation, Backup and Restore
- Experience in monitoring Informatica repository health through Informatica admin console
- Expertise in Master Data Management concepts, Methodologies and ability to apply this knowledge including MDM solutions
- Strong exposure in designing Data Marts and Data Warehouse using Erwin data modeling tool.
- 2 years of work experience in BI Tools (Actuate, SSRS, Tableau Reporting) and core Java
- Knowledge on the HIPAA transactions and ANSI X12 Code Set 837 (I/P/D), 835, 270 and 271
- Good explore and experience java Technology like OOP, Core java, JDBC
- Experience in Software Development Life Cycle (SDLC) and agile methodology.
- 9+ yrs of PL/SQL programming experience. Extensive experience with PL/SQL in designing, developing functions, procedures, triggers and packages.
- Used Informatica Administrator Console to manage the Domain, Nodes, Integration Service and Repository Services and granting user permissions.
- Experienced in creating stored procedures using PL/SQL and tuning queries to improve the performance.
- Used UNIX Shell scripts extensively to handle large number of upstream and downstream files and possess solid understanding of most popular UNIX commands grep, sed and awk
- Experienced on Informatica Powerconnect for SAP R/3 systems and worked in ABAP method for RFC/ BAPI method of mappings
- Experience in SAP BW data load using Informatica Power Connect and good Knowledge on PSA data model.
- Good experience in IDE tool with Data Cleansing and Data profiling by using IDQ and Data Analyst to examine different patterns of source data.
- Experienced to Profile Analysis Standardize Clean Integrate Score Carding Marge Match Reference Data from various source systems using Informatica Data Quality IDQ Toolkits.
- Worked with Address Doctor different algorithms in IDQ.
- Experienced IDE tool with Informatica Data Quality(IDQ) transformations like AddressDoctor, validator, Parser, Match, Exception, Merge, Standardizer and Human Task
- Experienced in Power Exchange data maps and used CDC techniques to capture the changed data that were used to pull the data from files on the Mainframe /VSAM files.
- Experienced on development and customization of Informatica mappings, mapplets, workflows with worklets & tasks using various transformations for loading the data from multiple sources to data warehouse.
- Created Python scripts to generate and transmit Flat files.
- Configured SSH on Dev/UAT/Production Unix boxes to securely transmit (SFTP) the Flat files to External/Internal teams.
- Proficiency in the areas of Access Management, Trouble Shooting and Technical Documentation.
- Having knowledge in other ETL Tools: - IBM Infosphere (Data stage 8.1), SSIS and OLAP.
- Extensively worked on Monitoring and Scheduling of Jobs using UNIX Shell Scripts
- Experienced with PMCMD to interact with Informatica Server from command line and execute the Shell script.
- Experience on migrating the environment from Informatica 8.x to 9.x.
- Extensive experiences to create Logical Physical Data Modeling on Relational OLTP Dimensional Data Modeling OLAP on Star schema for Fact Dimension tables using CA ERwin.
- Strong exposure on CA Autosys scheduling tool & JIL Coding.
- Excellent communication skills, documentation skills, team problem solving ability, analytical and programming skills in high speed, quality conscious, multi-tasked environment.
- Extensive knowledge OBIEE, SSRS, ALM, JIRA, SVN, share point tool & application for support project repository.
- Proficient in interacting with the business users. Pioneered in different load strategies from heterogeneous source systems to target. Successfully implemented SCD Type1/Type2 load Capture Data Changes and maintain the Data history.
- Experienced to identify the Bottlenecks of data load and Tuned the Bottlenecks for better performance.
- Excellent communication and interpersonal skills. Ability to grasp new concepts very quickly both technical and business related.
- Have knowledge Mulesoft message service for claims HL7, Peoplesoft, JMS skills.
- Have knowledge Hadoop, HDFS and Hive skills.
TECHNICAL SKILLS:
ETL Tools: Informatica Power Center 10.1/9.6/8.x (Repository Manager, Mapping Designer, Meta Data manager, Workflow Monitor), Informatica IDQ, B2B, IDR, MDM, DVO and Pentaho - Kettle
Data Modeling: Dimensional Data Modeling, Star Schema Modeling, Snowflake Modeling, Physical and Logical Data Modeling, Erwin tool
Database: Oracle Exadata 11g, SQL Server 20012, Teradata, Postgres 9.6.
DB Tools: TOAD, SQL*Plus, SQL Developer, SQL * Loader and SQL Assistant
Reporting Tool: Cognos 10.2, Actuate 11, SSRS, BIRT, Tableau 8.2
Programming: SQL, PL/SQL, Java, python, UNIX Shell Scripting
Environment: Windows 7/XP/2000, MS DOS, UNIX.
Others: Autosys, IDE tool, SSIS, Pentaho 5.2, JIRA, SVN, MS Excel, MS Word
PROFESSIONAL EXPERIENCE:
Confidential, South Carolina
ETL Architect/Lead
Responsibilities -
- Onsite coordinator for design and data modeling phase in the solution lifecycle, including architecture review, data analysis, design review, code development, and deployment .
- Build the team of 9 ETL developers (2 onsite and 7 offshore) on the process of gathering Lineage from Informatica objects and database tasks and support production applications.
- Developed Unix Scripts for automating workflow run, stored procedure execution and used Autosys for scheduling.
- Provide status update to the business reporting manager on the project milestones and highlight achievements by the team weekly. Prepare weekly status report
- By using Erwin tool for developing Fact & Dimensional tables, Logical and Physical models.
- Informatica code migrations using Deployment groups and also Export/Import xml.
- Created Python scripts to generate Flat files out of SQL tables using SQL BCP utility and transmit Flat files to downstream server using UNIX credential object.
- We are successfully implemented SSIS ETL project to Informatica 10.1.
- Used Informatica Metadata Manager and Custom Metadata Configurator to display the Lineage data in graphical form. work on edit or view the all workflow and DB connection from Metadata Manager
- Work with Informatica Vendor in case of any data integration or product related issues or defects and with respective BI Teams for Cognos, Tableau COE related admin activities, operations etc.
- We implemented Data schema by using Informatica data replication (IDR) tool.
- Work on performance issue as reported clients and achieved more than 80%.
- Understanding the complex logic business transformation and implement into ETL development and design.
- Work on SFTP Setups for secured file transfers by using Python
- Analyzing the Design pattern of various staging tables and Dimension and facts tables.
- Perform performance tuning in the load level to improve the performance.
- Monitored daily and monthly batches and helped Production support team with any questions, failures and other emergency updates.
- Developing CDC feature, functions, procedures, triggers and packages by using PL/SQL
- We implemented pattern, Null, ERROR and Exception handled for bad data process.
- Used the Address Doctor Geo-coding table to validate the address and performed exception handling reporting and monitoring.
- Created database layer and business layer, defining appropriate relationship between query subjects in cognos Framework Manager.
- Support Informatica MDM system- Data Mappings (landing, staging and Base objects), Data validation, Match and Merge rules, customizing/configuring applications.
- Implemented patterns with cleanse, parse, standardization, validation, scorecard, exception, notification and Human task by using Informatica Data Quality (IDQ).
- It is Agile environment with daily Scrum meetings to track progress and new Requirement.
Environment: Informatica 10.1- Meta Data, EIC, Power Center, IDR, IDQ, HDFS,, SSIS, UNIX, Oracle Exadata, oracle DRM Autosys 11.3, Remedy, putty
Confidential, Mount Laurel, NJETL Tech Lead
Responsibilities:
- Knowledge on Dodd-Frank Regulatory, Financial and Compliance reporting to several LOBs banks, finance, risk, compliance etc
- Support current application & worked closely with Business Users to gather the business requirements.
- Created Python scripts to generate and transmit Flat files.
- Create user credential and folder & privileges developer and business user
- Scheduling and Monitoring Informatica repository backups
- Provide analysis & solution if any performance related issues.
- Prepared Flow diagram, design document and support handbook.
- Worked with Production support team on the setup and knowledge transfer.
- Before project development create Data model by using Erwin tool for Fact & Dimensional tables, Logical and Physical models.
- Actively involved in fine tuning the long running automation jobs.
- Mentor and coach the employees to build a strong team.
- Actively involved in design and data modeling phase in the solution lifecycle, including architecture review, data analysis, design review, code development, and deployment .
- Coordinating the Business users and testing Team.
- Metadata Manager help the impact analysis of particular objects search.
- Create Business Process Flow diagrams showing interfaces among connected systems for each scenario in Visio
- Created the Conceptual Model for the Data Warehouse using Erwin Data Modeling tool
- Facilitated Daily Stand Up Calls and during Implementation phases
- Extensive experience with designing CDC, developing functions, indexes, procedures, triggers and packages.
- Prepared the DDL,DML Scripts for creating the tables and supporting table records
- We implemented file pattern, Null, ERROR and Exception handled for data process.
- Create user credential and privileges developer and business user
- Designed delta or Incremental load process for all the ETL jobs
- Experience in monitoring Informatica repository health through Informatica admin console.
- Performance Tuning of the mappings to handle increasing data volume
- Developed Shell Scripts for retrieving files from FTP server, achieving the source files, Concatenating files and finally to deliver them to remote shared drive
- Oracle 11g Exadata feature used for data compression implemented for Fact tables.
- Writing minus queries for data counts by using Informatica DVO
- Implemented patterns with cleanse, parse, standardization, validation, scorecard, exception, notification and Human task by using Informatica Data Quality (IDQ).
- Informatica Data Quality(IDQ) mappings by using IDE and integrated to the power center
- Build out best practices regarding data staging data cleansing and data transformation routines within the Informatica MDM solution
- Experience with Informatica MDM hub configurations - Data Mappings (landing, staging and Base objects), Data validation, Match and Merge rules, customizing/configuring Informatica data director (IDD) applications.
- Define and build best practices regarding creating Master data using Informatica MDM solution
- Created Datamaps via PowerExchange to connect VSAM COBOL copybook file reads.
- Upgraded Informatica from 8.6 to 9.6.1
- Production support exiting applications in production.
- Experience on Automation by using Autosys scheduling tool
- Monitoring the loads (workflows) whether the loads are scheduled to respective timings.
- Recovery or Re-start in loads in case of any failures.
- Two projects performance improved more than 60%.
- We are using python script for migration data from legacy system to new system.
- We are used python code for connect Database its improve 20% on current system.
Environment: Informatica Power Center 9.6, IDQ, Power Exchange, Meta Data, Informatica MDM, Oracle 11g, SharePoint, Toad, HP Quality Center, MS Office Suite, AutoSys, Erwin, UNIX Shell
Confidential, Phoenix, AZLead ETL Architect
Responsibilities:
- Worked as a Data migration consultant by converting various complex objects like Customer Master Data, Vendor Master Data and Joint Operations Agreements etc.
- Responsible for translating business requirements into technical specs
- Convert SSIS ETL project to Informatica as client request.
- Played a lead role in designing customer experience profile backend processes
- Created the Conceptual Model for the Data Warehouse using Erwin Data Modeling tool
- Experience in production support to make sure all issues are fixed in the respective turn-around times.
- Knowledge EDI claims in HIPAA transactions and ANSI X12 Code Set 837 (I/P/D), 835, 270 and 271 and HL7
- Knowledge on Coding and billing tools for ICD-10-CM/PCS, CPT, HCPCS
- Work on Protected health information (PHI) under US law is any information about health status, provision of health care.
- Designed & developed the reports using Cognos 8 Report Studio, Query Studio, Analysis Studio reports.
- Customized Cognos Connection with appropriate reports and security measures.
- Developed ETL mappings and workflows to load Input feeds using Informatica Power Center
- Responsible for optimizing SQL queries & Automation jobs based on requirement
- Responsible for developing unix shell, perl scripts to process PL/SQL programs, SQL scripts and Informatica workflows
- Utilized/development of Autosys to schedule jobs by using JIL code
- Responsible for Informatica Monitoring the loads (workflows) whether the loads are scheduled to respective timings
- Manage and coordinate onsite and offshore team. Arrange status meetings.
- Provide status update to the business reporting manager on the project milestones and highlight achievements by the team weekly. Prepare weekly status report
- Used Informatica repository manager to create folders and add users for the new developers.
- Developed complex mappings in Informatica to load the data from various sources
- Work on Teradata utilities like Fastload, Mload for bulk data loading.
- Explore in Teradata client tools BTEQ, SQL Assistant used for write SQL queries and testing.
- Implemented Exception Handling Mappings by using Data Quality (IDQ), Data Profiling, Data cleansing and data validation.
- Good explore in Informatica DVO for test the data validation by write SQL queries source and Target database and schedule tasks and reporting test result.
- Implemented data cleansing for Dimension tables by using Informatica Data Quality(IDQ)
- IDQ IDE tool used for Address and customer names and matching and data masks.
- JIRA system used for tracking bug/enhance of tracking.
- Designed the custom error control logic with in the pipeline to capture and load the bad records to a control table, and recover the workflow in the event of failures.
- Experience in defining and configuring landing tables, staging tables, base objects, lookups, query groups, queries/custom queries, packages, hierarchies and foreign key relationships.
- Created procedures to truncate data in the target before the session as per requirement.
- Written documentation to describe program development, logic, coding, testing, changes and corrections
- We are using python script for migration data from legacy system to new system.
- Implemented SOAP, Rest Web Services for Real Time process.
- Provided support during various phases of project and plan the agile process.
- Was involved in production support to make sure all issues are fixed in the respective turn-around times.
- Responsible for Monitoring the loads (workflows) whether the loads are scheduled to respective timings
- Create user credential and privileges developer and business user
Environment: Informatica Power Center 9.1, SSIS, IDQ, Oracle 10g Teradata 14, Actuate, Cognos 8.2, MS-SQL server 2012, and JIRA, Mulesoft, Toad, HP Quality Center, MS Office Suite
Confidential, Phoenix, AZDWH Analyst
Responsibilities:
- Responsible for gathering requirement of the project by directly interacting with client and made analysis accordingly.
- Monitoring the loads (workflows) whether the loads are scheduled to respective timings. Recovery or Re-start in loads in case of any failures.
- Coordinated the work flow between onsite and offshore teams.
- Experience in HIPAA transactions and ANSI X12 Code Set 837 (I/P/D), 835, 270 and 271
- Defined various facts and Dimensions in the data mart including Fact tables, Aggregate and Summary facts.
- Assessed the DB2 environment for implementation of the ETL solutions.
- Created the Conceptual Model for the Data Warehouse using Erwin Data Modeling tool
- Implemented SOAP web service for Report server authentication.
- Extracting, Scrubbing and Transforming data from Flat Files, Oracle, SQL Server, Teradata and then loading into DB2 database using Informatica.
- Worked on optimizing the ETL procedures in Informatica 8.6 version.
- Performance tuning of the Informatica mappings using various components like Parameter files, Variables and Dynamic Cache.
- Provided Knowledge Transfer to the end users and created extensive documentation on the design, development, implementation, daily loads and process flow of the mappings.
- Implementing logical and physical data modeling with STAR and SNOWFLAKE techniques using Erwin in Data warehouse as well as in Data Mart.
- Used Type 1 and Type 2 mappings to update Slowly Changing Dimension Tables.
- Involved in the performance tuning process by identifying and optimizing source, target, and mapping and session bottlenecks.
- Worked Database level tuning, SQL Query tuning for the Data warehouse and OLTP Databases.
- Used Informatica repository manager to create folders and add users for the new developers.
- Maintained stored definitions, transformation rules and targets definitions using Informatica repository manager.
- Negotiated with superiors to acquire the resources necessary to produce the project on time and within budget. Get resources onsite if required to meet the deadlines.
- Developed UNIX Shell scripts for calling the Informatica mappings and running the tasks on a daily basis.
- Wrote Oracle PL/SQL procedures and functions whenever needed.
- Created & automated UNIX batch scripts to run sessions on desired date & time for imports.
- Experience in production support to make sure all issues are fixed in the respective turn-around times.
Environment: Informatica Power Center 8.6, Pentaho, PL/SQL, DB2, Oracle 10g, Teradata 14, TOAD, Erwin, SSRS, Unix, SQL Server 2008, Query Surge, Windows XP, Visio 2003.
ConfidentialETL Analyst
Responsibilities:
- Participated in documenting the existing operational systems.
- Involved in the requirements gathering for the warehouse. Presented the requirements and a design document to the client.
- Created ETL jobs to load data from staging area into data warehouse.
- Analyzed the requirements and framed the business logic for the ETL process.
- Designed and developed complex aggregate, join, lookup transformation rules (business rules) to generate consolidated (fact/summary) data using Informatica Power center
- Designed and developed mappings using Source qualifier, Aggregator, Joiner, Lookup, Sequence generator, stored procedure, Expression, Filter and Rank transformations
- Extensive experience with PL/SQL in designing, developing functions, procedures, triggers and packages.
- Developed Informatica mappings, re-usable Sessions and Mapplets for data load to data warehouse.
- Designed and developed Informatica mappings and workflows; Identify and Remove Bottlenecks in order to improve the performance of mappings and workflows and used Debugger to test the mappings and fix the bugs
- Worked on 2 projects migrated from legacy system database& File system to RDBMS Oracle database.
- Have knowledge on Report development.
- Kettle used for development and carte used for execution of scripts
- Developed and maintained optimized SQL queries in the Data Warehouse.
Environment: Windows XP/NT, Informatica Power center 8.1, JIRA, UNIX, Oracle 10g, SQL, PL/SQL, Erwin, Actuate, SVN,TOAD, Web logic server
ConfidentialDICE Project
Responsibilities:
- Design and Construction of BIRT Reports (open source code provided by eclipse).
- Write Stored Procedures for application support.
- Identified the bottlenecks in the sources, targets, mappings, sessions and resolved the problems.
- Support and testing of DICE Portal application for client demos.
- Involved in creating Data Marts and altering and creating new dimensions and Facts
- Design of new Cube reports using Actuate Analytics Cube designer.
- Excel reports by using Actuate eSpreadsheet Reporting tool
- Migration older version to Latest version for Actuate.
Environment: Informatica, Actuate reports, Oracle 9, Java, Web sphere app server 5.1, PVCS
ConfidentialSoftware Analyst
Responsibilities:
- Extensively worked with the data modelers to implement logical and physical data modeling to create an enterprise level data warehousing.
- Created and Modified T-SQL stored procedures for data retrieval from MS SQL Server database.
- Automated mappings to run using UNIX shell scripts, which included Pre and Post-session jobs and extracted data from Transaction System into Staging Area.
- Extensively used Informatica Power Center to extract data from various sources and load in to staging database.
- Extensively worked with Informatica Tools - Source Analyzer, Warehouse Designer, Transformation developer, Mapplet Designer, Mapping Designer, Repository manager, Workflow Manager, Workflow Monitor, Repository server and Informatica server to load data from flat files, legacy data.
- Designed the mappings between sources (external files and databases) to operational staging targets.
- Involved in data cleansing, mapping transformations and loading activities.
- Involved in the process design documentation of the Data Warehouse Dimensional Upgrades. Extensively used Informatica for loading the historical data from various tables for different departments.
Environment: Informatica, JIRA, SVN, Actuate Reports, Web logic, PL/SQL, MS Access, SQL Server, Windows 2000, UNIX
ConfidentialSoftware Engineer
Responsibilities:
- Understanding the Client requirement for moving the data from source to target.
- Translating the Requirements into Design like HLD, LLD
- Involved in designing the dimensional model like
- Prepared the DDL Scripts for creating the tables and supporting table records
- Experienced working with Heath care Information for mapping data from legacy systems to target systems
- Extensively used Pentaho tools like Spoon, carte, kitchen, Transformations.
- Cleanse the source data, Standardize the Vendors address, Extract and Transform data with business rules, and built data module using Spoon Designer.
- Designed and developed complex Aggregate, Expression, Filter, Join, Switch, Lookup and Update transformation rules.
- Developed schedules to automate the update processes and sessions and batches.
- Analyze, design, construct and implement the ETL jobs using kitchen.
Environment: Pentaho 3.8, iReports, Actuate Reports, JBoss, Windows NT, PL/SQL, Excel, Oracle.