Etl Consultant Resume
0/5 (Submit Your Rating)
San Antonio, TexaS
SUMMARY
- Around eight years of experience in data warehousing involving in analysis, design, construction, testing and support phases of SDLC in waterfall and agile methodologies.
- Highly skilled in designing data warehouses and data marts, designing and developing ETL and Unix scripts, performance tuning, code optimization and designing reusability.
- Advanced programming abilities in multiple languages including SQL, Java and R. Possess excellent analytical and logical skills which always help me learn and adapt to different technologies.
- Excellent knowledge and working experience in Informatica Power Center Designer, Workflow Manager, Workflow Monitor, Repository Manager
- Implemented various transformation logics in Informatica including Type 1 & Type 2 slowly changing dimensions, reconciliation of data, statistical and target rating engines for performance evaluation etc.
- Developed reusable components in Informatica to process and derive analytical metrics for different performance measures
- Excellent Knowledge and working experience in Datastage Designer & Director
- Developed reusable components in Datastage to extract data from datasets and load it into different databases, to extract data from different databases and load it into datasets and to implement Type 1 and Type 2 Slowly Changing Dimension loads
- Proficient in applying different partitioning techniques in Datastage in order to accomplish efficient processing of data
- Excellent Knowledge in SQL
- Developed SQL scripts to perform end to end ETL dataflow testing
- Experience in developing Unix Shell Scripts
- Developed Data Validation script in Unix to check for the data quality of different metrics and alert the stake holders in case of quality breach
- Experience in Data Modeling and successfully designed data marts
- Involved in Code Optimization, Performance Tuning, Reusability etc.
- Conceptualized and developed a Informatica code review tool to check for the coding standards which will reduce a lot of manual review effort
- Actively involved in User Acceptance Testing and training of end users
- Exposure to SAP Business objects
TECHNICAL SKILLS
Software Tools: Informatica Power Center 8.6/9.x, Datastage 8.x/9.1/11.5 Control - M, SQL Squirrel, Aginity, SAP Business Objects
Programming: SQL, Java, R, Unix Shell Scripting
Databases: DB2, Netezza, Sql Server, Oracle, IBM Dash DB
Key Expertise: Data Profiling, Data Warehousing, Data Modeling, Data Analysis
PROFESSIONAL EXPERIENCE
ETL CONSULTANT
Confidential, SAN ANTONIO, TEXAS
Responsibilities:
- Identifying the acquisitions that have brought in the New Members, that have led to deepen the customer relationship within a specific Line of Business, that have contributed to open a relationship with the existing member in a new Line of Business, that have been closed by referral offers, that have contributed to make a particular product as a primary one, among other products which the member acquired from the competitors etc
- Identifying the instances where a Member Service representative helped a customer acquire multiple products in a single contact with the customer
- Identifying the acquisition assists made by the MSRs for those acquisitions that have happened over digital channels
- Deriving monthly reconciled numbers for all the above metrics after taking into the account the products that are lost within a month from the time of acquisition
- Associating the product financial information at the policy level down to the individual products within that policy, thereby knowing how much revenue is brought in by individual member service representative
- Developed reusable sequence and parallel jobs in Datastage to extract data from different databases and load the data into datasets and to extract data from datasets and load data into different databases
- Developed reusable sequence and parallel jobs in Datastage for type 1 and type 2 slowly changing dimensions
- Developed reusable job in Datastage by loading daily load statistics for different tables and performance metrics
- Developed data validation script in Unix to check for the data quality of different metrics and alert the IT team in case of quality breach
- Developed SQL scripts to perform end to end ETL dataflow testing
- Designed data model to build analytical data mart
- Worked on performance tuning of long running jobs by identifying bottle necks in Datastage and SQL queries before fully deploying the application in production
- Sync up with the Business Analysts to understand the Business requirements
- Coordinate with various teams to identify the source data systems
- Perform Data profiling on the source systems to identify the required data
- Design the Physical & Logical Data model and conduct walkthrough with the Data Modeler
- Design the Datastage ETL logic to integrate the data from various systems and to derive the defined performance metrics
- Coordinate with offshore to prepare the Source to Target Document, Low Level Design Documents and to develop the code
- Perform Code reviews and conduct Code Walkthrough with the stakeholders
- Review the Unit test cases and the test results to ensure they meet the business requirements
- Coordinate with the DBAs in getting the DB components created in all the environments
- Design the Datastage jobs cycle flow and coordinate with the scheduling team to get the jobs scheduled
- Coordinate with infrastructure team to migrate the Datastage and Unix code to different environments
- Coordinate with Offshore throughout the project and update them to carry out their tasks
- Support the production code throughout the warranty period
- Prepare Project transition document for the Maintenance team
ETL CONSULTANT
Confidential, SAN ANTONIO, TEXAS
Responsibilities:
- Developed statistical and target Rating engines in Informatica
- Developed solution to convert and store employee hierarchical data in a fashion so as to allow for seamless drill up and drill down the hierarchy layers and analyse different performance metrics
- Worked on performance tuning of jobs by identifying bottle necks in Informatica and Sql and fixing them
- Actively involved in user acceptance testing (UAT) and training of end users
- Gather Business Requirements from the Business Analysts
- Design the Physical & Logical Data model and conduct walkthrough with the Data Modeler
- Design the entire front end and the Informatica ETL Flow
- Prepare the Source to Target Document and Low Level Design Documents
- Review the Unix and Informatica components
- Conduct Code Walkthrough
- Review the Unit test cases and the test results to ensure they meet the business requirements
- Coordinate with the DBAs in getting the DB components created in all the environments
- Design the Informatica jobs cycle flow and coordinate with the scheduling team to get the jobs scheduled
- Coordinate with infrastructure team to migrate the Unix and Informatica code to different environments
- Coordinate with Offshore throughout the project and update them to carry out their tasks
- Support the production code throughout the warranty period
- Prepare Project transition document for the Maintenance team
ETL CONSULTANT
Confidential, SAN ANTONIO, TEXAS
Responsibilities:
- Understand the Source System Data Architecture by syncing up with the team in the nascent stages of the project on a regular basis
- Gather the requirements from the Business team
- Analyze the data in the source systems
- Design the Data Model to build the Data Mart for Reporting and Analytics after extensive analysis to simplify Report Building
- Coordinate with DBA in building the DB components
- Design the Datastage ETL Flow
- Coordinate with the Data Infrastructure team to set up the required infrastructure
- Develop the Unix and Datastage jobs and perform code reviews
- Coordinate with the scheduling team to get the jobs scheduled as desired
- Support the production code throughout the warranty period
ETL CONSULTANT
Confidential, SAN ANTONIO, TEXAS
Responsibilities:
- Understand the existing application to bring in similar data from the new system
- Coordinate with the vendor/supplier to understand the new system and Data architecture
- Design the Data model to build tables that house combined data from both the systems
- Coordinate with DBA in getting the DB components created in all environments
- Design the Datastage ETL flow to populate near real-time data into the combined tables by employing a robust mechanism to correct a defective load with minimal manual intervention
- Analyze the ETL and reporting impacts for downstream applications by bringing in the new data
- Brief the Business team about the related impacts and help them to understand any new data that can be of immense use to them
- Develop the Datastage and Unix code and perform frequent reviews
- Coordinate with the scheduling team to get the jobs scheduled as desired
- Support the production code throughout the warranty period
ETL DEVELOPER
Confidential
Responsibilities:
- Understand the Business Requirements by syncing with Onshore on regular basis
- Analyze the Operational Data by querying on different tables and identify the ones that are required to meet the Business Requirements
- Prepare the Design and Source to Target Documents
- Develop the Informatica ETL Code and the required infrastructure code in Unix Shell Scripting
- Conduct Peer Reviews as per the processes set by my company and conduct code walkthrough with the Onshore Team
- Perform Unit Testing and System Testing and log the test results
- Undertook the additional responsibility as being the Process POC for other projects and enabled the teams to perform as per the process standards set by my organization and ensured that they were logged into the system
- Developed reusable infrastructure code in Unix shell scripts required for the execution of Informatica jobs and the same may be reused for any new ETL cycle. The code will archive the Data folder in UNIX before every cycle run, generate the ETL NR by calling the DB sequence, generate the start date based on the end date of the previous cycle by querying on the infrastructure table and update the job execution status upon the completion of the cycle