Datawarehouse Engineer Resume Profile
Irvine, CA
Professional Summary:
- Over 8 years of IT Experience in analysis, design, development, implementation and troubleshooting of Data warehouse applications.
- Expertise in building Enterprise Data Warehouses EDW , Operational Data Store ODS , Data Marts, and Decision Support Systems DSS using Multidimensional and Dimensional modeling Star and Snowflake schema Concepts.
- Significant Multi-dimensional and Relational data modeling experience, Data Flow Diagrams, Process Models, ER diagrams with modeling tools like ERWIN VISIO.
- Involved in all phases of data warehouse project life cycle. Designed and developed ETL Architecture to load data from various sources like DB2 UDB, Oracle, Flat files, XML files, Sybase and MS SQL Server into Oracle, Teradata ,Netezza, XML, and SQL server targets.
- Demonstrated expertise in utilizing ETL tool Informatica power center 9.x/8.x/7.x/6.x/5.x for developing the Data warehouse loads as per client requirement.
- Strong experience in implementing CDC using Informatica Power Exchange 8.x/7.x.
- Extensive experience in developing mappings using various transformations like Source Qualifier, Expression, Lookup, Aggregator, Router, Rank, Filter and Sequence Generator transformations and various reusable components like Mapplets and reusable transformations.
- Excellent knowledge of Informatica Administration. Involved in grid management, creation and up gradation of repository contents, creation of folders and users and their permissions.
- Extensive knowledge in developing Teradata, Fast Export, Fast Load, Multi load and BTEQ scripts. Coded complex scripts and finely tuned the queries to enhance performance.
- Profound knowledge about the architecture of the Netezza database.
- Experience in writing PL/SQL procedures for processing business logic in the database. Tuning of SQL queries for better performance.
- Extensive experience in implementation of Data Cleanup procedures, transformations Scripts, Triggers, Stored Procedures and execution of test plans for loading the data successfully into the targets.
- Good experience in designing and developing audit, error identification and reconcile process to ensure the Data Quality of Data warehouse.
- Excellent knowledge in identifying performance bottlenecks and tuning the Informatica Load for better performance and efficiency.
- Experience in UNIX shell scripting, CRON, FTP and file management in various UNIX environments.
- Strong understanding of Data warehouse project development life cycle. Expertise in documenting all the phases of DWH projects.
Technical Skills:
Databases | Oracle 7.x/8.x/9i/10g/11g, SQL Server 2008/2005/2003/2000, Teradata V2R4 V2R5, DB2 UDB 7.2/8.0, SQL Assistant, My SQL 5.0/4.1Editors SQL Navigator, Toad |
ETL Tools | Informatica Power Center 9.1/8.6/8.5/8.1.2/8.1.1/7.1.1/ 6.2 / 5.1 And Power Mart 6.2 / 6.0 / 5.1 |
Data Modeling Tools | ERWIN r 7/r4, Visio 2007/2010, ER/Studio |
Programming Skills | Shell Scripting K-Shell, C-Shell , PL/SQL, SQL, PERL, FORTRAN, JAVA Eclipse IDE and Net Beans IDE , HTML, JAVA Script, J2EE, CSS |
Methodologies | Data Modeling Logical, Physical Dimensional Modeling - Star / Snowflake |
Reporting Tools | Cognos, Business Objects XI, MS SQL Server Reporting services 2005 SSRS , Developer 2000 Forms 5.0, Reports 6i , Crystal Reports 10, Crystal Reports 2008 |
Operating Systems | UNIX Sun-Solaris, HP/UX , Windows 95/98/00/NT/XP |
PROFESSIONAL EXPERIENCE:
Confidential
Informatica Developer/DataWareHouse Engineer
The Project mainly involves designing, developing, and maintaining the data in the data warehouse and delivering the data to various departments in the Web team according to the client requirements. The ETL process mainly involves extracting the data from Confidential, Flat Files, SQL Server and Mainframes with the target data warehouse as Netezza and SQL Server according to the client requirements
Responsibilities:
- Member of core ETL team involved in gathering requirements, performing source system analysis and development of ETL jobs to migrate the data from the source to the target DW
- Analyzed the business requirement document and created functional requirement document mapping all the business requirements.
- Worked in Agile Methodology.
- Created logical and physical data models for the star schema using ERWIN 4.1
- Used Erwin to reverse-engineer and refine business data models.
- Involved in designing the process flow for extracting the data across various source systems.
- Extensively involved in Data Extraction from Mainframe source VSAM , Oracle using Power Exchange when there was change in the source data sets.
- Prepared Data Architect document, Mapping Specification and Unit testing documents for ease of future maintainenance and support.
- Installed and configured Informatica PowerExchange CDC 9.1.0 and 9.0.1 for Oracle on UNIX platform.
- Designed and developed mapping using various transformations like Source Qualifier, Expression, Lookup, Aggregator, Router, Rank, Filter and Sequence Generator transformations.
- Created Workflows and used various tasks like Email, Event-wait and Event-raise, Timer, Scheduler, Control, Decision, Session in the workflow manager.
- Used Mapping Variables, Mapping Parameters in the Informatica Mappings to filter the daily data from the source systems.
- Made use of Post-Session success and Post-Session failure commands in the Session task to execute scripts needed for cleanup and update purposes.
- Tuned performance on Sources, targets and mappings and SQL queries in the mappings.
- Implemented Informatica push down optimization for utilizing the data base resources for better performance.
- Developed sessions using different types of partitions like round robin, hash key portioning for better performance.
- Redrafted non-ANSI-compliant Sybase procedures to ensure adherence to ANSI.
- Automated Unix shell scripts to verify the count of records added everyday due to incremental data load for few of the base tables in order to check for the data consistency.
- Involved in writing windows batch scripting.
- Worked with Netezza database Developed stored procedures on Netezza and SQL server for data manipulation and data warehouse population
- Involved in writing PL/SQL code in Oracle stored procedures, functions and packages to support applications front end and back end.
- Developed PL/SQL procedures, functions to facilitate specific requirement.
- Worked on SQL Loader for bulk load of data and used SQL tuner for tuning SQL.
- Optimized and Tuned SQL queries and PL/SQL blocks to eliminate Full Table scans to reduce Disk I/O and Sorts.
- Used the feature EXPLAIN PLAN to find out the bottlenecks in a given Query, thus improving the performance of the job.
- Automated the batch jobs using Unix shell scripts.
- Scheduled jobs using Autosys and monitored automated weekly jobs.
- Prepared Detail design documentation thoroughly for production support and release management department to use as hand guide for future production runs before code gets migrated.
- Also involved in conducting and leading the team meetings and providing status report to project manager.
Environment: Informatica Power Center 9.1/8.6, power exchange, Erwin r7, Oracle 11g, SQL, PL/SQL, DB2 8.0, MS SQL Server 2008, Flat Files, Autosys, Windows XP, UNIX, PL/SQL, SQL Loader, TOAD, ANSI SQL, Netezza 7.0.2
Confidential
Sr ETL Informatica Developer
Confidential is a for-profit education company that sells online schooling and curriculum to state and local governments. Its educational products and services are designed as alternatives to traditional bricks and mortar education for public school students from kindergarten to 12th grade. K12 is a publicly traded education management organization EMO that provides online education services to charter school student
Responsibilities:
- Interacted with both Technical, functional and business audiences across different phases of the project life cycle.
- Actively involved in gathering requirements and acquiring application knowledge from Business Managers Application SME's.
- Analyzed the business requirement document and created functional requirement document mapping all the business requirements.
- Translated the business requirements into technical specifications to build the Enterprise data warehouse.
- Involved in System Study analysis for logical/physical data model there by defining strategy for implementing Star Schema with Fact and Dimension tables.
- Used Informatica as the tool for extracting data from the flat files, relational databases and transforming the data according to the target database and performed loading.
- Identified and tracked the slowly changing dimensions from heterogeneous sources and determined the hierarchies in dimensions.
- Loaded data from various sources like DB2, Oracle, SQL Server, Flat files into Oracle data warehouse..
- Designed and developed various complex ETL mappings by making use of transformations like Source Qualifier, Joiner, Update Strategy, Connected Lookup and unconnected Lookup, Rank, Expression, Router, Filter, Aggregator and Sequence Generator transformations.
- Developed Informatica mappings, sessions and workflows as per the business rules and loading requirements and done some java transformations .
- Created Workflows using various tasks like sessions, events raise, event wait, decision, e-mail, command, worklets, Assignment and worked on scheduling of the workflows.
- Used mapping parameters and variables.
- Prepared mapping specification document, which gives the data flow and transformation logic for populating each column in the data warehouse table.
- Used debugger to analyze the data flow between source and target to fix the data issues.
- Analyzed requirements and created Database Triggers for validations written in PL/SQL.
- Extensively used Oracle PL/SQL programming in backend and front-end functions, procedures, packages to implement business rules, security efficiently and test cases for the project.
- Worked on SQL Loader for bulk load of data and used SQL tuner for tuning SQL
- Extensively wrote SQL Queries based on the given specifications
- Developed PL/SQL procedures, functions to facilitate specific requirement.
- Implemented audit and reconcile process to ensure Data warehouse is matching with the source systems in all reporting perspectives.
- Developed a UNIX Shell scripts which will send the reports to client over the network by using file transfer protocol FTP and generating the log file, which will keep the history for the FTP reports.
- Prepared ETL specification document providing the rules for initial and Delta load, and also business rules to calculate in each column in the target warehouse.
- Created the release requests for QA Builds to include all the release requirements and involved in the implementation of QA, UAT and Production releases.
- Involved in Unit testing, User Acceptance testing to check whether the data loads into target are accurate, which was extracted from different source systems according to the user requirements.
- Maintaining issue log of the issues during the UAT phase for future reference
- Actively involved in the production support and also transferred knowledge to the other team members.
- Provided data loading, monitoring, and system support and worked on data issues raised by end user during its production support phase
- Prepared the Standard Operating Procedure Knowledge Transfer document, which provides necessary information, required for the Maintenance and Operation of the application.
Environment: Informatica Power Center 8.1, Erwin 4.0, DB2, Oracle 9i, DB2,Flatfiles, UNIX, Windows XP Professional, UltraEdit-32,SQL,PL/SQL, MS Visio 2002, MS-Excel, Ms-Access, SQL Navigator, Clear Quest, Clear Case.
Confidential
ETL Developer
Confidential is a leading company dealing with insurance products, financial products and savings and banking. It offers insurance for autos, homes, property, life, mortgage protection, and businesses. It deals in auto insurance with different categories like accident forgiveness, deductible rewards, safe driving bonus and new car replacement, home insurance for renters, condominiums and mobile homes, life insurance and also business insurance.
Responsibilities:
- Actively involved in gathering requirements and acquiring application knowledge from Business Managers Application Owners.
- Prepared the High-level Design document to provide an overview of the technical design specification system required for Application Enhancements.
- Designed data model structure and E-R Modeling with all the related entities and relationship with each entities based on the rules provided by the business manager
- Implemented dimension model logical and physical data modeling in the existing architecture using Erwin.
- Involved in designing the process flow for extracting the data across various systems interacting.
- Extensively worked with ETL tools to extract data from various source including Oracle, Flat files, XML sources.
- Developed ETL routines using Informatica Power Center and created mappings involving transformations like Lookup, Aggregator, Ranking, Expressions, Mapplets, connected and unconnected stored procedures, SQL overrides usage in Lookups and source filter usage in Source qualifiers and data flow management into multiple targets using Routers.
- Used workflow manager for session management, database connection management and scheduled the jobs to run in the batch process.
- Troubleshoot the issues by checking sessions and workflow logs.
- Maintained warehouse metadata, naming standards and warehouse standards for future application development.
- Involved in analyzing the bugs, performance of PL/SQL and provided solutions to improve the same.
- Developed procedures to populate the customer data warehouse with transaction data, cycle and monthly summary data, and historical data.
- Optimize SQL queries for better performance.
- Written and used UNIX shell scripts extensively for scheduling and pre/post session management
- Involved in the performance tuning process by identifying and optimizing source, target, mapping and session bottlenecks.
- Prepared the Standard Operating Procedure Knowledge Transfer document, which provides necessary information, required for the Maintenance and Operation of the application.
- Provided data loading, monitoring, system support and general trouble shooting necessary for all the workflows involved in the application during its production support phase.
Environment: Informatica Power center 7.1.2, Oracle 9i, Erwin r4, XML File, Flat files, SQL, PL/SQL, TOAD 8.0, Autosys 4.0, Windows 2000 and UNIX.
Confidential
Informatica/Warehouse Developer
Confidential is a world leader in providing global network infrastructure products and services that enable the profitable delivery of high-speed Internet, video, data, and voice services to residential, business and mobile subscribers Confidential has its base application running on SAP R3 and this project developed data warehousing solution for the Analysis team in the business area of Software Systems Division. ADC data comes from different operational sources and stored in a Data mart and then data is transformed and loaded into a centralized data warehouse for various strategic business reports.
Responsibilities:
- Interacted with end-users and functional analysts to identify and develop BRD and transform it into technical requirements.
- Extensively used ETL to load data from Oracle database, XML files, and Flat files data also used Power Connect to import data from IBM Mainframes.
- Imported Sources and Targets to create Mappings based on business logic and developed Transformations using Power center Designer. Used Informatica Workflow Manager, Workflow Monitor to create sessions and batches.
- Extensively used Transformation Language functions in the mappings to produce the desired results.
- Partitioned the Sessions for better performance.
- Worked on all the transformations like Lookup, Aggregator, Expression, Filter, Router, Update Strategy, Stored Procedure and Sequence Generator.
- Created and ran pre-existing and debug sessions in the Debugger to monitor and test the sessions prior to their normal run in the Workflow Manager, Workflow Monitor.
- Used workflow manager and workflow monitor to schedule and monitor the workflows.
- Extensively made use of several features in TOAD to keep track of the various source, staging, target tables and used UltraEdit-32 to view the Session log files.
- Extensively worked on correcting and reloading rejected files using Reject Loader Utility.
- Scheduled the Scripts using Autosys Scheduler.
- Developed Shell Scripts to automate file manipulation and data loading procedures.
- Successfully moved the Sessions and Batches from the Development environment to Production environment.
- Completed documentation in relation to detailed work plans, mapping documents and high-level data models.
- Did Unit Testing and tuned for better performance.
Environment: Informatica Power center 6.1.1, Power Connect, Erwin 3.5, Oracle 9i, PL/SQL, MS Access 2000, MS SQL Server 2000, XML, Windows NT 4.0, Windows 2000, HP-UX, Sun Solaris 2.6, UNIX Shell Scripts, Unix Sync Sort.