Sr. Enterprise Data Architect Resume
Arlington, TX
SUMMARY
- 18 years of IT experience in Conceptualization, Source Data Analysis, Data Model Design, Development, Testing, Data Standardization, Integration and Production Support for Data warehouse and Business Intelligence systems using ETL, BI tools and Web applications on Unix and Big Data platforms in various verticals. Experience working in Financial, Healthcare, Human Resource, Claims, Education and Security/Identity Domains.
- Led Various Data Mart implementations including Data Model design and integrated it with the existing Data marts in the Warehouse.
- Extensive experience in conducting meeting with Business Users/BSA and capturing the business requirement, translating them into Technical Specification for the Data warehouse implementation.
- Highly proficient in Data Modeling retaining concepts of RDBMS, Logical and Physical Data Modeling and Multidimensional Data Model design techniques (data Granularity, Facts and dimensions, Star, Snow - Flake schemas, Data Vault and confirmed dimensions).
- Complete knowledge of data warehouse methodologies (Ralph Kimball, Bill Inmon and Dan Linstedt), ODS, EDW and Metadata Management.
- Extensive knowledge in designing and working with Data Vault Data Models with Hubs, Links and Satellites.
- Expert-level mastery in designing and developing complex ETL Informatica mappings using various transformations, to implement logic such as slowly changing dimensions, CDC, Point-in-Time data snapshot and incremental/Delta data load.
- Extensive experience in using ETL/ELT tools like Informatica PowerCenter, Power Exchange, Informatica Cloud, Data Stage and extracting data extraction from heterogeneous sources like Oracle, SQL Server, SAP, Flat File (CSV, XML, JSON), Salesforce, Big Data Databases like Hive, and Mainframe Database like DB2, ADABAS/VSAM Files.
- Expert-level mastery in designing and developing complex Informatica mappings using various transformations, to implement logic such as slowly changing dimensions, Point-in-Time data snapshot and incremental loading.
- Experience in source systems analysis, identifying data sources, data profiling and data cleanings using Informatica IDQ/IDQ.
- Experience in developing security structures for data warehouse to give data access to business users based on their granted roles/responsibilities.
- Expert in Performance tuning the existing ETL Loads by analyzing mappings to identify bottle necks and tuning SQL Queries, mappings, sessions to optimize performance. Provided training to the developers to develop better performing mappings.
- Extensive experience in using ETL tools like Informatica PowerCenter, Power Exchange, IDQ/IDE, IICS, MDM and SSIS.
- Extensive knowledge of implementing BIG DATA Solutions like YARN, HIVE, SQOOP, FLUME, SPARK, PYTHON, PIG and KAFKA.
TECHNICAL SKILLS
Data Warehousing: Informatica Power Center 10.2/9.6/8.6/7.1 , Informatica Cloud (IICS), Informatica Power Exchange, IDQ/IDE Data Profiling, Talend, MDM tools, SSIS, Data Virtualization
Relational Databases: Oracle Exadata/12c/11g/10g/9i/8i/7.x, MS SQL Server, DB2 UDB, ADABAS Mainframe
Big Data: HADOOP, HDFS, Spark, Python 3, Kafka, PIG, Hive, Sqoop, Flume, HBase, Cassandra, Snowflake.
NoSql Databases: HBase, Cassandra, MongoDB
DB tools: TOAD 11.x/9.x, SQL Developer, SQL* Plus, SQL * Loader, Developer 2000.
Scripting Languages: UNIX Shell Scripting, Perl, JavaScript.
Programming Languages: PL/SQL, SQL, T-SQL, Java (AWT, Swing, Servlets, EJB and JDBC), JavaScript, J2EE, C, C++, Visual Basic 5.0/6.0, Visual Studio, .NET.
BI Tools: Business Objects, Cognos BI, Micro strategy
Operating Systems: UNIX, AIX, Sun Solaris8, HP-UX, MS Windows NT/2000/XP
Data Modeling Tools: Erwin, Oracle Designer, Microsoft VISIO, UML, ER Studio.
Data Modeling Methods: Logical/Physical/Dimensional Database Design, Ralph Kimball/Bill Inmon methodology, Star/Snowflake Schemas, Data Vault, Staging/ODS, SDLC.
AWS Services: DynamoDB, EC2, S3, EBS, EMR, Kinesis
PROFESSIONAL EXPERIENCE
Confidential, Arlington, TX
Sr. Enterprise Data Architect
Responsibilities:
- Interacted with business users to gather and analyze the project requirements. Documented Technical Specifications for the Data Models/ETL processes, Reporting Parameters/Domains, Ad-hoc Views and Data warehouse Security.
- Translated business data requirements into Data Vault/Canonically data model (Hub, Link and Satellite) to store the data into the Date warehouse from various sources with historic data, also created Facts/Dimensions for Star Schema/Snowflake Schema for Presentation Layer/BI Data Analytics.
- Conducted meetings with BSAs to identify and fix gaps in the Business Requirement Document (BRD) and captured changes to the Technical Requirement Document (TRD).
- Created a Standard ETL Specification template that was used as a base line to provide the most important data elements and specification format that ETL was interested in knowing in order to complete the Informatica mappings effectively.
- Worked with the DA/ETL Lead in designing a Data Model/Process to identify the duplicate batch in CVL input file.
- Developed complex Informatica Mappings to Capture Point-In-Time Snapshot Data as of specific day of each month, Capture Incremental Changes for Historic Data reporting and to perform Advanced Analytics.
- Experience in working with Bigdata Sources/Targets to Load data into data Lake and utilize Tibco-Data Virtualization tool to provide data to the Business Users.
- To make the ETL Recovery Process Easier, Identified the cause why ETL was facing an Issue with the logic that was used to populate the effective date time stamp columns in the Data Vault. Made changes to data model design, PIXL views and Recommended ETL to parameterize the load date in the ETL Loads.
- Identified standardization issue in calculation of the Primary Hash Keys, between various ETL Technical Specifications for same HUB/LINK key. Proposed a Standard way of calculating the Hash Keys across the board and made it as an ETL Standard which was later used as a part of the Data Model standard.
- Understand and explain the interactions between systems, applications, and services within the environment, and evaluate the impact of changes or additions to the environment.
- Extensive experience on providing Production Support through the weekends to resolve issues like recovering SSP Accounts Data, which was caused due to missing data and erroneous values in the source dataset from the FIS system.
Environment: Informatica PowerCenter 10.2/9.1/8.6 , Power Exchange/CDC 9.0.1, Informatica IDE/IDQ/Developer Suite, Informatica Metadata Exchange, Oracle Exadata/12c/11g/10g/9i, SQL Server 2013, TOAD 11/9.x, Shell Scripting, PL/SQL, SharePoint, Erwin, Big Data Hadoop, Hive, Spark, Tibco Data Virtualization.
Confidential, Boston, MA
Sr. Enterprise Data Warehouse Developer/Architect
Responsibilities:
- Led two most important data mart implementations for HR and Financial (FIN) modules.
- Interacted with business users to gather and analyze the project requirement. Documented Technical Specification for the Data Models/ETL processes, Reporting Parameters/Domains, Adhoc Views and Data warehouse Security.
- Conducted Design discussions and meetings to review and refine the Data Mart Design to the lowest level of grain for each of the Dimensions involved.
- Designed and developed FIN/HR Star & Snowflake Schema and created Fact and Dimension Tables for the Data Warehouse using Data Modeling Tools like Erwin.
- Sourced heterogeneous data from ADABAS Mainframe database, SAP application, Oracle and SQL Server Databases into Enterprise Data Warehouse (EDW) using Power Exchange and SAP IDocs. Academic Data Warehouse (ADW)
- Created a new Data mart for the Student Enrolment and Registration in two phases to create adhoc reports, reusing some parts that already existed.
- Extracted legacy system data into ADW and cleansed data to following current ETL Standards using Informatica as ETL tool. Legacy system to SAP Migration Project
- Architected and implemented data Migration from Mainframe System, cleansed and modified the data into SAP.
- Extracted the data from SAP using IDOCS/SAP Tables back to Mainframe for the other applications to use the HR/FIN data.Data Integration Projects (PTS/FitRec/Schools)
- Provided Data Integration service to various departments able to create analytical reports and communicate data across.
- Analyzed existing mappings to identify bottle necks and tuned source SQLs, mappings, sessions to optimize performance.
- Provided training to the developers to develop better performing mappings.
Environment: Informatica PowerCenter 9.1/8.6, Power Exchange/CDC 9.0.1, Informatica IDE/IDQ/Developer Suite, Oracle 11g/10g/9i, SQL Server 2013, TOAD 11/9.x, Micro Strategy Reporting, SAP, ADABAS/VSAM Mainframe Database on zOS, AIX/UNIX, Redwood Scheduler, Shell Scripting, PL/SQL, SharePoint, and Windows-7/XP.
Confidential, Cambridge, MA
Senior Data Warehouse Architect/Consultant
Responsibilities:
- Gathered requirements from business users and analyze the source systems. Documented Technical Specification for Data Models/Mappings, Report Parameters/Domains, Adhoc Views and Data warehouse Security.
- Designed and developed a data mart using STAR Schema data model for the Capital Approval Processing System (CAPS), to fulfill Requirements for Adhoc and Canned Reports.
- Developed Informatica Mappings to Capture Point-In-Time (PIT) Data on a specific day of each month, this give reports an ability to get Historic Data and useful information for Advanced Analytics.
- Involved in designing the procedures for getting the data from heterogeneous source systems into Data Warehouse. The data was standardized before storing into various Business Unit tables.
- Considering the load dependencies, s essions, worklets and workflows were created. UNIX scripts were used to automate the workflows using pmcmd command.
- Designed Security Structures for securing the Data based on Row-Level and Column-Level security as per the Requirement.
- Analyzed existing Star Schemas and advised necessary changes to ETL processes for best practices to improve performance.ITIS Security Data mart implementation
- Converted legacy system into new Data mart following current ETL Standars using Informtica as ETL tool.
- Desinged and developed processes for ITIS Identity Management Imports and Exports Data to customers.Data Warehouse repository upgrade
- Led a team of developers to upgrade Data Warehouse repository from Informatica 5.1.2 to 7.1.2 and 8.6.
- Carried out impact analyses and Responsible for planning, implementing the data warehouse upgrade for Harvard Data Warehouse (HDW) repository from Informatica 5.1.2 to 7.1.2.
- Redesigned, Implemented, Integrated and scheduled the nightly, weekly and monthly Load processes with the existing Harvard FINSTAR Loads to improve performance.
- Created workflows and worklets. Used various tasks like Session, Event, Command, Decision, Control and Email in Workflow Manager. Wrote Pre-session and Post-session shell scripts for dropping, creating indexes for tables.
Environment: Informatica PowerCenter/Powermart 8.6/7.1.2/5.1.2 , Oracle 11g/10g/9i, TOAD 9.x, Oracle Financials, PeopleSoft, Maestro, Hyperion, Shell Scripting, PL/SQL, PVCS, and Sun Solaris UNIX, Windows-XP.
Confidential, Arlington Heights, IL
Senior Database Warehouse Design Analyst
Responsibilities:
- Designed and developed Star & Snowflake Schema and created Fact and Dimension Tables for the Data Warehouse using Data Modeling Tools.
- Developed various Informatica mappings and mapplets using Source qualifier, Aggregators, lookups, Filters, Sequence, Update strategy and stored procedure transformations.
- Worked with different types of partition techniques like key range, pass through, Round Robin and Hashing.
- Involved in extraction, transformation and loading of data into Oracle database from sources like flat files, Mainframe, SQL Server, PeopleSoft and MS- Access databases.
- Involved in unit testing of mappings, mapplets also involved in integration testing and user acceptance testing.
- Involved in analysis of existing mappings/sessions and Implemented performance tuning logic on targets, sources, mappings and sessions to provide maximum efficiency and performance.
- Worked on PL/SQL packages, stored procedures, triggers and functions to improve design and load performance.
- Created Informatica Mappings using transformations like Source Qualifier, Aggregator, Expression, Joiner, Connected and Unconnected lookups, Filters, Sequence Generator, Router and update strategy.
- Created and managed the global and local repositories and permissions using Repository Manager in Oracle Database. Created users and assigned privileges.
- Created and Monitored Sessions and Batches using Server Manager to load the data into the Target Database.
- Involved in analysis and performance of mappings/sessions. Increased performance by tuning the transformations and discussing database issues with the DBA.
- Used pmserver, pmcmd, pmrepserver command to execute sessions using UNIX shell scripts.
Environment: Informatica PowerCenter/ Powermart 7.1/6.x/5.x, Informatica Power Plug/Metadata Exchange, Oracle 8i, Shell Scripting, SQL, PL/SQL, TOAD, PVCS, Cognos and Sun Solaris UNIX, Windows NT.
Confidential
ETL/Informatica Developer
Responsibilities:
- Worked with Power Center Designer tool to develop mappings and mapplets to extract, transform and load data. Used Copy/Paste functions to reuse mapping, mapplets and transformations.
- Implemented performance tuning logic on targets, sources, mappings and sessions to provide maximum efficiency and performance. Created reusable transformations and mapplets to use in multiple mappings.
- Scheduled Sessions and Batches on Informatica Server using Informatica Server Manager/Workflow Manager.
- Developed various reports using Business Objects like drill up, cross down, cross tab etc.
- Used shortcuts to reuse objects without creating multiple objects in the repository and inherit changes made to the source automatically.
- Developed and supported the Extraction, Transformation and Load process (ETL) for Data Warehouse using Informatica 5.1.
- Involved in analysis and performance of mappings/sessions. Increased performance by tuning the transformations and discussing database issues with the DBA.
- Created and Monitored Sessions and Batches using Server Manager to load the data into the Target Database.
- Created shell scripts for pre/post session actions. Used “pmcmd” command to execute sessions using Perl scripts.
- Developed a database application for the salesperson marketing research to track the potentiality of doing Businesses in particular areas.
- The interface was made user friendly, The records were displayed based on different criteria’s like SIC code, State, Sales volume, Employee Size and random sampling techniques were used.
- Reports were generated from the results which were used to make major business decisions and also tracked sales person’s performance.
Environment: Informatica PowerCenter 5.1/4.7, Power Analyzer 3.5, TOAD, Oracle8x/7.3, SQL Server 7.0, HP UNIX, Mac OS.