Data Architect Resume
Renton, WA
PROFESSIONAL SUMMARY:
- Over 15 years of extensive experience in Information Technology with special emphasis on Design and Development of Data Models, Data Integration (ETL) in Banking, Finance, Health Care, Manufacturing, Telecom and Retail domains.
- 6 years of experience in architecting data warehouse ETL solutions and integrating technical components for large scale of data projects.
- 6 years of strong Data Modeling (OLTP & Dimensional) experience using Erwin 9x/8x, SQL Data Modeler and KDM tools.
- 11+ years of design, development & support experience of ETL Extract, Transform and Loading data methodology for data transformations and processing, in a corporate wide ETL Solution using Informatica PowerCenter 9.x/8.x/7.x and ODI 11g & BODS 4.0 .
- Around 7.0 years of experience in all the phases of Data Warehouse life cycle involving Requirement Gathering, Analysis & Design, Development and Testing to build Data Warehouse & Data Marts.
- 2 years of solid experience on design, deploy and support highly available and scalable distributed Cassandra and NOSQL database solutions for high transactions mission critical applications.
- Primary responsibility of involving feasibility study of business requirements, Understanding source systems, Architect the system and assisting the team members in developing the ETL components, Interfaces, checking their dependencies and load factors.
- Facilitates the design and drives the planning for the integration of all data warehouse components, with scalability, performance, and availability trends.
- Designing & developing OLTP data models using with CA Erwin Data Modeler for database systems.
- Designed and developed Datawraehousing Data Marts by following Star Schema and Snowflake Schema Methodology, using Data Modeling tool ERWIN.
- Created Conceptual, logical, physical (ER) and dimensional data models (Data warehousing data modelling), reports and diagrams based on the requirements.
- Documented logical, physical and dimensional data models (Data warehousing data modelling)
- Facilitates the adoption of the ETL best practices within the organization, including Data integration and evolution of development methodology.
- Drives and performing reviews the design and testing of data models, ETL Components data extracts/transforms and processes and performance & Error handling perspective
- Designed the DWH ETL architecture using multiple ETL tools like Informatica, ODI and BODS.
- Solid experience in Ralph Kimball Methodology, Logical/ Physical Modeling, Dimensional modeling, Star Schema, FACT tables and Dimension tables.
- Expertise inLogical and Physical Data Model design usingvarious modeling Tools likeErwin 7.3andSQL Developer.
- Worked on Teradata utilities like (FLOAD, MLOAD, Bteq and TPUMP) to feed an Operational Data Store and Data Warehouses.
- Preparing extensive documents on the ETL Design, Development, Deployment & daily loads of the mappings.
- Strong Data cleansing / Data Migration experience using ETL Informatica, ODI (Oracle Data integrator) and BODS (Business objects Data Services).
- Experience in Extraction, Transformation and Loading (ETL) data from various data sources into Data Marts and Data Warehouse using Informatica power center components (Repository Manager, Designer, Workflow Manager, Workflow Monitor and Informatica Administration Console).
- Strong Experience in developing Mappings, Mapplet, Sessions/Tasks, Worklets and Workflows using Workflow Manager tools - Task Developer, Workflow & Worklet Designer.
- Responsible for defining the testing procedures, test plans, error handling strategy and performance tuning for mappings, Jobs and interfaces.
- Involved on cluster management on Cassandra like add/bootstrap nodes, remove/ replace node to clusters.
- Maintaining and developing data models, structures, and procedures to ensure the integrity and performance of database components in Cassandra.
- Strong experience on complex inner workings of Cassandra, such as the gossip protocol, hinted handoffs, read repairs & Merkle trees.
- Configurations of Cassandra.yaml, Include file, heap dump directory, Commit log archive, Gossip settings and Change Data Capture (CDC) logging implementations.
- Analyze and resolve technical problems including performance tuning, debugging and troubleshooting Cassandra databases and associated applications..
- Strong implementation experience on Kafka topic creations, File Receiver, Kafka Receiver, and other customized receivers framework interfaces.
- Strong POC experience on Hadoop ecosystems like HDFS, Sqoop, Flume, Hive, Pig and NOSQL database HBase.
- Extensive experience in writing Unix/Linux Scripting for Data validations and data transfer.
- Provided assistance for optimization and performance tuning of Oracle, SQL Server and stored procedures.
- Developing Stored Procedures, Functions, Views and Triggers, Complex SQL queries using SQL Server and Oracle database.
- Working on solution architecture, design of dataflow for batch and streaming data processing using Bigdata concept build apps MSGRTR and DataRTR.
- Providing professional solution and services to clients in various Data Layer requirement segments for Bigdata Hadoop implementation.
- Involved in Development, Implementation and Support of ETL processes (3 Terra Byte) applications.
- Setting up Scheduling mechanism and appropriate tool selections based on project requirement functionality.
- Trained on Big Data Hadoop framework using Hadoop HDFS, Flume, Sqoop, Hive, Pig and Cassandra for large scale structured and unstructured data modeling, data processing, loading & Data Analytics and very much comfortable to execute POC/projects if situation demands.
- Solid understanding of Object-Oriented software application design and development - Java & Open Source.
- Very good proven experience in identifying, developing and maintaining relationships with technical vendors, Business Users and prime contractors in the Data Warehousing, Business Intelligence space.
TECHNICAL SKILLS:
DW Tools: ETL Informatica PC 9x, ODI 11g, BODS 4x, BO 4x & IDQ 8
Data Modeling Tools: Erwin 8x/9x, ERStudio, KDM & TOAD Data Modeler
Big Data: Cassandra 3.0.6, HDFS 2.2, Sqoop 1.4, Flume1.5, Kafka 0.8, Zookeeper, Hive 1.1, Pig 0.12, Spark 2.0.1, HBase & Solar 7.0
RDMS: Oracle, DB2, SQL Server, Netezza & Teradata
Programming Languages: SQL & PL/SQL
Scheduling Tools: AUTOSYS & TIVOLI
Cloud Computing: AWS & Rack Space.
Operating Systems: Windows NT/2000/XP, MS-DOS, UNIX & Linux
Tools: & Utilities: PUTTY, ClearCase & VSS
Web Related: HTML, DHTML & Node JS
Domains: BFSI, Health Care, Manufacturing, Retail, Telecom and Transports
PROFESSIONAL EXPERIENCE:
Confidential, Renton, WA
Data Architect
Environment: (with skill versions) Languages: SQL & PL/SQL Database/ETL Tool Oracle 11g, SQL Server 2012, Teradata & Informatica PC 9.6Responsibilities:
- Responsible for complete plaining, Legacy data analysis and classify the key business drivers for this project initiative.
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Identifying Legacy source system data to be Data quality check once loaded in staging environment and load finally into Pega & ODS for data analytics.
- Developing data models, structures, and procedures to ensure the integrity and performance of database components
- Designed data cleansing rules/data scrubbing techniques to ensure consistency amongst data sets using UDMF tool.
- Responsibilities included designing Data integration architecture and loading strategy using Informatica.
- Preparing best practices to help developers design ETL components like Mappings, Mapplets, Sessions/Tasks, Worklets and Workflows.
Confidential,Richardson, Texas
Data Architect
Environment: (with skill versions) Languages: SQL & PL/SQL mc Database: 11g & NOSQL Cassandra 2.2Responsibilities:
- Responsible for defining the complete analysis and classify the key business drivers for this project initiative.
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Created standard abbreviation document for logical, physical and data models.
- Created logical, physical and dimensional data models (Data warehousing data modelling)
- Developing data models, structures, and procedures to ensure the integrity and performance of database components
- Created entity relationship diagrams and multidimensional data models, reports and diagrams based on the requirements.
- Designed data cleansing/data scrubbing techniques to ensure consistency amongst data sets.
- Responsibilities included designing architecture and developing complex Informatica mappings including Type-II slowly changing dimensions.
- Developing complex Mappings, Mapplets, Sessions/Tasks, Worklets and Workflows using Workflow Manager tools -Task Developer, Workflow & Worklet Designer.
- Collected performance data for sessions and performance tuned by adjusting Informatica session parameters.
- Created pre-session and post-session shell scripts and mail-notifications.
- Configuration and Deployment of Cassandra NoSQL Data Store on Confidential &T Server.
- Analyzing and resolving technical problems on Cassandra performance tuning and debugging.
- Involved on cluster management on Cassandra like add/bootstrap nodes, remove/ replace node to clusters.
- Responsible for technology design, integration, problem analysis, data migration to Cassandra NoSQL on Confidential &T Server .
- Hands on writing Cassandra CQL programs and cluster configurations and management activities.
- Implementation of Kafka topic with File Receiver, Kafka Receiver, and other customized receivers framework interfaces.
Confidential
Data Architect
Environment: (with skill versions) Languages: No SQL Database: SQL Server & No SQL Cassandra Database WITH unixResponsibilities:
- Responsible for defining the complete data model and classify the key business drivers for this POC initiative.
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Created standard abbreviation document for logical, physical and CQL data models.
- Created logical, physical and dimensional data models (Data warehousing data modelling)
- Documented logical, physical and dimensional data models (Data warehousing data modelling)
Confidential
ETL/Data Architect
Environment: (with skill versions) Languages: SQL, PL/SQL Database: Oracle, PostgreSQL & Teradata with Unix
Responsibilities:
- Responsible for defining the complete data warehouse architecture (i.e. Data modeling, ETL process, Data Marts, EDW) and classify the key business drivers for the data warehouse initiative.
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Documented logical, physical and dimensional data models (Data warehousing data modelling)
- Created entity relationship diagrams and multidimensional data models, reports and diagrams based on the requirements.
- Suggest steps for best possible use of processes for new components.
- Involved with designing and developing of Pl/PgSQL Stored Procedures, Data Types, Triggers, Constraints and Constraints.
- Worked on optimizing postgresql.conf to enhance the performance of database loading and retrieving activities.
- Responsible for defining the testing procedures, test plans, error handling strategy and performance tuning for mappings, Jobs and interfaces.
- Performing code reviews for ETL mappings from a performance & Error handling perspective.
- Responsible for developing PostgreSQL Programming, Developing database objects including tables, triggers, stored procedures and views.
- Optimize postgresql.conf for performance improvement.
- Setting up Scheduling mechanism and appropriate tool selections based on project requirements.
Confidential
ETL / Data Architect
Environment: (with skill versions) Languages: SQL, PL/SQL Database: Oracle & SQL Server with Windows XP & UnixResponsibilities:
- Responsible for defining the complete data warehouse architecture (i.e. Data modeling, ETL process, Data Marts, EDW) and classify the key business drivers for the data warehouse initiative.
- Created entity relationship diagrams and multidimensional data models, reports and diagrams based on the requirements.
- Responsible for ETL Software Installation, Validating and configuring server environments.
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Performing code reviews for ETL mappings from a performance & Error handling perspective.
- Configuring configuration management processes and setting up versioned and global repositories in Informatica.
- Managed change control implementations and coordinating weekly, monthly & quarterly release processes.
- Setting up Scheduling mechanism and appropriate tool selections based on project requirements.
Confidential
ETL Architect
Environment: (with skill versions) Languages: SQL, PL/SQL Database: DB2 with UnixResponsibilities:
- Responsible for defining the complete data warehouse architecture (i.e. Data modeling, ETL process, Data Marts, EDW) and classify the key business drivers for the data warehouse initiative.
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Created entity relationship diagrams and multidimensional data models, reports and diagrams based on the requirements.-
- Suggest steps for best possible use of processes for new components.
- Performing code reviews for ETL mappings from a performance & Error handling perspective.
Confidential
ETL Tech Lead
Environment: (with skill versions) Languages: SQL, PL/SQL Database: Oracle with UNIXResponsibilities:
- Responsible for ETL Software Installation, Validating and configuring server environments.
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Suggest steps for best possible use of processes for new components.
- Responsible for defining the testing procedures, test plans, error handling strategy and performance tuning for mappings, Jobs and interfaces.
- Performing code reviews for ETL mappings from a performance & Error handling perspective.
- Managed change control implementations and coordinating weekly, monthly & quarterly release processes.
- Setting up Scheduling mechanism and appropriate tool selections based on project requirements.
Confidential
Team Lead
Environment: (with skill versions) Languages: SQL & PL/SQL Database: Oracle 9i with Sun Solaris 2.6Responsibilities:
- Feasibility study of business requirements and assisting the team members in developing the ETL components, checking their dependencies and load factors.
- Responsible for defining the design procedures, error handling strategy and performance tuning for mappings, Jobs and interfaces.
- Performing code reviews for ETL mappings from a performance & Error handling perspective.
- Involved in Development, Implementation and Support of ETL processes up to 3 Terra Byte applications.
- Configuring configuration management processes and setting up versioned and global repositories in Informatica & BODS
Confidential
ETL Consultant
Environment: (with skill versions) Languages: SQL & PL/SQL Database: Oracle with UNIX
Responsibilities:
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Suggest steps for best possible use of processes for new components.
- Configuring configuration management processes and setting up versioned and global repositories in Informatica.
- Involved with error handling & performance tuning strategy for informatica ETL process