Sr. Data Architect/ Data Modeler Resume
Atlanta, GA
SUMMARY
- Around 9+ years of Experience in Data Analysis, Data Modeling, Data Architect, Data Warehouse & Business intelligence professional with applied information Technology.
- Extensive experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like Erwin and ER Studio.
- Experienced performing structural modifications using Map - Reduce, analyzing data using Hive and visualizing in dashboards using Tableau and d3.js
- Experienced in Data Management solution that covers DWH/Data Architecture design, Data Governance Implementation and Big Data.
- Very good exposure to cloud and big data technologies like Azure, AWS, Hadoop, Hive, HDFS, NoSQL Database (MongoDB and Cassandra) and Sqoop.
- Experienced in all phases of the software development life cycle (SDLC), from requirements definition through implementation and supported my models in transformation and analysis phase
- Experienced in Data Modeling using Dimensional Data Modeling, Star Schema modeling, Fact and Dimensions tables including Physical, Logical data modeling and experienced with Business Process Modeling, Process Flow Modeling & Data flow modeling.
- Experienced with Data Conversion, Data Quality and Data Profiling, Performance Tuning and System Testing and implementing RDBMS features.
- Expertise in writing Complex SQLqueries to perform end-to-end ETLvalidations and support Ad-hoc business requests and expertise in implementing Security models for Dashboards, Row level, object Level, Role Level, and Dashboard Level.
- Expert in creating and developing Power BI Dashboards in to rich look.
- Excellent experience in Normalization (1NF, 2NF, 3NF and BCNF) and De-normalization techniques for effective and optimum performance in OLTP and OLAP environments.
- Expertise in OLTP/OLAP System Study, Analysis and E-R modeling, developing Database Schemaslike Star schema and Snowflakeschema used in relational, dimensional and multidimensional modeling.
- Excellent experience in Extract , Transfer and Load process using ETL tools like Data Stage , Informatica , Data Integrator and SSIS for Data migration and Data Warehousing projects.
- Experienced in the development of Data Warehouse, Business Intelligence architecture that involves dataintegration and the conversion of data from multiple sources and platforms.
- Experienced in writing, testing and implementing SQL/T-SQL/PL/SQL, Stored Procedures, Trigger and Functions and excellent experience in trouble shooting SQL queries, ETL jobs, data warehouse/data mart/data store models.
- Excellent knowledge with Talend MDM, Informatica MDM for creating matching and merge rules for MDMhub.
- Experienced in using Excel and MSAccess to dump the data and analyze based on business needs.
TECHNICAL SKILLS
Analysis and Modeling Tools: Erwin 9.x, 8.x 7.x, Sybase Power Designer, Oracle Designer, Rational Rose, ER/Studio, MS Visio
ETL Tools: Informatica Power Center 9.5/8.6.1/8.1 , Data Stage, Ab Initio, Talend
Cloud and BigData Tools: AWS S3, AWS Redshift, Azure Data Factory, Azure SQL, AWS EMR, Hadoop Framework, HiveQL, HDFS, Cassandra, and MongoDB
Languages: Python, SQL, PL/SQL, T-SQL, XML, HTML, UNIX Shell Scripting.
Databases: MS SQL Server 2016/2014/2010/2008 , Oracle12c/11g/10g/9i, DB2 V9.x, Teradata 15/14/V2R6.x and MS Access
Operating Systems: Windows, iOS, UNIX, Linux
Project Execution Methodologies: Ralph Kimball and Bill Inmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
Tools: & Software: TOAD, MS Office, BTEQ, Teradata SQL Assistant, Power BI.
Tools: MS-Office suite (Word, Excel, MS Project and Outlook), VSS
Packages: Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point Portal Server
PROFESSIONAL EXPERIENCE
Confidential, Atlanta GA
Sr. Data Architect/ Data Modeler
Responsibilities:
- Understand and analyze business data requirements and architect an accurate, extensible, flexible and logical data model andDefining and implementing conceptual, logical, and physical data modeling concepts.
- Defining Data Sources and data models, documenting actual data flows, data exchanges, and systems interconnections and interfaces. Ensuring these is aligned with the enterprise data model.
- Develop and optimize ETL processes by working closely with multiple data partners and stakeholders across the company to meet growing business needs.
- Design and build world class high-volume real-time data ingestion frameworks and automate various data sources into Bigdata technologies like Hadoop etc.
- Worked on Azure and Azure data lake and architecting a solution to load data, create data models also created Logical and Physical Enterprise Warehouse Data Model (Star Schema - Facts & Dimensions) for Azure SQL.
- Performed Data mapping between source systems to Target systems, logical data modeling, created class diagrams and ERdiagrams and used SQLqueries to filter data.
- Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
- Develop Python, Pyspark, HIVE scripts to filter/map/aggregate data and used scoop to transfer data to and from Hadoop.
- Develop and keep current, a high-level data strategy that fits with the Data Warehouse Standards and the overall strategy of the Company and identify and predict technology needs related to data and reporting for the organization, and propose technology solutions.
- Extensively used Erwin for developing data model using star schema methodologies and designed different type of STAR schemas using ERWIN with various Dimensions like time, services, customers and FACTtables.
- Analyze database infrastructure to insure compliance with customer security standards, database performance considerations, and reverse engineering of existing database environments.
- Used Hive and created Hive tables and involved in dataloading and writing HiveUDFs and worked on Hive for exposing data for further analysis and for generating transforming files from different analytical formats to text files.
- Designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using Erwin.
- Creation of BTEQ, Fast export, MultiLoad, TPump, Fast load scripts for extracting data from various production systems.
- Creation of database objects like tables, views, Materialized views, procedures, packages using Oracle tools like PL/SQL, SQL* Plus, SQL*Loader and Handled Exceptions.
- Extensively involved in creating Azure Analysis Cube (Tabular Model), for building an enterprise semantic data model which caters to self-service business intelligence for all the end market users.
- Worked on importing and exporting data from Oracle and DB2 into HDFS using Sqoop
- Created, optimized, reviewed and executed Teradata SQL test queries to validate transformation rules used in source to target mappings/source views, and to verify data in target tables.
- Used Pig as ETL tool to do transformations, event joins and some pre-aggregations before storing the data onto HDFS and provided ad-hoc queries and data metrics to the Business Users using Hive, Pig.
Environment: ERWIN 9.6/7, Informatica Power Mart (Source Analyzer, Data warehousing designer, Mapping Designer, Transformations), MS SQL Server 2016, Oracle 12c, SQL, Hive, Map Reduce, PIG, Sqoop, HDFS, Hadoop, Teradata, Netezza, PL/SQL, Informatica, SSIS, SSRS, Python, AzureSQL, Azure Data Lake Analytics (ADLA), Azure Data Flow (ADF), Azure Data Factory.
Confidential, Chicago IL
Sr. Data Modeler/Data Architect
Responsibilities:
- Understand end-to-end scope of work and design architectural solution accordingly and converting business requirement into technical requirements.
- Involved in designing logical, physical and dimensional database designs using reverse and forward engineering of ERWIN.
- Developed automated data pipelines from various external data sources (web pages, API etc) to internal Data Warehouse (SQL server, AWS), then export to reporting tools like Datorama by Python.
- Designed and implemented Hive and Pig UDF's for evaluation, filtering, loading and storing of data and imported and exporting data into HDFS and Hive using Sqoop
- Define real-time and batch data ingestion architecture using AWS Lambda approach, including Kakfa, Storm and Hbase for real-time as well as Sqoop and Hive for batch layer.
- Updated Python scripts to match training data with our database stored in AWS Cloud Search, so that we would be able to assign each document a response label for further classification.
- Worked on Amazon Redshift and architecting a solution to load data, create data models and involved in Data Profiling, Mapping and Integration from multiple sources to AWS S3.
- Architecting the complete data model for the application (OLTP) and reporting (OLAP) and used ERWIN's Database Generation for generating DDL, stored procedure and trigger code for your target database
- Provided Data Architecture solutions for multiple relational and dimensional models and involved in Data Warehouse Support - StarSchema and Dimensional modeling to help design DataMart and data warehouse.
- Designed both 3NF data models for ODS, OLTP systems and Dimensional Data Models using Star and Snow flake Schemas.
- Developing MapReduce using java for transforming raw log files (semi-structured data) into structured format for further analysis using HiveQL from the web based analysis tools.
- Developed Data Mapping, Data profiling, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Forward engineering the data model and creating the development Database modify and maintain all changes to the Data model and Database scripts.
- Managed definition and execution of data mapping , conversion and reconciliation processes, for data originating from a plethora of enterprise and SAP , leading into to ongoing data governance organization design.
- Define AWS architecture for implementing a completely cloud-based big data solution using EMR, S3, Lambda and Redshift.
- Identified and Analysis of various facts from the source system and business requirements to be used for the data warehouse (Kimball Approach).
- Used various sources to pull data into Power BI such as SQL Server, SAP BW, Oracle, SQL Azure etc.
- Work in team using ETL tool Informatica to populate the database, data transformation from the old database to the new database using Oracle and SQL Server.
- Executed Hive queries on Parquet tables stored in Hive to perform data analysis to meet the business requirements.
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata and resolved the data type inconsistencies between the source systems and the target system using the mapping documents and analyzing the database using SQL queries.
- Managing Development team, Code review, and Unit test case review and helping team with solutions for all technical difficulties (ex- SQL/DB Performance tuning).
- Developed several behavioral reports and data points creating complex SQL queries and stored procedures using SSRS and Excel and developed different kind of reports such as Drill down, Drill through, Sub Reports, Charts, Matrix reports, Parameterized reports and Linked reports using SSRS.
Environment: ERWIN 9.5, Informatica, Hadoop, Hive, HDFS, Sqoop, AWS Lambda, Cassandra, Power BI, MongoDB, JIRA, AWS S3, AWS Redshift, SQL, SSRS, Excel, Teradata, Oracle, MS SQL Server 2014, PL/SQL, Map Reduce, Python and Tableau.
Confidential
Sr. Data Modeler/Data Architect
Responsibilities:
- Involved in architecting system interfaces, understand interface requirement and design data model (Logical and Physical) using ER Studio and Deliverables are PDM, DDL scripts and STTM documents.
- Performed Reverse Engineering of the legacy application using DDL scripts in ER Studio, and developed Logical and Physical data models for Central Model consolidation.
- Complex Semantic Model is created using Azure Analysis Services Cubes over Azure SQL DW for design and developing Cubes.
- Defining data governance process: processes for data quality rules definition, review process, communication plan, templates, etc.
- Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
- Involved in performing extensive Back-End Testing by writing SQL queries and PL/SQL stored procedures to extract the data from SQL Database using Hadoop
- Strong ability in developing advanced ANSI SQL queries to extract, manipulate, and/or calculate information to fulfill data and reporting requirements including identifying the tables and columns from which data is extracted.
- Closely worked with BI team to write DAX queries for building the Cube in Azure Analysis Services.
- Analyzed, Designed, and Developed OBIEE Metadata repository (RPD) that consists of Physical Layer, Business Mapping and Model Layer and Presentation Layer.
- Responsible for all metadata relating to the EDW's overall data architecture, descriptions of data objects, access methods and security requirements and developed and automated multiple departmental Reports using Tableau and MS Excel.
- Review & analyzing SQL Server databases in terms of finding performance bottlenecks & prepare optimization strategies as well.
- Worked on all types of transformations that are available in Power BI query editor and wrote calculated columns, Measures query’s in Power BI desktop to show good data analysis techniques.
- Design and document the complete ETL specification document detailing the queries involved for extraction, transformation logic and loading methods.
- Used Azure SQL reporting services to create reports with tables, charts and maps.
- Create logical and physical data mo dels using best practices to ensure high data quality and reduced redundancy, Define, document, and articulate design goals and standards.
- Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data warehouse, data mart reporting system in accordance with requirements.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, flatfiles, SQLServer2010 with high volume data.
- Identifying functional and technical gaps and writing complex SQL's for generating reports and done performance Tuning of SQL and Database server with huge data volume across all the clients.
- Generated and DDL (Data Definition Language) scripts using ER Studio and assisted DBA in Physical Implementation of data Models.
Environment: ER Studio, MS Visio, Oracle 11g, Oracle Designer, CRM, Hadoop, Power BI, Teradata, SQL Server 2010, Oracle, SQL, PL/SQL,Hive, Map Reduce, JIRA, Windows7, ERP, and UNIX, Azure, AzureSQL and NoSQL Data Base, and SSIS ETL Tool.
Confidential
Sr. Data Modeler/Analyst
Responsibilities:
- Worked with Business users for requirements gathering, business analysis and project coordination and understood and translate business needs into data models supporting underwriting workstation services.
- Worked with the Application Development team to implement data strategies, build data flows and develop data models and designed and developed Use Cases, Activity Diagrams, Sequence Diagrams, OOD (Object oriented Design) using UML and Visio.
- Transformed Logical Data Model to Physical Data Model ensuring the Primary Key and Foreign key relationships in PDM, Consistency of definitions of Data Attributes and Primary Index considerations.
- Involved in Teradata utilities (BTEQ, Fast Load, Fast Export, Multiload, and Tpump) in both Windows and Mainframe platforms.
- Involved with Full Data warehouse Lifecycle Implementation upgrading the existing Legacy Data Warehouse to Enterprise Data warehouse using the Kimball’s Four Fixes approach by conforming the Non-conformed Dimensions, creating surrogate keys, delivering the atomic details and reducing redundancies and also designing from the scratch.
- Involved in the entire data Migration process from analyzing the existing data, cleansing, validating, translating tables, converting and subsequent upload into new platform.
- Generated and DDL (Data Definition Language) scripts using ER Studio and assisted DBA in Physical Implementation of data Models.
- Involved in writing T-SQL, working on SSIS, SSRS, SSAS, Data Cleansing, Data Scrubbing and Data Migration.
- Developed Conceptual, Logical and Physical data models for central model consolidation and used Normalization (1NF, 2NF & 3NF) and de-normalization techniques for effective performance in OLTP and OLAP systems.
- Generated ad hoc reports in Excel Power Pivot and shared them using Power BI to the decision makers for strategic planning and involved in developing Power BI reports and dashboards from multiple data sources using data blending.
- Worked on Performance Tuning of the database which includes indexes, optimizing SQL Statements and conducted data modeling JAD sessions and communicated data-related standards.
- Developed SQL Queries to fetch complex data from different tables in remote databases using joins, database links and Bulk collects.
- Used SSRS for generating Reports from Databases and Generated Sub-Reports, Drill down reports, Drill through reports and parameterized reports using SSRS.
- Used Model Mart of ER Studio for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Implemented Forward engineering to create tables, views and SQL scripts and mapping documents and worked on PL/SQL programming Stored Procedures, Functions, Packages and Triggers.
- Wrote DDL and DML statements for creating, altering tables and converting characters into numeric values.
- Involved in development and implementation of SSIS, SSRS and SSAS application solutions for various business units across the organization.
Environment: ER Studio, OLTP, Power BI, ETL, JIRA, Sql, PL/SQL, DB2, SSIS, SSRS, Queries, T-SQL, Oracle, Alteryx, Teradata, SSAS, DB2, Excel, Netezza and SSAS
Confidential
Sr. Data Analyst / Data Modeler
Responsibilities:
- Participated in all phases of project including Requirement gathering, Analysis, Design, Coding, Testing and Documentation.
- Designed ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and converted them to physical data model including capacity planning, object creation and aggregation strategies, partition strategies, Purging strategies as per business requirements.
- Developed Data models and ERD diagrams using Erwin.
- Created and Designed logical database schema layout structure to prepare for ETL processes, data mining, extraction, Analysis and Reporting System.
- Data analysis of existing data base and new sources to understand the data flow and business rule applied to different data bases by SQL
- Involved in Logical & Physical Database Layout Design using Erwin and Created and maintained Logical Data Model (LDM) / Physical Data Modeling. Included documentation of all entities, attributes, data relationships, primary and foreign key relationships, allowed values, codes, business rules, glossary terms, etc.
- Conducted meetings with the business and technical team to gather necessary analytical data requirements in JAD sessions and wrote requirements for ETL Data warehousing, database profiling.
- Developed the Data mart for the base data in Star Schema, Snow Flake Schema and Multi Star Schema and involved in developing the Data warehouse for the Database.
- Performed reverse Engineering on existing Data model and data mapping, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data within the Oracle database.
Environment: Erwin 7.x, SQL Server, Informatica Power Center, Oracle, Source Analyzer, Data warehouse designer, Workflow Manager, Workflow Monitor, PL/SQL.
Confidential
Data Analyst
Responsibilities:
- Created and maintained Database Objects (Tables, Views, Indexes, Partitions, Synonyms, Database triggers, Stored Procedures) in the data model.
- Actively involved in creating Physical and Logical models using Erwin.
- Presented the data scenarios via, Erwin logical models and excel mockups to visualize the database.
- Worked on building the data model using Erwin as per the requirements, discussion and approval of the model from the BA.
- Involved with Data Analysis primarily Identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats
- Designed and Developed logical & physical data models and Meta Data to support the requirements using Erwin
- Involved with Data Profiling activities for new sources before creating new subject areas in warehouse
- Experienced in creating UNIX scripts for file transfer and file manipulation.
- Developed ER and Dimensional Models using Power Designer advanced features. Created physical and logical data models using Erwin.
- Wrote PL/SQL statement and stored procedures in Oracle for extracting as well as writing data.
Environment: DB2, CA Erwin, Oracle 11g, MS-Office, SQL Architect, TOAD Benchmark Factory, SQL Loader, PL/SQL, SharePoint, Erwin, MS-Office, SQL Server 2008.