We provide IT Staff Augmentation Services!

Sr. Data Modeler Resume

4.00/5 (Submit Your Rating)

Houston, TX

SUMMARY

  • 9+ years of industry experience with solid understanding of Data Modeling, Data Analysis, architecture, Evaluating Data Sources, and strong understanding of Data Warehouse/Data Mart Design, Data Lake, ETL, BI, Data visualization, OLAP, OLTP and Client/Server applications.
  • Facilitated several organizations through their digital transformation journey by helping them migrate from on prem to Cloud solutions using MS Azure Stack and AWS.
  • Experience driving cross - functional analytics projects from beginning to end question formation, data model design, exploratory data analysis (EDA), validation, analysis, visualization, and presentation.
  • Solid understanding of Data Governance, Meta Data, Data Management, and control; Adept in Data warehouse and Data mart architecture.
  • Strong Experience in ER & Dimensional Data Modeling to deliver normalized ER & Star/Snowflake schemas using Erwin, ER Studio, EA Sybase power designer, SQL Server Enterprise manager and Oracle designer.
  • Proven skills in designed and maintained the Detail Design Document (DDD), Business requirement documents (BRD), Data Requirement Document (DRD), Data Flow Diagram (DFD), Data Management Plan Document, Data Dictionary, Meta Data Model, Logical and Physical Data Models, Full DDL, Alter DDL, Insert statement for all the applications.
  • Hands on experience in Normalization and De-Normalization techniques for optimum performance in relational and dimensional database environments and experience with modeling using Erwin in both forward and reverse engineering processes.
  • Extensive knowledge of Big-data, Hadoop, Hive, Sqoop, HDFS, NoSQL Databases such as MongoDB and Cassandra and other emerging technologies.
  • Sound knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Worked with Amazon Web Services (AWS) for a multitude of applications utilizing the Amazon Web Services focusing on high-availability, fault tolerance and auto-scaling and good experience and knowledge on AWS Redshift, RDS, AWS S3 and AWS Glue.
  • Excellent SQL programming skills and developed Stored Procedures, Triggers, Functions, Packages using SQL, PL/SQL.
  • Experience in Data transformation and Data mapping from source to target database schemas and also data cleansing.
  • Extensive experience in Text Analytics, generating data visualizations using R, Python and creating dashboards using tools like Tableau.
  • Expert in building Enterprise Data Warehouse or Data warehouse appliances from Scratch using both Kimball and Inmon Approach.
  • Extensive experience on usage of ETL & Reporting tools like Datastage, Informatica SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS), Tableau, Power BI and Microstrategy.
  • Proficient in Software Development Life Cycle (SDLC), Project Management methodologies, and Microsoft SQL Server database management and working with Agile and Waterfall data modeling methodologies.
  • Extensive experience in Data Visualization including producing tables, graphs, listings using various procedures and tools such as Tableau and PowerBI.
  • Strong background in data processing, data analysis with hands on experience in MS Excel, MS Access, Unix and Windows Servers.

TECHNICAL SKILLS

Data Modeling Tools: ERwin, ER Studio, and Power Designer.

Cloud Platform: AWS (S3 Bucket, RDS and Amazon Redshift) and MS Azure (Azure SQL, Azure DW, Azure ADF, Storage Blob, Databricks and Synapse)

Reporting Tools: SSRS, Power BI, Tableau and Microstrategy

Databases: Oracle 12c/11g/10g, Teradata R15/R14, MS SQL Server 2016/2014, DB2, MongoDB and Cassandra.

Programming Languages: SQL, T-SQL, PL/SQL, R and Python.

Operating System: Windows, Unix, Sun Solaris.

Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Agile, Waterfall Model.

BigData Tools: Hadoop, Hive, HDFS, Sqoop, and Spark

PROFESSIONAL EXPERIENCE

Confidential - Houston TX

Sr. Data Modeler

Responsibilities:

  • Interacted with technical and non-technical business user\'s to id business requirements and to translate into data requirements to build data model design.
  • Involved in logical and physical designs and transforming logical models into physical implementations and Created Entity/Relationship Diagrams, grouped and created the tables, validated the data, identified PKs for lookup tables.
  • Worked on Software Development Life Cycle (SDLC), testing methodologies, resource management and scheduling of tasks and used Agile Method for daily scrum to discuss the project related information.
  • Utilized Erwin's forward/reverse engineering tools and target database schema conversion process and designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using Erwin and created DDL scripts using Erwin and source to target mappings to bring the data from source to the warehouse.
  • Involved in Data Migration using SQL, SQL Azure, Azure Storage, and Azure Data Factory, SSIS, and PowerShell and created processes to load data from Azure Storage blob to Azure Synapse, to load from web API to Azure SQL and scheduled web jobs for daily loads.
  • Involved in collaborating with ETL teams to source data, perform data analysis to identify gaps and involved in loading the data from Source Tables to Operational Data Source tables using Transformation and Cleansing Logic.
  • Performed data transformations using PySpark and SparkSQL within Azure databricks notebook and setup loads into target.
  • Defined Validation rules in MDM system by analyzing excel sheet part master data and input from the business users and worked with MDM systems team with respect to technical aspects and generating reports.
  • Setup Delta Lake with Azure databricks to write only the incremental delta records into the refined folder of the Data Lake to be consume for the Azure Cognitive Search Canonical Model.
  • Clean data and processed third party spending data into maneuverable deliverables within specific format with Excel macros and python libraries such as NumPy, SQL Alchemy and matplotlib.
  • Developed automated data pipelines from various external data sources (web pages, API etc.) and used Spark and SparkSQL for data integrations, manipulations.
  • Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL access on Hadoop data and Created tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Performed data analysis and data profiling using complex SQL queries on various sources systems including Oracle and monitored the Data quality and integrity of data was maintained to ensure effective functioning of department.
  • Generated detailed report after validating the graphs using Python and adjusting the variables to fit the model and worked on Tableau for insight reporting and data visualization and presented the Dashboard to Business users and cross functional teams, define KPIs (Key Performance Indicators), and identify data sources.

Environment: Agile, Erwin 9.7, Python, Azure SQL, Azure DW, StorageBlob, pySpark, Synapse SQL, Teradata, Informatica, Tableau, Hive, HDFS, Sqoop, MDM, Spark, Oracle 12c, PL/SQL, SQL Server, SSIS and SSRS.

Confidential - Boston MA

Sr. Data Modeler / Data Architect

Responsibilities:

  • Participated in the design, development, and support of the corporate operation data store and enterprise data warehouse database environment and Building relationships and trust with key stakeholders to support program delivery and adoption of enterprise architecture.
  • Developed, managed, and validated existingDataModels including Logical and Physical Models of theDataWarehouse and source systems utilizing a 3NF model.
  • Worked with data ingestions from multiple sources into the Azure SQL data warehouse, developing purging scripts and routines to purge data on Azure SQL Server and Azure Blob storage.
  • Participated in big data architecture for both batch and real-time analytics and mapped data using scoring system over large data on HDFS and Worked on analyzingHadoopcluster and different bigdataanalytic tools including HBase database and Sqoop.
  • Implemented dimension model (logical and physical data modeling) in the existing architecture using ER/Studio.
  • Worked on predictive and what-if analysis using Python from HDFS and successfully loaded files to HDFS from Teradata and loaded from HDFS to HIVE.
  • Gathering the business requirements from customers and creatingdatamodels for different branches using MS access andER/Studio.
  • Analyzed escalated incidences within theAzure SQL database and involved in SQL andAzure SQL DWcode development usingT-SQL.
  • Cleansed thedataby eliminating duplicate and inaccuratedatain Python and usedPythonscripts to update the content in database and manipulate files.
  • Executed ad-hoc data analysis for customer insights using SQL using Amazon AWS Hadoop Cluster.
  • Worked onNormalizationandDe-Normalizationtechniques for both OLTP and OLAP systems.
  • Utilized Azure Data Factory for transforming and moving data from virtual machine to Data Factory, BLOB storage, and SQL Server.
  • Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata Database analysis, development, testing, implementation, and deployment.
  • DevelopedMDM integrationplan and hub architecture for customers, products and vendors, DesignedMDM solutionfor three domains.
  • Developed requirements, perform data collection, cleansing, transformation, and loading to populate facts and dimensions for data warehouse and Built database Model, Views and API’s usingPythonfor interactive web-based solutions.
  • Maintaining data mapping documents, business matrix and other data design artifacts that define technical data specifications and transformation rules
  • Developed ETL mappings, testing, correction and enhancement and resolveddataintegrity issues and coordinated multiple OLAP and ETL projects for variousdatalineage and reconciliation.
  • Worked in the capacity of ETL Developer (Oracle Data Integrator (ODI) / PL/SQL) to migrate data from different sources in to target Oracle Data Warehouse.
  • Worked with high volume datasets from various sources like SQL Server 2012, Oracle, DB2, and Text Files and involved in cleaning of largedatasets usingpython and created named sets, calculated member and designed scope in SSAS, SSIS, and SSRS.
  • MigratedSQLserver2008 toSQLServer2014 in Microsoft WindowsServer2003 and troubleshooting high availability scenarios involving Clustering, Database Mirroring, Log Shipping and Replication.
  • Written SQL scripts to test the mappings and Developed Traceability Matrix of Business Requirements mapped to Test Scripts to ensure any Change Control in requirements leads to test case update.
  • Involved in extensive Data validation by writing several complex SQL queries and involved in back-end testing and worked withdataquality issues.

Environment: ER/Studio, SSIS, SSRS, SAS, Azure SQL, Azure DW, Storage Blob, AD, Excel, MDM, PL/SQL, ETL, Python, Tableau, Hadoop, Hive, Pig, Aginity, Teradata SQL Assistant, Cassandra, PL/SQL, T-SQL, Cognos, DB2, Oracle11g, SQL, Teradata14.1, Informatica Power Center9.6, HBase.

Confidential -Austin, TX.

Sr. Data Modeler/Data Architect

Responsibilities:

  • Involved in Data Modeling role to review business requirement and compose source to target data mapping documents and involved in relational and dimensional Data Modeling for creating Logical and Physical design of the database and ER diagrams using data modeling tool like Erwin.
  • Involved in story-driven Agile development methodology and actively participated in daily scrum meetings and used methodology as the organization Standard to implement the data Models.
  • Created data models for AWS Redshift and Hive from dimensional data models and worked on Data modeling, Advanced SQL with Columnar Databases using AWS and driven the technical design of AWS solutions by working with customers to understand their needs.
  • Worked on development of data warehouse, Data Lake and ETL systems using relational and non-relational tools like SQL, No SQL.
  • Generated various dashboards as per the requirements, which were used by management to make key business decisions and developed and maintained data dictionary to create metadata reports for technical and business purpose and write Python scripts to parse JSON documents and load the data in database.
  • Involved in requirement gathering and database design and implementation of star-schema, snowflake schema/dimensional data warehouse using ERwin.
  • Involved in extensive Data validation by writing several complex SQL queries and involved in back-end testing and worked with data quality issues.
  • Migrated data from existing product into Informatica MDM hub and involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
  • Generated comprehensive analytical reports by running SQL queries against current databases to conduct Data Analysis and improved performance on SQL queries used Explain plan / hints /indexes for tuning created DDL scripts for database.
  • Created PL/SQL Procedures and Triggers. Worked on Normalization and De-Normalization techniques for both OLTP and OLAP systems and worked with data investigation, discovery and mapping tools to scan every single data record from many sources and performed Data mapping between source systems to Target systems, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data.
  • Participated in Data Acquisition with Data Engineer team to extract historical and real-time data by using Hadoop MapReduce and HDFS and developed Hive queries for analysis and exported the result set from Hive to MySQL using Sqoop after processing the data.
  • Designed Datastage ETL jobs for extracting data from heterogeneous source system, transform, and finally load into the Data Marts.
  • Created and published multiple dashboards and reports using Tableau server, Power BI and Microstrategy.

Environment: ERwin, Microsoft SQL Server, AWS S3, Redshift, RDS, MySQL, SAS, MDM, HDFS, HBase, Hive QL Queries, Sqoop, OLTP, OLAP, Metadata, MS Excel, QlikView, Tableau, Power BI and Microstrategy, SQL, T-SQL, Python, Spark and PL/SQL.

Confidential -Columbus, GA.

Sr. Data Modeler/ Data Analyst/ Data Warehousing

Responsibilities:

  • Understand the business process; gather business requirements; determine impact analysis based onERP and Created logical physicaldatamodels and MetaDatato support the requirements Analyzed requirements to develop design concept and technical approaches to find the business requirements by verifying Manual Reports.
  • Involved in fixing invalid mappings, testing of Stored Procedures and Functions, Unit and Integrating testing of Informatica Sessions, Batches and the Target Data.
  • Designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snowflake Schema.
  • Worked on data integration and workflow application on SSIS platform and responsible for testing all new and existing ETL data warehouse components.
  • Reverse Engineered the Data Models and identified the Data Elements in the source systems and adding new Data Elements to the existing data modelsand used SQL for querying the database in UNIX environment.
  • Involved in all the steps and scope of the project dataapproach to MDM and CreatedData Dictionary and Data Mapping from Sources to the Target in MDMDataModel.
  • End to End process involvement from gathering client business requirements, developing the dashboard inTableauand publishing the dashboard into server.
  • Developed Data Migration and Cleansing rules for the Integration Architecture (OLTP, ODS, DW)and performed Business Area Analysis and logical and physical data modeling for a Data Warehouse utilizing the Bill Inmon Methodology and also designed Data Mart application utilizing the Star Schema Dimensional Ralph Kimball methodology.
  • Extracted/Transformed/Loaded (ETL) design and implementation in areas related to Teradata utilities such as Fast Export and MLOAD for handling numerous tasks.
  • Implement functional requirements using Base/SAS,SAS/Macros,SAS/QL, UNIX, Oracle and CodingSASprograms with the use of BaseSASandSAS/Macros for Adhoc jobs requested by Users and DB2 and Upgrading the SQL Server Databases, Monitoring and Performances tuning and developed reports using Crystal Reports with T-SQL, MS Excel and Access.
  • Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration
  • MigratedSQLServer2005 databases toSQLServer2008R, 2008R2 databases also migrated to IBM DB2.
  • Worked on multiple Data Marts in Enterprise Data Warehouse Project (EDW) and involved in designing OLAP data models extensively used slowly changing dimensions (SCD).
  • Worked on all activities related to the development, implementation, administration and support of ETL processes for large-scale Data Warehouses using SQL Server SSIS.
  • Developed automated procedures to produce data files using Microsoft Integration Services (SSIS) and performeddataanalysis anddataprofiling using complex SQL on various sources systems including Oracle and Netezza
  • Developed ER and Dimensional Models using ER Studio advanced features and created physical and logicaldatamodels using ER Studio.
  • Used SQL Profiler for monitoring and troubleshooting performance issues in T-SQL code and stored procedures.
  • ImplementedAgileMethodology for building an internal application.
  • Extracted data from databases Oracle, Teradata, Netezza, SQL server and DB2 using Informatica to load it into a single repository for data analysis and used SQL on a wide scale for analysis, performancetuning and testing.

Environment: ER Studio, SQL Server 2012, SQL Server Analysis Services 2008, SSIS, SSRS 2008, Oracle 10g, Business Objects XI, Rational Rose,Tableau, ERP, Netezza, Teradata, Excel, Informatica MDM, Pivot tables, DB2, Datastage, MS Office, MS Visio, SQL, Rational Rose, T-SQL, UNIX, Agile, SAS, MDM, Shell Scripting, Crystal Reports 9.

Confidential -Dallas TX

Sr. Data Analyst/Data Modeler

Responsibilities:

  • Worked as a Data Analyst/Modeler to generate Data Models using SAP PowerDesigner and developed relational database system.
  • Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
  • Conducted user interviews, gathering requirements, analyzing the requirements using Rational Rose, Requisite pro RUP.
  • Developed logical data models and physical database design and generated database schemas using SAP PowerDesigner.
  • Analyzed the business requirements of the project by studying the Business Requirement Specification document.
  • Extensively worked on Data Modeling tools SAP PowerDesigner Data Modeler to design the data models.
  • Created ER diagrams using Power Designer modeling tool for the relational and dimensional data modeling.
  • Involved in data mapping document from source to target and the data quality assessments for the source data.
  • Responsible for data profiling and data quality checks to suffice the report requirements
  • Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
  • Developed and maintained data dictionary to create metadata reports for technical and business purpose.
  • Created SQL tables with referential integrity and developed queries using SQL, SQL*PLUS and PL/SQL.
  • Designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snowflake Schemas.
  • Designed the Database Tables & Created Table and Column Level Constraints using the suggested naming conventions for constraint keys.
  • Reversed Engineered the existing database structure to understand the existing data models so that any changes in corporate would synchronize with current model.
  • Involved in Normalization / De-normalization, Normal Form and database design methodology.
  • Conducted JAD Sessions with the SME, stakeholders and other management teams in the finalization of the User Requirement Documentation.
  • Wrote T-SQL statements for retrieval of data and Involved in performance tuning of T-SQL queries and Stored Procedures.
  • Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
  • Designed and Developed Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing.
  • Handled performance requirements for databases in OLTP and OLAP models and used excel sheet, flat files, CSV files to generated Tableau ad-hoc reports
  • Facilitated in developing testing procedures, test cases and User Acceptance Testing (UAT).
  • Involved in Data profiling and performed Data Analysis based on the requirements, which helped in catching many Sourcing Issues upfront.
  • Developed Data mapping, Data Governance, Transformation and Cleansing rules for the Data Management involving OLTP, ODS and OLAP.

Environment: SAP, Power Designer 16.6, OLTP, OLAP, T-SQL, SSIS, SQL Server, SQL, PL/SQL, Rational Rose, ODS

Confidential - Cleveland OH

Data Analyst

Responsibilities:

  • Worked as Data Analyst for requirements gathering, business analysis and project coordination.
  • Responsible for the analysis of business requirements and design implementation of the business solution.
  • Performed Data Analysis and Data validation by writing SQL queries using SQL assistant.
  • Translated business concepts into XML vocabularies by designing XML Schemas with UML
  • Gathered business requirements through interviews, surveys with users and Business analysts.
  • Worked on Data Mining and data validation to ensure the accuracy of the data between the warehouse and source systems.
  • Developed SQL Queries to fetch complex data from different tables in databases using joins, database links.
  • Involved in designing and developing SQL server objects such as Tables, Views, Indexes (Clustered and Non-Clustered), Stored Procedures and Functions in Transact-SQL.
  • Participated in JAD sessions, gathered information from Business Analysts, end users and other stakeholders to determine the requirements.
  • Performed Data analysis of existing data base to understand the data flow and business rules applied to Different data bases using SQL.
  • Performed data analysis and data profiling using complex SQL on various sources systems and answered complex business questions by providing data to business users.
  • Performed the detail data analysis, Identify the key facts and dimensions necessary to support the business requirements.
  • Generated Data dictionary reports for publishing on the internal site and giving access to different users.
  • Used MS Visio and Rational Rose to represent system under development in a graphical form by defining use case diagrams, activity and workflow diagrams.
  • Wrote a complex SQL, PL/SQL, Procedures, Functions, and Packages to validate data and testing process.
  • Worked in generating and documenting Metadata while designing OLTP and OLAP systems environment.
  • Worked in data management performing data analysis, gap analysis, and data mapping.
  • Established a business analysis methodology around the RUP (Rational Unified Process).
  • Developed stored procedures in SQL Server to standardize DML transactions such as insert, update and delete from the database.
  • Created SSIS package to load data from Flat files, Excel and Access to SQL server using connection manager.
  • Develop all the required stored procedures, user defined functions and triggers using T-SQL and SQL.
  • Produced report using SQL Server Reporting Services (SSRS) and creating various types of reports.

Environment: SQL, SQL server, PL/SQL, MS Visio, Rational Rose, SSIS, T-SQL, SSRS

We'd love your feedback!