We provide IT Staff Augmentation Services!

Sr. Data Architect/ Data Modeler Resume

Chicago, IL

SUMMARY:

  • 8+ years of IT industry experience in Application Design, Development, and Data Management - Data Governance, Data Architecture, Data Modeling, Data Warehousing and BI, Data Integration, Meta-data, Data and MDM.
  • Experience as Architect UML models and leverage the advanced executable code generators to target different domains.
  • Experience in Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data.
  • Experience in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
  • Experience with emerging technologies such Big Data, Hadoop, and NoSQL.
  • Strong experience in Normalization (1NF, 2NF, 3NF and BCNF) and Denormalization techniques for effective and optimum performance.
  • Experience in cloud development architecture on Amazon AWS, EC2, EC3, Elastic Search, Redshift and Basic on Azure.
  • Extensive experience on usage of ETL &Reporting tools like SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS).
  • Proficient in UML Modeling like Use Case Diagrams, Activity Diagrams, and Sequence Diagrams with Rational Rose and MS Visio.
  • Experienced in Technical consulting and end-to-end delivery with architecture, data modeling, data governance and design - development - implementation of solutions.
  • Solid hands on experience in creating and implementation of Conceptual, Logical and Physical Models for Online Transaction Processing and Online Analytical Processing. Efficient in developing Logical and Physical Data model and organizing data as per the business requirements using Erwin, ER Studio in both OLTP and OLAP applications.
  • Strong experience in analyzing/ Data Transformation of large amounts of data sets writing Pig scripts and Hive, AWS EMR, AWS RDS. Extensive knowledge in Hadoop stack components viz. Apache Hive, Pig Scripting, etc.
  • Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop, Flume.
  • Expertise lies in Data Modeling, Database design and implementation of Oracle, AWS Redshift databases and Administration, Performance tuning etc. Excellent experience in troubleshooting SQL queries, ETL jobs, data warehouse/data mart/data store models. Practical understanding of the Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables. Worked on data modeling using ERWIN tool to build logical and physical models.
  • Skillful in Data Analysis using SQL on Oracle, MS SQL Server, DB2 & Teradata.
  • Extensive experience in development of T-SQL, Oracle PL/SQL Scripts, Stored Procedures and Triggers for business logic implementation.
  • Decode the Teradata and SQL queries to find all the data attributes involved and document for the purpose of development.
  • Excellent understanding of Hub Architecture Style for MDM hubs the registry, repository and hybrid approach.
  • Mapping the Risk Data elements to the Authoritative Data Source and documenting the Schema, Database, Table details for data modelling purpose.
  • Good exposure on usage of NoSQL database.
  • Experienced in understanding the ETL framework metadata to understand the current state ETL implementation.

TECHNICAL SKILLS:

Data Modeling Tools: Erwin R6/R9, Rational System Architect, IBM Info sphere Data Architect, ER Studio and Oracle Designer .

Big Data: PIG, Hive, HBase, Spark, Sqoop, Flume.

Database Tools: Microsoft SQL Server12.0, Teradata 15.0, Oracle 12c/11G/9i, and MS Access

BI Tools: Tableau 7.0/8.2, Tableau server 8.2, Tableau Reader 8.1, SAP Business Objects, Crystal Reports

Packages: Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point.

Operating Systems: Windows, Centos, Sun Solaris, UNIX, Ubuntu Linux

Version Tool: VSS, SVN, CVS, SAP BO 4.1

Tools: & Utilities TOAD 9.6, Microsoft Visio 2010.

Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Waterfall Model.

Cloud Platforms: EMR, EC2, S3, Azure.

WORK EXPERIENCE:

Confidential, Chicago, IL

Sr. Data Architect/ Data Modeler

Responsibilities:

  • Working as a Sr. Data Architect/Modeler to generate Data Models using Erwin r9.64 and developed relational database system.
  • Design and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
  • Architect, researched, evaluated and deployed new tools, frameworks, and patterns to build sustainable Big Data platforms for our clients.
  • Designed the Logical Data Model using ERWIN 9.64 with the entities and attributes for each subject areas.
  • Used Tableau for BI Reporting and Data Analysis.
  • Develop and maintain data architecture, including master data and data quality, using Toad Data Modeler and Microsoft Master Data Manager (MDS) as well as Oracle Data Integrator.
  • Configured Hunk to read customer transaction data from Hadoop Ecosystems such as HDFS and Hive
  • Used Flume extensively in gathering and moving log data files from Application Servers to a central location in Hadoop Distributed File System (HDFS) for data science.
  • Involved in Normalization / Denormalization techniques for optimum performance in relational and dimensional database environments.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Working with project management, business teams and departments to assess and refine requirements to design/develop BI solutions using Azure.
  • Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
  • Collected large amounts of log data using Apache Flume and aggregating using PIG/HIVE in HDFS for further analysis.
  • Created Logical and Physical Data Model using IBM Data Architect tool.
  • Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL and BI.
  • Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data.
  • Designing Star Denormalized tables on Azure
  • Designing of Big Data platform technology architecture. The scope includes data intake, data staging, data warehousing, and high performance analytics environment.
  • Utilize U-SQL for data analytics/data ingestion of raw data in Azure and Blob storage
  • Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
  • Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Responsible for Dimensional Data Modeling and Modeling Diagrams using ERWIN.
  • Applied data analysis, data mining and data engineering to present data clearly.
  • Converting existing hive queries to Spark SQL queries to reduce execution time.
  • Demonstrated expertise utilizing ETL tools, including SQL Server Integration Services (SSIS), Data Transformation Services (DTS), and Data Stage and ETL package design, and RDBM systems like SQL Servers, Oracle, and DB2.
  • Review and Patch of Netezza and Oracle environments including DB2, OS and Server firmware.
  • Extensively used Crystal Reports SAP SE 14.2 for Data Reporting.
  • Gathered and analyzed existing physical data models for in scope applications and proposed the changes to the data models according to the requirements.
  • Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
  • Used Teradata Administrator and Teradata Manager Tools for monitoring and control the system.
  • Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Experience with SAP Business Object 4.1
  • Developed PL/SQL scripts to validate and load data into interface tables
  • Participated in maintaining data integrity between Oracle and SQL databases.
  • Participated in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi-dimensional models such as Star Schemas and Snowflake Schema.

Environment: Oracle 12c, MS-Office, SQL Architect, Spark, TOAD Benchmark Factory, Teradatav15, Hadoop, SQL Loader, SharePoint, ERwin r 9.64, DB2, MS-Office, SQL Server 2008/2012, Azure, HBase, Hive.

Confidential, Minneapolis, MN

Sr. Data Architect / Data Modeler

Responsibilities:

  • Involve in Data Architect role to review business requirement and compose source to target data mapping documents.
  • Actively involved in the Design and development of the Star schema data model.
  • Implemented slowly changing and rapidly changing dimension methodologies; created aggregate fact tables for the creation of ad-hoc reports.
  • Created and maintained surrogate keys on the master tables to handle SCD type 2 changes effectively.
  • Worked with the developers in deciding the application architecture.
  • Designed and implemented a Data Lake to consolidate data from multiple sources, using Hadoop stack technologies like SQOOP, HIVE/HQL.
  • Designed semantic layer data model. Conducted performance optimization for BI infrastructure.
  • Involved in the creation, maintenance of Data Warehouse and repositories containing Metadata.
  • Performed Hive programming for applications that were migrated to big data using Hadoop.
  • As an Architect implement MDM hub to provide clean, consistent data for a SOA implementation.
  • Installing and configuring the a 3-node Cluster in AWS EC2 Linux Servers.
  • Designed different type of STAR schemas like detailed data marts and Plan data marts, Monthly Summary data marts using ER studio with various Dimensions Like Time, Services, Customers and various FACT Tables.
  • Developed and maintained data dictionary to create metadata reports for technical and business purpose.
  • ETL processing using Pig & Hive in AWS EMR, S3
  • Developed strategies for data warehouse implementations, data acquisitions, provided technical and strategic advice and guidance to senior managers and technical resources in the creation and implementation for data architecture and data modelling.
  • Extensive Data validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
  • Data Profiling, Mapping and Integration from multiple sources to AWS S3.
  • Design and development of ETL routines to extract data from heterogeneous sources and loading to Actuarial Data Warehouse.
  • Participated in preparing Logical Data Models/Physical Data Models.
  • Identify source systems, their connectivity, related tables and fields and ensure data suitably for mapping.
  • Worked with BTEQ to submit SQL statements, import and export data, and generate reports in Teradata.
  • Worked on HL7 2.x file format (ADT and clinical messages) on MEDIFAX and a thorough understanding of how interface development projects work.
  • Developed company-wide data standards, data policies and data warehouse/business intelligence architectures.
  • Designed and documented Use Cases, Activity Diagrams, Sequence Diagrams, OOD (Object Oriented Design) using UML and Visio.
  • Performed data cleaning and data manipulation activities using NOSQL utility.
  • Designed and Developed Oracle PL/SQL Procedures and UNIX Shell Scripts for Data Import/Export and Data Conversions.
  • Written complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2.
  • Designing Logical data models and Physical Data Models using ER Studio.
  • Developing the Conceptual Data Models, Logical data models and transformed them to creating schema using ER Studio.

Environment: DB2, ER Studio, Oracle 11g, MS-Office, SQL Architect, Hadoop, Hive, Pig, TOAD Benchmark Factory, Sqoop, SQL Loader, AWS S3, PL/SQL, SharePoint, MS-Office, SQL Server 2008/2012

Confidential, Hartford, CT

Sr. Data Modeler/ Data Analyst

Responsibilities:

  • Understood source business process and data model and do Data Analysis and map the requirement to source system.
  • Generated comprehensive analytical reports by running SQL queries against current databases to conduct Data Analysis.
  • Performed Data Analysis, Data Migration and data profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Develop Logical and Physical data models that capture current state/future state data elements and data flows using Erwin.
  • Experienced in Designing Database with prominent activities like maintaining sequences, index, primary key, foreign key, manipulating columns and tables.
  • Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
  • Involved in preparing modeling estimates, DFD, Volumetric, defining enterprise standards for best practices.
  • Data architect for large-scale pharmacy data management (PBM) data migration project from SAS to Teradata for the Enterprise Information Management (EIM) program. Redesigned database schemas and created conceptual, logical and physical data models.
  • Performed Data modeling using Erwin Data Modeler. Identified objects and relationships and how those all fit together as logical entities, these are then translated into physical design using forward engineering
  • Identified the entities and relationship between the entities to develop Conceptual Model using ERWIN.
  • Worked with Hadoop eco system covering HDFS, HBase, YARN and Map Reduce
  • Enforced referential integrity in the OLTP data model for consistent relationship between tables and efficient database design.
  • Responsible for the development and maintenance of Logical and Physical data models, along with corresponding metadata, to support Applications.
  • Excellent experience and knowledge on data warehouse concepts and dimensional data modeling using Ralph Kimball methodology.
  • Worked with BTEQ to submit SQL statements, import and export data, and generate reports in Teradata.
  • Gathering the business requirements from the respective application team (Data Architect's / Application Owner's).
  • Extensively used SQL, Transact SQL and PL/SQL to write stored procedures, functions, packages and triggers.
  • Used forward engineering to create a physical data model with DDL that best suits the requirements from the Logical Data Model.
  • Worked with data compliance teams, Data governance team to maintain data models, Metadata, Data Dictionaries; define source fields and its definitions.

Environment: ERWIN 9.0.0, Windows 2010, UNIX, Oracle 12c, SQL SERVER 12.0, PL/SQL, DB2, Teradata SQL assistant, DQ analyzer, HDFS, HBase, YARN, IBM Optim Tool, IBM Rational Clear Case and Clear Quest, Toad 9.7

Confidential, Manhattan, New York

Sr. Data Modeler/Data Analyst

Responsibilities:

  • Database Design (Conceptual, Logical and Physical) for OLTP and OLAP systems.
  • Created and developed Slowly Changing Dimensions tables SCD2, SCD3 to facilitate maintenance of history.
  • Created documents for technical & business user requirements during requirements gathering sessions.
  • Turned SQL queries to make use of data base indexes, and analyzed the data base objects.
  • Created Logical and Physical EDW models and data marts.
  • Experienced in data migration and cleansing rules for the integrated architecture (OLTP, ODS, DW).
  • Managed all indexing, debugging and query optimization techniques for performance tuning using T-SQL.
  • Developed the logical and physical model from the conceptual model developed using a tool Erwin by understanding and analyzing business requirements.
  • Handled data loading operations from flat files to tables using NZLOAD utility.
  • Experienced in data cleansing for accurate reporting. Thoroughly analyzed the data and integrated different data sources to process matching functions.
  • Applied data naming standards, created the data dictionary and documented data model translation decisions and also maintained DW metadata.
  • Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
  • Extensively used Normalization techniques (up to 3NF).
  • Writing complex queries using Teradata SQL.
  • Worked with the ETL team to document the transformation rules for data migration from source to target systems.
  • Developed source to target mapping documents to support ETL design.

Environment: ERWIN r7.2, PL/SQL, MS SQL, MS Visio, Business Objects, Windows NT, Linux, Sybase Power Designer, Oracle 8i, SQL Server, Windows, MS Excel, Informatica.

Confidential

Data Modeler/Analyst

Responsibilities:

  • Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
  • Created logical data model from the conceptual model and it's conversion into the physical database design using Erwin.
  • Analyzed the data and provide resolution by writing analytical/complex SQL in case of data discrepancies.
  • Tuning and code optimization using different techniques like dynamic SQL, dynamic cursors, and tuning SQL queries, writing generic procedures, functions and packages.
  • Responsible for Relational data modeling (OLTP) using MS Visio (Logical, Physical and Conceptual).
  • Designed, developed and implemented solutions with data warehouse, ETL, data analysis, and BI reporting technologies.
  • Design and development of ETL processes using Informatica ETL tool for dimension and fact file creation
  • Create and execute test scripts, cases, and scenarios that will determine optimal system performance according to specifications.
  • Reverse Engineered DB2 databases and then forward engineered them to Teradata using Erwin.
  • Tested the database to check field size validation, check constraints, stored procedures and cross verifying the field size defined within the application with metadata.
  • Extensively worked on development of mappings with BODS Transformations like Map Operation, Table Comparison, History Preserving, Key Generation, Pivot, Reverse Pivot etc.
  • Responsible for design of logical and physical Data model for client's investment management ODS using dimensional modeling.

Environment: Erwin, Oracle 8i, Developer 2000 with Forms 5.0 and Reports 3.0, O, Windows XP, PL/SQL, MS-Access, Sql Server, MS Office, MS Visio, Informatica Power center 5.1, Teradata SQL Assistant

Hire Now