Sr. Data Architect/ Data Modeler Resume
Dorchester, MA
SUMMARY:
- 9+ years of IT industry experience in Application Design, Development, and Data Management - Data Governance, Data Architecture, Data Modeling, Data Warehousing and BI, Data Integration, Meta-data, Reference Data and MDM.
- Experience as Architect UML models and leverage the advanced executable code generators to target different domains.
- Strong Knowledge of Data Warehouse Architecture and Star Schema, Snow flake Schema, FACT and Dimensional Tables.
- Strong Experience in Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data.
- Experience in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
- Experience with emerging technologies such Big Data, Hadoop, and NoSQL.
- Strong experience in analyzing/ Data Transformation of large amounts of data sets writing Pig scripts and Hive, AWS EMR, AWS RDS. Extensive knowledge in Hadoop stack components viz. Apache Hive, Pig Scripting, etc.
- Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop, Flume.
- Hands on experience in Normalization (1NF, 2NF, 3NF and BCNF) Denormalization techniques for effective and optimum performance in OLTP and OLAP environments.
- Experience in cloud development architecture on Amazon AWS, EC2, EC3, Elastic Search, Redshift and Basic on Azure.
- Experience in BI/DW solution (ETL, OLAP, Data mart), Informatica, BI Reporting tool like Tableau and QlikView and also experienced leading the team of application, ETL, BI developers, Testing team.
- Good experience in working with different ETL tool environments like SSIS, Informatica and reporting tool environments like SQL Server Reporting Services (SSRS), Cognos and Business Objects.
- Proficient in UML Modeling like Use Case Diagrams, Activity Diagrams, and Sequence Diagrams with Rational Rose and MS Visio.
- Experienced in Technical consulting and end-to-end delivery with architecture, data modeling, data governance and design - development - implementation of solutions.
- Solid knowledge of Data Marts, Operational Data Store (ODS), OLAP, Dimensional Data Modeling with Ralph Kimball Methodology (Star Schema Modeling, Snow-Flake Modeling for FACT and Dimensions Tables) using Analysis Services. rganizing data as per the business requirements using Erwin, ER Studio in both OLTP and OLAP applications.
- Expertise lies in Data Modeling, Database design and implementation of Oracle, AWS Redshift databases and Administration, Performance tuning etc. Excellent experience in troubleshooting SQL queries, ETL jobs, data warehouse/data mart/data store models. Practical understanding of the Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables. Worked on data modeling using ERWIN tool to build logical and physical models.
- Skillful in Data Analysis using SQL on Oracle, MS SQL Server, DB2 & Teradata.
- Extensive experience in development of T-SQL, Oracle PL/SQL Scripts, Stored Procedures and Triggers for business logic implementation.
- Strong experience with architecting highly per formant databases using PostgreSQL, PostGIS, MySQL and Cassandra.
- Decode the Teradata and SQL queries to find all the data attributes involved and document for the purpose of development.
- Excellent understanding of Hub Architecture Style for MDM hubs the registry, repository and hybrid approach.
- Mapping the Risk Data elements to the Authoritative Data Source and documenting the Schema, Database, Table details for data modelling purpose.
- Good exposure on usage of NoSQL database.
- Experienced in understanding the ETL framework metadata to understand the current state ETL implementation.
TECHNICAL SKILLS:
Data Modeling Tools: IBM Info sphere Data Architect, ER Studio and Oracle Designer, Erwin R6/R9, Rational System Architect.
Big Data: HBase, PIG, Hive, Spark, Sqoop, Flume.
Cloud Platforms: Azure, AWS EMR, AWS RDS, EC2, S3.
Database Tools: Oracle 12c/11G/10g/9i, Microsoft SQL Server12.0, Teradata 15.0, and MS Access
BI Tools: SAP Business Objects, Tableau 7.0/8.2, Tableau server 8.2, Tableau Reader 8.1, Crystal Reports.
Tools: & Utilities: TOAD 9.6, Microsoft Visio 2010.
Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Waterfall Model.
Packages: SAP and Microsoft Visio, Share point, Microsoft Office 2010, Microsoft Project 2010.
Operating Systems: Windows, Centos, Sun Solaris, UNIX, Ubuntu Linux
Version Tool: VSS, SVN, CVS, SAP BO 4.1
WORK EXPERIENCE:
Confidential, Dorchester, MA
Sr. Data Architect/ Data Modeler
Responsibilities:
- Owned and managed all changes to the data models. Created data models, solution designs and data architecture documentation for complex information systems.
- Working as a Sr. Data Architect/Modeler to generate Data Models using Erwin r9.64 and developed relational database system.
- Architect, researched, evaluated and deployed new tools, frameworks, and patterns to build sustainable Big Data platforms for our clients.
- Translated the business requirements into workable functional and non-functional requirements Confidential detailed production level using Workflow Diagrams, Sequence Diagrams, Activity Diagrams and Use Case Modelling
- Designed the Logical Data Model using ERWIN 9.64 with the entities and attributes for each subject areas.
- Used Tableau for BI Reporting and Data Analysis.
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Used data vault modeling method which was adaptable to the needs of this project.
- Design and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Develop and maintain data architecture, including master data and data quality, using Toad Data Modeler and Microsoft Master Data Manager (MDS) as well as Oracle Data Integrator.
- Configured Hunk to read customer transaction data from Hadoop Ecosystems such as HDFS and Hive
- Used Flume extensively in gathering and moving log data files from Application Servers to a central location in Hadoop Distributed File System (HDFS) for data science.
- Involved in Normalization / Denormalization techniques for optimum performance in relational and dimensional database environments.
- Working with project management, business teams and departments to assess and refine requirements to design/develop BI solutions using Azure.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Collected large amounts of log data using Apache Flume and aggregating using PIG/HIVE in HDFS for further analysis.
- Created Logical and Physical Data Model using IBM Data Architect tool.
- Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL and BI.
- Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data.
- Designing Star Denormalized tables on Azure.
- Participated in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi-dimensional models such as Star Schemas and Snowflake Schema.
- Designing of Big Data platform technology architecture. The scope includes data intake, data staging, data warehousing, and high performance analytics environment.
- Utilize U-SQL for data analytics/data ingestion of raw data in Azure and Blob storage
- Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
- Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
- Responsible for Dimensional Data Modeling and Modeling Diagrams using ERWIN.
- Applied data analysis, data mining and data engineering to present data clearly.
- Converting existing hive queries to Spark SQL queries to reduce execution time.
- Demonstrated expertise utilizing ETL tools, including SQL Server Integration Services (SSIS), Data Transformation Services (DTS), and Data Stage and ETL package design, and RDBM systems like SQL Servers, Oracle, and DB2.
- Review and Patch of Netezza and Oracle environments including DB2, OS and Server firmware.
- Extensively used Crystal Reports SAP SE 14.2 for Data Reporting.
- Gathered and analyzed existing physical data models for in scope applications and proposed the changes to the data models according to the requirements.
- Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
- Used Teradata Administrator and Teradata Manager Tools for monitoring and control the system.
- Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
- Experience with SAP Business Object 4.1
- Developed PL/SQL scripts to validate and load data into interface tables
- Participated in maintaining data integrity between Oracle and SQL databases.
Environment: Oracle 12c, MS-Office, SQL Architect, Spark, TOAD Benchmark Factory, Teradatav15, Hadoop, SQL Loader, SharePoint, ERwin r 9.64, DB2, MS-Office, SQL Server 2008/2012, Azure, HBase, Hive.
Confidential, Dover, NH
Sr. Data Architect / Data Modeler
Responsibilities:
- Involve in Data Architect role to review business requirement and compose source to target data mapping documents.
- Designed and build relational database models and defines data requirements to meet the business requirements.
- Developing the Conceptual Data Models, Logical data models and transformed them to creating schema using ER Studio.
- Participated in data analysis and data dictionary and metadata management - Collaborating with business analysts, ETL developers, data quality analysts and database administrators.
- Worked with Data Steward Team for designing, documenting and configuring Informatica Data Director for supporting management of MDM data.
- Actively involved in the Design and development of the Star schema data model.
- Implemented slowly changing and rapidly changing dimension methodologies; created aggregate fact tables for the creation of ad-hoc reports.
- Created and maintained surrogate keys on the master tables to handle SCD type 2 changes effectively.
- Worked with the developers in deciding the application architecture.
- Designed and implemented a Data Lake to consolidate data from multiple sources, using Hadoop stack technologies like SQOOP, HIVE/HQL.
- Written complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2.
- Designing Logical data models and Physical Data Models using ER Studio.
- Designed semantic layer data model. Conducted performance optimization for BI infrastructure.
- Involved in the creation, maintenance of Data Warehouse and repositories containing Metadata.
- Performed Hive programming for applications that were migrated to big data using Hadoop.
- As an Architect implement MDM hub to provide clean, consistent data for a SOA implementation.
- Installing and configuring the a 3-node Cluster in AWS EC2 Linux Servers.
- Designed different type of STAR schemas like detailed data marts and Plan data marts, Monthly Summary data marts using ER studio with various Dimensions Like Time, Services, Customers and various FACT Tables.
- Developed and maintained data dictionary to create metadata reports for technical and business purpose.
- ETL processing using Pig & Hive in AWS EMR, S3
- Implemented Data Vault Modeling Concept solved the problem of dealing with change in the environment by separating the business keys and the associations between those business keys, from the descriptive attributes of those keys using HUB, LINKS tables and Satellites.
- Extensive Data validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
- Data Profiling, Mapping and Integration from multiple sources to AWS S3.
- Design and development of ETL routines to extract data from heterogeneous sources and loading to Actuarial Data Warehouse.
- Participated in preparing Logical Data Models/Physical Data Models.
- Identify source systems, their connectivity, related tables and fields and ensure data suitably for mapping.
- Worked with BTEQ to submit SQL statements, import and export data, and generate reports in Teradata.
- Worked on HL7 2.x file format (ADT and clinical messages) on MEDIFAX and a thorough understanding of how interface development projects work.
- Developed company-wide data standards, data policies and data warehouse/business intelligence architectures.
- Designed and documented Use Cases, Activity Diagrams, Sequence Diagrams, OOD (Object Oriented Design) using UML and Visio.
- Performed data cleaning and data manipulation activities using NOSQL utility.
- Designed and Developed Oracle PL/SQL Procedures and UNIX Shell Scripts for Data Import/Export and Data Conversions.
Environment: ER Studio, Oracle 11g, MS-Office, SQL Architect, Hadoop, Hive, Pig, TOAD Benchmark Factory, Sqoop, SQL Loader, AWS S3, PL/SQL, DB2, SharePoint, MS-Office, SQL Server 2014
Confidential, San Diego, CA
Sr. Data Modeler/ Data Analyst
Responsibilities:
- Understood source business process and data model and do Data Analysis and map the requirement to source system.
- Created and documented the models, data Workflow Diagrams, Sequence Diagrams, Activity Diagrams and field mappings of all existing system.
- Create Logical Data Model for Staging, ODS and Data Mart and Time dimension as well.
- Generated comprehensive analytical reports by running SQL queries against current databases to conduct Data Analysis.
- Performed Data Analysis, Data Migration and data profiling using complex SQL on various sources systems including Oracle and Teradata.
- Develop Logical and Physical data models that capture current state/future state data elements and data flows using Erwin.
- Designed Database with prominent activities like maintaining sequences, index, primary key, foreign key, manipulating columns and tables.
- Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
- Involved in preparing modeling estimates, DFD, Volumetric, defining enterprise standards for best practices.
- Performed Data modeling using Erwin Data Modeler. Identified objects and relationships and how those all fit together as logical entities, these are then translated into physical design using forward engineering
- Identified the entities and relationship between the entities to develop Conceptual Model using ERWIN.
- Worked with Hadoop eco system covering HDFS, HBase, YARN and Map Reduce
- Enforced referential integrity in the OLTP data model for consistent relationship between tables and efficient database design.
- Worked with BTEQ to submit SQL statements, import and export data, and generate reports in Teradata.
- Gathering the business requirements from the respective application team (Data Architect's / Application Owner's).
- Extensively used SQL, Transact SQL and PL/SQL to write stored procedures, functions, packages and triggers.
- Used forward engineering to create a physical data model with DDL that best suits the requirements from the Logical Data Model.
- Worked with data compliance teams, Data governance team to maintain data models, Metadata, Data Dictionaries; define source fields and its definitions.
Environment: ERWIN, Windows, UNIX, Oracle10g, SQL SERVER 12.0, PL/SQL, DB2, Teradata SQL assistant, DQ analyzer, HDFS, HBase, YARN, IBM Optim Tool, IBM Rational Clear Case and Clear Quest, Toad 9.7
Confidential, Baltimore, MD
Sr. Data Modeler/Data Analyst
Responsibilities:
- Analyzed the business requirements by dividing them into subject areas and understood the data flow within the organization
- Attended and participates in information and requirements gathering sessions.
- Database Design (Conceptual, Logical and Physical) for OLTP and OLAP systems.
- Created and developed Slowly Changing Dimensions tables SCD2, SCD3 to facilitate maintenance of history.
- Created documents for technical & business user requirements during requirements gathering sessions.
- Turned SQL queries to make use of data base indexes, and analyzed the data base objects.
- Created Logical and Physical EDW models and data marts.
- Experienced in data migration and cleansing rules for the integrated architecture (OLTP, ODS, DW).
- Managed all indexing, debugging and query optimization techniques for performance tuning using T-SQL.
- Developed the logical and physical model from the conceptual model developed using a tool Erwin by understanding and analyzing business requirements.
- Handled data loading operations from flat files to tables using NZLOAD utility.
- Experienced in data cleansing for accurate reporting. Thoroughly analyzed the data and integrated different data sources to process matching functions.
- Applied data naming standards, created the data dictionary and documented data model translation decisions and also maintained DW metadata.
- Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
- Extensively used Normalization techniques (up to 3NF).
- Writing complex queries using Teradata SQL.
- Worked with the ETL team to document the transformation rules for data migration from source to target systems.
- Developed source to target mapping documents to support ETL design.
Environment: MS Visio, Business Objects, Informatica, ERWIN r7.2, PL/SQL, MS SQL, Windows NT, Linux, Sybase Power Designer, Oracle 9i, SQL Server, Windows, MS Excel.
Confidential
Data Modeler/Analyst
Responsibilities:
- Reviewed functional requirements and use cases to determine the necessary data requirements.
- Participated in creating realistic project plans with detailed tasks and ensuring their timely execution.
- Designed and implemented business intelligence to support sales and operations functions to increase customer satisfaction
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
- Created logical data model from the conceptual model and it's conversion into the physical database design using Erwin.
- Analyzed the data and provide resolution by writing analytical/complex SQL in case of data discrepancies.
- Tuning and code optimization using different techniques like dynamic SQL, dynamic cursors, and tuning SQL queries, writing generic procedures, functions and packages.
- Responsible for Relational data modeling (OLTP) using MS Visio (Logical, Physical and Conceptual).
- Designed, developed and implemented solutions with data warehouse, ETL, data analysis, and BI reporting technologies.
- Design and development of ETL processes using Informatica ETL tool for dimension and fact file creation
- Create and execute test scripts, cases, and scenarios that will determine optimal system performance according to specifications.
- Reverse Engineered DB2 databases and then forward engineered them to Teradata using Erwin.
- Tested the database to check field size validation, check constraints, stored procedures and cross verifying the field size defined within the application with metadata.
- Extensively worked on development of mappings with BODS Transformations like Map Operation, Table Comparison, History Preserving, Key Generation, Pivot, Reverse Pivot etc.
- Responsible for design of logical and physical Data model for client's investment management ODS using dimensional modeling.
Environment: Erwin, Oracle 8i, Developer 2000 with Forms 5.0 and Reports 3.0, O, Windows XP, PL/SQL, MS-Access, Sql Server, MS Office, MS Visio, Informatica Power center 5.1, Teradata SQL Assistant.