We provide IT Staff Augmentation Services!

Sr. Data Architect/data Modeler Resume

Minneapolis, MN

SUMMARY:

  • Over 9+ years of experience as a Data Architecture, Data Analyst and Data Modeler with solid understanding of Business Requirements Gathering, Business Process Mapping Evaluating Data Sources and Data mapping, data profiling, Hadoop ecosystem, AWS, Data Analytics, Data Warehousing and ETL.
  • Experienced in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
  • Solid experience in development of Conceptual, Logical and Physical Models for Online Transaction Processing and Online Analytical Processing (OLTP & OLAP) and Power Design.
  • Experiences with data modeling with NoSQL databases such as MongoDB for document models, AWS Redshift etc.
  • Excellent experience in trouble shooting SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Experienced with distributed data warehousing and/or data mining systems, using one or more Big Data/NoSQL technologies (Hadoop, Hive, HBase, Pig, Cassandra, MongoDB)
  • Hands on experience with modeling using ERWIN, ER Studio, MS Visio in both forward and reverse engineering cases and skillful in Data Analysis using SQL on Oracle, MS SQL Server, Netezza, DB2 & Teradata
  • Well versed in Normalization / De normalization techniques for optimum performance in relational and dimensional database environments.
  • Experience in working with Business Intelligence and Enterprise Data Warehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum and Amazon Redshift (AWS).
  • Good experience and understanding of Teradata SQL Assistant, Teradata Administrator and data load/ export utilities like BTEQ, Fast Load, Multi Load, Fast Export.
  • Hands on experience with various Data Architect and ETL Architect, subsystem and patterns, including Change Date Capture, Slow Change Dimension, Data Cleansing, auditing and validation, etc. Solid experience with data governance, data steward, data quality concepts and implementations and expertise in performing User Acceptance Testing (UAT) and conducting end user sessions.
  • Strong background in various Data Modeling tools using Erwin, ER Studio and Power Designer.
  • Experience in integration of various relational and non - relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, and Netezza database.
  • Extensive knowledge and experience in producing tables, reports, graphs and listings using various procedures and handling large databases to perform complex data manipulations.
  • Experienced in data transformation, data mapping from source to Confidential database schemas and data cleansing procedures using Informatica Power Center, Talend and Pentaho.
  • Good in Data warehouse loads, determining hierarchies, building various logics to handle Slowly Changing Dimensions.
  • Strong experience in Normalization (1NF, 2NF, 3NF and BCNF) and De-normalization techniques for effective and optimum performance in OLTP and OLAP environments.
  • Expertise in Informatica Power Center and Data Quality (IDQ) tools.
  • Excellent experience in trouble shooting SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Expertise in SQL Server Analysis Services (SSAS) and SQL Server Reporting Services (SSRS)
  • Trained end users, prepare related documentation (requirements, materials, process and data flows, use cases, functional design, etc), and addressed critical questions from the user base involving the CRM
  • Good experience with use of Access queries, excel functions V-Lookup, formulas, Pivot Tables, etc.

TECHNICAL SKILLS:

Operating Systems: Linux, HP UNIX, Sun Solaris, and Windows 2000/XP/Vista

MS Office Tools: MS Word, MS Excel, MS Access, MS PowerPoint, MS Visio

Languages: SQL, PL/SQL, JSON, UNIX script, XML, Scala, R

Data warehousing Tools: Business Objects, Cognos, Microstrategy, Tableau, Informatica, TalendPentaho, Datastage

Hadoop Ecosystem: Hadoop Framework, HiveQL Queries, Pig, HBase, MongoDB, Sqoop and spark, scala

Modeling Tools: ERWIN r9.6/r 9.5/9.1, Embarcadero, ER/Studio, MSVisio, Sybase Power Designer

Databases: Netezza, Teradata R14, R13, R12, Oracle 12c/11g/10g, DB2, SQL, Oracle, MS Access, SQL Server 2000, MongoDB

AWS: AWS S3, AWS Redshift, AWS EMR.

Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC)Waterfall Model

ETL Tools: Informatica Power Center 9x/8x, Pentaho, SSIS.

Reporting Tools: Crystal reports XI, Business Intelligence, SSRS, Business Objects 5.x /6.x, Tableau.

WORK EXPERIENCE:

Confidential, Minneapolis, MN

Sr. Data Architect/Data Modeler

Responsibilities

  • Designed architecture collaboratively to develop methods of synchronizing data coming in from multiple source systems.
  • Researched, evaluated, architect, and deployed new tools, frameworks and patterns to build sustainable Big Data platforms for our clients.
  • Implemented Agile Methodology for building Integrated Data Warehouse, involved in multiple sprints for various tracks throughout the project lifecycle.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Worked on Normalization and De-normalization concepts and design methodologies like Ralph Kimball and Bill Inmon's Data Warehouse methodology.
  • Developed prototype solutions to verify capabilities for new systems development, enhancement, and maintenance of MDM
  • Reviewed the Conceptual EDW (Enterprise Data Warehouse) Data Model with Business Users, App Dev. and Information Architects to make sure all the requirements are fully covered.
  • Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
  • Using Python, SQL, Hadoop ecosystem for extracting data and building predictive models.
  • Worked on NoSQL databases including HBase, Mongo DB, and Cassandra. Implemented multi- datacenter and multi-rack Cassandra cluster.
  • Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
  • Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow Flake Schemas.
  • Involved in Normalization/De-normalization techniques for optimum performance in relational and dimensional database environments.
  • Worked on Amazon Redshift, AWS & Azure and architecting a solution to load data create data models and run BI on it.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop.
  • Developed ETL processes extracted data daily and loaded data into a SSIS based Decision Support Warehouse.
  • Responsible for Metadata Management, keeping up to date centralized metadata repositories using Erwin modeling tools.
  • Driven the technical design of AWS solutions by working with customers to understand their needs.
  • Conducted numerous POCs (Proof of Concepts) to efficiently import large data sets into the database from AWS S3 Bucket.
  • Worked on analyzing source systems and their connectivity, discovery, data profiling and data mapping.
  • Driven the technical design of AWS solutions by working with customers to understand their needs
  • Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
  • Collected large amounts of log data using Apache Flume and aggregating using PIG in HDFS for further analysis.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Designed and architecting AWS Cloud solutions for data and analytical workloads such as warehouses, Big Data, data lakes, real-time streams and advanced analytics
  • Interacted with End-users for gathering Business Requirements and Strategizing the DataWarehouse processes
  • Implemented Python scripts to import/export JSON file, which contains the customer survey information and/or asset information, to/from the database.
  • Write complex Netezza views to improve performance and push down the load to database rather than doing it in the ETL tool.
  • Involved in data model reviews with internal data architect, business analysts, and business users with explanation of the data model to make sure it is in-line with business requirements.
  • Created DDL scripts using ER Studio and source to Confidential mappings to bring the data from source to the warehouse.
  • Worked with MapReduce frameworks such as Hadoop and associated tools (pig, Sqoop, etc)
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
  • Generated the frame work model from IBM data Architect for the Cognos reporting team.

Environment: ER Studio V17, Netezza, SQL Server 2016, Taradata15, OLAP, OLTP, UNIX, MDM, Hadoop, Hive, Pig, HBase, HDFS, SAP, AWS, Redshift, EMR, S3, Apache Flume, Ralph Kimball and Bill Inmon's, PL/SQL, BTEQ, Python.

Sr. Data Architect/Data Modeler

Confidential, Bethesda, MD

Responsibilities:

  • Provide data architecture support to enterprise data management efforts, such as the development of the enterprise data model, Master and data.
  • Lead the strategy, architecture and process improvements for data architecture and data management, balancing long and short-term needs of the business.
  • Building relationships and trust with key stakeholders to support program delivery and adoption of enterprise architecture.
  • Defined best practices for data modeling and extraction and ensure architectural alignment of the designs and development.
  • Implemented logical and physical relational database and maintained Database Objects in the data model using Erwin 9.6
  • Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
  • Working in Hadoop Environment using pig, Sqoop, Hive, HBase, MongoDB, and detailed understanding of map reduce programs
  • Participated in integration of MDM (Master Data Management) Hub and data warehouses.
  • Responsible of designing the whole data warehouse system including ODS, DWH and data marts.
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
  • Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
  • Used Normalization methods up to 3NF and De-normalization techniques for effective performance in OLTP systems.
  • Extensively using Agile methodology as the Organization Standard to implement the data Models.
  • Performed the Data Mapping, Data design ( Data Modeling) to integrate the data across the multiple databases in to EDW.
  • Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Used Load utilities (Fast Load & Multi Load) with the mainframe interface to load the data into Teradata and Data reconciliation activities between Source and EDW Teradata databases.
  • Involved in designing and developing Data Models and Data Marts that support the Business Intelligence Data Warehouse.
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
  • Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections.
  • Transforming staging area data into a STAR schema (hosted on Amazon Redshift) which was then used for developing embedded Tableau dashboards.
  • Maintained metadata ( data definitions of table structures) and version controlling for the data model.
  • Created stored procedures, functions, database triggers and packages as per the business needs for developing ad-hoc and robust reports.
  • Used Spark Data frames, Spark-SQL, Spark MLLib extensively and developing and designing POC's using Scala, Spark SQL and MLlib libraries.
  • Developed SQL scripts for creating tables, Sequences, Triggers, views and materialized views.
  • Scheduling the SSIS packages execution in SQL Server Agent and tracking the success or failure of the execution of SSIS Packages and configuring the email notifications through SQL Server Agent.
  • Developed multiple MapReduce jobs in java for Data Cleaning and pre-processing analyzing data in PIG.
  • Worked in Data Analysis, data profiling and data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, and Oracle)
  • Coordinating with DBA team to implement physical models & to setup development, test, staging & production environments for DDL & DML Generation & usage.
  • Used Teradata OLAP functions like RANK, ROW NUMBER, QUALIFY, CSUM and SAMPLE.

Environment: Erwin 9.6, MDM, AWS, Hadoop, Spark, Sqoop, Hive, HDFS, OLAP, OLTP, ODS, PL/SQL, Tableau, Teradata R15, Teradata SQL Assistant 15.0, Oracle 12c, SQL, PL/SQL, Flat Files, Informaitca, XML, SSIS, SSRS, AWS Redshift, MongoDB, HBase, AWS S3.

Sr. Data Modeler/Data Analyst

Confidential, San Francisco, CA

Responsibilities:

  • Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
  • Participated in the design, development, and support of the corporate operation data store and enterprise data warehouse database environment.
  • Designed the Logical Data Model using ER/Studio with the entities and attributes for each subject areas and Worked on ER Studio for multiple Operations across in both OLAP and OLTP applications.
  • Used the Agile Scrum methodology to build the different phases of Software development life cycle.
  • Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
  • Independently coded new programs and design Tables to load and test the program effectively for the given POC's using Big Data/Hadoop.
  • Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Responsible for full data loads from production to AWS Redshift staging environment and Worked on migrating of EDW to AWS using EMR and various other technologies.
  • Design of ODS layer, Dimensional modeling using Kimball Methodologies, of the Data Warehouse sourced from MDM Base tables and other Transactional systems.
  • Extensively used Aginity Netezza work bench to perform various DML, DDL etc operations on Netezza database.
  • Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
  • Extracting Mega Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports
  • Worked with the ETL team to document the transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
  • Worked with Netezza and Oracle databases and implemented various logical and physical data models for them.
  • Developed complex SQL scripts for Teradata database for creating BI layer on DW for Tableau reporting.
  • Involved in mapping the data elements from the User Interface to the Database and help identify the gaps and Prepared of Data Dictionary/Meta Data of the data elements.
  • Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Generated preliminary Source to Confidential mapping by writing a ER/Studio Macro in the ER/Studio VBA platform document and handed over to BAs for further transformation analysis.
  • Submitted the DDL and Security document by writing an ERStudio Macro to the SVN repository and communicated to the DBA for implementation.
  • Extracting Mega Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports.
  • Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
  • Designed and Developed Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing
  • Extensively worked on documentation of Data Model, Mapping Transformation and Scheduling jobs.
  • Worked extensively with Business Objects XI Report Developers in solving critical issues of defining hierarchy, loops and Contexts.

Environment: ER/Studio, Oracle 12c, Informatica, DB2, MDM AWS Redshift, AWS S3, Hadoop, Hive, Netezza, MongoDB, HBase, Sqoop, Spark, HDFS, MapReduce, OLAP, OLTP, PL/SQL, Flat Files, SQL, XML, Teradata, Tableau, SSRS, SSIS and Cassandra.

Sr. Data Modeler/Data Analyst

Confidential, New York, NY

Responsibilities:

  • Provided a consultative approach with business users, asking questions to understand the business need and deriving the data flow, logical, and physical data models based on those needs.
  • Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
  • Specifies overall Data Architecture for all areas (Healthcare payer domain) and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
  • Created Physical & logical data model from the conceptual model and it's conversion into the physical database with the DDL's using forward engineering options in Erwin.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture.
  • Completed enhancement for MDM (Master data management) and suggested the implementation for hybrid MDM (Master Data Management).
  • Handled importing data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
  • Worked on designing, implementing and deploying into production an Enterprise data warehouse
  • Developed SQl process using SSIS with Various Control Flow, Data Flow tasks and Store Procedures for Work Oracle Validation process .
  • Worked on importing and exporting data from Oracle and Teradata into HDFS and HIVE using Sqoop.
  • Designed and Developed Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing.
  • Used forward engineering to create a Physical Data Model with DDL that best suits the requirements from the Logical Data Model.
  • Responsible for data profiling and data quality checks to suffice the report requirements gathered above and provide an ETL Mapping.
  • Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS).
  • Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow Flake Schemas.
  • Worked in using Teradata tools like Fast Load, Multi Load, Tpump, FastExport, Teradata Parallel Transporter (TPT) and BTEQ.
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
  • Used SQL for Querying the database in UNIX environment
  • Worked on the metadata management and part of data governance team which created the Data
  • Creating or modifying the T-SQL queries as per the business requirements.
  • Developed various QlikView Data Models by extracting and using the data from various sources files Excel, Flat Files and Big data.
  • Primarily responsible for Tableau customization for statistical dashboard to monitor sales effectiveness and also used Tableau for customer marketing data visualization.
  • Developed the performance tuning of the database by using EXPLAIN PLAN, TKPROF utilities and also debugging the SQL code.

Environment: Erwin 9.1, MapReduce, ODS, MDM, OLAP, OLTP, Pig, Oracle 12, Hadoop, HDFS, Sqoop, Hive, NoSQL, Teradata R13, Netezza, PL/SQL, MS-Visio, T-SQL, SSIS, SSRS, SQL, Unix, Tableau, Informatica, PL/SQL.

Data Analyst

Confidential, Pleasanton, CA

Responsibilities:

  • Performed data analysis and profiling of source data to better understand the sources.
  • Created logical data model from the conceptual model and it's conversion into the physical database design using Erwin.
  • Worked with DBA's to create a best-fit Physical Data Model from the logical data model.
  • Designed Sources to Targets mappings from SQL Server, Excel/Flat files, Xml Files to Teradata using Informatica Power Center Data cleansing, integrating, and matching using Informatica Data Quality IDQ
  • Redefined many attributes and relationships in the reverse engineered model and cleansed unwanted tables/columns as part of data analysis responsibilities.
  • Interacted with the database administrators and business analysts for data type and class words.
  • Conducted design sessions with business analysts and ETL developers to come up with a design that satisfies the organization's requirements.
  • Worked on enterprise logical data modeling project (in third normal form) to gather data requirements for OLTP enhancements. Converted third normal form ERDs into dimensional ERDs for data warehouse effort.
  • Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
  • Created ER Diagrams, Data Flow Diagrams, grouped and created the tables, validated the data, identified PK/ FK for lookup tables.
  • Created 3NF business area data modeling with de-normalized physical implementation data and information requirements analysis using Erwin tool.
  • Developed Star Schema and Snowflake Schema in designing the Logical Model into Dimensional Model.
  • Assisted the ETL team to document the transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
  • Implement necessary DQ rules in IDQ Analyst while profiling the data.
  • Involved in extensive data analysis on the Teradata and Oracle systems querying and writing in SQL and TOAD.
  • Used SQL joins, aggregate functions, analytical functions, group by, order by clauses and interacted with DBA and developers for query optimization and tuning.
  • Conducted several Physical Data Model sessions with the ETL Developers. Worked with them on day-to-day basis to resolve any questions on Physical Model. .

Environment: CA Erwin 9.1, Oracle11g, SQL server 2005, IBM DB2, Informatica Power Center, IDQ, SQL BI 2008, Oracle BI, Visual Studio, SSIS&SSRS, Tibco Spotfire, SQL server management studio 2012.

Hire Now