We provide IT Staff Augmentation Services!

Sr. Data Architect/data Modeler Resume

3.00/5 (Submit Your Rating)

Wallingford, CT

SUMMARY:

  • Above 9+ years of IT experience as Data Architect/Data Modeler & Data Analyst in the Architecture, Design and Development.
  • Expertise in Data Governance, Data Integration, Metadata, Data and MDM.
  • Experience in Hadoop ecosystem components like Hadoop MapReduce, HDFS, HBase, Oozie, Hive, Sqoop, Pig, Flume including their installation and configuration.
  • Expertise in designing Star schema, Snowflake schema for Data Warehouse, ODS architecture by using tools like Erwin data modeler, Power Designer, and E - R Studio.
  • Strong experience with different project methodologies including Agile Scrum Methodology and Waterfall methodology.
  • Experience in setting up connections to different RDBMS Databases like Oracle, SQL Server, DB2, Teradata according to users requirement.
  • Experience in Data Modeling techniques employing Data warehousing concepts like Star/Snowflake Schema.
  • Experience in Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data.
  • Good knowledge of Data Marts, Operational Data Store (ODS), Dimensional Data Modeling with Ralph Kimball Methodology using Analysis Services.
  • Experience in working with Business Intelligence and Enterprise Data Warehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum, Amazon Redshift and Azure Data Warehouse.
  • Strong experience with architecting highly performance databases using PostgreSQL, PostGIS, MYSQL and Cassandra.
  • Strong experience in Normalization (1NF, 2NF, 3NF and BCNF) and Denormalization techniques for effective and optimum performance.
  • Good understanding and hands on experience in setting up and maintaining NoSQL Databases like Cassandra, MongoDB, and HBase
  • Expertise in Database Performance Tuning using Oracle Hints, Explain plan, TKPROF, Partitioning and Indexes
  • Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Worked with BI Reporting teams which use tools such as Micro-Strategy, Business Objects and SSRS as well as development of ETL mapping and scripts.
  • Experienced in Data Extraction/Transformation/Loading (ETL), Data Conversion and Data Migration by using SQL Server Integration Services (SSIS) and PL/SQL Scripts.
  • Experience in importing and exporting data from different relational databases like MySQL6.x, Netezza, Oracle into HDFS and Hive using Sqoop.
  • Experience in SAS interacts with third party software such as databases (DB2, SQL SERVER, Oracle), web application servers, strong knowledge on SAS HPA architecture.
  • Experience in Teradata RDBMS using Fast load, Fast Export, Multi load, T pump, and Teradata SQL Assistance and BTEQ Teradata utilities.
  • Good experienced in Normalization for OLTP and De-normalization of Entities for Enterprise Data Warehouse.
  • Extensive experience on usage of ETL & Reporting tools like SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS)
  • Experience in testing, data validation and writing SQL and PL/SQL statements - Stored Procedures, Functions, Triggers and packages.
  • Experience with SQL Server and T-SQL in constructing Temporary Tables, Table variables, Triggers, user functions, views, Stored Procedures.

TECHNICAL SKILLS:

Data Modeling Tools: Erwin r9.6/r9.5, ER Studio 9.7/9.0, Sybase Power Designer.

Big Data: Hadoop, HDFS, Hive, Pig, HBase, Sqoop, Flume, Kafka.

Database Tools: Oracle 12c/11g, Teradata 15/14, Netezza, Microsoft SQL Server 2014/2016, and MS Access, PostgreSQL.

Quality Assurance Tools: Win Runner, Load Runner, Test Director, Quick Test Pro, Quality Center, Rational Functional Tester.

Reporting tools: SQL Server Reporting Services (SSRS), Tableau, Crystal Reports, Business Objects, MicroStrategy, Business Objects 5.1, Cognos 6.5/7.0

ETL Tools: SSIS, Pentaho, Informatica9.6.

Programming Languages: Java, Base SAS and SAS/SQL, SQL, T-SQL, HTML, Java Script, CSS, UNIX shells scripting, PL/SQL.

Operating Systems: Microsoft Windows 8/7, UNIX, Linux, Redhat

Tools & Software: TOAD, SQL *PLUS, SQL*LOADER, MS Office, BTEQ, Teradata SQL Assistant

Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Agile, Waterfall Model.

PROFESSIONAL EXPERIENCE:

Confidential - Wallingford, CT

Sr. Data Architect/Data Modeler

Responsibilities:

  • Responsible for the data architecture design delivery, data model development, review, approval and Data warehouse implementation.
  • Created logical and physical 3NF relational models based on XML data extracts from transportation logistics application (ER/Studio)
  • Used Star Schema and Snowflake Schema methodologies in building and designing the Logical Data Model into Dimensional Models
  • Lead a team responsible for design and implementation of ODS for transportation and logistics management system
  • Documented logical, physical, relational and dimensional data models. Designed the Data Marts in dimensional data modeling using Star and Snowflake schemas.
  • Attended numerous s to understand the Healthcare Domain and the concepts related to the project (Healthcare Informatics).
  • Collaborated with other data modelers to understand and implement best practices within the organization.
  • Involved in Data Architecture, Data profiling, data mapping and Data architecture artifacts design.
  • Developed strategies for warehouse implementation, data acquisition, and archive recovery.
  • Used M-LOAD, Fast-load and T-pump loading to migrate data from Oracle to Teradata.
  • Designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using ER Studio 9.7.
  • Driven the technical design of AWS solutions by working with customers to understand their needs.
  • Conducted numerous POCs (Proof of Concepts) to efficiently import large data sets into the database from AWS S3 Bucket.
  • Worked on analyzing source systems and their connectivity, discovery, data profiling and data mapping.
  • Driven the technical design of AWS solutions by working with customers to understand their needs
  • Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
  • Collected large amounts of log data using Apache Flume and aggregating using PIG in HDFS for further analysis.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Designed and architecting AWS Cloud solutions for data and analytical workloads such as warehouses, Big Data, data lakes, real-time streams and advanced analytics
  • Interacted with End-users for gathering Business Requirements and Strategizing the Data Warehouse processes
  • Write complex Netezza views to improve performance and push down the load to database rather than doing it in the ETL tool.
  • Involved in data model reviews with internal data architect, business analysts, and business users with explanation of the data model to make sure it is in-line with business requirements.
  • Created DDL scripts using ER Studio and source to target mappings to bring the data from source to the warehouse.
  • Worked with MapReduce frameworks such as Hadoop and associated tools (pig, Sqoop, etc)
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
  • Generated the frame work model from IBM data Architect for the Cognos reporting team.

Environment: ER/Studio 9.7, Teradata 15, Amazon Redshift, AWS, Oracle 12c, ODS, OLAP, OLTP, Hadoop, MapReduce, HDFS, SQOOP, Flume, Agile, OLAP, SAP Kafka, Pig, Oozie, Cassandra, MDM, Informatica 9.6, NoSQL, Unix.

Confidential - Boston, MA

Sr. Data Architect/Data Modeler

Responsibilities:

  • Responsible for the data architecture design delivery, data model development, review, approval and Data warehouse implementation.
  • Designed and developed the conceptual then logical and finally physical data models to meet the needs of reporting.
  • Involved in designing and developing Data Models and Data Marts that support the Business Intelligence Data Warehouse.
  • Implemented logical and physical relational database and maintained Database Objects in the data model using Erwin 9.5
  • Responsible for Big data initiatives and engagement including analysis, brainstorming, POC, and architecture.
  • Used Agile Methodology of Data Warehouse development using Kanbanize.
  • Worked with Hadoop eco system covering HDFS, HBase, YARN and Map Reduce.
  • Performed the Data Mapping, Data design (Data Modeling) to integrate the data across the multiple databases in to EDW.
  • Designed both 3NF Data models and dimensional Data models using Star and Snowflake schemas.
  • Involved in Normalization/Denormalization techniques for optimum performance in relational and dimensional database environments.
  • Developed Master data management strategies for storing data.
  • Worked with Data Stewards and Business analysts to gather requirements for MDM Project.
  • Involved in Testing like Unit testing, System integration and regression testing.
  • Worked with SQL Server Analysis Services (SSAS) and SQL Server Reporting Service(SSRS).
  • Worked on Data modeling, Advanced SQL with Columnar Databases using AWS.
  • Perform reverse engineering of the dashboard requirements to model the required data marts.
  • Developed Source to Target Matrix with ETL transformation logic for ETL team.
  • Cleansed, extracted and analyzed business data on daily basis and prepared ad-hoc analytical reports using Excel and T-SQL
  • Created Data Migration and Cleansing rules for the Integration Architecture (OLTP, ODS, DW).
  • Handled performance requirements for databases in OLTP and OLAP models.
  • Conducted meetings with business and development teams for data validation and end-to-end data mapping.
  • Responsible for Metadata Management, keeping up to date centralized metadata repositories using Erwin modeling tools.
  • Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Sql database.
  • Lead data migration from legacy systems into modern data integration frameworks from conception to completion.
  • Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from legacy DB2 and SQL Server 2014 database systems..
  • Managed the meta-data for the Subject Area models for the Data Warehouse environment.
  • Generated DDL and created the tables and views in the corresponding architectural layers.
  • Handled importing of data from various data sources, performed transformations using Map Reduce, loaded data into HDFS and Extracted the data from My SQL into HDFS using Sqoop
  • Involved in performing extensive Back-End testing by writing SQL queries and PL/SQL stored procedures to extract the data from SQL Database.
  • Participate in code/design reviews and provide input into best practices for reports and universe development.
  • Involved in Netezza Administration Activities like backup/restore, performance tuning, and Security configuration
  • Involved in the validation of the OLAP, Unit testing and System Testing of the OLAP Report Functionality and data displayed in the reports.
  • Created a high-level industry standard, generalized data model to convert it into logical and physical model at later stages of the project using Erwin and Visio
  • Participated in Performance Tuning using Explain Plan and TKPROF.
  • Involved in translating business needs into long-term architecture solutions and reviewing object models, data models and metadata.

Environment: Erwin 9.5, HDFS, AWS, HBase, Hadoop, Metadata, MS Visio, SQL Server 2016, Agile, PL/SQL, ODS, OLAP, OLTP, flat files, MDM.

Confidential - Cincinnati, OH

Sr. Data Architect/Data Modeler

Responsibilities:

  • Developed strategies for data acquisitions, archive recovery, and implementation of databases and working in a data warehouse environment, which includes data design, database architecture, and Metadata and repository creation.
  • Designed and implemented scalable Cloud Data and Analytical architecture solutions for various public and private cloud platforms using Azure
  • Designed and developed data architecture solutions in big data architecture or data analytics.
  • Evaluate architecture patterns, Define best patterns for data usage, data security, data compliance, Define concept models, logical & physical data model.
  • Involved with Key-value data modeling, data load process and classify the key business drivers for the data management initiative.
  • Preparing extensive documents on the ETL Design, Development, Testing, Deployment & daily loads of the mappings and interfaces.
  • Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data
  • Implemented Agile Methodology for building Integrated Data Warehouse, involved in multiple sprints for various tracks throughout the project lifecycle.
  • Involved in writing the PL/SQL validation scripts to identify the data inconsistencies in the sources.
  • Worked with Business Analysts to design weekly reports using Cognos.
  • Participated with key management resources in the strategic analysis and planning requirements for Data Warehouse/Data Mart reporting and data mining solutions.
  • Coordinated the database design on the DB infrastructure objects and performance/storage related changes with DBA on Oracle 11g platform.
  • Identified project requirement gaps using various data profiling techniques.
  • Translated business requirements into working logical and physical data models for Staging, Operational Data Store and Data marts applications.
  • Generated SQL scripts and implemented the relevant databases with related properties from keys, constraints, indexes & sequences.
  • Extensively worked with SSIS packages for Data Migration from source systems.
  • Implemented metadata management as one part of data governance.
  • Created entity relationship diagrams and multidimensional data models, reports and diagrams based on the requirements.
  • Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
  • Designed data cleansing/data scrubbing techniques to ensure consistency amongst data sets.

Environment: Erwin 9.1, Hive, Metadata, OLAP, OLTP, ODS, SSIS, Data Marts, PL/SQL, Cognos, Hadoop, HDFS, Oracle 10g, Data Stage

Confidential - Dallas, TX

Sr. Data Analyst

Responsibilities:

  • Analyze the OLTP Source Systems and Operational Data Store and research the tables/entities required for the project.
  • Designing the measures, dimensions and facts matrix document for the ease while designing.
  • Created data flowcharts and attribute mapping documents, analyzed the source meaning to retain and provide proper business names following the very stringent FTB's data standards.
  • Developed several scripts to gather all the required data from different databases to build the LAR file monthly.
  • Developed numerous reports to capture the transactional data for the business analysis.
  • Developed complex SQL queries to bring data together from various systems.
  • Organized and conducted cross-functional meetings to ensure linearity of the phase approach.
  • Collaborated with a team of Business Analysts to ascertain capture of all requirements.
  • Created multiple reports on the daily transactional data which involves millions of records.
  • Used Joins like Inner Joins, Outer joins while creating tables from multiple tables.
  • Created Multiset, temporary, derived and volatile tables in Teradata database.
  • Implemented Indexes, Collecting Statistics, and Constraints while creating tables.
  • Utilized ODBC for connectivity to Teradata via MS Excel to retrieve automatically from Teradata Database.
  • Developed various ad hoc reports based on the requirements
  • Designed & developed various Ad hoc reports for different teams in Business (Teradata and Oracle SQL, MS ACCESS, MS EXCEL)
  • Developed SQL Queries to fetch complex data from different tables in remote databases using joins, database links and formatted the results into reports and kept logs.
  • Involved in writing complex SQL queries using correlated sub queries, joins, and recursive queries.
  • Delivered the artifacts within the time lines and excelled in the quality of deliverables.
  • Validated the data during UAT testing.
  • Performing source to target Mapping.
  • Involved in Metadata management, where all the table specifications were listed and implemented the same in Ab Initio metadata hub as per data governance.
  • Developed Korn Shell scripts to parallel extract and process data from different sources simultaneously to streamline performance and improve execution time in a parallel process for better time, resource management and efficiency.
  • Used Teradata utilities such as TPT (Teradata Parallel Transporter), FLOAD (Fastload) and MLOAD (Multiload) for handling various tasks.
  • Developed Logical data model using Erwin and created physical data models using forward engineering.

Environment: Erwin 8.0, Teradata 13, TOAD, Oracle 10g/11g, MS SQL Server 2008, Teradata SQL Assistant, XML Files, Flat files

Confidential - Columbus, GA

Sr. Data Analyst/Data Modeler

Responsibilities:

  • Involved with Data profiling for multiple sources and answered complex business questions by providing data to business users.
  • Developed data transformation and cleansing rules for migration using ETL tools.
  • Performed data mining on data using very complex SQL queries and discovered pattern.
  • Created DML code and statements for underlying & impacting databases.
  • Created dimensional model for the reporting system by identifying required dimensions and facts using Erwin.
  • Responsible for different Data mapping activities from Source systems to SQL Server.
  • Assisted in the oversight for compliance to the Enterprise Data Standards
  • Worked in importing and cleansing of data from various sources like Oracle, flat files, SQL Server with high volume data
  • Performed Data Analysis and data profiling using complex SQL on various sources systems including Oracle 10g and SQL Server2008.
  • Maintained warehouse metadata, naming standards and warehouse standards for future application development.
  • Data analysis of existing data base to understand the data flow and business rule applied to different data bases by SQL
  • Consolidated existing SSRS reports to Micro Strategy reports without losing the functionality.
  • Created stored procedures for extracting XML Message, simulating data for carrying out Post session loads. Using SQL Servers Enterprise manager for writing Queries and exploring data.
  • Created stored procedures to populate sample data and carrying out test load with multi- terabyte databases.
  • Developed detailed ER diagram and data flow diagram using modeling tools following the SDLC structure.
  • Implemented the Data management strategies across the various database domains and products with respect to each business unit.
  • Involved with Data Extraction, Modification, Validation, Analysis, Management and Reporting.
  • Worked very close with Data Architectures and DBA team to implement data model changes in database in all environments.

Environment: ERWIN 8.0, MS SQL Server 2008, SSIS, SSRS, SQL Developer, SQL, Oracle 10g, Flat Files, MS Excel. MS Excel 2007, PL/SQL, XML, My SQL

Confidential - Houston, TX

Data Analyst/Data Modeler

Responsibilities:

  • Worked as a Data Modeler/Data Analyst to generate Data Models using Erwin and developed relational database system.
  • Designed and developed Use Cases, Activity Diagrams, Sequence Diagrams, OOD (Object oriented Design) using UML and Visio.
  • Extensively used SQL for Data Analysis and to understand and documenting the data behavior within and between two companies.
  • Worked extensively with Business Objects Report developers in creating data marts and develop reports to cater the existing business needs.
  • Designed Mapping Documents and Mapping Templates for SSIS and Informatica ETL developers.
  • Implemented the Slowly changing dimension scheme for most of the dimensions.
  • Created data masking mappings to mask the sensitive data between production and test environment.
  • Extensively use of Triggers, Table Spaces, Pre/Post SQL, Sequences, Materialized Views, Procedures and Packages in Data Models.
  • Conduct Design discussions and meetings to come out with the appropriate Data Mart at the lowest level of grain for each of the Dimensions involved
  • Designed a STAR schema for the detailed data marts and Plan data marts involving shared dimensions (Conformed).
  • Worked on Informatica Data Quality to resolve customers address related issues.
  • Performed Data Analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata, to ensure accuracy of the data between the warehouse and source systems.
  • Conducted meetings with the business and technical team to gather necessary analytical data requirements in JAD sessions.
  • Partitioned tables based on the requirement and developed purging and archival rules.
  • Creating Data mappings, Tech Design, loading strategies for ETL to load newly created or existing tables.
  • Used forward engineering to create a physical data model with DDL that best suits the requirements from the Logical Data Model.
  • Involved in preparing the design flow for the Data stage objects to pull the data from various upstream applications and do the required transformations and load the data into various downstream applications
  • Worked on data profiling and data validation to ensure the accuracy of the data between the warehouse and source systems.
  • Helped the testing team in creating the test plans and test scripts. Assisted the users in UAT testing by providing test scenarios and test data.

Environment: Erwin 7.0, Oracle 9i, Teradata 12, SQL, PL/SQL, Informatica, OOD, MS-ACCESS, MS Excel, MS Visio, OOD, SSIS

We'd love your feedback!