We provide IT Staff Augmentation Services!

Sr. Data Architect/data Modeler Resume

2.00/5 (Submit Your Rating)

Union, NJ

SUMMARY:

  • Over 11+ years of extensive experience in Data Analysis, Data Modeling, System Analysis, Data Architecture and Development, Testing and Deployment of business applications
  • Experienced in Data Modeling & Data Analysis experience using Dimensional Data Modeling and Relational Data Modeling, Star Schema/Snowflake Modeling, FACT & Dimensions tables, Physical & Logical Data Modeling.
  • Experience with distributed data warehousing and/or data mining systems, using one or more Big Data/NoSQL technologies (Hadoop, Hive, HBase, Pig, Cassandra, MongoDB)
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
  • Experience in Technical consulting and end - to-end delivery with architecture, data modeling, data governance and design - development - implementation of solutions.
  • Extensive experience in developing and driving strategic direction of SAP operating system (SAP ECC) and SAP business intelligence (SAP BI) system
  • Experience in designing, building and implementing complete Hadoop ecosystem comprising of Map Reduce, HDFS, Hive, Impala, Pig, Sqoop, Oozie, HBase, MongoDB, and Spark.
  • Experience in various Teradata utilities like Fastload, Multiload, BTEQ, and Teradata SQL Assistant.
  • Experience in working with Business Intelligence and Enterprise Data Warehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum, Amazon Redshift and Azure Data Warehouse
  • Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
  • Experienced in Data Scrubbing/Cleansing, Data Quality, Data Mapping, Data Profiling, Data Validation in ETL
  • Experienced in creating and documenting Metadata for OLTP and OLAP when designing a systems.
  • Performeddataanalysis anddataprofiling using complex SQL on various sources systems including Oracle and Teradata.
  • Excellent Knowledge of Ralph Kimball and BillInmon's approaches toDataWarehousing.
  • Excellent in performing data transfer activities between SAS and various databases and data file formats like XLS, CSV, DBF, MDB etc.
  • Experience in working with Excel Pivot and VBA macros for various business scenarios.
  • Experience in Data modeling using ER diagram, Dimensional data modeling, Conceptual/Logical/Physical Modeling using 3NormalForm (3NF), Star Schema modeling, Snowflake modeling using tools like ER/Studio, CA Erwin, Sybase Power Designer for both forward and reverse engineering.
  • Expert in BI reporting and Data Reporting tools like Pentaho and SAP BI.
  • Excellent experience in troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Proficient withDataAnalysis, mapping source and target systems fordatamigration efforts and resolving issues relating todatamigration.
  • Excellent understanding and working experience of industry standard methodologies like System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE Methodologies.
  • Advanced expertise in Spotfire visualization design and development, SQL and Oracle/SQL Server.
  • Expert in building Enterprise Data Warehouse or Data warehouse appliances from Scratch using both Kimball and Inmon Approach.
  • Experience in various databases Design of development and Production environment involving Oracle, SQL server, Netezza, MY SQL, DB2, MS Access, Teradata etc.
  • Hands on experience in Normalization & De-normalization techniques design considerations upto 3NF Form for OLTP Databases and Models.
  • Strong experience in writing SQL and PL/SQL, Transact SQL programs for Stored Procedures, Triggers and Functions.
  • Expertise in analyzing and documenting business requirement documents (BRD) and functional requirement documents (FRD) along with Use Case Modeling and UML.
  • Experience in UNIX shell scripting, Perl scripting and automation of ETL Processes.
  • Expertise in designing complex Mappings and have expertise in performance tuning and slowly-changing Dimension Tables and Fact tables
  • Extensively worked with Teradata utilities BTEQ, Fast Export, and Multi Load to export and load data to/from different source systems including flat files.

TECHNICAL SKILLS:

DataModeling Tools: Erwin 9.6/9.5, ER/Studio 9.7/9.0, Sybase Power DesignerBig Data Technologies: Hadoop, Hive, HDFS, HBase, Flume, Sqoop, Spark, Pig, Impala, MapReduce.

Programming Languages: SQL, PL/SQL, UNIX shell Scripting, PERL, AWK, SED

Databases: Oracle 12c/11g, Teradata R15/R14, MS SQL Server 2014/2016Testing and defect tracking Tools: HP/Mercury (Quality Center, Win Runner, Quick Test Professional, Performance Center, Requisite, MS Visio & Visual Source Safe

Operating System: Windows, Unix, Sun Solaris

ETL/Datawarehouse Tools: Informatica 9.6/9.1, SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau, Pentaho

Tools & Software: TOAD, MS Office, BTEQ, Teradata SQL Assistant

Other Tools: Teradata SQL Assistant, Toad 9.7/8.0, DB Visualizer 6.0, Microsoft Office, Microsoft Visio, Microsoft Excel, Microsoft Project

Project Execution Methodologies: Ralph Kimball and BillInmondatawarehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)

PROFESSIONAL EXPERIENCE

Confidential, Union, NJ

Sr. Data Architect/Data Modeler

Responsibilities:

  • Provide data architecture support to enterprise data management efforts, such as the development of the enterprise data model, Master and Reference data.
  • Lead the strategy, architecture and process improvements for data architecture and data management, balancing long and short-term needs of the business.
  • Implemented logical and physical relational database and maintained Database Objects in the data model using Erwin 9.6
  • Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
  • Participated in integration of MDM (MasterDataManagement) Hub anddatawarehouses.
  • Responsible of designing the whole data warehouse system including ODS, DWH and data marts.
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
  • Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
  • Used Normalization methods up to 3NF and De-normalization techniques for effective performance in OLTP systems.
  • Extensively using Agile methodology as the Organization Standard to implement thedataModels.
  • Performed theDataMapping,Datadesign(DataModeling) to integrate thedataacross the multiple databases in to EDW.
  • Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Used Load utilities (Fast Load & Multi Load) with the mainframe interface to load the data into Teradata.
  • Data reconciliation activities between Source and EDW Teradata databases.
  • Changed the session properties to override setting values included source/target table name, schema name, source query, connection string, update strategy and log detail tracing level.
  • Involved in designing and developing Data Models and Data Marts that support the Business Intelligence Data Warehouse.
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
  • Involved in Planning, Defining and Designing data base using Erwin on business requirement and provided documentation.
  • Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections.
  • Designed and created queries to extract the desired data and runs the report. Formatted the data and validated the results.
  • Transforming staging area data into a STAR schema (hosted on AmazonRedshift) which was then used for developing embedded Tableau dashboards
  • Translating business requirements into SAS code for use within internal systems and models
  • Building relationships and trust with key stakeholders to support program delivery and adoption of enterprise architecture.
  • Maintained metadata(datadefinitions of table structures) and version controlling for thedatamodel.
  • Created stored procedures, functions, database triggers and packages as per the business needs for developing ad-hoc and robust reports.
  • Defined best practices fordatamodeling and extraction and ensure architectural alignment of the designs and development.
  • Used Spark Data frames, Spark-SQL, Spark MLLib extensively and developing and designing POC's using Scala, Spark SQL and MLlib libraries.
  • Working in Hadoop Environment using pig, Sqoop, Hive, HBase and detailed understanding of map reduce programs
  • Involved in integration of various relational and non-relational sources such as Oracle, XML and Flat Files.
  • Developed SQL scripts for creating tables, Sequences, Triggers, views and materialized views.
  • Scheduling theSSISpackages execution in SQL Server Agent and tracking the success or failure of the execution ofSSISPackages and configuring the email notifications through SQL Server Agent.
  • Developed multiple MapReduce jobs in java for Data Cleaning and pre-processing analyzing data in PIG.
  • Analyze existing source system with the help ofDataProfiling and source system datamodels thus creating individualdatamodels for various domains/subject areas for the proposeddatawarehouse solution.
  • Worked in Data Analysis, data profiling and data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, and Oracle)
  • Coordinating with DBA team to implement physical models & to setup development, test, staging & production environments for DDL & DML Generation & usage.
  • Used Teradata OLAP functions like RANK, ROW NUMBER, QUALIFY, CSUM and SAMPLE.

Environment: Erwin 9.6, MDM, AWS, Hadoop, Spark, Sqoop, Hive, HDFS, OLAP, OLTP, ODS, PL/SQL, Tableau, Teradata R15, Teradata SQL Assistant 15.0, Oracle 12c, SQL, PL/SQL, Flat Files.

Confidential, Chevy Chase, MD

Sr. Data Architect/Data Modeler

Responsibilities:

  • Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
  • Participated in the design, development, and support of the corporate operation data store and enterprise data warehouse database environment.
  • Designed the Logical Data Model using ER/Studio 9.0 with the entities and attributes for each subject areas.
  • Used the Agile Scrum methodology to build the different phases of Software development life cycle.
  • Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
  • Independently coded new programs and design Tables to load and test the program effectively for the given POC's using Big Data/Hadoop.
  • Translating high level business requirements into solution, infrastructure architectures, involved enterprise architects to align strategic view.
  • Worked on ER Studio for multiple Operations across in both OLAP and OLTP applications.
  • Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Oracle, and DB2 database
  • Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Responsible for full data loads from production to AWS Redshift staging environment.
  • Designed the Fact and Dimension table for Data Marts using ERStudio 9.0.
  • Design of ODS layer, Dimensional modeling using Kimball Methodologies, of the Data Warehouse sourced from MDM Base tables and other Transactional systems.
  • Extensively used Aginity Netezza work bench to perform various DML, DDL etc operations on Netezza database.
  • Worked on migrating of EDW to AWS using EMR and various other technologies.
  • Designed and produced client reports using Excel, Access, Tableau and SAS.
  • Created logical and physical data model using Cassandra’s model
  • Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
  • Extracting Mega Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports
  • Worked with the ETL team to document the transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
  • Worked with Netezza and Oracle databases and implemented various logical and physical data models for them.
  • Generated comprehensive analytical reports by running SQL queries against current databases to conductdataanalysis.
  • Developed complex SQL scripts for Teradata database for creating BI layer on DW for Tableau reporting.
  • Worked on debugging and identifying the unexpected real-time issues in the production serverSSIS packages.
  • Involved in mapping the data elements from the User Interface to the Database and help identify the gaps.
  • Prepared ofDataDictionary/MetaDataof thedataelements.
  • Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Generated preliminary Source to Target mapping by writing a ER/Studio Macro in the ER/Studio VBA platform document and handed over to BAs for further transformation analysis.
  • Submitted the DDL and Security document by writing a ERStudio Macro to the SVN repository and communicated to the DBA for implementation.
  • Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
  • Designed and Developed Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing
  • Extensively worked on documentation of Data Model, Mapping Transformation and Scheduling jobs.
  • Worked extensively with Business Objects XI Report Developers in solving critical issues of defining hierarchy, loops and Contexts.

Environment: ER/Studio 9.0, Oracle 11g, DB2, MDM AWS,, Hadoop, Hive, Netezza, Sqoop, Spark, HDFS, MapReduce, OLAP, OLTP, PL/SQL, Flat Files.

Confidential - New York, Ny

Sr. Data Architect/Data Modeler

Responsibilities:

  • Worked extensively with Micro strategy report-developers in creating data marts and development of reports (Operational and Regulatory).
  • Coordinating the data execution and loading in ETL with the Big Data Framework like HDFS, Hive, HBase etc.
  • Involved in complete end to end Software Development lifecycle (SDLC) of projects including requirement analysis, effort estimation, Capacity Planning, design, development, unit-test, system
  • Used SQL to extract data from different databases such as Oracle and Amazon Redshift.
  • Created a high-level generalized data model to convert it into logical and physical model at later stages of the project using Erwin.
  • Worked on a MapReduce Hadoop platform to implement Big data solutions using Hive, MapReduce, shell scripting and Pig.
  • Worked on designing the whole data warehouse architecture from scratch, from ODS to data marts
  • Worked on the enhancement of architecture of existing Data warehouse and Application databases
  • Performed Extracting, Transforming and Loading (ETL) data from Excel, Flat file, Oracle to MS SQL Server by using SSIS services
  • Performed Data analysis, statistical analysis, generated reports, listings and graphs using SAS Tools-SAS/Base, SAS/Macros and SAS/Graph, SAS/SQL, SAS/Connect, SAS/Access.
  • Worked with various Teradata14 tools and utilities like Teradata Viewpoint, Multi Load, ARC, Teradata Administrator, BTEQ
  • Involved in several facets of MDM implementations including Data Profiling, metadata acquisition and data migration.
  • Worked extensively in both Forward Engineering as well as Reverse Engineering usingdatamodeling tools.
  • Involved in the creation, maintenance ofDataWarehouse and repositories containing Metadata.
  • Resolved thedatatype inconsistencies between the source systems and the target system using the Mapping Documents and analyzing the database using SQL queries.
  • Extensively used both Star Schema and Snow flake schema methodologies in building and designing the logicaldatamodel in both Type1 and Type2Dimensional Models.
  • Worked with DBA group to create Best-Fit PhysicalDataModel from the LogicalDataModel using Forward Engineering.
  • DevelopedDataMigration and Cleansing rules for the Integration Architecture(OLTP, ODS, DW).
  • Enabled theSSISpackage configuration to make the flexibility to pass the connection strings to connection managers and values to package variables explicitly based on environments.
  • Tuned SQL statements and analysis query performance issues in Teradata
  • Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata14.1 Database.
  • Created PL/SQL Tables, Collections, Records and Partitions.
  • Used Dynamic SQL and triggers for faster data access and to incorporate business logic.
  • Identified the reference code and profiled them into the format of reference value to be used in the ODS model and future reference value lookup model.
  • Used Erwin model mart for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
  • Used Python scripts to update the content in database and manipulate files
  • Write complex SQL queries in ANSI SQL and PostgreSQL
  • Coordinated with DBA's and generated SQL code from the data models and generated DDL scripts using ERWIN.
  • Developed SQL Queries to fetch complexdatafrom different tables in remote databases using joins, database links and Bulk collects.

Environment: ERWIN r9.5, Netezza, Oracle11g, Taradata14.1, T-SQL, DB2, SSIS, LINUX, MDM, PL/SQL, Excel, MS Access, ETL, Hadoop, OLTP and OLAP, SAP, OLAP, OLTP, ODS, Python.

Confidential - Chicago, IL

Sr. Data Analyst/Data Modeler

Responsibilities:

  • Created Physical Data Analyst from the Logical Data Analyst using Compare and Merge Utility in ER Studio and worked with the naming standards utility.
  • Developed normalized Logical and Physical database models for designing an OLTP application.
  • Extensively used Star Schema methodologies in building and designing the logicaldatamodel into Dimensional Models
  • Creation of database objects like tables, views, Materialized views, procedures, packages using Oracle tools like PL/SQL, SQL*Loader and Handled Exceptions.
  • Enforced referential integrity in the OLTP data model for consistent relationship between tables and efficient database design.
  • Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
  • Utilized SDLC and Agile methodologies such as SCRUM.
  • Involved in administrative tasks, including creation of database objects such as database, tables, and views, using SQL, DDL, and DML requests.
  • Worked on Data Analysis, Data profiling, and Data Modeling, data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
  • Loaded multi format data from various sources like flat-file, Excel, MS Access and performing file system operation.
  • Used T-SQL stored procedures to transfer data from OLTP databases to staging area and finally transfer into data marts.
  • Worked on Physical design for both SMP and MPP RDBMS, with understanding of RDMBS scaling features.
  • Wrote SQL Queries, Dynamic-queries, sub-queries and complex joins for generating Complex Stored Procedures, Triggers, User-defined Functions, Views and Cursors.
  • Wrote simple and advanced SQL queries and scripts to create standard and ad hoc reports for senior managers.
  • Performed ETL SQL optimization designed OLTP system environment and maintained documentation of Metadata.
  • Involved withDataAnalysis primarily IdentifyingDataSets, SourceData, Source MetaData, Data Definitions andDataFormats
  • Worked with developers on data Normalization and De-normalization, performance tuning issues, and provided assistance in stored procedures as needed.
  • Used Teradata for OLTP systems by generating models to support Revenue Management Applications that connect to SAS.
  • Created SSIS Packages for import and export of data between Oracle database and others like MS Excel and Flat Files.
  • Worked in the capacity of ETL Developer (Oracle Data Integrator (ODI) / PL/SQL) to migrate data from different sources in to target Oracle Data Warehouse.
  • Designed and Developed PL/SQL procedures, functions and packages to create Summary tables.
  • Involved in creating tasks to pull and push data from Salesforce to Oracle Staging/Data Mart.
  • Created VBA Macros to convert the Excel Input files in to correct format and loaded them to SQL Server.
  • Helped the BI, ETL Developers in understanding the Data Model, data flow and the expected output for each model created

Environment: ER/Studio 8.0, Oracle 10g Application Server, Oracle Developer Suite, PL/SQL, T-SQL, SQL Plus, SSIS, Teradata 13, OLAP, OLTP, SAS, MS Excel.

Confidential, NJ

Data Analyst/Data Modeler

Responsibilities:

  • Gathered Business requirements by organizing and managing meetings with business stake holders, Application architects, Technical architects and IT analysts on a scheduled basis
  • Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
  • Performed Normalization, Indexes Tuned and Optimized the existing database design.
  • Performed reverse engineering using the Erwin Data Modeling tool.
  • Performed various tasks that included data massaging, data extraction from flat files, data transformations into various file formats, validations, loading data and checking for any data integrity errors.
  • Responsible for the development and maintenance of Logical and Physical data models, along with corresponding metadata, to support Applications.
  • Conducted meetings with the business and technical team to gather necessary analytical data requirements in JAD sessions.
  • Analyzed all existing SSIS packages, SQL Server objects & new functional specs.
  • Identified and tracked the slowly changing dimensions, heterogeneous sources an determined the hierarchies in dimensions.
  • Created Mappings, Tech Design, loading strategies for ETL to load newly created or existing tables.
  • Worked extensively with MicroStrategy Report developers in creating data marts and develop reports
  • Created databases for OLAP Metadata catalog tables using forward engineering of models in Erwin.
  • Enforced referential integrity in the OLTPdatamodel for consistent relationship between tables and efficient database design.
  • Involved in importing/exporting large amounts ofdatafrom files to Teradata and vice versa.
  • Designed Mapping Documents and Mapping Templates for Data Stage ETL developer.
  • Designed, Build the Dimensions, cubes with Star schema and Snow Flake Schema using SQL Server Analysis Services (SSAS).
  • Wrote PL/SQL statement, stored procedures and Triggers in DB2 for extracting as well as writing data.
  • Created a Data Mapping document after each assignment and wrote the transformation rules for each field as applicable
  • Analyzed OLTP source systems and Operational Data Store and research the tables/entities required for the project. Designing the measures, dimensions and facts matrix document for the ease while designing.
  • Tuning all database via indexing of tables, MS SQL Server 2005 configuration parameters and stored procedures SQL code optimization.
  • Used forward engineering to create a physical data model with DDL that best suits the requirements from the Logical Data Model.

Environment: Erwin 7.3, SSIS, Metadata, OLAP, OLTP, Data Mart, PL/SQL, T-SQL, DB2, SQL Plus, Microsoft SQL Server 2005, Flat files.

We'd love your feedback!