Sr. Data Architect/data Modeler Resume
Union, NJ
SUMMARY:
- Over 11+ years of extensive experience in Data Analysis, Data Modeling, System Analysis, Data Architecture and Development, Testing and Deployment of business applications
- Experienced in Data Modeling & Data Analysis experience using Dimensional Data Modeling and Relational Data Modeling, Star Schema/Snowflake Modeling, FACT & Dimensions tables, Physical & Logical Data Modeling.
- Experience with distributed data warehousing and/or data mining systems, using one or more Big Data/NoSQL technologies (Hadoop, Hive, HBase, Pig, Cassandra, MongoDB)
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
- Experience in Technical consulting and end - to-end delivery with architecture, data modeling, data governance and design - development - implementation of solutions.
- Extensive experience in developing and driving strategic direction of SAP operating system (SAP ECC) and SAP business intelligence (SAP BI) system
- Experience in designing, building and implementing complete Hadoop ecosystem comprising of Map Reduce, HDFS, Hive, Impala, Pig, Sqoop, Oozie, HBase, MongoDB, and Spark.
- Experience in various Teradata utilities like Fastload, Multiload, BTEQ, and Teradata SQL Assistant.
- Experience in working with Business Intelligence and Enterprise Data Warehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum, Amazon Redshift and Azure Data Warehouse
- Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
- Experienced in Data Scrubbing/Cleansing, Data Quality, Data Mapping, Data Profiling, Data Validation in ETL
- Experienced in creating and documenting Metadata for OLTP and OLAP when designing a systems.
- Performeddataanalysis anddataprofiling using complex SQL on various sources systems including Oracle and Teradata.
- Excellent Knowledge of Ralph Kimball and BillInmon's approaches toDataWarehousing.
- Excellent in performing data transfer activities between SAS and various databases and data file formats like XLS, CSV, DBF, MDB etc.
- Experience in working with Excel Pivot and VBA macros for various business scenarios.
- Experience in Data modeling using ER diagram, Dimensional data modeling, Conceptual/Logical/Physical Modeling using 3NormalForm (3NF), Star Schema modeling, Snowflake modeling using tools like ER/Studio, CA Erwin, Sybase Power Designer for both forward and reverse engineering.
- Expert in BI reporting and Data Reporting tools like Pentaho and SAP BI.
- Excellent experience in troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
- Proficient withDataAnalysis, mapping source and target systems fordatamigration efforts and resolving issues relating todatamigration.
- Excellent understanding and working experience of industry standard methodologies like System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE Methodologies.
- Advanced expertise in Spotfire visualization design and development, SQL and Oracle/SQL Server.
- Expert in building Enterprise Data Warehouse or Data warehouse appliances from Scratch using both Kimball and Inmon Approach.
- Experience in various databases Design of development and Production environment involving Oracle, SQL server, Netezza, MY SQL, DB2, MS Access, Teradata etc.
- Hands on experience in Normalization & De-normalization techniques design considerations upto 3NF Form for OLTP Databases and Models.
- Strong experience in writing SQL and PL/SQL, Transact SQL programs for Stored Procedures, Triggers and Functions.
- Expertise in analyzing and documenting business requirement documents (BRD) and functional requirement documents (FRD) along with Use Case Modeling and UML.
- Experience in UNIX shell scripting, Perl scripting and automation of ETL Processes.
- Expertise in designing complex Mappings and have expertise in performance tuning and slowly-changing Dimension Tables and Fact tables
- Extensively worked with Teradata utilities BTEQ, Fast Export, and Multi Load to export and load data to/from different source systems including flat files.
TECHNICAL SKILLS:
DataModeling Tools: Erwin 9.6/9.5, ER/Studio 9.7/9.0, Sybase Power DesignerBig Data Technologies: Hadoop, Hive, HDFS, HBase, Flume, Sqoop, Spark, Pig, Impala, MapReduce.
Programming Languages: SQL, PL/SQL, UNIX shell Scripting, PERL, AWK, SED
Databases: Oracle 12c/11g, Teradata R15/R14, MS SQL Server 2014/2016Testing and defect tracking Tools: HP/Mercury (Quality Center, Win Runner, Quick Test Professional, Performance Center, Requisite, MS Visio & Visual Source Safe
Operating System: Windows, Unix, Sun Solaris
ETL/Datawarehouse Tools: Informatica 9.6/9.1, SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau, Pentaho
Tools & Software: TOAD, MS Office, BTEQ, Teradata SQL Assistant
Other Tools: Teradata SQL Assistant, Toad 9.7/8.0, DB Visualizer 6.0, Microsoft Office, Microsoft Visio, Microsoft Excel, Microsoft Project
Project Execution Methodologies: Ralph Kimball and BillInmondatawarehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
PROFESSIONAL EXPERIENCE
Confidential, Union, NJ
Sr. Data Architect/Data Modeler
Responsibilities:
- Provide data architecture support to enterprise data management efforts, such as the development of the enterprise data model, Master and Reference data.
- Lead the strategy, architecture and process improvements for data architecture and data management, balancing long and short-term needs of the business.
- Implemented logical and physical relational database and maintained Database Objects in the data model using Erwin 9.6
- Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
- Participated in integration of MDM (MasterDataManagement) Hub anddatawarehouses.
- Responsible of designing the whole data warehouse system including ODS, DWH and data marts.
- Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
- Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
- Used Normalization methods up to 3NF and De-normalization techniques for effective performance in OLTP systems.
- Extensively using Agile methodology as the Organization Standard to implement thedataModels.
- Performed theDataMapping,Datadesign(DataModeling) to integrate thedataacross the multiple databases in to EDW.
- Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
- Used Load utilities (Fast Load & Multi Load) with the mainframe interface to load the data into Teradata.
- Data reconciliation activities between Source and EDW Teradata databases.
- Changed the session properties to override setting values included source/target table name, schema name, source query, connection string, update strategy and log detail tracing level.
- Involved in designing and developing Data Models and Data Marts that support the Business Intelligence Data Warehouse.
- Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
- Involved in Planning, Defining and Designing data base using Erwin on business requirement and provided documentation.
- Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections.
- Designed and created queries to extract the desired data and runs the report. Formatted the data and validated the results.
- Transforming staging area data into a STAR schema (hosted on AmazonRedshift) which was then used for developing embedded Tableau dashboards
- Translating business requirements into SAS code for use within internal systems and models
- Building relationships and trust with key stakeholders to support program delivery and adoption of enterprise architecture.
- Maintained metadata(datadefinitions of table structures) and version controlling for thedatamodel.
- Created stored procedures, functions, database triggers and packages as per the business needs for developing ad-hoc and robust reports.
- Defined best practices fordatamodeling and extraction and ensure architectural alignment of the designs and development.
- Used Spark Data frames, Spark-SQL, Spark MLLib extensively and developing and designing POC's using Scala, Spark SQL and MLlib libraries.
- Working in Hadoop Environment using pig, Sqoop, Hive, HBase and detailed understanding of map reduce programs
- Involved in integration of various relational and non-relational sources such as Oracle, XML and Flat Files.
- Developed SQL scripts for creating tables, Sequences, Triggers, views and materialized views.
- Scheduling theSSISpackages execution in SQL Server Agent and tracking the success or failure of the execution ofSSISPackages and configuring the email notifications through SQL Server Agent.
- Developed multiple MapReduce jobs in java for Data Cleaning and pre-processing analyzing data in PIG.
- Analyze existing source system with the help ofDataProfiling and source system datamodels thus creating individualdatamodels for various domains/subject areas for the proposeddatawarehouse solution.
- Worked in Data Analysis, data profiling and data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, and Oracle)
- Coordinating with DBA team to implement physical models & to setup development, test, staging & production environments for DDL & DML Generation & usage.
- Used Teradata OLAP functions like RANK, ROW NUMBER, QUALIFY, CSUM and SAMPLE.
Environment: Erwin 9.6, MDM, AWS, Hadoop, Spark, Sqoop, Hive, HDFS, OLAP, OLTP, ODS, PL/SQL, Tableau, Teradata R15, Teradata SQL Assistant 15.0, Oracle 12c, SQL, PL/SQL, Flat Files.
Confidential, Chevy Chase, MD
Sr. Data Architect/Data Modeler
Responsibilities:
- Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
- Participated in the design, development, and support of the corporate operation data store and enterprise data warehouse database environment.
- Designed the Logical Data Model using ER/Studio 9.0 with the entities and attributes for each subject areas.
- Used the Agile Scrum methodology to build the different phases of Software development life cycle.
- Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
- Independently coded new programs and design Tables to load and test the program effectively for the given POC's using Big Data/Hadoop.
- Translating high level business requirements into solution, infrastructure architectures, involved enterprise architects to align strategic view.
- Worked on ER Studio for multiple Operations across in both OLAP and OLTP applications.
- Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Oracle, and DB2 database
- Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
- Responsible for full data loads from production to AWS Redshift staging environment.
- Designed the Fact and Dimension table for Data Marts using ERStudio 9.0.
- Design of ODS layer, Dimensional modeling using Kimball Methodologies, of the Data Warehouse sourced from MDM Base tables and other Transactional systems.
- Extensively used Aginity Netezza work bench to perform various DML, DDL etc operations on Netezza database.
- Worked on migrating of EDW to AWS using EMR and various other technologies.
- Designed and produced client reports using Excel, Access, Tableau and SAS.
- Created logical and physical data model using Cassandra’s model
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
- Extracting Mega Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports
- Worked with the ETL team to document the transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
- Worked with Netezza and Oracle databases and implemented various logical and physical data models for them.
- Generated comprehensive analytical reports by running SQL queries against current databases to conductdataanalysis.
- Developed complex SQL scripts for Teradata database for creating BI layer on DW for Tableau reporting.
- Worked on debugging and identifying the unexpected real-time issues in the production serverSSIS packages.
- Involved in mapping the data elements from the User Interface to the Database and help identify the gaps.
- Prepared ofDataDictionary/MetaDataof thedataelements.
- Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
- Generated preliminary Source to Target mapping by writing a ER/Studio Macro in the ER/Studio VBA platform document and handed over to BAs for further transformation analysis.
- Submitted the DDL and Security document by writing a ERStudio Macro to the SVN repository and communicated to the DBA for implementation.
- Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
- Designed and Developed Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing
- Extensively worked on documentation of Data Model, Mapping Transformation and Scheduling jobs.
- Worked extensively with Business Objects XI Report Developers in solving critical issues of defining hierarchy, loops and Contexts.
Environment: ER/Studio 9.0, Oracle 11g, DB2, MDM AWS,, Hadoop, Hive, Netezza, Sqoop, Spark, HDFS, MapReduce, OLAP, OLTP, PL/SQL, Flat Files.
Confidential - New York, Ny
Sr. Data Architect/Data Modeler
Responsibilities:
- Worked extensively with Micro strategy report-developers in creating data marts and development of reports (Operational and Regulatory).
- Coordinating the data execution and loading in ETL with the Big Data Framework like HDFS, Hive, HBase etc.
- Involved in complete end to end Software Development lifecycle (SDLC) of projects including requirement analysis, effort estimation, Capacity Planning, design, development, unit-test, system
- Used SQL to extract data from different databases such as Oracle and Amazon Redshift.
- Created a high-level generalized data model to convert it into logical and physical model at later stages of the project using Erwin.
- Worked on a MapReduce Hadoop platform to implement Big data solutions using Hive, MapReduce, shell scripting and Pig.
- Worked on designing the whole data warehouse architecture from scratch, from ODS to data marts
- Worked on the enhancement of architecture of existing Data warehouse and Application databases
- Performed Extracting, Transforming and Loading (ETL) data from Excel, Flat file, Oracle to MS SQL Server by using SSIS services
- Performed Data analysis, statistical analysis, generated reports, listings and graphs using SAS Tools-SAS/Base, SAS/Macros and SAS/Graph, SAS/SQL, SAS/Connect, SAS/Access.
- Worked with various Teradata14 tools and utilities like Teradata Viewpoint, Multi Load, ARC, Teradata Administrator, BTEQ
- Involved in several facets of MDM implementations including Data Profiling, metadata acquisition and data migration.
- Worked extensively in both Forward Engineering as well as Reverse Engineering usingdatamodeling tools.
- Involved in the creation, maintenance ofDataWarehouse and repositories containing Metadata.
- Resolved thedatatype inconsistencies between the source systems and the target system using the Mapping Documents and analyzing the database using SQL queries.
- Extensively used both Star Schema and Snow flake schema methodologies in building and designing the logicaldatamodel in both Type1 and Type2Dimensional Models.
- Worked with DBA group to create Best-Fit PhysicalDataModel from the LogicalDataModel using Forward Engineering.
- DevelopedDataMigration and Cleansing rules for the Integration Architecture(OLTP, ODS, DW).
- Enabled theSSISpackage configuration to make the flexibility to pass the connection strings to connection managers and values to package variables explicitly based on environments.
- Tuned SQL statements and analysis query performance issues in Teradata
- Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata14.1 Database.
- Created PL/SQL Tables, Collections, Records and Partitions.
- Used Dynamic SQL and triggers for faster data access and to incorporate business logic.
- Identified the reference code and profiled them into the format of reference value to be used in the ODS model and future reference value lookup model.
- Used Erwin model mart for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Used Python scripts to update the content in database and manipulate files
- Write complex SQL queries in ANSI SQL and PostgreSQL
- Coordinated with DBA's and generated SQL code from the data models and generated DDL scripts using ERWIN.
- Developed SQL Queries to fetch complexdatafrom different tables in remote databases using joins, database links and Bulk collects.
Environment: ERWIN r9.5, Netezza, Oracle11g, Taradata14.1, T-SQL, DB2, SSIS, LINUX, MDM, PL/SQL, Excel, MS Access, ETL, Hadoop, OLTP and OLAP, SAP, OLAP, OLTP, ODS, Python.
Confidential - Chicago, IL
Sr. Data Analyst/Data Modeler
Responsibilities:
- Created Physical Data Analyst from the Logical Data Analyst using Compare and Merge Utility in ER Studio and worked with the naming standards utility.
- Developed normalized Logical and Physical database models for designing an OLTP application.
- Extensively used Star Schema methodologies in building and designing the logicaldatamodel into Dimensional Models
- Creation of database objects like tables, views, Materialized views, procedures, packages using Oracle tools like PL/SQL, SQL*Loader and Handled Exceptions.
- Enforced referential integrity in the OLTP data model for consistent relationship between tables and efficient database design.
- Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
- Utilized SDLC and Agile methodologies such as SCRUM.
- Involved in administrative tasks, including creation of database objects such as database, tables, and views, using SQL, DDL, and DML requests.
- Worked on Data Analysis, Data profiling, and Data Modeling, data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
- Loaded multi format data from various sources like flat-file, Excel, MS Access and performing file system operation.
- Used T-SQL stored procedures to transfer data from OLTP databases to staging area and finally transfer into data marts.
- Worked on Physical design for both SMP and MPP RDBMS, with understanding of RDMBS scaling features.
- Wrote SQL Queries, Dynamic-queries, sub-queries and complex joins for generating Complex Stored Procedures, Triggers, User-defined Functions, Views and Cursors.
- Wrote simple and advanced SQL queries and scripts to create standard and ad hoc reports for senior managers.
- Performed ETL SQL optimization designed OLTP system environment and maintained documentation of Metadata.
- Involved withDataAnalysis primarily IdentifyingDataSets, SourceData, Source MetaData, Data Definitions andDataFormats
- Worked with developers on data Normalization and De-normalization, performance tuning issues, and provided assistance in stored procedures as needed.
- Used Teradata for OLTP systems by generating models to support Revenue Management Applications that connect to SAS.
- Created SSIS Packages for import and export of data between Oracle database and others like MS Excel and Flat Files.
- Worked in the capacity of ETL Developer (Oracle Data Integrator (ODI) / PL/SQL) to migrate data from different sources in to target Oracle Data Warehouse.
- Designed and Developed PL/SQL procedures, functions and packages to create Summary tables.
- Involved in creating tasks to pull and push data from Salesforce to Oracle Staging/Data Mart.
- Created VBA Macros to convert the Excel Input files in to correct format and loaded them to SQL Server.
- Helped the BI, ETL Developers in understanding the Data Model, data flow and the expected output for each model created
Environment: ER/Studio 8.0, Oracle 10g Application Server, Oracle Developer Suite, PL/SQL, T-SQL, SQL Plus, SSIS, Teradata 13, OLAP, OLTP, SAS, MS Excel.
Confidential, NJ
Data Analyst/Data Modeler
Responsibilities:
- Gathered Business requirements by organizing and managing meetings with business stake holders, Application architects, Technical architects and IT analysts on a scheduled basis
- Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
- Performed Normalization, Indexes Tuned and Optimized the existing database design.
- Performed reverse engineering using the Erwin Data Modeling tool.
- Performed various tasks that included data massaging, data extraction from flat files, data transformations into various file formats, validations, loading data and checking for any data integrity errors.
- Responsible for the development and maintenance of Logical and Physical data models, along with corresponding metadata, to support Applications.
- Conducted meetings with the business and technical team to gather necessary analytical data requirements in JAD sessions.
- Analyzed all existing SSIS packages, SQL Server objects & new functional specs.
- Identified and tracked the slowly changing dimensions, heterogeneous sources an determined the hierarchies in dimensions.
- Created Mappings, Tech Design, loading strategies for ETL to load newly created or existing tables.
- Worked extensively with MicroStrategy Report developers in creating data marts and develop reports
- Created databases for OLAP Metadata catalog tables using forward engineering of models in Erwin.
- Enforced referential integrity in the OLTPdatamodel for consistent relationship between tables and efficient database design.
- Involved in importing/exporting large amounts ofdatafrom files to Teradata and vice versa.
- Designed Mapping Documents and Mapping Templates for Data Stage ETL developer.
- Designed, Build the Dimensions, cubes with Star schema and Snow Flake Schema using SQL Server Analysis Services (SSAS).
- Wrote PL/SQL statement, stored procedures and Triggers in DB2 for extracting as well as writing data.
- Created a Data Mapping document after each assignment and wrote the transformation rules for each field as applicable
- Analyzed OLTP source systems and Operational Data Store and research the tables/entities required for the project. Designing the measures, dimensions and facts matrix document for the ease while designing.
- Tuning all database via indexing of tables, MS SQL Server 2005 configuration parameters and stored procedures SQL code optimization.
- Used forward engineering to create a physical data model with DDL that best suits the requirements from the Logical Data Model.
Environment: Erwin 7.3, SSIS, Metadata, OLAP, OLTP, Data Mart, PL/SQL, T-SQL, DB2, SQL Plus, Microsoft SQL Server 2005, Flat files.