Sr. Data Architect/data Modeler Resume
Union, NJ
SUMMARY:
- Above 7+ years of Experience working as a Data Architect/Data Modeler and Data Analyst with emphasis on Data Mapping, Data Validation in Data Warehousing Environment.
- Highly proficient in Data Modeling retaining concepts of RDBMS using 3NormalForm (3NF) and Multidimensional Data Modeling Schema (Star schema, Snow - Flake Modeling, Facts and dimensions).
- Experienced in developing Entity-Relationship ER diagrams and modeling Transactional Databases and Data Warehouse using tools like ERWIN, ER/Studio and Power Designer.
- Excellent experience with IBM Ionosphere utilizing MDM, Data profiling and data Modeling.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
- Strong hands on experience using Teradata utilities like BTEQ, Fast-Load, Multi-Load, Fast-Export, Tpump, Teradata Manager and Visual Explain.
- Work on Background process in oracle Architecture. Also drill down to the lowest levels of systems design and construction.
- Excellent understanding of Hub Architecture Style for MDM hubs the registry, repository and hybrid approach.
- Expertise in developing transactional enterprise data models that strictly meet normalization rules, as well as Enterprise Data Warehouses using Kimball and Inmon Data Warehouse methodologies
- Strong experience in Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export through the use of multiple ETL tools such as Ab Initio and Informatica Power Center
- Experience in writing SQL queries and optimizing the queries in Oracle, SQL Server, Netezza, Teradata and Big Data.
- Experienced in generating and documenting Metadata while designing OLTP and OLAP systems environment.
- Excellent experience in troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
- Expert level understanding of using different databases in combinations for Data extraction and loading, joining data extracted from different databases and loading to a specific database.
- Experienced in generating and documenting Metadata while designing OLTP and OLAP systems environment
- Strong Database experience using Oracle, XML, DB2, Teradata, SQL server, Big data and NoSQL.
- Capabilities to provide AWS operations and deployment guidance and best practices throughout the lifecycle of a project.
- Excellent experience on Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and Fast Export.
- Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
- Excellent experience in Data mining with querying and mining large datasets to discover transition patterns and examine financial data.
- Good experience in data transformation, data mapping from source to target database schemas and also data cleansing.
- Experience in working with business intelligence and data warehouse software, including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum Database, Amazon Redshift, or Azure Data Warehouse
- Expert in implementing the projects from end to end & in providing the Architectural with emphasis on requirements analysis, design, coding, testing and documentation.
- Additionally experienced in NameNode where Hadoop stores all the file location information in HDFS and tracks the file data across the cluster or multiple machines.
- Experience in conducting Joint Application Development (JAD) sessions with SMEs, Stakeholders and other project team members for requirement gathering and analysis.
- Good experience in handling Data Dictionaries and Warehousing duties
TECHNICAL SKILLS:
Data Modeling: Erwin 9.6/9.5, Toad, ER studio 9.7, Star-Schema Modeling, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables
Languages: PL SQL, T-SQL, Unix Shell scripting, XML
Database: Oracle 12c/11g, MS SQL Server2016/2014, DB2, Teradata 14/15, DB2, Netezza, Cassandra.
Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume, Splunk
Testing Tools: Win Runner, Load Runner, Test Director, Mercury Quality Center, Rational Clear Quest
BI Tools: Tableau 7.0/8.2, Pentaho 6, SAP Business Objects, Crystal Reports
ETL/Data warehouse Tools: Informatica, SAP Business Objects XIR3.1/XIR2, Talend, Pentaho
Operating System: UNIX, Windows 8/7, Linux, Red Hat
Other Tools: TOAD, BTEQ, MS-Office suite (Word, Excel, Project and Outlook)
PROFESSIONAL EXPERIENCE
Confidential, Union, NJ
Sr. Data Architect/Data Modeler
Responsibilities
- Massively involved in Data Architect role to review business requirement and compose source to target data mapping documents.
- Researched, evaluated, architect, and deployed new tools, frameworks, and patterns to build sustainable Big Data platforms for our clients
- Involved in Normalization/De-normalization techniques for optimum performance in relational and dimensional database environments.
- Implemented logical and physical relational database and maintained Database Objects in the data model using ER/Studio 9.7
- Worked on Amazon Redshift, AWS & Azure and architecting a solution to load data, create data models and run BI on it.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle12c, HBase, Mongo DB PL/SQL Server with high volume data.
- Used Ab Initio DQE for data quality solution for enterprise-level data processing and data management systems
- Worked on NoSQL databases including HBase, Mongo DB, and Cassandra. Implemented multi-data center and multi-rack Cassandra cluster.
- Applied Data Governance rules (primary qualifier, Class words and valid abbreviation in Table name and Column names).
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and extracted the data from Oracle into HDFS using Sqoop.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow Flake Schemas
- Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL access on Hadoop data
- Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
- Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark.
- Used Meta Stage to maintain the metadata for different Data Warehouse environments and projects.
- Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
- Primarily responsible for Tableau customization for statistical dashboard to monitor sales effectiveness and also used Tableau for customer marketing data visualization.
- Involved in making screen designs, Use Cases and ER diagrams for the project using ER/Studio 9.7.
- Extracted data from IBM Cognos to create automated visualization reports and dashboards on Tableau.
- Performed Data mapping between source systems to Target systems, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data
- Used Agile Methodology of Data Warehouse development using Kanbanize.
- Developed triggers, stored procedures, functions and packages using cursors and ref cursor concepts associated with the project using PL/SQL
- Generate DDL scripts for database modification, Teradata, Macros, Views and set tables.
- Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS
- Developed the performance tuning of the database by using EXPLAIN PLAN, TKPROF utilities and also debugging the code.
Environment: ER/Studio 9.7, Oracle12c, Hive, Amazon Redshift, AWS, MapReduce, Hadoop, Cassandra, HBase, Mongo DB, Pig, Agile, NoSQL, PL/SQL, OLAP, OLTP, SQL, IBM Cognos, Ab Initio, Tableau, Crystal Reports 2008, HDFS.
Confidential, Baltimore, MD
Sr. Data Architect/Modeler
Responsibilities
- Developing full life cycle software including defining requirements, prototyping, designing, coding, testing and maintaining software.
- Extensively worked on Mapping Variables, Mapping Parameters, Workflow Variables and Session Parameters.
- Extracting data from various source systems like Oracle, SQL Server and flat files as per the requirements.
- Worked as Architect and designed conceptual, logical and physical models and build data marts using hybrid Inmon and Kimball DW methodologies
- Involved in writing Shell Scripts to accumulate the MTD source file Collaboration with Architects and Managers for review of solutions and data strategy
- Used data virtualization tool connect multiple heterogeneous sources without requirement of physically moving the data.
- Extensively involved in analyzing various data formats using industry standard tools and effectively communicate them with business users and SME's.
- Worked on data warehousing, ETL, SQL, scripting and big data (MPP + Hadoop).
- Worked on Normalization and De-Normalization techniques for both OLTP and OLAP systems
- Created reports from Greenplum (Pivotal) which references the positions/transactions for each customer's monthly invoice for all jurisdiction (CFTC, ESMA, CANADA, ASIC & MAS) and make it available on the Portal .
- Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
- Involved in Netezza Administration Activities like backup/restore, performance tuning, and Security configuration.
- Create architectural documents such as, the e2e data flow for the IBM Information data warehouse system.
- Identified security loopholes, established data quality assurance and addressed data governance.
- Designed Physical Data Model (PDM) using IBM Info sphere Data Architect data modeling tool and Oracle PL/SQL.
- Used push down optimization in Informatica to call Greenplum GPLoad functions
- Lead the design and modeling of tactical architectures for development, delivery, and support of projects.
- As a Architect implement MDM hub to provide clean, consistent data for a SOA implementation.
- Called Greenplum Business Rules, Data Rules and Transform Rules functions using Informatica Stored Procedure Transformation.
- Developed and maintained policies, standards, and guidelines to ensure that a consistent framework is applied across the company.
- Involved in all the steps and scope of the project reference data approach to MDM, Creating a Data Dictionary and Mapping from Sources to the Target in MDM Data Model.
- Promoted the use of a shared infrastructure, application roadmap, and documentation of interfaces to improve information flow and reduce costs.
- Architect and build meta data repository to describe Digital business data, technical data, and processes.
- Designed solutions for multiple large data warehouses with a good understanding of cluster and parallel architecture as well as high-scale or distributed RDBMS and/or knowledge on NoSQL platforms.
- Developed Ad-Hoc Queries, Views and functions in Greenplum in order to make data accessible for Business Analyst and Managers.
Environment: Erwin 9.6, PL/SQL, ODS, Hadoop, MS SQL Server 2014, flat files, Oracle 12c, MDM, Information Analyzer, Informatica, IBM Infosphere
Confidential, Omaha, NB
Sr. Data Architect/Modeler
Responsibilities
- Implemented Dimensional Modeling using Star and Snow Flake Schema, Identifying Facts and Dimensions, Physical and logical data modeling using Erwin
- Developing full life cycle software including defining requirements, prototyping, designing, coding, testing and maintaining software.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow flake Schemas.
- Responsible for creating OLAP cubes for deep through analysis using SSAS .
- Used data profiling automation to uncover the characteristics of the data and the relationships between data sources before any data-driven.
- Involved in the design and development of user interfaces and customization of Reports using Tableau and OBIEE.
- Created a detailed plan to migrate the EDW from Oracle to Greenplum.
- Collaborated with product teams to understand and translate requirements, ensuring specifications and architecture design meet broader technical strategic goals.
- Collaborated with engineering teams throughout the development process in ensuring best practices and designing scalable analytics infrastructure and data transformation / aggregation processes.
- Implemented Dimensional model for the Data Mart and responsible for generating DDL scripts using Erwin.
- Created a high-level industry standard, generalized data model to convert it into logical and physical model at later stages of the project using Erwin.
- Performed Greenplum database versions upgrades on QA, development and production environments.
- Designed Metadata Repository to store data definitions for entities, attributes & mappings between data warehouse and source system data elements.
- Created native Oracle functions to Greenplum DDLs for Business Objects to use
- Providing extensive technical, strategic advice and guidance of the highest level to senior managers and technical resources in the creation and implementation of new data standards and databases.
- Worked on data manipulation and analysis & accessed raw data in varied formats with different methods and analyzing and processing data.
- Perform administrative tasks, including creation of database objects such as database, tables, and views, using SQL DCL, DDL, and DML requests.
- Used BTEQ script to create a sample tables. Redefine the partitioning of a populated table.
- Involved with data cleansing/scrubbing and validation.
- Performed dicing and slicing on data using Pivot tables to acquire the churn rate pattern and prepared reports as required.
- In depth analyses of data report was prepared weekly, biweekly, monthly using MS Excel, SQL & UNIX.
- Performed literature searches and ad-hoc data collection based on requests.
- Developing complex mappings to extract data from diverse sources including flat files, RDBMS tables, legacy system files, XML files, Applications and Teradata.
- Created SSIS Packages using Pivot Transformation, Execute SQL Task, Data Flow Task, etc to import data into the data warehouse.
Environment: Erwin r9.5, BTEQ, PL/SQL, Oracle 11g, Teradata R14, Teradata SQL Assistant 14.0, Flat Files, SSAS, Pivot, OBIEE
Confidential, Bellevue, WA
Sr. Data Architect/Data Modeler
Responsibilities:
- Developed new Data architecture, including Transformation and Reporting Layers, missing in legacy system
- Established data architecture strategy, best practices, standards, and roadmaps.
- Ensured adherence with data standards and best practices through the establishment of data governance processes.
- Management of data including definition, usage, and quality via architecture repositories like data dictionary, data models, and data quality logs.
- Designing the Data Marts in dimensional Data modeling using star and snowflake schemas.
- Gathered requirements, application designing, development, implementation and enhancement, team leadership and project management.
- Collaborated with DBA's, Business Analysts and Data Stewards conducted design review sessions to validate the developed models and logical mapping.
- Extensively involved in analyzing various data formats using industry standard tools and effectively communicate them with business users and SME's.
- Facilitated project kickoff session with Business and SMEs to understand the requirements which also involved Project Manager, Development Lead, Data Modeler, Data Mapper, Architect etc.
- Developed OLTP system by designing Logical and eventually Physical Data Model from the Conceptual Data Model and used Erwin tool to develop a Conceptual Model based on business requirements analysis.
- Designed and documented conceptual, logical, physical data models and proof of design for enterprise data management.
- Used reverse engineering for a wide variety of RDBMS, including MS Access, Oracle and Teradata to connect to existing database and create graphical representation using Erwin.
- Designed and implemented high performance engines using extraction of data pre-aggregated in Redshift into RDSPostgres using postgres fdw and dblinkPostgresplugins, led offshore developers
- Heavily engaged in the database change request process between Data Management and User Groups, the process was to change the logical model and then change the physical model and then generate the corresponding DDL script for implementation.
Environment: Erwin r9.4, Oracle11g, Teradata15, Rational Rose, Requisite Pro. SQL Server 2014, SAS, UNIX, Tableau 9.1.2, Hadoop, Hive, Pig, Map Reduce, Spark, PL/SQL
Confidential, Boston, MA
Sr. Data Modeler/Data Analyst
Responsibilities:
- Gathered requirements, analyzed and wrote the design documents.
- Developed logical data models and physical database design and generated database schemas using ER Studio.
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and DB2.
- Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
- Worked with data investigation, discovery and mapping tools to scan every single data record from many sources. Create and Monitor workflows using workflow designer and workflow monitor.
- Involved in extensive Data validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues. Identify & record defects with required information for issue to be reproduced by development team.
- Document all data mapping and transformation processes in the Functional Design documents based on the business requirements.
- Generated and DDL (Data Definition Language) scripts using ER Studio and assisted DBA in Physical Implementation of data Models.
- Developed, managed and validated existing data models including logical and physical models of the data warehouse and source systems utilizing a 3NFmodel.
- Prepared High Level Logical Data Models using Erwin, and later translated the model into physical model using the Forward Engineering technique.
- Involved in Data mapping specifications to create and execute detailed system test plans. The data mapping specifies what data will be extracted from an internal data warehouse, transformed and sent to an external entity.
- Analyzed business requirements, system requirements, data mapping requirement specifications, and responsible for documenting functional requirements and supplementary requirements in Quality Center.
Environment: ER Studio8.0, Microsoft Visio, MS SQL Server 2012, DB2 Oracle10g/11g, workflow designer
Confidential, Providence, RI
Sr. Data Modeler/Data Analyst
Responsibilities:
- Involved in analysis of Business requirement, Design and Development of High level and Low level designs, Unit and Integration testing
- Created the conceptual model for the data warehouse using Erwin data modeling tool.
- Reviewed the conceptual EDW (Enterprise Data Warehouse) data model with business users, App Dev and Information architects to make sure all the requirements are fully covered.
- Worked on designing the OLAP Model, Dimension model for BI Reporting sourcing from SAP Transactions.
- Used Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Designed and documented Use Cases, Activity Diagrams, Sequence Diagrams, OOD (Object Oriented Design) using UML and Visio.
- Reverse engineered existing data bases to understand the data flow and business flows of existing systems and to integrate the new requirements to future enhanced and integrated system. Designing Partitioned strategy in data model.
- Used and supported database applications and tools for extraction, transformation and analysis of raw data
- Data modeling was performed using ERWIN tool to build logical and physical models.
- Used Erwin for reverse engineering to connect to existing database and ODS to create graphical representation in the form of Entity Relationships and elicit more information.
- Gathered reporting and analysis requirements and translated into reporting structures / data models, including aggregate tables, relational and dimensional (star-schema) marts
- Forward Engineered the physical data model and generated DDL script using Forward Engineering option in Erwin.
- Analyzed the impact on the Enterprise data warehouse and Down streams.
- Developed PL/SQL scripts to validate and load data into interface tables
Environment: Erwin 7.x, Microsoft Visio, OOD, SAP, OLAP, ODS, MS SQL Server 2008, DB2 Oracle 10g