Sr. Data Architect/data Modeler Resume
Warren, NJ
SUMMARY:
- Above 9+ years of experience as Sr. Data Architect/Modeler with Data Analytics Professional in System Analysis, Data Architecture and Development, Testing and Deployment of business applications.
- Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop and Flume.
- Hands - on experience in architecting and data modeling for AWS Redshift, AWS Oracle RDS, AWS PostgreSQL and AWS Aurora.
- Good understanding and hands on experience with AWS S3 and EC2.
- Experience in designing stunning visualizations using Tableau software and publishing and presenting Dashboards, Storyline on Web and Desktop platforms.
- Experienced in generating and documenting Metadata while designing OLTP and OLAP systems environment.
- Experience in writing SQL queries and optimizing the queries in Oracle, SQL Server, Netezza, Teradata and Big Data.
- Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
- Hands on experience in Normalization and De-Normalization techniques upto 3NF for optimum performance in relational and dimensional database environments.
- Strong background in various Data Modeling tools using ERWIN, ER/Studio and Power Designer.
- Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
- Experience in metadata design, real time BI Architecture including Data Governance for greater ROI.
- Experience in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
- Strong Database experience using Oracle, XML, DB2, Teradata, SQL server, Big data and NoSQL.
- Proficient in using Python, SQL, Hadoop ecosystem for extracting data and building predictive models.
- Extensive ETL testing experience using Informatica 9x/8x, Talend, Pentaho.
- Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export
- Experience in conducting Joint Application Development (JAD) sessions with SMEs, Stakeholders and other project team members for requirement gathering and analysis.
- Strong hands on experience using Teradata utilities like BTEQ, Fast-Load, Multi-Load, Fast-Export, Tpump, Teradata Manager and Visual Explain.
- Strong validation experience of data models by different measures such as AUC, ROC, and confusion matrix.
- Proficient in project implementations using various Software Development Life Cycle (SDLC) methodologies like Waterfall, Agile (SCRUM) and RUP.
- Experienced in Client-Server application development using Oracle, PL/SQL, SQL PLUS, SQL Developer, TOAD, SQL LOADER.
- Expertise in SQL Server Analysis Services (SSAS) to deliver Online Analytical Processing (OLAP) and data mining functionality for business intelligence applications.
- Extensive experience in SSIS Packages, SSRS reports and SSAS cubes on production server.
- Experience in designing Enterprise Data Warehouses, Data Marts, Reporting data stores (RDS) and Operational data stores (ODS).
- Experienced on creation of reports, dashboards, and ad-hoc analysis using the reporting tool Tableau
TECHNICAL SKILLS:
Data Modeling Tools: Erwin r9.6/9.5, ER/Studio 9.7, Sybase Power Designer
Languages: SQL, PL/SQL, ASP, Visual Basic, XML, Python, SQL, T-SQL, SQL Server, C, C++, JAVA, HTML, UNIX shell scripting, PERL.
Big Data Tools: Hadoop, Hive, Spark, Pig, HBase, Sqoop, Flume.
Database: Oracle 11g/12c, MS Access, SQL Server 2016/2014, Sybase and DB2, Teradata14/15, Hive
BI Tools: Tableau 7.0/8.2, Tableau server 8.2, Tableau Reader 8.1,SAP Business Objects, Crystal Reports
Operating Systems: Microsoft Windows 8/7 and UNIX.
Applications: Toad for Oracle, Oracle SQL Developer, MS Word, MS Excel, MS Power Point, Teradata, Designer 6i
Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Waterfall Model
Project Execution Methodologies:: Ralph Kimball and BillInmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
PROFESSIONAL EXPERIENCE:
Confidential, Warren, NJ
Sr. Data Architect/Data Modeler
Responsibilities:
- Massively involved in Data Architect role to review business requirement and compose source to target data mapping documents.
- Installed and Configured Open Source Software like Pig, Hive, HBase, Flume and Sqoop.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow Flake Schemas
- Worked with Netezza and Oracle databases and implemented various logical and physical data models for them.
- Involved in Normalization/De-normalization techniques for optimum performance in relational and dimensional database environments.
- Worked on Amazon Redshift, AWS & Azure and architecting a solution to load data, create data models and run BI on it.
- Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL access on Hadoop data
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture.
- Applied Data Governance rules for primary qualifier, Class words and valid abbreviation in table name and Column names.
- Worked on Tableau for insight reporting and data visualization
- Performed POC for Big data solution using Cloudera Hadoop for data loading and data querying
- Created OLAP data architecture, analytical data marts, and cubes optimized for reporting.
- Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
- Cleaned and maintained the database by removing and deleting old data.
- Involved in Logical modeling using the Dimensional Modeling techniques such as Star Schema and Snow Flake Schema.
- Worked on Hadoop ecosystem, hive queries, MongoDB, Cassandra, Pig, and Apache Strom.
- Loaded and transformed large sets of structured, semi structured and unstructured data using Hadoop/Big Data concepts.
- Designed ER diagrams (Physical and Logical using Erwin) and mapping the data into database objects.
- Created SSIS packages for different data loading operations for many applications.
- Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from legacy Oracle and SQL Server database systems.
- Developed LINUX Shell scripts by using NZSQL/NZLOAD utilities to load data from flat files to Netezza database.
- Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark.
- Used Ab-Initio DQE for data quality solution for enterprise-level data processing and data management systems
Environment: Erwin9.6, Informatica, Big Data, LINUX, Teradata, SQL, Oracle, AWS, Netezza, Tableau, Hadoop, Hive, OLAP, OLTP, Flume, NZSQL/NZLOAD, MongoDB, Cassandra.
Confidential, Cincinnati, OHSr. Data Architect/Data Modeler
Responsibilities:
- As a Architect implement MDM hub to provide clean, consistent data for a SOA implementation.
- Gathered and translated business requirements into detailed, production-level technical specifications, new features, and enhancements to existing technical business functionality.
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
- Created and maintained Database Objects (Tables, Views, Indexes, Partitions, Synonyms, Database triggers, Stored Procedures) in the data model.
- Worked with other Data Architects to design New Data Mart to design the Google Analytics data reports.
- Worked on analyzing Hadoop cluster and different big data analytic tools including Pig, HBase database and Sqoop.
- Connected to Amazon Redshift through Tableau to extract live data for real time analysis.
- Created Rich dashboards using Tableau Dashboard and prepared user stories to create compelling dashboards to deliver actionable insights.
- Created Hive architecture used for real time monitoring and HBase used for reporting
- Worked for map reduce and query optimization for Hadoop hive and HBase architecture
- Worked with DBA group to create Best-Fit Physical Data Model from the Logical Data Model using Forward engineering using Erwin.
- Generate DDL scripts for database modification, Teradata, Macros, Views and set tables.
- Build and maintain scalable data pipelines using the Hadoop ecosystem and other open source components like Hive and HBase.
- Performed data management and fulfilling ad-hoc requests according to user specifications by utilizing data management software programs and tools like Perl, TOAD, MS Access, Excel and SQL.
- Worked on Naming standards for Table/Column/Index/Constraints names thru Erwin Macros and Master Abbreviations file.
- Worked in Data Analysis, data profiling and data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
- Worked on Physical design for both SMP and MPP RDBMS, with understanding of RDMBS scaling features.
- Performed Researching and deploying new tools, frameworks and patterns to build a sustainable Big data platform.
- Extensively used agile methodology as the Organization Standard to implement the data Models.
- Involved in performing extensive Back-End testing by writing SQL queries and PL/SQL stored procedures to extract the data from SQL Database.
- Involved in the validation of the OLAP Unit testing and System Testing of the OLAP Report Functionality and data displayed in the reports.
- Applies architectural and technology concepts to address scalability, security, reliability, maintainability and sharing of enterprise data
- Designed Metadata Repository to store data definitions for entities, attributes & mappings between data warehouse and source system data elements.
Environment: Erwin9.6, Big Data, OLTP, OLAP, SMP, Teradata R13, Teradata SQL Assistant, Hadoop, Hive, Pig, HBASE, DB2, Big Data, Agile, MS-Office, Oracle, SQL Server.
Confidential, Charlotte, NCSr. Data Architect/Data Modeler
Responsibilities:
- Worked as a Data Modeler/Architect to generate Data Models using Erwin and developed relational database system.
- Lead Architectural Design in Big Data, Hadoop projects and provide for a designer that is an idea-driven.
- Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration
- Designed Physical Data Model (PDM) using Erwin and Oracle PL/SQL.
- Designed Normalization up to 3NF and performed Forward & Reverse Engineering using the Erwin.
- Built relationships and trust with key stakeholders to support program delivery and adoption of enterprise architecture.
- Involved in writing T-SQL, working on SSIS, SSRS, SSAS, Data Cleansing, Data Scrubbing and Data Migration.
- Developing full life cycle software including defining requirements, prototyping, designing, coding, testing and maintaining software.
- Involved in Teradata utilities (BTEQ, Fast Load, Fast Export, Multiload, and Tpump) in both Windows and Mainframe platforms.
- Developed Business Intelligence architecture using Microsoft and Tableau products.
- Provided guidance and solution concepts for multiple projects focused on data governance and master data management.
- Performed extensive data profiling and data analysis for detecting and correcting inaccurate data from the databases and track the data quality.
- Developed Source to Target Matrix with ETL transformation logic for ETL team
- Participated with key management resources in the strategic analysis and planning requirements for Data Warehouse/Data Mart reporting and data mining solutions.
- Managed the meta-data for the Subject Area models for the Data Warehouse environment.
- Created data masking mappings to mask the sensitive data between production and test environment.
- Worked Extensively with DBA and Reporting team for improving the Report Performance with the Use of appropriate indexes and Partitioning.
- Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data
- Worked with Hadoop eco system covering HDFS, HBase, YARN and Map Reduce.
- Generated and DDL (Data Definition Language) scripts using ERWIN and assisted DBA in Physical Implementation of Data Models.
- Responsible for technical data governance, enterprise wide data modeling and database design.
- Developed Data mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
Environment: Erwin 9.5, Tableau, MDM, QlikView, PL/SQL, HDFS, Teradata 13, JSON, HADOOP (HDFS), MapReduce, PIG, Spark, AWS.
Confidential, Chicago, ILSr. Data Modeler/Data Analyst
Responsibilities:
- Worked as a Data Modeler/Analyst to generate Data Models using Erwin and developed relational database system.
- Created logical data model from the conceptual model and it's conversion into the physical database design using Erwin
- Interacted with users for verifying User Requirements, managing Change Control Process, updating existing Documentation.
- Developed dimensional model for Data Warehouse/OLAP applications by identifying required facts and dimensions.
- Developed Data Migration and Cleansing rules for the Integration Architecture (OLTP, ODS, DW)
- Worked very close with Data Architectures and DBA team to implement data model changes in database in all environments.
- Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
- Performed data cleaning and data manipulation activities using NZSQL utility.
- Generated DDL (Data Definition Language) scripts using Erwin 8 and supported the DBA in Physical Implementation of data Models.
- Maintained warehouse metadata, naming standards for future application development.
- Produced PL/SQL statement and stored procedures in SQl for extracting as well as writing data.
- Extensively made use of Triggers, Table Spaces, Pre/Post SQL, Sequences, Materialized Views, Procedures and Packages in Data Models.
- Performed data analysis and data profiling using SQL queries on various sources systems including Oracle and SQL Server 2008
- Used Erwin for reverse engineering to connect to existing database and ODS to create graphical representation in the form of Entity Relationships and elicit more information.
- Facilitated meetings with the business and technical team to gather necessary analytical data requirements.
- Development of database objects like Tables, views and materialized views etc using SQL.
- Assisted in designing test plans, test scenarios and test cases for integration, regression and user acceptance testing.
Environment: ERWIN 9.0, OLAP, OLTP, SSIS, ODS, PL/SQL, Metadata, SQL Server 2008, Oracle9i
Confidential, Bangalore, INData Analyst/Data Modeler
Responsibilities:
- Involved with all the phases of Software Development Life Cycle (SDLC) methodologies throughout the project life cycle.
- Analyzed Data sources and requirements and business rules to perform logical and physical Data Modeling
- Created a logical design and physical design in ER/Studio.
- Communicated with users and business analysts to gather requirements. Involved in business process modeling using UML through Rational Rose.
- Reverse Engineered the Data Models and identified the Data Elements in the source systems and adding new Data Elements to the existing data models.
- Worked on data profiling and data validation to ensure the accuracy of the data between the warehouse and source systems.
- Executed the UNIX shell scripts that invoked SQL loader to load data into tables.
- Developed Star Schema and Snowflake Schema in designing the Logical Model into Dimensional Model.
- Creation of database objects like tables, views, Materialized views, procedures, packages using Oracle tools like PL/SQL, SQL* Plus, SQL*Loader and Handled Exceptions.
- Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections.
- Extensively used ER Studio for developing data model using star schema methodologies.
- Participated in several JAD (Joint Application Design/Development) sessions to track end to end flow of attributes starting from source screens to all the downstream systems.
- Involved in Data Profiling, Data Cleansing and make sure the data is accurate and analyzed when it is transferring from OLTP to Data Marts and Data Warehouse.
- Involved in data extraction, validation, analysis of the data and the store in data marts.
- Performed Data Analysis and data profiling using complex SQL on various sources systems including Oracle 8i and Teradata, to ensure accuracy of the data between the warehouse and source systems.
- Involved in completing the Data Dictionary, Data Lineage and Data Flow diagrams for Meta data
- Involved in Performance tuning by leveraging oracle explain utility and SQL tuning.
- Involved in creating Sessions, worklets and Workflows and scheduling workflows using Workflow Manager.
Environment: ER Studio, Star Schema, Oracle 9i/8i, Teradata Oracle SQL Developer, PL/SQL, Business Objects, OLAP, OLTP, Workflow Manager.