We provide IT Staff Augmentation Services!

Sr. Data Architect/data Modeler Resume

5.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY:

  • Having more than 9 years of experience focused on Data Architecture, Data Management, implementing Data Warehouse & Business Intelligence solutions, Data Modeling/Data Analysis, Data Profiling, Data Migration, Data Conversion, Data Quality, Data Governance, Data Integration, MDM, NoSQL and Metadata Management Services and Configuration Management.
  • Experienced on data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing.
  • Experienced in Dimensional Data Modeling experience using Data modeling, Relational Data modeling, ER/Studio, Erwin, and Sybase Power Designer, Star Join Schema/Snowflake modeling, FACT & Dimensions tables, Conceptual, Physical & logical data modeling.
  • Experienced in writing Pig Latin scripts, MapReduce jobs and HiveQL.
  • Expertise in the Data Analysis, Design, Development, Implementation and Testing using DataConversions, Extraction, Transformation and Loading (ETL) and SQLServer, ORACLE and other relational and non - relational databases.
  • Experience with Amazon Web Services, including AWS Lambda, ELK, AWS EC2, AWS S3 etc.
  • Extensively worked on ERWIN tool with all features like REVERSE Engineering, FORWARD Engineering, SUBJECTAREA, DOMAIN, Naming Standards Document etc.
  • Experienced in importing and exporting the data using Sqoop from HDFS to RelationalDatabase systems/mainframe and vice-versa.
  • Experienced in Hadoop ecosystem components like Hadoop MapReduce, HDFS, HBase, MongoDB, Oozie, Hive, Sqoop, Pig, and Zookeeper.
  • Familiar with Kimball DW/BI modeling principles and knowledgeable in Data warehouse modeling for different kind of business.
  • Strong experience in using Excel and MSAccess to dump the data and analyze based on business needs and experienced with Integration Services (SSIS), Reporting Service (SSRS) and Analysis Services (SSAS).
  • Expertise in Normalization to 3NF/De-normalization techniques for optimum performance in relational and dimensional database environments.
  • Experienced in carrying out SoftwareDevelopmentLifeCycle (SDLC) in relational and object methodologies.
  • Extensive experienced on ERModeling, DimensionalModeling (StarSchema, SnowflakeSchema) and Data warehousing and OLAP tools.
  • Expertise in data base programming (SQL, PLSQL) MSAccessOracle12c/11g/10g/9i, XML, DB2, Informix, Teradata,, Data base tuning and Query optimization.
  • Expertise in performing data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Experience in testing Business Intelligence reports generated by various BI Tools like Cognos and Business Objects.
  • Experienced in logical/physical database design and review sessions to determine and describe dataflow and data mapping from source to target databases coordinating with End Users, BusinessAnalysts, DBAs and Application Architects.
  • Extensive experience in development of T-SQL, OLAP, PL/SQL, Stored Procedures, Triggers, Functions, Packages, performance tuning and optimization for business logic implementation.
  • Expertise in SQL queries, PL/SQL Packages, SQL plus, Stored Procedures, Functions, Triggers and Performance Analysis, Creating Partitions, Indexes, Aggregating tables when required.
  • Expertise in Visio, Process FlowDiagrams, ActivityDiagrams, CrossFunctionalDiagram, SwimLaneDiagrams, UseCaseDiagrams.
  • Expertise in scheduling JAD (JointApplicationDevelopment) with End Users, stake Holders, Subject Matter Experts, Developers and Testers.
  • Expertise in Data modeling (Dimensional&Relational) concepts like Star-SchemaModeling, SnowflakeSchemaModeling, Fact and Dimension tables.
  • Expertise in writing Stored Procedures, Functions, Nested Functions, building Packages and developing Public and Private Sub-Programs using PL/SQL and providing Documentation.
  • Expertise in loading data by using the Teradata loader connection, writing Teradata utilities scripts (Fastload, Multiload) and working with loader logs.
  • Experienced in developing TSQL scripts and stored procedures to perform various tasks and multiple DDL, DML, and DCL activities to carry out business requirements
  • Strong RDBMS concepts and well experience in creating database Tables, Views, Sequences, triggers, joins taking the Performance and Reusability into consideration.
  • Efficient in Extraction, Transformation and Loading (ETL) data from spread sheets, database tables using Microsoft data transformation service (DTS)
  • Extensive knowledge in software testing methodology and developing Test Plans, TestProcedures, TestCaseDesign and Execution, Modification Requests.
  • Strong in conceptualizing and communicating enterprise data architecture frameworks for global enterprises for inter operation of data warehouses, middleware, and web applications.

TECHNICAL SKILLS:

Data Modeling Tools: Erwin 9x, 8x, and 7x, ER Studio and Oracle Designer, Power Designer.

OLAP Tools:: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9

ETL Tools:: SSIS, Informatica Power Centre.

Programming Languages:: Java, Base SAS and SAS/SQL, SQL, T-SQL, HTML, Java Script, CSS, UNIX shells scripting, PL/SQL.

Database Tools:: Microsoft SQL Server 2010/2012/2014/2016, Teradata 14/15, Oracle 12c/11g/10g, and MS Access, PostgerSQL, Netezza.

Web technologies:: HTML, DHTML, XML, JavaScript

Reporting Tools:: Business Objects, Crystal Reports, Cognos and Tableau.

Packages:: Microsoft Office, Microsoft Project, SAP and Microsoft Visio, Share point Portal Server.

Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX

Quality Assurance Tools:: Quick Test Pro, Win Runner, Load Runner, Quality Center.

Big Data: Hadoop, HDFS, Hive, Pig, HBase, Sqoop, MongoDB, Cassandra.

Other Tools:: MS-Office suite (Word, Excel, Project and Outlook), BTEQ, Teradata, SQL Assistant, Aginity

PROFESSIONAL EXPERIENCE:

Confidential, Chicago IL

Sr. Data Architect/Data Modeler

Responsibilities:

  • Collaborates with cross-functional team in support of business case development and identifying modeling method(s) to provide business solutions. Determines the appropriate statistical and analytical methodologies to solve business problems within specific areas of expertise.
  • Document all data mapping and transformation processes in the Functional Design documents based on the business requirements and queried the databases, wrote test validation scripts and performed the System testing.
  • Owned and managed all changes to the data models. Created data models, solution designs and data architecture documentation for complex information systems.
  • Generating Data Models using Erwin9.6 and developed relational database system and involved in Logical modeling using the Dimensional Modeling techniques such as Star Schema and Snow Flake Schema.
  • Guide the full lifecycle of a Hadoop solution, including requirements analysis, platform selection, technical architecture design, application design and development, testing, and deployment
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
  • Created data models for AWS Redshift, Hive and HBase from dimensional data models.
  • Worked on NoSQL databases including HBase, Mongo DB, and Cassandra. Implemented multi-datacenter and multi-rack Cassandra cluster.
  • Define Big Data strategy, including designing multi-phased implementation roadmaps and handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS
  • Manage timely flow of business intelligence information to users and Involved in Normalization and De-Normalization of existing tables for faster query retrieval.
  • Involved in making screen designs, UseCases and ERdiagrams for the project using ERWIN and Visio.
  • Analyze the Business information requirements and research the OLTP source systems to identify the measures, dimensions and facts required for the reports.
  • Performed Datamapping between source systems to Target systems, logicaldata modeling, created classdiagrams and ERdiagrams and used SQLqueries to filter data
  • Lead design of high-level conceptual and logicalmodels that facilitate a cross-system/cross functional view of data requirements
  • Involved in designing and developing Data Models and Data Marts that support the Business Intelligence Data Warehouse.
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
  • Maintaining conceptual, logical and physicaldatamodels along with corresponding metadata.
  • Done data migration from an RDBMS to a NoSQL database, and gives the whole picture for data deployed in various data systems.
  • Designed and developed the data dictionary and Meta data of the models and maintain them.
  • Involved in DataWarehouse Support - StarSchema and Dimensionalmodeling to help design datamarts and datawarehouse
  • Developed LINUXShell scripts by using NZSQL/NZLOAD utilities to load data from flat files to Netezza database.
  • Responsible for Metadata Management, keeping up to date centralized metadata repositories using Erwin modeling tools.
  • Developed triggers, stored procedures, functions and packages using cursors and ref cursor concepts associated with the project using Pl/SQL
  • Experienced with Informatica Power Center in designing and developing complex ETL packages .
  • Installed and configured Informatica components for both Power Center and Big Data editions, including high availability; managing server activations and deactivations for all environments; ensuring that all systems and procedures adhere to organizational best practices
  • Prepared documentation for all entities, attributes, data relationships, primary and foreign key structures, allowed values, codes, business rules, glossary evolve and change during the project
  • Exported the patterns analyzed back to Teradata using Sqoop.
  • Troubleshoot test scripts, SQL queries, ETL jobs, and data warehouse/data mart/data store models.
  • Used Normalization methods up to 3NF and De-normalization techniques for effective performance in OLTP systems.
  • Developed complex Stored Procedures for SSRS (SQL Server Reporting Services) and created database objects like tables, indexes, synonyms, views, materialized views etc.
  • Developed the performance tuning of the Oracle database by using EXPLAINPLAN, TKPROFutilities and also debugging the SQLcode.
  • Perform analyses such as regression analysis, logistic regression, discriminate analysis, cluster analysis using SAS programming.
  • Excellent understanding of an Approach to MDM to creating a data dictionary, Using Informatica or other tools to do mapping from sources to the Target MDM Data Model
  • Used Meta data tool for importing metadata from repository, new job categories and creating new data elements.

Environment: Erwin r9.6, Oracle 12c, SQL Server 2016, Hive, AWS, S3, AWS Redshifit, NoSQL, Teradata, Netezza, PL/SQL, MS-Visio, Informatica, T-SQL, SQL, Crystal Reports 2008, Java, HDFS, PIG, SSRS, SSIS, Metadata, SQL.

Confidential, Dallas TX

Sr. Data Architect/ Data Modeler

Responsibilities:

  • Involved in the Architectural role to design and develop EDW program initiative and involved in design and conversion of business requirements to Conceptual Data Model, logical Data Model and Physical Data Model for subject areas.
  • Developed strategic oversight and planning of data models and database design within the CRM platform.
  • Worked with key stakeholders to understand the data requirements and translates strategic requirements into usable enterprise information architecture.
  • Design database, data models, ETL processes, data warehouse applications and business intelligence (BI) reports through the use of best practices and tools, including Erwin, SQL, SSIS, SSRS and OLAP, OLTP.
  • Conducted a full impact analysis and remediation of data warehouse migration from Teradata to SQL Server with Data Vault architecture.
  • Created Hive queries that helped analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Transformed Logical Data Model to Physical Data Model ensuring the Primary Key and Foreign Key relationships in PDM, Consistency of definitions of Data Attributes and Primary Index Considerations.
  • Involved with Data Analysis primarily Identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats and Validated the data of reports by writing SQL queries in PL/SQL Developer against ODS.
  • Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website and managed and reviewed Hadoop log files.
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS
  • Developed Star and Snow flake schemas based dimensional model to develop the data warehouse.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Designed the ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin
  • Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website and Processed the data using HQL (like SQL) on top of Map-reduce.
  • Implemented Hybrid Architecture comprising Inmon and Kimball approaches implementing Inmon at integrated layer and Kimball at the Semantic layer which is the best suited for Teradata.
  • Developed and reviewed custom SQL queries with use of joins clauses (inner, left, right, full) in Tableau desktop to validate and reconcile data.
  • Involved with all the phases of Software Development Life Cycle (SDLC) methodologies throughout the project life cycle.
  • Extensively involved in the Physical/logicalmodeling and development of Reporting DataWarehousing System.
  • Performing reverse engineering of physicaldatamodels from databases and SQLscripts.
  • Used Normalization (1NF, 2NF&3NF) and De-normalization techniques for effective performance in OLTP and OLAP systems.
  • Worked on Data modeling, Advanced SQL with Columnar Databases using AWS
  • Optimized ROLAP-style analytics, providing a Kimball style star-schema with a multi-dimensional view of all data and worked with Ralf Kimball's Change Data Capture (CDC) system, "diff compare", comparing with Hexadecimal (using MD5).
  • Created Complex SQL Queries using Views, Indexes, Triggers, Roles, Stored procedures and User Defined Functions Worked with different methods of logging in SSIS.
  • Worked in importing and cleansing of data from various sources like DB2, Oracle, flat files onto SQL Server with high volume data.
  • Involved in Migrating the data model from one database to Teradata database and prepared a Teradata staging model.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports by our BI team.
  • Created complex SQL server and Oracle stored procedures and triggers in support of the applications with Error handling and Created stored procedures and functions using Dynamic SQL and T- SQL.
  • Used Teradata SQL Assistant, Teradata Administrator, PMON and data load/export utilities like BTEQ, Fast Load, Multi Load, Fast Export, Tpump on UNIX/Windows environments and running the batch process for Teradata.
  • Responsible for creating Hive tables, loading data and writing hive queries.
  • Involved in Performance tuning the SQL Server Queries, which are used in SSRS Reports and developed Ad Hoc Report Model for specific users is SSRS.

Environment: Erwin 9.5, Teradata V14, Teradata SQL Assistant, Informatica Power Centre, Oracle 11g, Netezza, SQL Server 2014, Tableau, Mainframes, SQL, PL/SQL, XML, Hive, Hadoop, PIG, Hadoop, AWS, Redshift, S3, T-SQL, SSRS, SSIS, XML.

Confidential, Charlotte, NC

Sr. Data Modeler/Data Analyst

Responsibilities:

  • Worked with business users to gather requirements and create data flow, process flows and functional specification documents.
  • Developed Data Mapping, Data Governance and transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Developed, enhanced and maintained Snow Flakes Schemas within data warehouse and data mart with conceptual data models.
  • Created different views using Tableau Desktop that were published the business stake holders for analysis and customization using filters and actions.
  • Worked on enhancements to the Data Warehouse model using ER Studio as per the Business reporting requirements.
  • Designed 3rd normal form target data model and mapped to logical model.
  • Involved in extensive Data validation using SQLqueries and back-end testing and used SQL for Querying the database in UNIX environment
  • Involved with Data Analysis primarily Identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats
  • Involved in data analysis and creating data mapping documents to capture source to target transformation rules.
  • Used ER Studio and Visio to create 3NF and dimensional data models and published to the business users and ETL / BIteams.
  • Involved in Datamapping specifications to create and execute detailed system test plans. The datamapping specifies what data will be extracted from an internal data warehouse, transformed and sent to an external entity.
  • Developed Informatica SCD type-I, Type-II and Type III mappings and tuned them for better performance. Extensively used almost all of the transformations of Informatica including complex lookups, Stored Procedures, Update Strategy, mapplets and others.
  • Performed SAS programming using SAS Base and SAS Macros for loading data from ORACLE tables to SAS datasets and for generating reports on daily and fortnight basis.
  • Creating or modifying the T-SQL queries as per the business requirements and worked on creating role playing dimensions, factlessFact, snowflake and starschemas.
  • Designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using ER Studio.
  • Using ER Studio modeling tool, publishing of a data dictionary, review of the model and dictionary with subject matter experts and generation of data definition language.
  • Extracted data from databases Oracle, Teradata, Netezza, SQL server and DB2 using Informatica to load it into a single repository for Data analysis.
  • Involved in development and implementation of SSIS, SSRS and SSAS application solutions for various business units across the organization.
  • Managed full SDLC processes involving requirements management, workflow analysis, source data analysis, data mapping, metadata management, data quality, testing strategy and maintenance of the model.
  • Participated in integration of MDM (Master Data Management) Hub and data warehouses.
  • Created complex SSIS packages and SSRS Reports from Stored Procedures, Views and by writing complex T-SQL.
  • Designed and developed various analytical reports from multiple data sources by blending data on a single worksheet in Tableau Desktop.
  • Created and modified several UNIX shell Scripts according to the changing needs of the project and client requirements.
  • Identified and tracked the slowly changing dimensions, heterogeneous sources and determined the hierarchies in dimensions.
  • Wrote complex SQL queries for validating the data against different kinds of reports generated by Business Objects.
  • Used Teradata utilities such as Fast Export, MLOAD for handling various tasks.
  • Analysis of functional and non-functional categorized data elements for dataprofiling and mapping from source to target data environment. Developed working documents to support findings and assign specific tasks.
  • Developed SAS programs to support Statistical programming activities and created new SAS programs using SAS MACROS to improve ease and speed of modifications as well as consistency of results.
  • Involved in fixing invalid mappings, testing of Stored Procedures and Functions, Unit and Integrating testing of Informatica Sessions, Batches and the Target Data.
  • Involved in the validation of the OLAPUnittesting and SystemTesting of the OLAPReport Functionality and data displayed in the reports

Environment: ER Studio, Informatica Power Center 8.1/9.1, SAS, Power Connect/ Power exchange, Oracle 11g, Main frames,DB2 MS SQL Server 2012, SQL,PL/SQL, XML, Windows NT 4.0, Unix Shell Scripting, Teradata, Netezza, Aginity.

Confidential, New York NY

Sr. Data Modeler/Data Analyst

Responsibilities:

  • Created and maintained Logical and Physicalmodels for the data mart. Created partitions and indexes for the tables in the datamart.
  • Performed data profiling and analysis applied various data cleansing rules designed data standards and architecture/designed the relational models.
  • Maintained metadata (data definitions of table structures) and version controlling for the data model.
  • Did data modeling and involved with design and development of conceptual, logical and physical data models using All Fusion Data Modeler (Erwin)
  • Worked with the ETL team to document the transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
  • Complex mappings were developed to fetch data from data warehouse and required performance tuning done to get the required records within very less time.
  • Developed SQLscripts for creating tables, Sequences, Triggers, views and materializedviews
  • Worked on query optimization and performance tuning using SQL Profiler and performance monitoring.
  • Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings.
  • Utilized Erwin's forward/reverse engineering tools and target database schema conversion process.
  • Extracted, performed validation and generated SAS data sets from Teradata, applied SQL Pass through Facility.
  • Performed Data validation, Data cleansing, Data integrity, Data Quality checking before delivering data to operations, Business, Financial analyst by using Oracle, Teradata.
  • Worked on creating enterprise wide Model EDM for products and services in Teradata Environment based on the data from PDM. Conceived, designed, developed and implemented this model from the scratch.
  • Results are performed using SAS programming and using techniques such as SAS Macro language, and statistical procedures (e.g., PROC FREQ, PROC REPORT)
  • Created Rich dashboards using Tableau Dashboard and prepared user stories to create compelling dashboards to deliver actionable insights.
  • Facilitated in developing testing procedures, test cases and User Acceptance Testing (UAT) and applied data naming standards, created the data dictionary and documented data model translation decisions and also maintained DW metadata.
  • Write SQLscripts to test the mappings and Developed Traceability Matrix of Business Requirements mapped to Test Scripts to ensure any Change Control in requirements leads to test case update.
  • Responsible for development and testing of conversion programs for importing Data from text files into map Oracle Database utilizing PERL shell scripts &SQL*Loader.
  • Involved in extensive DATA validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
  • Developed and executed load scripts using Teradata client utilities MULTILOAD, FASTLOAD and BTEQ.
  • Developed and executed SAS SQL queries for merging, concatenating and updating large volumes of data.
  • Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS).
  • Worked with the ETL team to document the Transformation Rules for Data Migration from OLTP to Warehouse Environment for reporting purposes.
  • Created SQLscripts to find dataquality issues and to identify keys, data anomalies, and data validation issues.

Environment: Erwin, MS SQL Server 2008, DB2, Oracle SQL Developer, PL/SQL, Business Objects, Erwin, MS office suite, Windows XP, TOAD, SQL*PLUS, SQL*LOADER, Teradata, Netezza, Tableau, SQL Assistant, Informatica, XML.

Confidential, Malvern PA

Data Analyst

Responsibilities:

  • Created stored procedures, functions, Database triggers and packages as per the Business Needs for developing Ad-hoc and robust reports. Incorporated Dynamic SQL, for generation of the where clauses dynamically based on the lexical parameters passed.
  • Suggested various changes in the physicalmodel to support the business requirements.
  • Developed an application using Base SAS, Proc SQL and Proc Contents to generate concise data profiling reports.
  • Developed server jobs to load the data from flat files, text files, tag text files and MSSQL.
  • Generated comprehensive Analytical reports by running SQL Queries against current Databases to conduct Data Analysis.
  • Utilized shared containers for code reusability and for implementing the predefined business logic.
  • Involved in All the phases of the development like Analysis, Design, Coding, Unit Testing, System Testing and UAT
  • Developed data load programs using SAS Enterprise Guide to insert data periodically into the SAS server in UNIX.
  • Wrote, executed, performance tuned SQL Queries for Data Analysis& Profiling.
  • Wrote complex SQL queries using joins, sub queries and correlated sub queries.
  • Exported large SAS data sets to external files to be sent to external vendors using SAS and SAS/Macro.
  • Wrote PL/SQL stored procedures, functions and packages and triggers to implement business rules into the application.
  • Developed shellscripts to invoke back end SQL and PL/SQL programs
  • Performed unit testing to check the validity of the data at each stage.
  • Created ETL workflows in SAS/Data Integration Studio to populate the analytical data marts for business reporting.
  • Used Data Stage Director to debug the jobs and to view the error log to check for errors.
  • Implemented best practices in the development environment (code standards, code migration).
  • Used SAS PROC SQL Pass through Facility to work with Teradata database.
  • Written complex SQLs using joins, sub queries and correlated sub queries and expertise in SQL Queries for cross verification of data.
  • Used Informatica features to implement Type I & II changes in slowly changing dimension tables.
  • Created and ran workflows and Worklets using Workflow Manager to load the data into the target database.
  • Performance tuning of SQLQueries, Sources, Targets and sessions

Environment: SQL, T-SQL, SAS, SSRS, SSIS, DB2, Oracle 9i, MS SQL Server, PL/SQL, Toad, UNIX Shell Scripting.

We'd love your feedback!