Sr. Data Architect/modeler / Engineer Resume
Peoria, IL
SUMMARY
- Over 9 Years in IT Experience, Versatile IT Professional Strong IT experienced in Data Architecture, Data Modeling, Data Engineer and Big Data Reporting Design and Development.
- Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export.
- Experience in Azure Cloud, Software Development, Analysis Datacenter Migration, Azure Data Factory (ADF) V2. Managing Database, Azure Data Platform services (Azure Data Lake(ADLS), Data Factory(ADF), Data Lake Analytics, Stream Analytics, Azure SQL DW, HDInsight/Data bricks, NoSQL DB),SQL Server, Oracle, Data Warehouse etc.
- Experienced in Technical consulting and end - to-end delivery with architecture, data modeling, data governance and design - development - implementation of solutions.
- Extensive experience Logical and physical database designing like Tables, Constraints, Index, etc. using Erwin, ER Studio, TOAD Modeler and SQL Modeler.
- Practical understanding of the Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
- Proficiency in Big Data Practices and Technologies like HDFS, MapReduce, Hive, Pig, HBase, Sqoop, Oozie, Flume, Spark, Kafka.
- Experienced in integration of various relational and non-relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, COBOL, XML and Flat Files, to Netezza database.
- Experience in BI/DW solution (ETL,OLAP, Data mart), Informatica, BI Reporting tool like Tableau and Qlikview and also experienced leading the team of application, ETL, BI developers, Testing team.
- Work on Background process in oracle Architecture. Also drill down to the lowest levels of systems design and construction.
- Skillful in Data Analysis using SQL on Oracle, MS SQL Server, DB2 & Teradata.
- Expertise on Relational Data modeling (3NF) and Dimensional data modeling.
- Skillful in Data Analysis using SQL on Oracle, MS SQL Server, DB2 & Teradata.
- Logical and physical database designing like Tables, Constraints, Index, etc. using Erwin, ER Studio, TOAD Modeler and SQL Modeler.
- Heavy use of Access queries, V-Lookup, formulas, Pivot Tables, etc. Working knowledge of CRM Automation Salesforce.com, SAP.
- Data Warehousing Full life-cycle project leadership, business-driven requirements, capacity planning, gathering, feasibility analysis, enterprise and solution architecture, design, construction, data quality, profiling and cleansing, source-target mapping, gap analysis, data integration/ETL, SOA, ODA, data marts, Inman/Kimball methodology, Data Modeling for OLTP, canonical modeling, Dimension Modeling for data ware house star/snowflake design.
- Have Knowledge on Apache Spark withCassandra.
- Experience in Dimensional Data Modeling, Star / Snowflake schema, FACT & Dimension tables.
- Good understanding and hands on experience with Azure, AWS S3 and EC2.
- Experienced on Hadoop Ecosystem and Big Data components including Apache Spark, Scala, Python, HDFS, Map Reduce, KAFKA.
- Good experience on programming languages Python, Scala.
- Heavy use of Access queries, V-Lookup, formulas, Pivot Tables, etc. Working knowledge of CRM Automation Salesforce.com SAP.
- Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
- Assist in creating communication materials based on data for key internal /external audiences.
TECHNICAL SKILLS
Data Modeling Tools: Erwin R6/R9, Rational System Architect, IBM Infosphere Data Architect, ER Studio and Oracle Designer.
Big Data Technologies: Pig, Hive, Spark, Scala, Kafka
Cloud Platforms: AWS, Azure.
ETL/Data warehouse Tools: Informatica 9.6/9.1/8.6.1/8.1 , SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau 8.2, Pentaho.
Database Tools: Microsoft SQL Server 12.0, Teradata 15.0, Oracle 11g/9i/12c and MS Access.
BI Tools: Tableau 7.0/8.2, Tableau server 8.2, Tableau Reader 8.1,SAP Business Objects, Crystal Reports Packages Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point Portal Server.
RDBMS: Microsoft SQL Server14.0, Teradata 15.0, Oracle 12c/11g/10g/9i, and MS Access
Version Tool: GIT, SVN
Project Execution Methodologies: Agile, Ralph Kimball and BillInmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
Tools: OBIE 10g/11g/12c, SAP ECC6 EHP5, Go to meeting, Docusign, Insidesales.com, Share point, Mat-lab.
Operating System: Windows, Unix, Sun Solaris
PROFESSIONAL EXPERIENCE
Confidential, Peoria, IL
Sr. Data Architect/Modeler / Engineer
Responsibilities:
- Responsible for developing and supporting a data model and architecture that supports and enables the overall strategy of expanded data deliverables, services, process optimization and advanced business intelligence.
- Worked with Data Vault Methodology Developed normalized Logical and Physical database models.
- Extensively involved in writing SQL queries (Sub queries and Join conditions), PL/SQL programming.
- Understanding the existing business processes and Interacting with Super User and End User to finalize their requirement.
- Design and Develop of Logical and physical Data Model of Schema Wrote PL/SQL code for data Conversion in there Clearance Strategy Project.
- Worked with Architecture team to get the metadata approved for the new data elements that are added for this project.
- Performed SAP Hybris Migration from Oracle to HANA predominantly using SAP HANA SDI.
- Extensively used Erwin r9.6 for Data modeling. Created Staging and Target Models for the Enterprise Data Warehouse.
- Oracle Golden Gate: learned fundamentals of golden gate and designed the replication processes using golden gate.
- Designed Distributed algorithms for identifying trends in data and processing them effectively.
- Used Spark and Scala for developing machine learning algorithms which analyses click stream data.
- Worked on to setup for the various Jenkins CICD pipeline configurations for all the micro services.
- Worked on Machine Learning Algorithms Development for analyzing click stream data using Spark and Scala. Implemented Oracle Golden gate in non-prod and prod environments.
- Created applications using Kafka, which monitors consumer lag within Apache Kafka clusters. Used in production by multiple report suites.
- Involved in Creation of tables, partitioning tables, Join conditions, correlated sub queries, nested queries, views, sequences, synonyms for the business application development.
- Upgraded Oracle 11.1.0.7 to Oracle 11.2.0.4 by using Oracle Golden gate with near to zero downtime.
- Enterprise Data Architect with focus on integration of Big Data, Data Lakes and relational databases with primary focus on Data Warehouses, Data Marts, Business Intelligence and Analytics
- Used Data Services to move data in to HIVE and Hadoop Clusters.
- Used ERWIN forward engineer to generate schema in a DB2 environment stored dividend information data models in a DB2 data warehouse.
- Responsible for backup and recovery of all DB2 physical objects.
- Explored the R statistical tool to provide data analysis on peer feedback data on leadership principles.
- Developed HIVE and Map-reduce tools to design and manage HDFS data blocks and data distribution methods.
- Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ERWIN.
- Worked with Data Steward Team for designing, documenting and configuring Informatica Data Director for supporting management of MDM data.
- Worked on a POC to compare processing time of Impala with Apache Hive for batch applications to implement the former in project.
- Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
- Worked on AWS and architecting a solution to load data, create data models.
- Migration of on premise data (Oracle/ SQL Server/ DB2/ MongoDB) to Azure Data Lake Store(ADLS) using Azure Data Factory(ADF V1/V2).
- Installed SAP HANA, SAP HANA Data Provisioning (DP) Agent and configured it in Oracle Server in SLES servers in POC environments.
- Developed KAFKA and consumers, HBase clients, SPARK and Hadoop Map Reduce jobs along with components on HDFS, Hive.
- Creation of BTEQ, Fast export, Multi Load, TPump, Fast load scripts for extracting data from various production systems.
- Created data models for AWS and Hive from dimensional data models and involved in the Configuration of Hadoop Ecosystems with developers to read data transaction from HDFS and Hive.
- Generated Sybase SQL scripts that include creating tables, Primary Keys, and Foreign Keys.
- Created complex stored procedures, Functions, Triggers Indexes, Tables, Views and SQL joins for applications.
- Troubleshooting performance issues and fine-tuning queries and stored procedures.
- Generated reports to retrieve data using database code objects, such as Stored Procedures, views, functions and multiple T-SQL.
- Used ERWIN forward engineer to generate schema in DB2 ORACLE SQL/SERVER environments .Data Models were stored in DB2 data warehouse.
- Selecting the appropriate AWS service based on data, compute, database, or security requirements.
- Developed several behavioral reports and data points creating complex SQL queries and stored procedures using SSRS and Excel.
- Designed DB2Architecture for Modeling a Data Warehouse by using tools like Erwin9.6.1,PowerDesigner and E-RStudio.
- Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS).
- Worked on the CICD setup for all the micro services on Profile, Service Availability, Cart, Pricing and promotion, Product Configuration, Foundation framework and product listing.
- Generated reports using Global Variables, Expressions and Functions using SSRS.
- Created Hive tables to process Advisor performance data in HDFS and exported it to downstream databases for various type of analyses using Sqoop.
- Exploring NoSQL options for current back using Azure Cosmos DB (SQL API).
- Developed Hive scripts (using partitions, joins, buckets) to process the data for analysis (like quaterly growths of products, area wise sales, predict prices for new products based on market needs etc).
- Implemented Data Vault Modeling Concept solved the problem of dealing with change in the environment by separating the business keys and the associations between those business keys, from the descriptive attributes of those keys using HUB, LINKS tables and Satellites.
- Designed ER diagrams (Physical and Logical using Erwin) and mapping the data into database objects and identified the Facts and Dimensions from the business requirements and developed the logical and physical models using Erwin.
Environment: Erwin, Informatica, AWS, SSRS, JDBC, Cassandra, NOSQL, Hive, Pig, Spark, Scala, Python, Hadoop, MySQL, PostgreSQL, SQL Server.
Confidential, Minneapolis MN
Sr. Data Modeler
Responsibilities:
- Worked as a Data Architect / Modeler to generate Data Models using Erwin and developed relational database system and massively involved in Data Architect role to review business requirement and compose source to target data mapping documents.
- Worked on CI/CD technologies both for Confidential - native application and non- Confidential applications, And build Automation tools; Chef, Jenkins, Puppet.
- Designed and developed Data models for Database (OLTP), the Operational DataStore (ODS), Data warehouse (OLAP), and federated databases to support client enterprise Information Manageme.
- Developed data pipeline using Kafka, Sqoop, Hive to ingest Customer transactional data and behavioral data into HDFS for processing and analysis.
- Created Communities, Domains, Assets, hierarchies in Collibra.
- Upgraded Oracle database to 12.1.0.2 by using various techniques (Golden gate, Data Pump, Data guard).
- Assisted with review of Oracle Golden gate replication architecture.
- Enterprise Data Architect with focus on integration of Big Data, Data Lakes and relational databases with primary focus on Data Warehouses, Data Marts, Business Intelligence and Analytics.
- Worked on Azure transformation projects and Azure architecture decision making Architect and implement ETL and data movement solutions using Azure Data Factory(ADF), SSIS.
- Migration of on premise Mongo DB to Cosmos DB using Azure Data Factory .
- Worked in Regulatory Compliance IT team where worked as Data Architect role which involved Data Profiling, Data Modeling, ETL Architecture & Oracle DBA.
- Extensively used Erwinr9 for Data modeling. Created Staging and Target Models for the Enterprise Data Warehouse.
- Designed and implemented SAP HANA data modeling using Attribute View, Analytic View, and Calculation View.
- Developed tools using Python, Shell scripting, XML to automate some of the menial tasks. Interfacing with supervisors, artists, systems administrators and production to ensure production deadlines are met.
- Migrated on-premises clusters to Microsoft Azure Cloud and enabled data scientists perform machine learning and advanced analytics by utilizing Azure Data Factory Pipelines, Data Lakes, Blobs, Catalogs, Keyvaults, HDInsight, Databricks, Azure ML Studio, PowerShell, Automations, Runbooks, CI/CD tools (Chef, Jenkins, Ansible, and Kubernetes, Docker, container orchestration), DevOps.
- Responsible for Big data initiatives and engagement including analysis, brainstorming, POC, and architecture.
- Designed the Logical Data Model using ERWIN 9.64 with the entities and attributes for each subject areas.
- Developed the Talend mappings using various transformations, Sessions and Workflows. Teradata was the target database, Source database is a combination of Flat files, Oracle tables, Excel files and Teradata database.
- Involved in creating Hive tables, and loading and analyzing data using hive queries Developed Hive queries to process the data and generate the data cubes for visualizing Implemented.
- Implemented Join optimizations in Pig using Skewed and Merge joins for large datasets schema.
- Designed and developed a Data Lake using Hadoop for processing raw and processed claims via Hive and Informatica.
- Developed and implemented different Pig UDFs to write ad-hoc and scheduled reports as required by the Business team.
- Involved in the process of adding a new Datacenter to existingCassandraCluster.
- Used Polybase for ETL/ELT process withAzureDataWarehouse to keepdatain Blob Storage with almost no limitation ondatavolume.
- Utilized Sqoop, Kafka, Flume and Hadoop File System API's for implementing data ingestion pipelines.
- Created the template SSIS package that will replicate about 200 processes to load thedatausingAzureSQL.
- Implemented Data Integrity and Data Quality checks in Hadoop using Hive and Linux scripts.
- Involved in loading data from LINUX file system to HDFS Importing and exporting data into HDFS and Hive using Sqoop Implemented Partitioning, Dynamic Partitions, Buckets in Hive.
- Designed and developed architecture fordataservices ecosystem spanning Relational, NoSQL, and BigDatatechnologies.
- Exploring NoSQL options for current back usingAzureCosmos DB (SQL API)
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Involved in Normalization / De normalization techniques for optimum performance in relational and dimensional database environments.
- Performance tuning and stress-testing of NoSQL database environments in order to ensure acceptable database performance in production mode.
- Implemented Spark solution to enable real time reports fromCassandradata.
- Implemented strong referential integrity and auditing by the use of triggers and SQL Scripts.
- Designed and developed T-SQL stored procedures to extract, aggregate, transform, and insert data.
- Created and maintained SQL Server scheduled jobs, executing stored procedures for the purpose of extracting data from DB2 into SQL Server.
- Developed SQL Stored procedures to query dimension and fact tables in data warehouse.
- Worked on SQL Server Reporting Services (SSRS) to author, manage, and deliver both paper-based and interactive Web-based reports.
- Used Talend to Extract, Transform and Load data into Netezza Data Warehouse from various sources like Oracle and flat files.
- Performed Hive programming for applications that were migrated to big data using Hadoop
- Deployed SSRS reports to Report Manager and created linked reports, snapshots, and subscriptions for the reports and worked on scheduling of the reports.
- Generated parameterized queries for generating tabular reports using global variables, expressions, functions, and stored procedures using SSRS.
- Created External and Managed tables in Hive and used them appropriately for different PIG scripts required for reporting.
- Focused on architecting NoSQL databases like Mongo, Cassandra and Cache database.
- Point in time Backup and recovery in MongoDB using MMS. Data modeling for data from RDBMS to and MongoDB for optimal reads and writes.
- Docker files created as part of the EP containers application to implement CI/CD process.
- Involved in designing Logical and Physical data models for different database applications using the Erwin.
- Reverse engineered some of the databases using Erwin.
- Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, SQL Server, and Oracle).
- Coordinating with Client and Business Analyst to understand and developOBIEEreports.
Environment: DB2, CA Erwin 9.6, Oracle 12c, MS-Office, SQL Architect, TOAD Benchmark Factory, SQL Loader, PL/SQL, SharePoint, ERwin r9.64, Talend, MS-Office, Redshift, SQL Server 2008/2012, Hive, Pig, Hadoop, Spark, Azure.
Confidential, Denver, CO
Sr. Data Analyst/Modeler
Responsibilities:
- Performedas a Data Analysis, Data Migration and data profiling using complex SQL on various sources systems including Oracle andTeradata.
- Used forward engineering to generate DDL from the Physical Data Model and handed it to the DBA.
- Achieved 45% growth for memory and compute utilization using Spark and Hive using optimization techniques.
- Exposure to spark batch processing using Scala and spark streaming using the Kafka data pipeline.
- Developed more than 20 jobs using the IBM Data Stage ETL tool for data warehouse consolidation.
- Have working knowledge of 4 to 5 Amazon Web Services, such as S3, EMR, Athena, Redshift, and Glue.
- Created automated scripts for regression testing using Transact-SQL and Python.
- Managing code reviews for more than 110 deployment activities using SQL and other tools.
- Generated comprehensive analytical reports by running SQL queries against current databases to conductData Analysis.
- Created a list of domains in Erwin and worked on building up the data dictionary for the company.
- Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
- Performed logical data model design including normalization/de-normalization referential integrity, data domains; primary and foreign key assignments and data element definitions as applied to both relational and dimensional modeling.
- Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required.
- Created dashboard SSRS reports under report server projects and publishing SSRS reports to the reports server.
- UseSAS, SQL, XML, PL/SQL and Windows batch programming techniques to code the technical specifications apply business logic and produce automated reporting solutions.
- Preparation of business (Collibra) and technical metadata (IBM Infosphere)
- Modified SSIS packages for manipulating the data Source coming from As400 to Back Office system. Altered the stored procedures and joins to efficient usage of mappings.
- Created various SSRS dashboard reports for commission, sales and profit analysis.
- Developed Complex Stored Procedures, Views and Temporary Tables as per the requirement.
- Created and developed the stored procedures, triggers to handle complex business rules, history data and audit analysis.
- Creation of custom Visual force pages, apex triggers and apex classes, REST API Web services.
- Developed several Apex Triggers, Classes and Apex API.
- Wrote SOQL and SOSL statements within custom controllers, extensions and triggers.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, flat files, SQL Server 2005 with high volume data.
- Designed ETL process using Talend Tool to load from Sources to Targets through data Transformations.
- Perform logical and physical OLAP / OLTP schema design.
- Designed STAR schema for the detailed data marts and plan data marts consisting of confirmed dimensions.
- Used Erwin model mart for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Used Model Manager Option in Erwin to synchronize the data models in Model Mart approach.
- Created views and dashboards on end client'sdata. Producedpowerful dashboards telling story behind the data in an easy to understand format such as pie, bar, geo, and line charts that are viewed daily by senior Management.
Environment: Windows, Erwin, OLAP, OLTP, Teradata, SQL SERVER, SSRS, Informatica Power Center 6.1/7.1, QTP 9.2, Test Director 7.x, Load Runner 7.0, Oracle11g, UNIX AIX 5.2, PERL, Shell Scripting.
Confidential, Richmond, VA
Data Analyst
Responsibilities:
- Analyzed the physicaldatamodel to understand the relationship between existing tables. Cleansed the unwanted tables and columns as per the requirements as part of the duty being aDataAnalyst.
- Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
- Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ER Studio.
- Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark .
- Created DDL scripts for implementing Data Modeling changes. Created ER Studio reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
- Developed the Talend jobs and make sure to load the data into HIVE tables & HDFS files and develop the Talend jobs to integrate with Teradata system from HIVE tables
- Developed Contracting Business Process Model Workflows (current / future state) usingBizagi Process Modeler software.
- Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
- Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
- Extensively used ER Studio as the main tool for modeling along with Visio
- Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
- Configured & developed the triggers, workflows, validation rules & having hands on the deployment process from one sandbox to other.
- Managed Logical and Physical Data Models in ER Studio Repository based on the different subject area requests for integrated model.
- Created automatic field updates via workflows and triggers to satisfy internal compliance requirement of stamping certain data on a call during submission.
- Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark.
- Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
- Developed enhancements toMongo DBarchitecture to improve performance and scalability.
- Forward Engineering the Data models, Reverse Engineering on the existing Data Models and Updates the Data models.
Environment: Oracle SQL Developer, OracleDataModeler, Teradata14, SSIS, Business Objects, SQL Server 2008, ER/Studio Windows, MS Excel.
Confidential, Atlanta, GA
Data Modeler/ Analyst
Responsibilities:
- Develop Logical and Physicaldatamodels that capture current state/future statedataelements and dataflows using Erwin / Star Schema.
- Analyzed database performance with SQL Profiler and Optimized indexes to significantly improve performance.
- Automated load run on Informatica sessions through UNIX Corn, PL/SQL scripts and implemented pre and post-session scripts, also automated load failures with successful notification through email.
- Coordinated with ETL team, DB administrators& BI teams to elevate thedatamodel changes in the system
- Completed High-Level design documents which included aBIZAGIdiagram of several processes, macro processes, use cases, activities, interfaces, business rules, gaps and premises
- Prepared scripts for model anddatamigration from DB2 to the new Appliance environments.
- ImplementedDataArchiving strategies to handle the problems with large volumes ofdataby moving inactivedatato another storage location that can be accessed easily.
- Worked closely hand in hand with the Business Analytics manager, who was also a part of the design/datamodeling team.
- Experience in Project development and coordination with onshore-offshore ETL/BI developers & Business Analysts.
- Led enterprise logicaldatamodeling project (in third normal form) to gatherdatarequirements for OLTP enhancements.
- Modeled multiple new LOBs (Financialdata) into existing Financial ServicesEDW, with modifications to existing and new entities.
- Converted third normal form ERDs into dimensional ERDs fordatawarehouse effort.
- Involved in mapping spreadsheets that will provide theDataWarehouse Development (ETL) team with source to targetdatamapping, inclusive of logical names, physical names,datatypes, domain definitions, and corporate meta-datadefinitions.
- Converted physical database models from logical models, to build/generate DDLscripts.
- Maintained warehouse metadata, naming standards and warehouse standards for future application development.
- Extensively used ETL to loaddatafrom DB2, Oracle databases.
- Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
- Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
- Expertise and worked on Physical, logical and conceptual data model
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snow flake Schemas
- Wrote and executed unit, system, integration and UAT scripts in a data warehouse projects.
- Extensively used ETL methodology for supporting data extraction, transformations and loading processing, in a complex EDW using Informatica.
- Worked and experienced on Star Schema, DB2 and IMS DB.
- Worked on Optimization of the application and the designing of the database tables with the right partitioning keys using the DPF feature of hash partitioning and range partitioning.
- Performed cross-platform database migration for development & Production databases running on Sun OS to LINUX using utilities such as db2move and db2look.
Environment: ERWIN, UNIX, Oracle, PL/SQL, DB2, Teradata SQL assistant, DQ analyzer