Sr. Data Architect/modeler Resume
Township, NJ
SUMMARY:
- Experienced Data Architect, Data modeling and Data Analysis Professional with specialization in implementing End - To-End Business Intelligence and Data Warehousing solutions with over 9 + years of hands on experience.
- Experienced in integration of various relational and non-relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, COBOL, XML and Flat Files, to Netezza database.
- Working knowledge in Analysis Services (SSAS) and Integration Services (SSIS).
- Work on Background process in oracle Architecture. Also drill down to the lowest levels of systems design and construction.
- Extensive experience in development of T-SQL, DTS, OLAP, PL/SQL, Stored Procedures, Triggers, Functions, Packages, performance tuning and optimization for business logic implementation.
- Experience in Dimensional Data Modeling, Star / Snowflake schema, FACT & Dimension tables.
- Good understanding and hands on experience with MS Azure, AWS S3 and EC2.
- Good experience on programming languages Python, Scala.
- Work on Background process in oracle Architecture. Also drill down to the lowest levels of systems design and construction.
- Experienced in Technical consulting and end-to-end delivery with architecture, data modeling, data governance and design - development - implementation of solutions.
- Highly skilled in writing Parameterized Queries for generating parameterized reports and Sub reports using global variables, Expressions, and Functions, Sorting the data, Defining Data sources and Subtotals for the reports using SSRS
- Experienced in big data analysis and developing data models using Hive, PIG, and Map reduce, SQL with strong data architecting skills designing data-centric solutions.
- Excellent development experience SQL, Procedural Language(PL) of databases like Oracle, Teradata, Netezza and DB2
- Skillful in Data Analysis using SQL on Oracle, MS SQL Server, DB2 & Teradata.
- Expertise on Relational Data modeling (3NF) and Dimensional data modeling.
- Practical understanding of the Data modeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
- Skillful in Data Analysis using SQL on Oracle, MS SQL Server, DB2 & Teradata.
- Logical and physical database designing like Tables, Constraints, Index, etc. using Erwin, ER Studio, TOAD Modeler and SQL Modeler.
- Heavy use of Access queries, V-Lookup, formulas, Pivot Tables, etc. Working knowledge of CRM Automation Salesforce.com, SAP.
- Strong analytical, logical, Communication and problem solving skills and ability to quickly adapt to new technologies by self-learning.
- Experience in BI/DW solution (ETL,OLAP, Data mart), Informatica, BI Reporting tool like Tableau and Qlikview and also experienced leading the team of application, ETL, BI developers, Testing team.
- Logical and physical database designing like Tables, Constraints, Index, etc. using Erwin, ER Studio, TOAD Modeler and SQL Modeler.
- Hands on experience with big data tools like Hadoop, Spark, Hive, Pig, SparkSql.
- Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, HBase, Pig, Sqoop, and Flume.
- Experienced in Data Modeling & Data Analysis experience using Dimensional Data Modeling and Relational Data Modeling, Star Schema/Snowflake Modeling, FACT & Dimensions tables, Physical & Logical Data Modeling.
- Excellent knowledge in preparing required project documentation and tracking and reporting regularly on the status of projects to all project stakeholders.
TECHNICAL SKILLS:
Data Modeling Tools::
Erwin R6/R9, Rational System Architect, IBM Infosphere Data Architect, ER Studio and Oracle Designer.
Tools: OBIE 10g/11g/12c, SAP ECC6 EHP5, Go to meeting, Docusign, Insidesales.com, Share point, Mat-lab.
ETL/Data warehouse Tools::
Informatica 9.6/9.1/8.6.1/8.1, SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau 8.2, Pentaho
Big Data Technologies: Pig, Hive, Spark, Scala, HBase, Sqoop
Database Tools: Microsoft SQL Server12.0, Teradata 15.0, Oracle 11g/9i/12c and MS Access
BI Tools: Tableau 7.0/8.2, Tableau server 8.2, Tableau Reader 8.1,SAP Business Objects, Crystal Reports Packages: Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point Portal Server
Version Tool: GIT, SVN
Project Execution Methodologies::
Agile, Ralph Kimball and BillInmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
Quality Assurance Tools: Win Runner, Load Runner, Test Director, Quick Test Pro, Quality Center, Rational Functional Tester.
RDBMS: Microsoft SQL Server14.0, Teradata 15.0, Oracle 12c/11g/10g/9i, and MS Access
Operating System::
Windows, Unix, Sun Solaris
PROFESSIONAL EXPERIENCE:
Confidentia,l Township NJ
Sr. Data Architect/Modeler
- Working in Regulatory Compliance IT team where worked as Data Architect role which involved Data Profiling, Data Modeling, ETL Architecture & Oracle DBA .
- Implemented Join optimizations in Pig using Skewed and Merge joins for large datasets schema.
- Designed and developed a Data Lake using Hadoop for processing raw and processed claims via Hive and Informatica.
- Developed and implemented different Pig UDFs to write ad-hoc and scheduled reports as required by the Business team.
- Design of Big Data platform technology architecture. The scope includes data intake, data staging, data warehousing, and high performance analytics environment.
- Used Polybase for ETL/ELT process with Azure Data Warehouse to keep data in Blob Storage with almost no limitation on data volume.
- Created the template SSIS package that will replicate about 200 processes to load the data using Azure SQL
- Involved in loading data from LINUX file system to HDFS Importing and exporting data into HDFS and Hive using Sqoop Implemented Partitioning, Dynamic Partitions, Buckets in Hive.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
- Responsible for Big data initiatives and engagement including analysis, brainstorming, POC, and architecture.
- Supported agency data management programs and data governance activities to align with the enterprise architecture and in corporate data strategy and data governance principles into program, project, and IT portfolio management processes.
- Researched, evaluated, architect, and deployed new tools, frameworks and patterns to build sustainable Big Data platforms for the clients.
- Worked with Hadoop eco system covering HDFS, HBase, Yarn and MapReduce.
- Involved in the Data Architecture and in developing the overall ETL architecture, including solution architecture as needed.
- Designed the Logical Data Model using ERWIN 9.64 with the entities and attributes for each subject areas.
- Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
- Designed and developed a Data Lake using Hadoop for processing raw and processed claims via Hive and Informatica.
- Involved in designing and architecting data warehouses and data lakes on high performance (Teradata) and big data (Hadoop - Hive and HBase) databases.
- Generate big data Hadoop modeling for Hive database tables. Conceptual, logical and physical data models are designed with maintaining data governance and naming standards.
- Responsible for Big data initiatives and engagement including analysis, brainstorming, POC, and architecture.
- Hand on Cloud computing using Microsoft Azure with various BI Technologies.
- Involved in creating Hive tables, and loading and analyzing data using hive queries Developed Hive queries to process the data and generate the data cubes for visualizing Implemented.
- Developed Data Mapping, Data Governance, Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Exploring NoSQL options for current back using Azure Cosmos DB (SQL API)
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Involved in Normalization / De normalization techniques for optimum performance in relational and dimensional database environments.
- Performance tuning and stress-testing of NoSQL database environments in order to ensure acceptable database performance in production mode.
- Involved in writing Liquid base scripts and generating SQL's.
- Implemented strong referential integrity and auditing by the use of triggers and SQL Scripts.
- Created and managed database objects (tables, views, indexes, etc.) per application specifications. Implemented database procedures, triggers and SQL scripts for development teams.
- Perform routine management operations, including configuration and performance analysis for mongodb. Diagnosing Performance Issues for mongodb.
- Creating MongoDB data set backups using system-level file snapshot tool, such as LVM or native storage appliance tools.
- Point in time Backup and recovery in MongoDB using MMS. Data modeling for data from RDBMS to and MongoDB for optimal reads and writes.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Involved in designing Logical and Physical data models for different database applications using the Erwin.
- Data modeling, Design, implement, and deploy high-performance, custom applications Confidential scale on Hadoop /Spark.
- Implemented Data Integrity and Data Quality checks in Hadoop using Hive and Linux scripts.
- Reverse engineered some of the databases using Erwin.
- Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, SQL Server, and Oracle).
- Coordinating with Client and Business Analyst to understand and develop OBIEE reports.
- Designed and developed T-SQL stored procedures to extract, aggregate, transform, and insert data.
- Created and maintained SQL Server scheduled jobs, executing stored procedures for the purpose of extracting data from DB2 into SQL Server.
- Developed SQL Stored procedures to query dimension and fact tables in data warehouse.
- Experience with SQL Server Reporting Services (SSRS) to author, manage, and deliver both paper-based and interactive Web-based reports.
- Performed Hive programming for applications that were migrated to big data using Hadoop
- Generated parameterized queries for generating tabular reports using global variables, expressions, functions, and stored procedures using SSIS.
- Created External and Managed tables in Hive and used them appropriately for different PIG scripts required for reporting.
- Developed, and scheduled variety of reports like cross-tab, parameterized, drill through and sub reports with SSRS.
- Generate comprehensive analytical reports by running SQL queries against current databases to conduct data analysis pertaining to various loan products.
- Focused on architecting NoSQL databases like Mongo, Cassandra and Cache database.
Environment: DB2, CA Erwin 9.6, Oracle 12c, MS-Office, SQL Architect, TOAD Benchmark Factory, SQL Loader, PL/SQL, SharePoint, ERwin r9.64, Talend, MS-Office, Redshift, SQL Server 2008/2012, Hive, Pig, Hadoop, Spark, Azure.
Confidential, New YorkCity, NY
Sr. Data Architect/Modeler
- Responsible for technical Data governance, enterprise wide Data modeling and Database design.
- Creation of BTEQ, Fast export, Multi Load, TPump, Fast load scripts for extracting data from various production systems.
- Wrote JIL scripts to manage the scheduled loading of sample date into the various Databases and respective objects, through a combination of shell and SQL scripts.
- Generated Sybase SQL scripts that include creating tables, Primary Keys, and Foreign Keys.
- Created complex stored procedures, Functions, Triggers Indexes, Tables, Views and SQL joins for applications.
- Troubleshooting performance issues and fine-tuning queries and stored procedures.
- Generated reports to retrieve data using database code objects, such as Stored Procedures, views, functions and multiple T-SQL.
- Configure Collibra Data Governance Center DGC based on client requirements and specifications.
- Experience with AWS ecosystem (EC2, S3, RDS, Redshift).
- Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS).
- Generated reports using Global Variables, Expressions and Functions using SSIS.
- Involved with multiple project teams of technical professionals through all phases of the SDLC using technologies including Oracle, Erwin, DataStage, Data Warehousing and Cognos.
- Responsible for developing and supporting a data model and architecture that supports and enables the overall strategy of expanded data deliverables, services, process optimization and advanced business intelligence.
- Worked with Data Vault Methodology Developed normalized Logical and Physical database models
- Worked with Architecture team to get the metadata approved for the new data elements that are added for this project.
- Extensively used Erwin r9.6 for Data modeling. Created Staging and Target Models for the Enterprise Data Warehouse.
- Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ERWIN.
- Developed Hive scripts (using partitions, joins, buckets) to process the data for analysis (like quaterly growths of products, area wise sales, predict prices for new products based on market needs etc).
- Used Data Services to move data in to HIVE and Hadoop Clusters.
- Developed HIVE and Map-reduce tools to design and manage HDFS data blocks and data distribution methods.
- Implemented Data Vault Modeling Concept solved the problem of dealing with change in the environment by separating the business keys and the associations between those business keys, from the descriptive attributes of those keys using HUB, LINKS tables and Satellites.
- Worked on AWS Redshift and RDS for implementing models and data on RDS and Redshift.
- Designed ER diagrams (Physical and Logical using Erwin) and mapping the data into database objects and identified the Facts and Dimensions from the business requirements and developed the logical and physical models using Erwin.
- Developed mapping spreadsheets for (ETL) team with source to target data mapping with physical naming standards, data types, volumetric, domain definitions, and corporate meta-data definitions.
- Worked with Data Steward Team for designing, documenting and configuring Informatica Data Director for supporting management of MDM data.
- Worked on a POC to compare processing time of Impala with Apache Hive for batch applications to implement the former in project.
- Developed different kind of reports such as Drill down, Drill through, Sub Reports, Charts, Matrix reports, Parameterized reports and Linked reports using SSRS.
- Created Hive tables to process Advisor performance data in HDFS and exported it to downstream databases for various type of analyses using Sqoop.
- Developed application interface workflows using Groovy scripting in conjunction with the out of the box Collibra REST API's.
Environment: AWS Redshift, RDS, SSRS, Big Data, JDBC, Cassandra, NOSQL, Hive, Pig, Spark, Scala, Python, Hadoop, MySQL, PostgreSQL, SQL Server, Erwin, Informatica.
Confidential, Pittsburg, KS
Designation: Sr. Data Analyst/Modeler
- Involved in Data Architecture, Data profiling, Data analysis, data mapping and Data architecture artifacts design.
- Modified SSIS packages for manipulating the data Source coming from As400 to Back Office system. Altered the stored procedures and joins to efficient usage of mappings.
- Developed Complex Stored Procedures, Views and Temporary Tables as per the requirement.
- Created and developed the stored procedures, triggers to handle complex business rules, history data and audit analysis.
- Evaluated technical decisions and deliverables for compliance and quality analysis Designed, developed and deployed Apex Classes, Controller Classes and Apex Triggers, SOQL, Packages, Visual force Packages for various functional needs of the application.
- Creation of custom Visual force pages, apex triggers and apex classes, REST API Web services.
- Developed several Apex Triggers, Classes and Apex API.
- Wrote SOQL and SOSL statements within custom controllers, extensions and triggers.
- Create process models and enhance existing business process models using Bizagi Modeler.
- Worked with the ETL team to document the transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
- Performed as a Data Analysis, Data Migration and data profiling using complex SQL on various sources systems including Oracle and Teradata.
- Used forward engineering to generate DDL from the Physical Data Model and handed it to the DBA.
- Generated comprehensive analytical reports by running SQL queries against current databases to conduct Data Analysis.
- Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
- Performed logical data model design including normalization/de-normalization referential integrity, data domains; primary and foreign key assignments and data element definitions as applied to both relational and dimensional modeling.
- Generated a separate MRM document with each assignment and shared it on SharePoint along with the PDF of updated data models.
- Responsible for all metadata relating to the EDW’s overall data architecture, descriptions of data objects, access methods and security requirements.
- Perform logical and physical OLAP / OLTP schema design.
- Designed STAR schema for the detailed data marts and plan data marts consisting of confirmed dimensions.
- Used Erwin model mart for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Used Model Manager Option in Erwin to synchronize the data models in Model Mart approach.
- Created views and dashboards on end client's data. Produced powerful dashboards telling story behind the data in an easy to understand format such as pie, bar, geo, and line charts that are viewed daily by senior Management.
- Defined data management policies including enterprise naming and definition standards, specification of business rules, data quality analysis, standardized calculations and summarization definitions, and retention criteria.
- Created a list of domains in Erwin and worked on building up the data dictionary for the company.
- Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required.
- Modified cube and dimensions, deployed KPIs on SSRS and created different metrics, added new measure groups and dimensions.
- Development of reports using different SSIS Functionalities like sort prompts and cascading parameters, Multi Value Parameters.
- Created dynamic SSRS reports with the help of parameters which changes according to the user input.
- Created dashboard SSRS reports under report server projects and publishing SSRS reports to the reports server.
- Use SAS, SQL, XML, PL/SQL and Windows batch programming techniques to code the technical specifications apply business logic and produce automated reporting solutions.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, flat files, SQL Server 2005 with high volume data.
- Integrated Spotfire visualization into client's Salesforce environment.
- Created data masking mappings to mask the sensitive data between production and test environment.
Environment: Windows, Erwin, OLAP, OLTP, Teradata, SQL SERVER, SSRS, Informatica Power Center 6.1/7.1, QTP 9.2, Test Director 7.x, Load Runner 7.0, Oracle11g, UNIX AIX 5.2, PERL, Shell Scripting.
Confidential, Tampa, FL
Data Analyst/Modeler
- Created automatic field updates via workflows and triggers to satisfy internal compliance requirement of stamping certain data on a call during submission.
- Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
- Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark.
- Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
- Analyzed the physical data model to understand the relationship between existing tables. Cleansed the unwanted tables and columns as per the requirements as part of the duty being a Data Analyst.
- Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
- Worked on different data formats such as Flat files, SQL files, Databases, XML schema, CSV files
- Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
- Extensively used ER Studio as the main tool for modeling along with Visio
- Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
- Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ER Studio.
- Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark .
- Trained Spotfire tool and gave guidance in creating Spotfire Visualizations to couple of colleagues
- Developed Contracting Business Process Model Workflows (current / future state) using Bizagi Process Modeler software.
- Created a Data Mapping document after each assignment and wrote the transformation rules for each field as applicable
- Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
- Configured & developed the triggers, workflows, validation rules & having hands on the deployment process from one sandbox to other.
- Managed Logical and Physical Data Models in ER Studio Repository based on the different subject area requests for integrated model.
- Developed enhancements to Mongo DB architecture to improve performance and scalability.
- Forward Engineering the Data models, Reverse Engineering on the existing Data Models and Updates the Data models.
- Performed data cleaning and data manipulation activities using NZSQL utility.
- Analyzed and understood the architectural design of the project in a step by step process along with the data flow .
- Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
Environment: Oracle SQL Developer, Oracle Data Modeler, Teradata14, SSIS, Business Objects, SQL Server 2008, ER/Studio Windows, MS Excel.
Confidential
Data Analyst
- Maintained warehouse metadata, naming standards and warehouse standards for future application development.
- Extensively used ETL to load data from DB2, Oracle databases.
- Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
- Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
- Expertise and worked on Physical, logical and conceptual data model.
- Worked with Business users during requirements gathering and prepared Conceptual, Logical and Physical Data Models.
- Wrote PL/SQL statement, stored procedures and Triggers in DB2 for extracting as well as writing data.
- Optimized the existing procedures and SQL statements for the better performance using EXPLAIN PLAN, HINTS, SQL TRACE and etc. to tune SQL queries.
- The interfaces were developed to be able to connect to multiple databases like SQL server and oracle.
- Assisted Kronos project team in SQL Server Reporting Services installation.
- Developed SQL Server database to replace existing Access databases.
- Attended and participated in information and requirements gathering sessions
- Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
- Designed and created web applications to receive query string input from customers and facilitate entering the data into SQL Server databases.
- Performed thorough data analysis for the purpose of overhauling the database using SQL Server.
- Designed and implemented business intelligence to support sales and operations functions to increase customer satisfaction.
- Converted physical database models from logical models, to build/generate DDL scripts.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snow flake Schemas
- Wrote and executed unit, system, integration and UAT scripts in a data warehouse projects.
- Extensively used ETL methodology for supporting data extraction, transformations and loading processing, in a complex EDW using Informatica.
- Worked and experienced on Star Schema, DB2 and IMS DB.
Environment: Oracle, PL/SQL, DB2, ERWIN, UNIX, Teradata SQL Assistant, Informatica, OLTP, OLAP, Data Marts, DQ analyzer.