Data Modeler /data Analyst Resume
Canonsburg, PA
PROFESSIONAL SUMMARY:
- Data modeler with Around 5+ years of experience in data analysis and modeling, with excellent understanding of Data Warehouse, ETL Developer and Data Mart designing.
- Extensive ETL tool experience working with IBM Data Stage and worked on DataStage client tools like DataStage Designer, DataStage Director and DataStage Administrator.
- Experienced in scheduling sequence, parallel and server jobs using DataStage Director, UNIX scripts and scheduling tools.
- Implemented Hive tables and HQL Queries for the reports.
- Design suitable data model and semantic layers for the Data Warehouse/Analytical Reporting/Operational Databases and define underlying data sources and transformation requirements
- Build/Provision Docker containers.
- Experience in developing pipelines in spark using Scala and python.
- Extensive experience in Informatica cloud services and creation and maintenance of database objects like tables, views, materialized views, indexes, constraints, primary keys, sequence, synonyms and database Link
- Utilized of Informatica IDQ 8.6.1 to complete initial data profiling and matching/removing duplicate data.
- Good experience in Data Stage Administration, Information Server (IS).
- Experience in Data cleansing by matching user introduced data with database data, removing duplicates, and extraction of relations from source systems using Quality Stage.
- Prepare data lineage documents.
- Experience in Data Enrichment and Re - Engineering Using Quality Stage and DataStage.
- Expert in Database Design, Data modeling, Development, Implementation, ETL and Reporting in SQL Server 2005/2008/2008 R2/2012 with expertise on data normalization.
- Good Experience with Django, a high-level Python Web framework.
- Used Python data frames to work with Google Big Query and extracted data tables on daily basis.
- Conversant with all phases of the Software Development Life Cycle (SDLC) especially Agile, Scrum, involving business process analysis, requirements gathering and analysis, SharePoint technologies, detailed design, development, testing and post implementation support.
- Having experience in designing customized interactive dashboards in Tableau using Marks, action, filters, parameter, security concepts, calculations and relationships.
- Hands on experience using query tools like TOAD, SQL Developer, PLSQL developer and Teradata SQL Assistant.
- Zabbix for monitoring docker host servers.
- Strong knowledge of Data Warehouse concepts and technologies such as ETL processes, dimensional modeling, Star and Snowflake Schemas, reporting tools and surrogate key generation.
- Highly proficient in creating database objects like tables, indexes, views, user defined functions, stored procedures, triggers, cursors, data integrity and constraints.
- Developed data governance policies as they fit the USMC needs, lending from my experience from storage sizing.
- Developed different types of reports using data from KUSTO.
- Strong expertise in Data modeling with good understanding of Relational Database Management Systems (RDBMS).
- Highly skilled in configuring and creating SSIS solutions for ETL and Business Intelligence Process for Data Integration and Migration services and managing Event Handlers, Package Configurations, Logging, System and User-defined Variables for SSIS Packages.
- Extensive experience in Data Extraction, Transformation and Loading (ETL) using DTS package in MS SQL Server 2005/2008/2008 R2/2012 SQL tools like SSIS, Data Transformation Services (DTS).
TECHNICAL SKILLS:
Database Architecture, Data Analysis,: Enterprise Data Warehouse, Database Design and Modeling, Data Integration and Migration, ETL Architecture and Design, Data Warehouse, OLTP, OLAP, Python, docker, Java 8, Azure.
Modeling Tools: Erwin 9.x, Rational Rose, ER/Studio, MS Visio, SAP Power designer, Embarcadero.
Databases: My SQL Server 2012/2008R2/2005, Oracle 12c/11g/10g/9i/8i, Teradata MS Access
Programming Languages: SQL, PL/SQL, UNIX shell scripting, PERL
Operating Systems: Windows, UNIX, MS DOS, Sun Solaris.
Web technologies: HTML, DHTML, XML, CSS.
Scripting Languages: VBScript, JavaScript, UNIX Shell Script.
Tools: MS-Office suite (Word, Excel, MS Project and Outlook), TOAD, BTEQ, Fast Load, Multi Load, Fast Export., Redshift.
PROFESSIONAL EXPERIENCE:
Confidential, Canonsburg, PA
Data Modeler /Data Analyst
Responsibilities:
- Performed System Study and Requirements Analysis, prepared Data Flow Diagrams, Entity Relationship Diagrams, Data Diagrams, Table Structures.
- Tested and deployed Google Analytics and Google Tag Manager accounts for Health and Human services public web sites.
- Cleaning and Blending multiple data sources to allow for different views on application data in a single dashboard Familiar with DAX Functions Such as SUMX, CALCULATE, ALLEXCEPT, FILTER Functions
- Worked Extensively with Databases, Tableau and Azure.
- Experience with JIRA, Confluence tools in the requirement management, sprint planning and the execution of sprints.
- Responsible for creating Hive tables, loading data and writing hive queries.
- Created Logical/Physical Data models in 3NF in the Warehouse area of Enterprise Data Warehouse.
- Performed Data analysis using Python Pandas.
- Developed jobs using Data Stage Designer to develop processes for extracting, transforming, integrating, and loading data from Mainframe COBOL, DB2 into desired target.
- Extracted the data from Teradata into HDFS using Sqoop.
- Expert in Creating and Deploying SSRS Reports on the portal for user accessibility.
- Performed integration testing between API, Nebula website and Nebula Hololens Application.
- Involved in the development of new Tableau Dashboards and publishing them to Tableau Server.
- Working on Cloud computing using Microsoft Azure with various BI Technologies and exploring NoSQL options for current back using Azure Cosmos DB (SQL API).
- Solution is deployed in a multi-cloud infrastructure with applications and services spanning on-premise as well as multi-tenant AWS and GCP cloud data centers .
- Experience in creating Data Governance Policies, Business Glossary, Data Dictionary, Data, Metadata, Data Lineage, and Data Quality Rules.
- Working knowledge in Google DataProc, Kubernates, Google cloud Machine learning.
- Used Scala and Python to convert Hive/SQL queries into RDD transformations in Apache Spark.
- Extensively used Informatica Client tools- Source Analyzer, Warehouse Designer, Mapping Designer.
- Experience in Software Configuration management using Github, IBM Rational ClearCase and Microsoft Visual SourceSafe (VSS).
- Designed ETL Process using ETL tools (Informatica) and Implementation of Data Movement, Error Capturing & Reporting, Initial & Delta Load, Implemented Change Data Capture methodology.
- Created DDL scripts for implementing Data Modeling changes. Created ERWIN crystal reports in HTML, RTF format depending upon the requirement.
- Good at Data Warehouse techniques -Dimensional data Modeling, Star Schema and Snowflake Schema
- Designed and developed star schema, snowflake schema and created fact tables and dimension tables for the warehouse and data marts using Erwin.
- Responsible for fetching real time data using Kafka and processing using Spark streaming with Scala.
- Provided guidance to ETL team to translate the data mapping document into a high level design document and also during the creation of ETL jobs.
- Created Power BI Dashboards Consuming Azure Cosmos DB and Converting the unstructured data into Consumable way to present it to the end users.
- Designed and developed ETL processes using DataStage to load data from Teradata, Flat Files to staging database and from staging to the target Data Warehouse database.
- Creating and Tracking defects in IBM Rational ClearQuest, Remedy Action Request System.
- Worked on troubleshooting, performance tuning and performances monitoring for enhancement of DataStage jobs and builds across Development, QA and PROD environments
- Created custom Google Analytics reports and dashboards focusing on KPIs to improve website traffic and user experience.
- Involved designing and developing packages for a Data Warehousing and Data Migrations projects using Integration services (SSIS) on different Data Sources.
- Created set of dashboards/reports for talent planning, performance management, competencies, and Executive profiles using OBIEE.
- Defined corporate Metadata definitions for all Enterprise Data Supported databases (operational source systems, data stores and data marts).
Environment: SQL Server Java, IBM Mainframe, XHTML, BigQuery, CSS, JSP, Hadoop, Oracle 11g, Data Warehousing, Data Mart, Metadata, Data stage, Azure, Quality Stage, SQL Server, Google Analytics, Visual Source Safe, IBM Rational Clear Quest, OBIEE, Windows XP, Oracle 10g, and Control M, Nebula, Spring Boot.
Confidential, Manhattan, NY
Data Modeler /Data Analyst
Responsibilities:
- Interacted with business users and analysts and gathered and documented the technical and business metadata.
- Created various standard/reusable jobs in DataStage using various active and passive stages like Sort, Lookup, Filter, Join, Transformer, aggregator, Change Capture Data, Sequential file, Datasets.
- Worked as a Data Modeler/Analyst to generate Data Models using Erwin and developed relational database system.
- Performed several ad-hoc data analyses in Azure Data bricks Analysis Platform on KANBAN board.
- Confluence is used for pulling the latest schedules dates for different phases
- Designed semantic layer data model. Conducted performance optimization for BI infrastructure.
- Used IBM Cognos Framework Manager for importing and retrieving Meta data from database and Query Studio for Meta data manipulation and leveraging client queries and requirements.
- Worked on multiple web Analytics projects involving end-to-end, Adobe Analytics development, support, and testing and data analysis.
- Created Project Expense Statements in all Data Marts and other Financial Reports using SSRS.
- Worked on Informatica Data Integration Tools - Such as Repository Manager, Designer, Workflow Manager, Workflow Monitor and Scheduled workflows using Workflow Manager.
- Delivered final source to target mapping and insert scripts to the Hadoop Developers.
- Designed and developed a customizable data management system using Hadoop to interface with the current RBAC system.
- Advanced skills on data cleaning, manipulation, Data migration and creation of Analytical Data mart with SAS/DI Studio.
- Developed and maintained data dictionary to create metadata reports for technical and business purpose.
- Metadata mapping of data when data are transformed from operational environment to data warehouse environment.
- Performed transformations using various SSIS tasks such as conditional split, derived column, that performed data scrubbing, including data validation checks during staging, before loading the data into the data warehouse.
- Created Views to reduce database complexities for the end users.
- Developed Unit Test Plans for SSIS packages for checking the functionality of each component.
- Involved in design, code, and deploy new data extractions using SSIS, and designed and produced documentation of data transformations for all extractions.
- Integrated data using SAS Data Integration Studio (SAS ETL Tool) for an end-to-end SAS data-mart solution for analytics and reporting.
- Provided guidelines and recommendations to the projects on their metadata, master data and data strategy and the selection of tools based on the approved set of tools within the EAD group
- Define the list codes and code conversions between the Source Systems and the Data Mart.
- Experienced in designing the data mart and creation of cubes.
- Extensive experience in working with Datastage Designer for developing jobs and Datastage Director to view the log file for execution errors.
- Designed database, tables and views structure for the new data mart.
- Developed complex T-SQL code such as Stored Procedures, functions, triggers, Indexes and views for the application.
- Assist the various stakeholders by providing them with process and program related data as requested, design layout of reports and determine the best way to present data to end users
- Performed transformations using various SSIS tasks such as conditional split, derived column, that performed data scrubbing, including data validation checks during staging, before loading the data into the data warehouse.
- Used ETL to implement the slowly changing transformation to maintain historical data in data warehouse.
Environment: ETL, SQL Server, Azure, BIRST, Java, XHTML, CSS, JSP, Oracle 11g, Data Warehousing, stakeholders, Data Mart, Metadata, SQL Developer, Toad, HP Quality Center, Unix, Windows XP, Oracle 10g and Teradata.
Confidential
Data Analyst
Responsibilities:
- Designed and developed the jobs using DataStage Designer for extracting, cleansing, transforming, integrating and loading data using various stages like Aggregator, Funnel, Change Capture, Change Apply and copy.
- Worked with DataStage Director to schedule, monitor and analyze performance of individual stages and run DataStage jobs.
- Gathered business requirements, definition and design of the data sourcing and data flows, data quality analysis, working in conjunction with the data warehouse architect.
- Designed SSIS Packages to transfer data from various sources like Text Files, XML (Extensible Markup Language) Files, SQL Server, Excel and Access to SQL Server using Business Intelligence Development Studio.
- Worked on DTS Packages, Import and Export Wizard for transferring data from Heterogeneous Database (Oracle and Text files and Excel files and MS word 2007 files) to SQL Server.
- Developed ETL process using SSIS with Various Control Flow, Data Flow task and Store Procedures for Work Order Validation process.
- Used Lookup transform to create the SSIS package for inserting the new entry from excel master source spreadsheet to the SQL Server table and update the old data.
- Responsible for the deployment of the reports on the reporting server and scheduled subscription for the windows file share.
- Designed different SharePoint views of libraries/lists which filtering the content depending on the need of users with different roles.
- Developed different types of reports like Tabular Reports, Matrix Reports, Drill-Down, Cross Tab Reports, Ad hoc reports, Drill through reports.
- Involved in writing T-SQL programming to implement Stored Procedures and Functions for different tasks and created Triggers to enforce data and referential integrity.
- Build packages and handled slowly changing dimensions to maintain the historic data.
- Responsible for optimizing all indexes, SQL queries, stored procedures to improve the quality of software.
- Evaluate business performance with key performance indicators ( KPI) using SQL Server Analysis Services.
- Used SQL Server profiler for auditing and analyzing the events which occurred during a particular time frame.
- Deployed reports on Microsoft Office SharePoint (Moss).
- Build the Dimensions, cubes with star schema and snowflake schema using SQL Server Analysis Services (SSAS) and created SSRS reports.
Environment: ETL, java, HTML, Java Script, Erwin, SQL Server, Data Warehousing, Data Mart, Metadata, UNIX Shell SQL Developer 8.0, Visual Source Safe 2005, HP Quality Center 10.0, UNIX, Windows XP, Oracle