We provide IT Staff Augmentation Services!

Sr. Data Modeler/data Analyst Resume

4.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY:

  • Has more than 9 years of experienceover Data Analysis and Data modeling Professional with specialization in implementing End - To-End Business Intelligence and Data Warehousing solutions.
  • Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop, Flume.
  • Good understanding and experienced on AWS, big data concepts and Hadoop ecosystem and experience over data Transformation using Pig scripts in AWS EMR, AWS RDS.
  • Expertise lies in Data Modeling, Database design and implementation of Oracle, Teradata, SQL Server, DB2, NoSQL Databases like MongoDB, HBase, AWS Redshift databases and Administration, Performance tuning etc.
  • Experienced in Azure SQL, Azure Data Lake, Data Lake Analytics, Data Lake Factory; convert SSIS package into Azure Data Factory and automatically run.
  • Expertise in transforming business requirements into analytical models, designing algorithms, building models, developing data mining and reporting solutions that scale across a massive volume of Structured and unstructured data.
  • Experience working with data modeling tools like Erwin, Power Designer and ER Studio and experience in Architecture, Design and Development of large Enterprise Data Warehouse (EDW) and Data-marts for Confidential user-base consumption.
  • Strong experience in analyzing/ Data Transformation of large amounts of data sets writing Pig scripts and Hive, AWS EMR, AWS RDS.
  • Experienced in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
  • Experienced providing solutions in Online Transactional and Data warehousing systems in Oracle projects using Oracle PLSQL, Unix Shell Scripting, ETL tools and Batch Design.
  • Solid knowledge and experience of Data Marts, Operational Data Store (ODS), OLAP, Dimensional Data Modeling with Ralph Kimball Methodology (Star Schema Modeling, Snow-Flake Modeling for FACT and Dimensions Tables) using Analysis Services.
  • Well versed in Normalization / De-normalization techniques for optimum performance in relational and dimensional database environments and strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export.
  • Experienced in various Teradata utilities like Fastload, Multiload, BTEQ,Tpump and Teradata SQL Assistant.
  • Develop and manage SQL and Python bases for data cleansing and data analysis using Git version control and experience in data analysis using Hive, PigLatin, and Impala.
  • Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server and Teradata and excellent in creating various artifacts for projects which include specification documents, data mapping and data analysis documents.
  • Experienced in report generation using SSRS and creating various types of reports like drill down, Parameterized, Cascading, Conditional, Table, Matrix, Chart, and Sub Report
  • Extensive experience on ETL using Informatica (Power Center/ Power Mart) (Designer, Workflow Manager, Workflow Monitor and Server Manager), Talend and SSIS.
  • An excellent team player& technically strong person who has capability to work with business users, project managers, team leads, architects and peers, thus maintaining healthy environment in the project.
  • Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
  • Experienced in ETL design, development and maintenance using Oracle SQL, PL/SQL, TOAD SQL Loader, and Database Management System (RDBMS).

TECHNICAL SKILLS:

Data Modeling Tools: Erwin R9.x, Rational System Architect, ER Studio and Power Designer.

ETL/Data warehouse Tools: Informatica 9.6/9.1/8.6.1, SAP Business Objects XIR3.1/XIR2, Web Intelligence, SSIS, TableauDatabase Tools: Microsoft SQL Server12.0, Teradata 15.0, Oracle 12c/11g/10g and MS Access.

Big Data Technologies: Pig, Hive, Spark, Scala, Sqoop, MongoDB, Cassandra, HBase, Kafka.

BI Tools: Tableau 7.0/8.2/10.x, Tableau server 8.2, Tableau Reader 8.1, SAP Business Objects, Crystal Reports Packages: Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point Portal Server

Cloud Platforms: AWS, Azure, AWS RDS, AWS S3 and AWS EMR, Azure Data Factory, Azure Blob.

Tools: OBIE 10g/11g/12c, SAP ECC6 EHP5, Go to meeting, Docusign, Insidesales.com, Share point, Mat-lab.

Operating System: Windows, Unix, Sun Solaris

RDBMS: Microsoft SQL Server14.0, Teradata 15.0, Oracle 12c/11g/10g/9i, and MS Access

Version Tool: GIT, SVN

Project Execution Methodologies: Agile, Ralph Kimball and BillInmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)

WORK EXPERIENCE:

Confidential, Chicago, IL

Sr. Data Modeler/Data Analyst

Responsibilities:

  • Owned and managed all changes to the data models. Created data models, solution designs and data architecture documentation for complex information systems and using Erwin r9.64 and developed relational database system.
  • Translated the business requirements into workable functional and non-functional requirements at detailed production level using Workflow Diagrams, Sequence Diagrams, Activity Diagrams and Use Case Modeling and used data vault modeling method which was adaptable to the needs of this project.
  • Designed the Logical Data Model using ERWIN 9.64 with the entities and attributes for each subject areas and responsible for Dimensional Data Modeling and Modeling Diagrams using ERWIN and used Tableau for BI Reporting and Data Analysis.
  • Converted SSIS package into Azure Data Factory and automatically run on a daily basis and used Azure SQL DW, Azure Data Lake, and Data Lake Analytics to save, retrieve and analyze data.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
  • Developed Data Model -Conceptual/l Logical/ Physical DM for ODS & Dimensional delivery layer in Azure SQL Data Warehouse
  • Used Flume extensively in gathering and moving log data files from Application Servers to a central location in Hadoop Distributed File System (HDFS) for data science.
  • Involved in Normalization / De-normalization techniques for optimum performance in relational and dimensional database environments.
  • Working with project management, business teams and departments to assess and refine requirements to design/develop BI solutions using Azure.
  • Utilize Python libraries for data processing (pandas) and representation (Plotly) while overseeing automation efforts across banking domains leveraging Azure, Python, and big data tools.
  • Collected large amounts of log data using Apache Flume and aggregating using PIG/HIVE in HDFS for further analysis and loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data.
  • Designing Star Denormalized tables on Azure and utilize U-SQL for data analytics/data ingestion of raw data in Azure and Blob storage
  • Participated in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi-dimensional models such as Star Schemas and Snowflake Schema.
  • Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Converting existing hive queries to Spark SQL queries to reduce execution time.
  • Demonstrated expertise utilizing ETL tools, including SQL Server Integration Services (SSIS), Data Transformation Services (DTS), and Data Stage and ETL package design, and RDBM systems like SQL Servers, Oracle, and DB2.
  • Used the Data Warehousing Life Cycle to identify data elements from the source systems, performed data analysis to come up with data cleansing and integration rules for the ETL process and creating Data mappings, Tech Design, loading strategies for ETL to load newly created or existing tables.
  • Monitored NoSQL data clusters and the monitoring utilities surrounding implementations while constructing statistical modeling of operational performance.
  • Involved in extensive data validation by writing several complexes SQL queries and involved in back-end testing and worked with data quality issues and perform administrative tasks, including creation of database objects such as database, tables, and views, using SQL DCL, DDL, and DML requests.
  • Created advanced chart visualizations in Tableau using Dual Axis, Box Plots, Bullet Graphs, Tree maps, Bubble Charts, Pie Chart, Gantt chart, Histograms.

Environment: Oracle 12c, MS-Office, SQL, Spark, TOAD Benchmark Factory, Teradatav15, Hadoop, SQL Loader, SharePoint, ERwin r 9.64, DB2, MS-Office, SQL Server 2012/2016, Azure, HBase, Hive, Hadoop, Python, SparkSQL, MongoDB, Azure Data Factory, ETL SSIS, Tableau and SSRS.

Sr. Data Modeler/ Data Analyst

Confidential, Dallas, TX

Responsibilities:

  • Designed and implemented a Data Lake to consolidate data from multiple sources, using Hadoop stack technologies like SQOOP, HIVE/HQL.
  • Developed, managed and validated existing data models including logical and physical models of the Data Warehouse and source systems utilizing a 3NF model and used Star schema and Snowflake Schema methodologies in building and designing the Logical Data Model into Dimensional Models.
  • Performed as a Data Analysis, Data Modeling, Data Migration and data profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Created data models for AWS Redshift and Hive from dimensional data models and worked on Data modeling, Advanced SQL with Columnar Databases using AWS.
  • Developed Map Reduce programs to cleanse the data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Managed Logical and Physical Data Models in ER Studio Repository based on the different subject area requests for integrated model. Developed Data Mapping, Data Governance, and transformation and cleansing rules involving OLTP, ODS and enforced referential integrity in the OLTP data model for consistent relationship between tables and efficient database design.
  • Worked on MongoDB database concepts such as locking, transactions, indexes, sharding, replication, schema design and developed enhancements to MongoDB architecture to improve performance and scalability.
  • Responsible for full data loads from production to AWS Redshift staging environment and Worked on migrating of EDW to AWS using EMR and various other technologies.
  • Developed different kind of reports such as Drill down, Drill through, Sub Reports, Charts, Matrix reports, Parameterized reports and Linked reports using SSRS and developed several behavioral reports and data points creating complex SQL queries and stored procedures using SSRS and Excel.
  • Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data.
  • Strong ability in developing advanced ANSI SQL queries to extract, manipulate, and/or calculate information to fulfill data and reporting requirements including identifying the tables and columns from which data is extracted.
  • Worked with Full life cycle of Data Lake, Data Warehouse with Big data technologies like Spark, Hadoop.
  • Used Data Vault technique and achieved many advantages of Data Vault approach some of them are simplified the data ingestion process, removed the cleansing requirement of a star schema and easily allowed for the addition of new data sources without disruption to existing schema
  • Used SQL tools like Teradata SQL Assistant and TOAD to run SQL queries and validate the data in warehouse and used Teradata 14 utilities such as Fast Export, Mload for handling various tasks data migration/ETL from OLTP Source Systems to OLAP Confidential Systems.
  • Extensively used ETL methodology for supporting data extraction, transformations and loading processing, in a complex EDW using Informatica and utilized Informatica toolset (Informatica Data Explorer, and Informatica Data Quality) to analyze legacy data for data profiling.
  • Performed POC's on NoSQL databases and implemented NoSQL databases like Mongo, Cassandra in Dev/Test environment.
  • Created and maintained surrogate keys on the master tables to handle SCD type 2 changes effectively and created several UNIX shell scripts to implement new features in the modules and update the existing one.
  • Developed and configured on Informatica MDM hub supports the Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS).

Environment: ER Studio, Oracle 12c, Informatica, MDM, SQL, Teradata 14, MongoDB, HBase, AWS Redshift, AWS EMR, AWS S3, AWS EC2, SSRS, Hadoop, Hive, Spark, ETL, UNIX, Teradata SQL Assistant, Sqoop, Tableau, Python, HDFS, T-SQL, OLAP/OLTP, Netezza, SQL Server 2012, DB2, Control M and Netezza.

Sr. Data Modeler/Data Analyst

Confidential, Minneapolis, MN

Responsibilities:

  • Participated in the design, development, and support of the corporate operation data store and enterprise data warehouse database environment.
  • Documented a whole process of working with Tableau Desktop, installing Tableau Server and evaluating Business Requirements.
  • Participated in big data architecture for both batch and real-time analytics and mapped data using scoring system over large data on HDFS
  • Implemented dimension model (logical and physical data modeling) in the existing architecture using ER/Studio.
  • Involved in Database using Oracle, XML, DB2, Teradata 14.1, Netezza, SQL server, Big Data and NoSQL based MongoDB and Cassandra.
  • Developed, managed and validated existing Data Models including Logical and Physical Models of the Data Warehouse and source systems utilizing a 3NFmodel.
  • Worked on predictive and what-if analysis using Python from HDFS and successfully loaded files to HDFS from Teradata, and loaded from HDFS to HIVE.
  • Gathering the business requirements from customers and creating data models for different branches using MS access and ERwin.
  • Designed and deployed scalable, highly available, and fault tolerant systems on AWS.
  • Worked on NoSQL databases including HBase, Mongo DB, and Cassandra. Implemented multi-datacenter and multi-rack Cassandra cluster.
  • Designed Source to Confidential mapping from primarily Flat files, SQL Server, Oracle 11g, Netezza using Informatica Power Center9.6.
  • Cleansed the data by eliminating duplicate and inaccurate data in Python and used Python scripts to update the content in database and manipulate files.
  • Executed ad-hoc data analysis for customer insights using SQL using Amazon AWS Hadoop Cluster. Worked on Normalization and De-Normalization techniques for both OLTP and OLAP systems.
  • Designed Logical data model and Physical Conceptual data documents between source systems and the Confidential data warehouse.
  • Extensively used ERwinfor developing data model using star schema methodologies.
  • Worked on Data Modeling using Dimensional Data Modeling, Star Schema/Snow Flake schema, and Fact & Dimensional, Physical & Logical data modeling.
  • Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata Database analysis, development, testing, implementation and deployment.
  • Built database Model, Views and API's using Python for interactive web based solutions.
  • Developed requirements, perform data collection, cleansing, transformation, and loading to populate facts and dimensions for data warehouse
  • Created mapreduce running over HDFS for data mining and analysis using R and Loading & Storage data to Pig Script and R for MapReduce operations.
  • Designed the schema, configured and deployed AWS Redshift for optimal storage and fast retrieval of data.
  • Created, managed, and modified logical and physical data models using a variety of data modeling philosophies and techniques including Inmon or Kimball
  • Managed the Master Data Governance queue including assessment of downstream impacts to avoid failures
  • Worked in the capacity of ETL Developer (Oracle Data Integrator (ODI) / PL/SQL) to migrate data from different sources in to Confidential Oracle Data Warehouse.
  • Created and Configured Workflows, Work lets, and Sessions to transport the data to Confidential warehouse Netezza tables using Informatica Workflow Manager.
  • Involved in cleaning of large data sets using python andCreated named sets, calculated member and designed scope in SSAS, SSIS, and SSRS.
  • Worked on Teradata SQL queries, Teradata Indexes, MDM Utilities such as Mload, Tpump, Fast load and Fast Export.
  • Migrated SQL server 2008 to SQL Server 2014 in Microsoft Windows Server 2003 and troubleshooting high availability scenarios involving Clustering, Database Mirroring, Log Shipping and Replication.
  • Involved in Troubleshooting and quality control of data transformations and loading during migration from Oracle systems into Netezza EDW.
  • Created SSIS packages to load data from different sources such as Excel, Flat file, DB2 to SQL server Data warehouse and SQL Server, PL/SQL Transactional database.

Environment: ERwin, SSIS, SSRS, SAS, Netezza, Excel, MDM, PL/SQL, ETL, Python, Tableau, Hadoop, Hive, Pig, Mongo DB, Aginity, Teradata SQL Assistant, Cassandra, PL/SQL, T-SQL, Cognos, DB2, Oracle11g, SQL, Teradata14.1, Informatica Power Center9.1, AWS Redshift, HBase.

Data Modeler/Data Analyst

Confidential, NYC, NY

Responsibilities:

  • Created conceptual, logical and physical relational models for integration and base layer; created logical and physical dimensional models for presentation layer and dim layer for a dimensional data warehouse in Power Desinger.
  • Involved in reviewing business requirements and analyzing data sources form Excel/Oracle SQL Server for design, development, testing, and production rollover of reporting and analysis projects.
  • Analyzing, designing, developing, implementing and maintaining ETL jobs using IBM Info sphere Data stage and Netezza.
  • Extensively worked in Client-Server application development using Oracle 10g, Teradata 14, SQL, PL/SQL, Oracle Import and Export Utilities.
  • Coordinated with DB2 on database build and table normalizations and de-normalizations.
  • Conducted brain storming sessions with application developers and DBAs to discuss about various de-normalization, partitioning and indexing schemes for Physical Model.
  • Involved in several facets of MDM implementations including Data Profiling, metadata acquisition and data migration.
  • Extensively used SQL Loader to load data from the Legacy systems into Oracle databases using control files and used Oracle External Tables feature to read the data from flat files into Oracle staging tables.
  • Involved in extensive Data validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
  • Used SSIS to create ETL packages to validate, extract, transform and load data to data warehouse databases, data mart databases, and process SSAS cubes to store data to OLAP databases
  • Strong understanding of Data Modeling (Relational, dimensional, Star and Snowflake Schema), Data analysis, implementations of Data warehousing using Windows and UNIX.
  • Extensively worked with Netezza database to implement data cleanup, performance tuning techniques.
  • Created ETL packages using OLTP data sources (SQL Server 2008, Flat files, Excel source files, Oracle) and loaded the data into Confidential tables by performing different kinds of transformations using SSIS.
  • Migrated SQL server 2008 to SQL Server 2008 R2 in Microsoft Windows Server 2008 R2 Enterprise Edition.
  • Developing reusable objects like PL/SQL program units and libraries, database procedures and functions, database triggers to be used by the team and satisfying the business rules.
  • Performed data validation on the flat files that were generated in UNIX environment using UNIX commands as necessary.
  • Worked with NZ Load to load flat file data into Netezza, DB2 and Architect to identify proper distribution keys for Netezza tables.

Environment: Power Designer, Teradata, Oracle10g, PL/SQL, MDM, SQL Server 2008, ETL, Netezza, DB2, SSIS, SSRS, SAS, SPSS, Datastage, Informatica, SQL, T-SQL, UNIX, Netezza, Aginity, SQL assistance etc.

Confidential, Atlanta, GA

Data Analyst

Responsibilities:

  • Analyzed the physical data model to understand the relationship between existing tables. Cleansed the unwanted tables and columns as per the requirements as part of the duty being a Data Analyst.
  • Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
  • Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ER Studio.
  • Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark
  • Trained Spotfire tool and gave guidance in creating Spotfire Visualizations to couple of colleagues
  • Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
  • Developed Contracting Business Process Model Workflows (current / future state) using Bizagi Process Modeler software.
  • Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
  • Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
  • Extensively used ER Studio as the main tool for modeling along with Visio and worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
  • Configured & developed the triggers, workflows, validation rules & having hands on the deployment process from one sandbox to other.
  • Managed Logical and Physical Data Models in ER Studio Repository based on the different subject area requests for integrated model.
  • Created automatic field updates via workflows and triggers to satisfy internal compliance requirement of stamping certain data on a call during submission.
  • Worked on Metadata Repository (MRM) for maintaining the definitions and mapping rules up to mark.
  • Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
  • Developed enhancements to Mongo DB architecture to improve performance and scalability.
  • Forward Engineering the Data models, Reverse Engineering on the existing Data Models and Updates the Data models.
  • Performed data cleaning and data manipulation activities using NZSQL utility and analyzed and understood the architectural design of the project in a step by step process along with the data flow

Environment: Oracle SQL Developer, Oracle Data Modeler, Teradata14, SSIS, Business Objects, SQL Server 2008, ER/Studio Windows, MS Excel.

We'd love your feedback!