Sr. Data Analyst/ Modeler Resume
Juno Beach, FL
SUMMARY
- Over 8years of experience in Information Technology in Data Analysis, Data Modeling, Design and development of Databases for business applications in Data Warehousing Environments
- Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server 2008 and Teradata.
- Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
- Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
- Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
- Expertise lies in Data Modeling, Database design and implementation of Oracle, AWS Redshift databases and Administration, Performance tuning etc.
- Extensive years of experience in Talend Open Studio and Talend Integration Suite.
- Experience in analyzing data using Hadoop Ecosystem including HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop, Flume.
- Experienced working with Excel Pivot and VBA macros for various business scenarios.
- Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export.
- Data Transformation using Pig scripts in AWS EMR, AWS RDS.
- Experience working with data modeling tools like Erwin, Power Designer and ER Studio.
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
- Experience in data analysis using Hive, PigLatin, Impala.
- Well versed in Normalization / De - normalization techniques for optimum performance in relational and dimensional database environments.
- Good understanding of AWS, big data concepts and Hadoop ecosystem.
- Experienced in various Teradata utilities like Fastload, Multiload, BTEQ, and Teradata SQL Assistant.
- Expert in writing SQL queries and optimizing the queries in Oracle, SQL Server 2008 and Teradata.
- Develop and manage SQL, Python and R code bases for data cleansing and data analysis using Git version control
- Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
- Extensive ETL testing experience using Informatica 8.6.1/8.1 (Power Center/ Power Mart) (Designer, Workflow Manager, Workflow Monitor and Server Manager)
- Excellent in creating various artifacts for projects which include specification documents, data mapping and data analysis documents.
- An excellent team player& technically strong person who has capability to work with business users, project managers, team leads, architects and peers, thus maintaining healthy environment in the project.
TECHNICAL SKILLS
Analysis and Modeling Tools: IBM Infosphere, Oracle Designer, Erwin 9.6/9.5, ER/Studio 9.7, Sybase Power Designer.
Database Tools: Oracle 12c/11g, MS Access, Microsoft SQL Server 2014/2012 Teradata 15/14
Big Data Technologies: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume.
Cloud Platform: AWS, EC2, S3, SQS, Azure.
OLAP Tools: Business Objects, Tableau, SAP BO, SSAS, Crystal Reports 9.
Operating System: Windows, Dos, Unix, Linux.
Reporting Tools: Business Objects, Crystal Reports.
Tools: & Software’s: TOAD, MS Office, BTEQ, Teradata SQL Assistant.
ETL Tools: SSIS, Pentaho, Informatica Power 9.6, SAP Business Objects XIR3.1/XIR2, Web Intelligence.
Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, Have worked on C++, UNIX, PL/SQL etc.
PROFESSIONAL EXPERIENCE
Confidential, Juno Beach, FL
Sr. Data Analyst/ Modeler
Responsibilities:
- Worked as a Sr. Data Analyst/Data Modeler I was responsible for all data related aspects of a project.
- Provided solutions on ingesting the data into the new Hadoop big data platform by designing data models for multiple features to help analyze the data on graph databases.
- Applied business rules in modeling Data Marts and data profiling to model and new data structures.
- Delivered scope, requirements, and design for transactional and data warehouse system which included Oracle DB, SQL server, and Salesforce database.
- Generated computed tables in Power BI by using Dax and Mquery
- Performed importing data from various sources to the Cassandra cluster using Python APIs.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Developed and implemented different Pig UDFs to write ad-hoc and scheduled reports as required by the Business team.
- Involved in Normalization / De normalization techniques for optimum performance in relational and dimensional database environments.
- Involved in loading data from LINUX file system to HDFS Importing and exporting data into HDFS and Hive using Sqoop Implemented Partitioning, Dynamic Partitions, Buckets in Hive.
- Used SSRS to create reports, customized Reports, on-demand reports, ad-hoc reports and involved in analyzing multi-dimensional reports in SSRS.
- Creating dimensional data models based on hierarchical source data and implemented on Teradata achieving high performance without special tuning.
- Focused on architecting NoSQL databases like Mongo, Cassandra and Cache database.
- Perform routine management operations, including configuration and performance analysis for mongodb. Diagnosing Performance Issues for mongodb.
- Involved in designing Logical and Physical data models for different database applications using the Erwin.
- Data modeling, Design, implement, and deploy high-performance, custom applications at scale on Hadoop /Spark.
- Design and implementation of an Azure data lake
- Extensive knowledge in Data loading using PL/ SQL Scripts and SQL Server Integration Services (SSIS).
- Work in team using ETL tool Informatica to populate the database, data transformation from the old database to the new database using Oracle and SQL Server.
- Ensured that data architecture tasks were executed within deadlines.
Environment: DB2, CA Erwin 9.6, Oracle 12c, MS-Office, TOAD Benchmark Factory, SQL Loader, PL/SQL, SharePoint, Erwin r9.64, Talend, MS-Office, SQL Server 2008/2012, Hive, Pig, Hadoop, Spark, Azure.
Confidential - Chicago, IL
Data Analyst/Modeler
Responsibilities:
- Developed and maintained the data definitions, data models, data flow diagrams, metadata management, business semantics, and metadata workflow management.
- Integrated 40 data sources in one data repository utilizing modeling tools (ER Studio) and ETL tool (PL/SQL).
- Involve in data cleaning procedure by removing old, corrupted or irrelevant data in consultation with the teams.
- Worked with Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data and conventional RDBMS.
- Involved in Relational and Dimensional Datamodeling for creating Logical and Physical Design of Database and ER Diagrams with all related entities and relationship with each entity based on the rules provided by the business manager using ER Studio.
- Worked on Normalization and De-normalization concepts and design methodologies like Ralph Kimball and Bill Inmon's Data Warehouse methodology.
- Developed Up-selling and Cross-selling reports using DAX functions and Mquery.
- Use database design and database modeling concepts to ensure data accessibility and security
- Designed both 3NF data models for OLTP systems and dimensional data models using Star and Snow Flake Schemas.
- Responsible for delivering and coordinating data-profiling, data-analysis, data-governance, data-models (conceptual, logical, physical), data-mapping, data-lineage and reference data management.
- Configured authorization to Cassandra cluster using Password Authenticator and Kerberos for Hadoop cluster
- Worked on SQL Server concepts SSIS (SQL Server Integration Services), SSAS (Analysis Services) and SSRS (Reporting Services).
- Generated and DDL (Data Definition Language) scripts using ER Studio and assisted DBA in Physical Implementation of Data Models.
- Extensively worked on creating the migration plan to Amazon web services (AWS).
- Extracted Mega Data from AWS, and Elastic Search engine using SQL Queries to create reports.
- Completed enhancement for MDM (Master data management) and suggested the implementation for hybrid MDM (Master Data Management).
- Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
- Exported data from HDFS environment into RDBMS using Sqoop for report generation and visualization purpose.
- Generated comprehensive analytical reports by running SQL queries against current databases to conduct Data Analysis.
- Worked with NoSQL databases like HBase in creating HBase tables to load large sets of semi-structured data coming from various sources.
- Development of Data stage design concepts, execution, testing and deployment on the client server
- Validated the data of reports by writing SQL queries in PL/SQL Developer against ODS.
- Involved in user training sessions and assisting in UAT (User Acceptance Testing).
Environment: ER Studio, AWS, OLTP, Teradata r15, Sqoop 1.4, Cassandra 3.11, HDFS, Linux, Shell, scripts, NoSQL, SSIS, SSAS, HBase 1.2, MDM.
Confidential - Portland, ME
Sr. Data Analyst /Modeler
Responsibilities:
- Worked on OLAP for data warehouse and data mart developments using Ralph Kimball methodology as well as OLTP models (3NF) and interacting with all the involved stakeholders and SME's to derive the solution..
- Designed the ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin
- Designed 3rd normal form target data model and mapped to logical model.
- Involved in extensive DATA validation using SQL queries and back-end testing
- Generated DDL statements for the creation of new ER/studio objects like table, views, indexes, packages and stored procedures.
- Design MOLAP/ROLAP cubes on Teradata Database using SSAS.
- Used SQL for Querying the database in UNIX environment
- Creation of BTEQ, Fast export, Multi Load, TPump, Fast load scripts for extracting data from various production systems.
- Working along with ETL team for documentation of transformation rules for data migration from OLTP to warehouse for purpose of reporting.
- Created views and extracted data from Teradata base tables and uploaded data to oracle staging server from Teradata tables, using fast export concept.
- Migrated SAS job to Talend workflow, to create report on claims and cornerstone.
- Worked RDS for implementing models and data on RDS.
- Developed mapping spreadsheets for (ETL) team with source to target data mapping with physical naming standards, data types, volumetric, domain definitions, and corporate meta-data definitions.
- Designing Star schema and Snow Flake Schema on Dimensions and Fact Tables
- Worked with Data Vault Methodology Developed normalized Logical and Physical database models.
- Transformed Logical Data Model to Physical Data Model ensuring the Primary Key and Foreign key relationships in PDM, Consistency of definitions of Data Attributes and Primary Index considerations.
- Wrote and running SQL, BI and other reports, analyzing data, creating metrics/dashboards/pivots/etc.
- Gather and analyze business data requirements and model these needs. In doing so, work closely with the users of the information, the application developers and architects, to ensure the information models are capable of meeting their needs.
Environment: SQL Server, Erwin9.1, Oracle, Informatica, RDS, Big Data, JDBC, NOSQL, Star schema, Snow Flake Schema, Python, MySQL, PostgreSQL .
Confidential - Long Beach, CA
Sr. Data Modeler /Analyst
Responsibilities:
- Involved in the projects from requirement analysis to better understand the requirements and support the development team with a better understanding of the data.
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture involving OLTP, ODS and OLAP.
- Involved in Data Architecture, Data profiling, Data analysis, data mapping and Data architecture artifacts design.
- Responsible for Relational data modeling (OLTP) using MS Visio (Logical, Physical and Conceptual).
- Analyzed the data and provide resolution by writing analytical/complex SQL in case of data discrepancies.
- Involved in logical and Physical Database design & development, Normalization and Data modeling using Erwin and SQL Server Enterprise manager.
- Prepared ETL technical Mapping Documents along with test cases for each Mapping for future developments to maintain Software Development Life Cycle (SDLC).
- Designed OLTP system environment and maintained documentation of Metadata.
- Worked on Amazon Redshift and AWS a solution to load data, create data models.
- Created dimensional model for the reporting system by identifying required dimensions and facts using
- Used Reverse Engineering to connect to existing database and create graphical representation (E-R diagram)
- Using Erwin modeling tool, publishing of a data dictionary, review of the model and dictionary with subject matter experts and generation of data definition language.
- Involved in design, development of Talend mappings.
- Coordinated with DBA in implementing the Database changes and also updating Data Models with changes implemented in development, QA and Production.
- Created and execute test scripts, cases, and scenarios that will determine optimal system performance according to specifications.
- Worked Extensively with DBA and Reporting team for improving the Report Performance with the Use of appropriate indexes and Partitioning
- Implemented error handling in Talend to validate the data integrity and data completeness for the data from flat file.
- Worked with medical claim data in the Oracle database for Inpatient/Outpatient data validation, trend and comparative analysis.
- Used Load utilities (Fast Load & Multi Load) with the mainframe interface to load the data into Teradata.
- Optimized and updated UML Models (Visio) and Relational Data Models for various applications.
Environment: Erwin9.0, Oracle11g, SQL Server 2010, Teradata14, XML, OLTP, PL/SQL, Linux, UNIX, Mload, BTEQ, UNIX shell scripting
Confidential
Data Analyst / Modeler
Responsibilities:
- Performed Data Analysis, Data Migration and data profiling using complex SQL on various sources systems including Oracle and Teradata.
- Logical and physical database models to design OLTP system for applications using Erwin.
- Forward engineering to create a physical data model with DDL that best suits the requirements from the logical data model using Erwin for effective model management of sharing, dividing and reusing model information.
- Worked with BTEQ to submit SQL statements, import and export data, and generate reports in Teradata.
- Translated business requirements into working logical and physical data models for Data Warehouse, Data marts and OLAP applications.
- Identified the entities and relationship between the entities to develop Conceptual Model using ERWIN.
- Involved in the creation, maintenance of Data Warehouse and repositories containing Metadata.
- Wrote and executed unit, system, integration and UAT scripts in a Data Warehouse projects.
- Extensively used SQL, Transact SQL and PL/SQL to write stored procedures, functions, packages and triggers.
- Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data Warehouse, and data mart reporting system in accordance with requirements.
- Excellent experience and knowledge on Data Warehouse concepts and dimensional data modelling using Ralph Kimball methodology.
- Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ERWIN
- Created and maintained Logical Data Model (LDM) for the project.
- Includes documentation of all entities, attributes, data relationships, primary and foreign key structures, allowed values, codes, business rules, glossary terms, etc.
Environment: Oracle, MS Visio, PL-SQL, OLTP, OLAP, Informatica 9.x, SQL, PL/SQL