Sr. Data Analyst/ Data Modeler/ Data Architect Resume
Brentwood, TN
SUMMARY:
- Above 9+ years of IT experience as Data Architect/Modeler and Data Analyst in design, development, testing and maintenance of data warehouse, business intelligence and operational data systems.
- Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the big data as per the requirement.
- Specialization in Data Modeling, Data warehouse design, Building conceptual Architect, Data Integration and Business Intelligence Solution.
- Worked with different type systems as source and target, including RDBMS, Queue, Flat file, XML, Teradata, Netezza, Hadoop, MYSQL.
- Experienced in designing, building and implementing complete Hadoop ecosystem comprising of Map Reduce, HDFS, Hive, Impala, Pig, Sqoop, Oozie, HBase, MongoDB, and Spark.
- Have ETL Experience with Oracle Data Integrator (ODI) is a plus.
- Coded custom data mappings and data analyzers for multiple elastic indices.
- Designing the architecture of the data warehouse after understanding and analyzing the legacy databases.
- Experience in working with business intelligence and data warehouse software, including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum, Amazon Redshift and Azure Data Warehouse.
- 4 years experience of developing BI solutions for enterprise processes such as financials, supply - chain, HCM, CRM etc.
- Good understanding and hands on experience in setting up and maintaining NoSQL Databases like Cassandra, Mongo DB, and HBase.
- Familiar with Kimball DW/BI modeling principles and knowledgeable in Data warehouse modeling for different kind of business.
- Experience with JAD sessions for requirements gathering and writing functional specifications, queries.
- Good Experience in designing and developing advanced .NET custom UI components.
- Good understanding of Teradata SQL Assistant, Teradata Administrator and data load/ export utilities like BTEQ, Fast Load, Multi Load, Fast Export.
- Proficient in Data Analysis, Data Validation, Data Lineage Data Cleansing, Data Verification and identifying data mismatch
- Having good knowledge in Normalization (1NF, 2NF and 3NF) and De-Normalization techniques for optimum performance on XML data, Relational and Dimensional databases environment.
- Experienced in Data loading using PL/SQL, SQL Server Integration Services packages (SSIS)
- Expertise in Data Management Data Governance, Data Integration, Metadata, Reference Data and MDM
- Experience in developing Map Reduce Programs using Apache Hadoop for analyzing the Big data as per the requirement.
- Hands on experience with Agile and Waterfall data modeling methodologies.
- Good experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like ERWIN, ER Studio and Power-Designer.
- Experience in developing database schemas like Star Schema and Snowflake Schema for Dimensional Modeling and identifying Facts and Dimensions. .
- Extensive experience in development of T-SQL, Oracle PL/SQL Scripts, Stored Procedures and Triggers for business logic implementation.
- Expertise in Database Performance Tuning using Oracle Hints, Explain plan, TKPROF, Partitioning and Indexes
- Experience in working with business users/SMEs (Subject Matter Experts) as well as senior management.
TECHNICAL SKILLS:
Database Tools: Microsoft SQL Server 2014/12 Teradata 15/14, Oracle 12c/11g MS Access, Poster SQL, Netezza.
OLAP Tools: Tableau 9.3, SAP BO, SSAS, Business Objects, and Crystal Reports 9.
Analysis and Modeling Tools: Erwin 9.6/9.5, Sybase Power Designer, Oracle Designer, ER/Studio 9.7, Star-Schema, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables.
Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume.
Reporting Tools: Business Objects, Crystal Reports
Tools & Software: TOAD, MS Office, BTEQ, SQL Assistant
ETL Tools: SSIS, Pentaho, Informatica Power 9.6 etc.
Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, Have worked on C++, UNIX, PL/SQL etc.
Web technologies: HTML, DHTML, XML, JavaScript.
Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX Windows 95, 98, Windows NT, Windows XP, 7.
PROFESSIONAL EXPERIENCE:
Confidential, Brentwood, TN
Sr. Data Analyst/ Data Modeler/ Data Architect
Responsibilities:
- Responsible for the data architecture design delivery, data model development, review, approval and Data warehouse implementation.
- Designed/modified Forms, derivations, validation procedures in Oracle Clinical.
- Independently coded new programs and design Tables to load and test the program effectively for the given POC's using Big Data/Hadoop.
- Performed Hive programming for applications that were migrated to big data using Hadoop.
- Experienced in Data warehouse Migration /Modernization with special skills on Oracle Exadata and Informatica.
- Involved in converting Hive/SQLqueries into Spark transformations using SparkRDDs, Python.
- Experience in programming with the .NET Framework using technology ASP.NET, ADO.NET, VB.NET, Multithreading, C# with Web Services, IIS, JavaScript, HTML, DHTML, CSS, XML, XSL.
- Used talend reusable components context variable and global Map variables.
- Understanding of Extract/Transform/Load (ETL) processes to support Data Conversion / Migration / Consolidation.
- Led Audit projects and team members throughout each phase of the audit process including budget planning, work paper review, and strategic analysis.
- Worked on SQL querying skills using tools like SAS and Toad/ SQL server /Oracle
- Developed various QlikView Data Models by extracting and using the data from various sources files Excel, Flat Files and Big data.
- Developed and managed all related extraction, transformation and load of data as per the defined standard.
- Utilized Informatica toolset (Informatica Data Explorer, and Informatica Data Quality) to analyze legacy data for data profiling.
- Installed, upgraded, maintained and support Oracle Instances on Exadata.
- Predictive Modeling (R, SAS, SQL) - build models and educating business partners on the virtues of predictive analytics for their business units.
- Managed and retrieved data from a variety of sources and perform statistical analysis on it.
- Installation and Configuration of other Open Source Software like Pig, Hive, HBase, Flume and Sqoop.
- Worked on Hadoop (Cloudera), Spark, Kafka, Pig, Hive, Flume ElasticSearch, Logstash, Kibana.
- Accountable for data consolidation and cleansing validation, processes.
- Supported Audit reviews as required.
- Research, create and present risk management presentations for client facility, corporate and field staff on topics related to clinical documentation, clinical risk management, and risk.
- Responsible for daily backup of client's major data management system (Clinical Trial Database/Clinical DataManagement System).
- Designed and set up for TMS and Laboratory Module Setup in Oracle Clinical
- Worked on Teradata database and Developed scripts for loading the data into the base tables in EDW using Fastload, Multiload and BTEQ utilities of Teradata.
- Cleansed, extracted and analyzed business data on daily basis and prepared ad-hoc analytical reports using Excel and T-SQL
- Created Data Migration and Cleansing rules for the Integration Architecture (OLTP, ODS, DW).
- Develop ongoing test automation using Ansible, Python based framework using Ansible to Setup/teardown of ELK stack (ElasticSearch, Logstash, Kibana).
- Handled performance requirements for databases in OLTP and OLAP models.
- Conducted meetings with business and development teams for data validation and end-to-end data mapping.
- Wrote Audit reports with recommendations for improvements to management resulting in improved efficiency, decreased expenses, and reduction of risks.
- Created entity relationship diagrams and multidimensional data models, reports and diagrams based on the requirements.
- Responsible for Metadata Management, keeping up to date centralized metadata repositories using Erwin modeling tools.
- Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Oracle database.
- Developed Master data management strategies for storing reference data
- Involved in Data Architecture, Data profiling, Data analysis, data mapping and Data architecture artifacts design.
- Lead data migration from legacy systems into modern data integration frameworks from conception to completion.
- Created listings and reports as per requirement in Oracle Clinical
- Developed and Implemented logical and physical data models using enterprise modeling tools Erwin 9.6.
- Working on Amazon Redshift and AWS and architecting a solution to load data, create data models and run BI on it.
- Involved in Netezza Administration Activities like backup/restore, performance tuning, and Security configuration
- Coordinated the database design on the DB infrastructure objects and performance/storage related changes with DBA on Oracle 12c platform.
- Identified project requirement gaps using various data profiling techniques.
- Performing code reviews for ETL mappings from a performance & Error handling perspective.
Environment: Erwin9.6, Oracle Exadata 11c, QlikView, ODS, OLAP, OLTP, Hadoop, MapReduce, HDFS, Star Schema, Snowflake Schema, MDM, NoSQL, Business Objects, Agile, DB2, Hadoop, Cassandra, SAP.
Confidential, Dallas, TX
Sr. Data Architect/ Data Modeler/ Data Analyst
Responsibilities:
- Involved in software development with Big Data technologies such as Hadoop, SQOOP, Flume, Kafka, Hive, Pig, Oozie, Storm, Cassandra & Apache NiFi.
- Created dimensional model for the reporting system by identifying required dimensions and facts using ER Studio 9.
- Generated data models to make extracts for SAS sourcing data from Enterprise Data Warehouse.
- Created Talend Mappings to populate the data into Staging, Dimension and Fact tables.
- Experience with data gathering, database consolidation and/ or reporting design & implementation.
- Integrated, migrated and improved data across platforms to produce consistent, accurate and reliable information with SAS Data Management software.
- Helping manager in handling risk assessment and subsequently created contingency and mitigation plans.
- Worked in Data Warehouse and Business Intelligence Projects along with the team of Informatica, Talend (ETL).
- Setup/teardown of ELK stack (ElasticSearch, Logstash, Kibana) for OSS functional testing.
- Big Data Platforms (Cloudera,Hortonworks etc.) and advanced analytics technologies (predictive as well as prescriptive - SAS,R,Python, Machine Learning etc.)
- Experienced with building complex data extraction, transformation, and loading (ETL) pipelines into structured databases, data warehouses, or data processing systemsd.
- Interface with Oracle Corp’s Platinum Services team to plan, schedule and coordinate patching of all Exadata environments.
- Experience in building and migrating large scale data warehouse systems with heavy knowledge on Oracle Exadata.
- Led Audit projects and team members throughout each phase of the audit process including budget planning, work paper review, and strategic analysis.
- Created data models for AWS Redshift and Hive from dimensional data models.
- Worked in NoSQL database on simple queries and writing Stored Procedures for Normalization and De-normalization.
- Experienced in exploratory analysis and predictive modelling techniques such as Linear Regression and K-Means clustering.
- Data Mining/Heuristic Search (VB, SQL) - scoring algorithm built to find profitable/unprofitable segments of business.
- Experienced in developing and understanding of Object Oriented Principles (OOP), multi-tier web application design, .Net Patterns and Relational Database (RDBMS) principles.
- Produced and enforced data standards. Maintain a repository of data architecture artifacts and procedures.
- Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
- Full life cycle of Data Lake, Data Warehouse with Big data technologies like Spark, Hadoop.
- Experienced in processes and suggest new ones that most effectively anticipate, manage and reduce risk and its participants.
- Providing advice regarding analyzers, tokenizers and Elastic Search mappings.
- Used push down optimization in Informatica to call Greenplum GPLoad functions
- Involved in writing Shell Scripts to accumulate the MTD source file Collaboration with Architects and Managers for review of solutions and data strategy
- Collaborating with business users to define the key business requirements and translate them into process and technical solutions.
- Developing strategies for data acquisitions, archive recovery, and implementation of databases and working in a data warehouse environment, which includes data design, database architecture, and Metadata and repository creation.
- Designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using ER Studio.
- Migrated reference data from existing product into Informatica MDM hub
- Worked on Data modeling, Advanced SQL with Columnar Databases using AWS
- Created entity relationship diagrams and multidimensional data models, reports and diagrams based on the requirements.
- Ensured high-quality data and understand how data is generated out experimental design and how these experiments can produce actionable, trustworthy conclusions.
- Responsible for technical data governance, enterprise wide data modeling and database design.
- Design and develop Business Intelligence architecture using Microsoft and Tableau products.
- Performed data validation on the flat files that were generated in UNIX environment using UNIX commands as necessary.
- Involved in the creation, maintenance of Data Warehouse and repositories containing Metadata.
- Working as a lead member and contributing on architectural level inputs and suggestions for critical ETL component design.
Environment: ER Studio 9, NoSQL, AWS, Tableau, HBase, Oracle 11g,, Hadoop, ODS, Oracle 12c, ETL,, MDM, PL/SQL, OLAP, OLTP.
Confidential, Irving, TX
Sr. Data Architect/ Data Modeler
Responsibilities:
- Collaborating with business users to define the key business requirements and translate them into process/technical solutions.
- Involved in data model reviews with internal data architect, business analysts, and business users with explanation of the data model to make sure it is in-line with business requirements.
- Designing and implementing data processing pipelines with a combination of the following technologies: Hadoop, Map Reduce, Spark, Hive, Kafka, Avro, SQL and NoSQL data warehouses.
- Performed logical data modeling, physical Data Modeling (including reverse engineering) using the ERWIN Data Modeling tool.
- Performed Data Analysis and data profiling using complex SQL on various sources systems including Oracle 10g/9i and Teradata.
- Created several Master Data Models (MDM) that unify data for critical concepts across applications.
- Performed data profiling and developed strategies to improve quality of data and developed backup and security strategies
- Created standard abbreviation document for logical, physical and CQL data models.
- Created data masking mappings to mask the sensitive data between production and test environment.
- Designed and Developed logical & physical data models and Meta Data to support the requirements using ERWIN.
- Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
- Extracted data from Oracle and upload to Teradata tables using Teradata utilities FASTLOAD & Multiload.
- Worked with supporting business analysis and marketing campaign analytics with data mining, data processing, and investigation to answer complex business questions.
- Involved in the creation, maintenance of Data Warehouse and repositories containing Metadata.
- Expert in developing SQL and stored procedures
- Developed test methodology and test strategy and reference data architecture for Analytics
- Responsible for defining the testing procedures, test plans, error handling strategy and performance tuning for mappings, Jobs and interfaces.
- Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Designed data cleansing/data scrubbing techniques to ensure consistency amongst data sets.
Environment: Erwin r9., Oracle 10g, DB2, Hadoop,, ODS, SQL Assistant, MS Visio, Spark,, UNIX, Flat files, ETL, PL/SQL, Metadata, OLAP, OLTP
Confidential, Cary, NC
Sr. Data Analyst/ Data Modeler
Responsibilities:
- Participated in JAD sessions involving the discussion of various reporting needs.
- Reverse Engineering the existing data marts and identified the Data Elements (in the source systems), Dimensions, Facts and Measures required for reports.
- Conduct Design discussions and meetings to come out with the appropriate Data Warehouse at the lowest level of grain for each of the Dimensions involved.
- Transformed project data requirements into project data models using ERStudio.
- Provided analysis and worked with developers in database design and table structure, indexes, views, stored procedures and planned the SSIS data flow.
- Excellent report creation skills using Microsoft Reporting Services (SSRS) with proficiency in using Report Designer as well as Report Builder.
- Created fully normalized till 3rd Normalization technique using ER Studio.
- Wrote PL/SQL Stored Procedures, Functions, Views, Packages and Triggers to implement business rules into the application.
- Provisions were made for ad-hoc querying and reporting using Crystal Reports.
- Validated and updated the appropriate LDM's to process mappings, screen designs, use cases, business object model, and system object model as they evolve and change.
- Conduct Design reviews with the business analysts and content developers to create a proof of concept for the reports.
- Created Use Case Diagrams, Activity Diagrams, Sequence Diagrams, Dataflow Diagrams and Conceptual Data Models in MS Visio.
- Worked extensively with VISIO and Erwin to do reverse engineering of Database for reporting
- Worked with the Implementation team to ensure a smooth transition from the design to the implementation phase.
- Interact with the business partners, functional analysts, SMEs and scrum masters to gather business and technical requirements and understanding the flow of the process.
- Performed data profiling in the source systems that are required for Data Marts.
- Worked with internal architects and assisted in the development of current and target state enterprise data architectures.
- Created logical, physical dimensional models using ERWIN.
- Created physical data models using forward engineering principles.
- Implemented Star Schema methodologies in designing and modeling the logical data model into Dimensional Models.
- Performed Gap Analysis, Impact Analysis and documented it in an Excel Sheet.
- Implemented Normalization techniques and build the tables as per the requirements given by the business users.
- Provided source to target mappings to the ETL team to perform initial, full, and Incremental loads into the target data mart.
- Prepared Test cases so as to check the transformation logics against the source tables.
- Identified and streamlined complex queries which were causing iterations and effecting database and system performance.
Environment: Erwin 8.0, Oracle 10g, PL/SQL, Metadata, OLAP, OLTP, Informatica 8.6, Teradata V2R12, Teradata SQL Assistant, OOD, MS Visio.
Confidential
Sr. Data Analyst/ Data Modeler
Responsibilities:
- Designed ER diagrams, logical model and physical database for Oracle and Teradata as per business requirements using Erwin.
- Worked on data profiling and data validation to ensure the accuracy of the data between the warehouse and source systems.
- Performed Data Analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata, in order to ensure accuracy of the data between the warehouse and source systems.
- Gathered business requirements through interviews, survey with users and business analysts.
- Prepared functional requirements to define the system rules by creating Use Cases, Interface design and data dictionaries.
- Analyzed various resource files being used to generate the report and documented data mapping and data source interface documents to assist the development team.
- Used SQL for querying and analysis purposes on various source tables and conditions applied and Wrote SQL joins, sub queries.
- Used SQL Server Reporting Services to schedule reports to be generated on predetermined time. Built up and maintain a strong data warehousing system for the organization
- Involved in business process modeling using UML through Rational Rose.
- Implement Cognos FM, cube and report development to support business user to identify opportunities for improvement in operations and processes
- Reverse Engineered DB2 databases and then forward engineered them to SQL Server 2000 using Erwin.
- Developed logical data model using Erwin and created physical data models using forward engineering in generating DDL scripts and creating indexing strategies
- De-normalized the database to put them into the star schema of the data warehouse
- Understood existing data model and documented suspected design affecting the performance of the system.
- Involved in portioning of cubes in order to increase the performance of cubes (SSAS).
- Identify the potential customers from the given data base and counsel the organization on ways of converting them to loyal customers as a part of CRM.
- Worked with Flat Files (Pipe Delimited) sources and implemented error handling routines.
Environment: DB2, SQL Server 2000, Hyperion Essbase, Informatica Power Center, Erwin, Microsoft Visio, Rational Requisite Pro, Rational Rose, Windows 2003 Server.
