We provide IT Staff Augmentation Services!

Lead Data Modeler/ Data Analyst Resume

5.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY:

  • Over 10+ years of Industry experienced in IT with solid understanding of Data Modeling, Data Analysis, data architecture, Evaluating Data Sources and strong understanding of Data Warehouse/Data Mart Design, ETL, BI, OLAP, OLTP, Client/Server applications.
  • Expert in data management vision, goals, priorities, design principles, and operating policies in support of the business goals of the organization.
  • Experienced in writing SQL queries and optimizing the queries in Oracle, SQL Server, and Netezza, Teradata.
  • Experienced in Dimensional Data Modeling using Data Modeling, Relational Data modeling, ER/Studio, ERwin, and Sybase Power Designer, Star Join Schema/Snowflake modeling, FACT & Dimensions tables, Conceptual, Physical & logical Data Modeling.
  • Experienced in Management and implementation of database models, data flow diagrams, database schemas, db scripts, DTD schemas, structures and data standards to support a robust data management infrastructure.
  • Experienced in Data Analysis and Data Profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Extensive experience in Normalization (1NF, 2NF, 3NF and BCNF) and De - normalization techniques for improved database performance in OLTP, OLAP and Data Warehouse/Data Mart environments.
  • Experienced on Metadata definition, implementation and maintenance, new business rules identification and implementation to data rules, transformation program library maintenance, XML file generation and data quality.
  • Experienced using MapReduce and Big data work on Hadoop and other NO SQL platforms
  • Experienced in designing standards for using normalized data structures, de-normalized structures, and dimensional structures. Defines common design patterns for modeling various types of relationships.
  • Experienced in Batch processes, Import, Export, Backup, Database Monitoring tools and Application support.
  • Experienced in big data analysis and developing data models using Hive, PIG, and Map reduce, SQL with strong data architecting skills designing data-centric solutions.
  • Experienced in Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and Fast Export.
  • Experienced in using the databases like DB2, Teradata and its utilities, Netezza, Oracle, SQL Server Integration Services (SSIS).
  • Experienced in data from various data sources/business systems including MS Excel, MS Access, Flat Files etc to SQL Server using SSIS using various features like data conversion etc.
  • Experienced in Oracle, Netezza, and Teradata, SQL Server, and DB2 database architecture.
  • Expertise in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Extensive experience in development of T-SQL, DTS, OLAP, PL/SQL, Stored Procedures, Triggers, Functions, Packages, performance tuning and optimization for business logic implementation.
  • Good knowledge of Data Marts, Operational Data Store (ODS), OLAP, Dimensional Data Modeling with Ralph Kimball Methodology (Star Schema Modeling, Snow-Flake Modeling for FACT and Dimensions Tables) using Analysis Services.
  • Excellent in performing data transfer activities between SAS and various databases and data file formats like XLS,CSV,DBF,MDB etc.
  • Experienced in ER Studio and Dimensional Models using ERwin advanced features, Conceptual, logical and physical data models using ERwin.
  • Experienced in development and support knowledge on Oracle, SQL, PL/SQL, T-SQL queries.
  • Experienced in Databases queries and analysis use IBM Netezza with data warehousing appliances.
  • Experienced in Data Scrubbing/Cleansing, Data Quality, Data Mapping, Data Profiling, Data Validation in ETL
  • Excellent Knowledge of RalphKimball and BillInmon's approaches to Data Warehousing.
  • Excellent knowledge in developing Informatica Mappings, Mapplets, Sessionss, Workflows and Worklets for data loads from various sources such as Oracle, Flat Files, DB2, SQL Server etc.
  • Experienced in writing UNIX shell scripting and hands on experienced with scheduling of shell scripts using Control-M.
  • Extensive experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like ERWIN, ER Studio.

TECHNICAL SKILLS:

Analysis and Modeling Tools:: Erwin r9.6/r9.5/9.1/r8.x, Sybase Power Designer, Oracle Designer, BP win ER/Studio, MS Access 2000, Star-Schema, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables.

OLAP Tools:: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9.

Oracle:: Oracle12c/11g/10g/9i/8.x, R2 database servers with RAC, ASM, Data Guard, Grid Control and Oracle Golden Gate(Oracle Enterprise Manager),Oracle Data Guard, SQL, Net, SQL Loader and SQL PLUS, AWR,ASH, ADDM, Explain Plan.

ETL Tools:: SSIS, Pentaho, Informatica Power Center 9.7/9.6/9.5/9.1 etc.

Programming Languages:: Java, Base SAS, SSIS and SAS/SQL, SQL, T-SQL, HTML/ XHTML/ HTML4.0.1/ HTML3.2, Java Script, CSS3/CSS2/CSS1, UNIX shells scripting, PL/SQL.

Database Tools:: Microsoft SQL Server 2014/2012/2008/2005, Teradata, and MS Access, Poster SQL, Netezza, SQL Server, Oracle.

Reporting Tools:: Business Objects, Crystal Reports and SSRS.

Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX Windows 95, 98, Windows NT, Windows XP, 7.

Tools & Software: TOAD, MS Office, BTEQ, Teradata 15/14.1/14/13.1/13, SQL Assistant

Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume.

Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, Have worked on C++, UNIX, PL/SQL etc.

PROFESSIONAL EXPERIENCE:

Confidential, Chicago IL

Lead Data Modeler/ Data Analyst

Responsibility:

  • Define data standards / reference architectures for use by project teams in developing technical designs and architecting & implementing enterprise & business segment OLTP solutions.
  • Gathers and understands client needs, finding key areas where technology leverage is possible to improve business processes, defines architectural approaches and develops technology proofs. Communicates technology direction.
  • Provide creative and efficient solutions to challenging data management and product development problems.
  • Involved in Conceptual , Logical and physical database design , modeling & implementation of enterprise.
  • Handled data collection, analysis, interpretation and presentation to management and other team members gathered via a wide range of available means and methods from users and business partners associated with Supply Chain, claims, billing activities.
  • Created new database tables, views, indexes, functions, procedures, and synonyms in support of many user groups within Confidential. Analyzed existing queries and procedures for processing improvements. Managed daily reporting needs as a part of the data warehouse team. Support included review and modification to source data ETL processes.
  • Worked on Data Modeling using Dimensional Data Modeling, Star Schema/Snow Flake schema, and Fact & Dimensional, Physical & Logical data modeling.
  • Create logical and physical data models using Erwin to meet the needs of the organization's information systems and business requirements.
  • Involved in Teradata utilities (BTEQ, Fast Load, Fast Export, Multiload, and Tpump) in both Windows and Mainframe platforms.
  • Performed extensive task which include Data Quality Analysis, Data Profiling, Data Mapping, Data lineage and Data Standardization data structures, data base design, data warehouses, business intelligence/analytic tools, SQL, ETL tools, and data integration methods.
  • Designed the schema, configured and deployed AWS Redshift for optimal storage and fast retrieval of data and used Spark Data frames, Spark-SQL, Spark MLLib extensively and developing and designing POC's using Scala, Spark SQL and MLlib libraries.
  • Performed PL/SQL code development and implemented Oracle based data warehouses for use in multiple dynamic client environments. Provided Oracle, SQL and UNIX based coding standards and leadership to other consultants and contractors.
  • Predominantly using Python and AWS (Amazon web services), and MySQL along with NoSQL (mongodb) databases for meeting end requirements and building scalable real time system.
  • Expert level UNIX shell script development for automating ETL processes, data loads, instance data migration & database or file system backup and restores.
  • Involved in designing and scheduling complex SSIS Packages for transferring data manually from multiple data sources to SQL server.
  • Involved in integration of various relational and non-relational sources such as DB2, Teradata15, Oracle, SFDC, Netezza, SQL Server, COBOL, XML and Flat Files.
  • Developed LINUX Shell scripts by using NZSQL/NZLOAD utilities to load data from flat files to Netezza database.
  • Responsible for managing and reviewing Hadoop log files. Designed and developed data management system using MySQL.
  • Created SSIS packages to load data from different sources such as Excel, Flat file, DB2 to SQL server 2014 Data warehouse and SQL Server, PL/SQL Transactional database.
  • Managed and created and altered Databases, Tables, Views, Indexes and Constraints with business rules using T-SQL.
  • Performed cluster analysis and page-rank on newswire articles using Hadoop framework in python to interpret the importance of keywords connecting documents.
  • Analyze the required Information e.g. Patients Information, Primary Carrier Info, Producer details, Revenue Code, Billed Amount for each Revenue Code, Diagnosis Codes.
  • Using SQL to extract data from different databases such as Oracle and Amazon Redshift.
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDF
  • Worked on SQL Server concepts SSIS (SQL Server Integration Services), SSAS (Analysis Services) and SSRS (Reporting Services).
  • Used the Agile Scrum methodology to build the different phases of Software development life cycle.
  • Designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snow flake Schemas.
  • Extensively used Aginity Netezza work bench to perform various DML, DDL operations on Netezza database.

Environment: Erwin r9.6, Netezza, Oracle12c, Taradata15, T-SQL, SQL Server 2014, 3NF, DB2, SSIS, SSRS, LINUX, UNIX, MDM, PL/SQL, Python, AWS Redshift, AWS S3, AWS EMR, ETL Informatica, Aginity, Teradata SQL Assistant, Metadata, Hadoop, Hive, Pig, Mapreduce, SQL, Tableau, BO, Netezza, SAS.

Confidential, Dallas TX

Lead Data Modeler/Data Analyst

Responsibility:

  • Lead in architecting, designing, modeling, developing, maintaining and deploying complex data models, and data integrations utilizing SQL server (SSIS) based ETL processes for new and existing systems.
  • Worked with strategic and tactical requirements of the organization to determine the best form of data persistence to fulfill business requirement.
  • Provide and curate a data framework, governance and roadmap for data management and build and maintain an Enterprise Data Dictionary to enforce data standards and polices
  • Determined database structural requirements by analyzing client operations, applications, and programming, reviewing objectives with clients, evaluating current systems
  • Designed ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical data models (capacity planning, object creation and aggregation strategies) as per business requirements.
  • Involved in documentation of Data Modeler/Data Analyst and ETL specifications for Data warehouse Erwin r9.5.
  • Worked with cloud based technology like Redshift, S3, AWS, EC2 Machine, etc.
  • Coordinated new data development ensuring consistency and integration with existing data warehouse structure.
  • Created and maintained Logical Data Model (LDM) for the project includes documentation of all entities, attributes, data relationships, primary and foreign key structures, allowed values, codes, business rules, glossary terms, etc.
  • Analysis of functional and non-functional categorized data elements for data profiling and mapping from source to target data environment.
  • Involved in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi-dimensional models such as Star Schemas and Snow Flake Schema for developing Cubes using MDM.
  • Used forward engineering to create a Physical Data Model with DDL that best suits the requirements from the Logical Data Model
  • Executed Hive queries on Parquet tables stored in Hive to perform data analysis to meet the business requirements and worked on importing and exporting data from Oracle and DB2 into HDFS and HIVE using Sqoop.
  • Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata14.1, Oracle, and Database, analysis, development, testing, implementation and deployment.
  • Created Mappings for Initial load from MS SQL server 2012 to Netezza while performing data cleansing.
  • Worked with cloud based technology like Redshift, S3, AWS, EC2 Machine, etc. and extracting the data from the Oracle financials and the Redshift database.
  • Loaded multi format data from various sources like flat-file, Excel, MS Access, and Oracle using Netezza and performing file system operation.
  • Developed ETL processes extracted data daily from Mainframe DB2 and ORACLE Databases, and loaded data into a SSIS based Decision Support Warehouse.
  • Performed Extracting, Transforming and Loading (ETL) data from Excel, Flat file, Oracle to MS SQL Server by using SSIS services
  • Involved in extensive DATA validation by writing several complex SQLqueries and Involved in back-end testing and worked with data quality issues.
  • Worked on Normalization and De-Normalization techniques for both OLTP and OLAP systems.
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
  • Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
  • Created PL/SQL Tables, Collections, and Records, Partitions and Used Dynamic SQL and triggers for faster data access and to in corporate business logic.
  • Responsible for creating Hive tables, loading data and writing hive queries.
  • Worked with various performance tuning techniques using tools provided by Oracle like EXPLAIN PLAN, SQL Trace, TKPROF, Oracle hints, DBMS PROFILER, DBMS STATS.
  • Responsible for determining the bottlenecks and fixing the bottlenecks with performance tuning using Netezza Database.

Environment: Erwin r9.5, Netezza, Oracle11g, AWS Redshift, AWS S3, Taradata14.1, T-SQL, SQL Server 2012, SSRS, SSIS, LINUX, MDM, PL/SQL, SPSS, ETL, OLTP and OLAP, Aginity, SQL Assistance, Metadata, Hadoop, DB2, UNIX, Excel, SAS, Hive, Pig, MapReduce, Mongodb, HBase etc.

Confidential, Chandler AZ

Sr. Data Modeler/ Data Analyst

Responsibility:

  • Analyzed across multiple data domains and define strategies to reduce data redundancy , improve data availability, and accessibility by partnering with the technical teams and deliver a maintainable and reusable data architecture
  • Maintained the end-to-end vision of the data flow diagram and develop logical data models into one or more physical data repositories.
  • Documented logical data integration (ETL) strategies for data flows between disparate source/target systems for structured and unstructured data into common data lake and the enterprise information repositories
  • Designed and implemented data solutions using Master Data Management principles and tools.
  • Facilitated and maintained enterprise information architecture for unified access to structured and unstructured data across multiple data sources including data warehouse and sensor/machine data
  • Implemented dimension model (logical and physical data modeling) in the existing architecture using ER Studio.
  • Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata14 Database.
  • Involved in Data modeling using ER Studio identified objects and relationships and how those all fit together as logical entities, these are then translated into physical design using forward engineering ER Studio tool.
  • Created ERD diagrams using ER Studio and implemented concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
  • Provided technical guidance for re-engineering functions of Teradata warehouse operations into Netezza.
  • Designed and Developed Oracle PL/SQL Procedures LINUX and UNIX Shell Scripts for Data Import/Export and Data Conversions.
  • Involved in SQL Server and T-SQL in constructing Tables, Normalization and De-normalization techniques on database Tables.
  • Worked on Cube structure optimization for MDM query performance in Analysis Services (SSAS).
  • Worked in generating and documenting Metadata while designing OLTP and OLAP systems environment.
  • Performed SQL and PL/SQL Performance tuning and Application Tuning using various tools like TKPROF, AUTOTRACE and DBMS SQLTUNE.
  • Handled data loading operations from flat files to tables using NZLOAD utility and extensively used NZSQL and NZLOAD utilities.
  • Developed complex TeradataSQL code in BTEQ script using OLAP and Aggregate functions to name few and reviewed and implemented the naming standards for the entities, attributes, alternate keys, and primary keys for the logical model.
  • Created SSIS Packages to schedule the stored procedures and monitoring the scheduled jobs and Involved in monitoring the weekly job.
  • Manipulating, cleansing & processing data using Excel, and SQL. Responsible for loading, extracting and validation of client data.
  • Performed Forward Engineering data models for Schema generation, Reverse Engineered on the Existing Data Models to accommodate new requirements.
  • Designed and developed star schema model for target database using ER Studio Data modeling.

Environment: Erwin r9.5, Netezza, Oracle11g, Taradata14.1, T-SQL, SQL Server 2012, SSRS, SSIS, LINUX, MDM, PL/SQL, SPSS, ETL, OLTP and OLAP, Aginity, SQL Assistance, Metadata, DB2, UNIX, Excel, SAS, Hive, Pig, Mapreduce, Mongodb, HBase etc.

Confidential - Baltimore MD

Sr. Data Modeler/ Data Analyst

Responsibility:

  • Extensively used Erwin r9.1 to design Logical/Physical Data Models, relational database design, forward/reverse engineering, publishing Data Model to acrobat files, created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data Model in model mart, created naming convention files, co-coordinated with DBAs to apply the Data Model changes.
  • Involved in requirement analysis, ETL design and development for extracting data from the source systems like Taradata13.1, DB2, Sybase, Oracle 9i, flat files and loading into Netezza.
  • Involved in Designing Star Schema (identification of facts, measures and dimensions), Snowflake Schema for Data Warehouse, ODS architecture by using tools like Data Model, Erwin.
  • Involved in Normalization and De-Normalization of existing tables for faster query retrieval.
  • Performed Verification, Validation, and Transformations on the Input data (Text files, XML files) before loading into target database.
  • Involved in Ralph Kimball and Bill Inman Methodologies (Star Schema, Snow Flake Schema).
  • Coded using Teradata Analytical functions, BTEQ SQL of TERADATA, write UNIX scripts to validate, format and execute the SQLs on LINUX environment.
  • Worked in importing and cleansing of data from various sources like Teradata, Oracle, Netezza flat files, SQL Server with high volume data.
  • Created SSIS Reusable Packages to extract data from Multi formatted Flat files, Excel, XML files into Database and Billing Systems
  • Creating various type of reports such as drill down & drill through reports, Matrix reports, Sub reports and Charts using SQL Server Reporting Services (SSRS).
  • Implemented naming Standards and Warehouse Metadata for fact and dimension of Logical and Physical Data Model.
  • Created new database objects like Procedures, Functions, Packages, Triggers, Indexes and Views using T-SQL in SQL Server.
  • Created complex Stored Procedures and PL/SQL blocks with optimum performance using Bulk Binds (BULK COLLECT & FORALL), Inline views, Reference cursors, cursor variables, dynamic SQL, v-arrays, external tables, nested tables, etc.
  • Created SSIS Packages using SSIS Designer for exporting heterogeneous data from OLE DB Source, Excel Spreadsheets to SQL Server.
  • Designing and developing SQL Server Database, Tables, and Indexes, Stored procedures, Views, User Defined Functions and other T- SQL statements.
  • Involved in purpose of this project is to migrate the Current Optum Rx Data Warehouse from the Iseries database environment to a Netezza appliance.

Environment: Netezza, Oracle9i, Taradata13.1, T-SQL, SQL Server, DB2, SSIS, LINUX, ERWIN r9.1, MDM, PL/SQL, ETL, SSRS, SSIS, Cognos, SAS, SPSS, Excel, Pivot tables, Shell scripting, UNIX etc.

Data Analyst

Confidential, Durham NC

Responsibility:

  • Involved in designing/developing Logical Data Analyst & Physical Data Analyst using Erwin r8.x DM.
  • Worked with DB2 Enterprise, Oracle8.x Enterprise, Teradata13, Mainframe sources, Netezza Flat files, and datasets operational sources.
  • Worked with various process improvements, normalization, de-normalization, data extraction, data cleansing, and data manipulation.
  • Performed data management projects and fulfilling ad-hoc requests according to user specifications by utilizing data management software programs and tools like TOAD, MS Access, Excel, XLS and SQL Server.
  • Created SQL scripts to load the custom data into Development, Test and production Instances using Import/Export. Created scripts to create custom Tables and Views.
  • Designed, Build the Dimensions, cubes with star schema and Snow Flake Schema using SQL Server Analysis Services (SSAS).
  • Involved in Integration of various data sources like DB2, Oracle, Netezza, SQL Server, SAP BW, XML Files.
  • Worked with requirements management, workflow analysis, source data analysis, data mapping, Metadata management, data quality, testing strategy and maintenance of the model.
  • Used SQL Profiler for troubleshooting, monitoring, and optimization of SQL Server and non production database code as well as T-SQL code from developers and QA.
  • Involved in data from various sources like Oracle Database, XML, Flat Files, CSV files and loaded to target warehouse.
  • Designed the ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin.
  • Designed and created MDX queries to retrieve data from cubes using SSIS.
  • Created SSIS Packages using SSIS Designer for exporting heterogeneous data from OLE DB Source, Excel Spreadsheets to SQL Server.
  • Extensively worked in SQL, PL/SQL, SQL Plus, SQL Loader, Query performance tuning, DDL scripts, database objects like Tables, Views Indexes, Synonyms and Sequences.
  • Developed and supported the extraction, transformation and load process (ETL) for a Data.

Environment: ERWIN8.x, Netezza, Oracle8.x, SQL, PL/SQL, SQL Plus, SQL Loader, Informatica, CSV, Taradata13, T-SQL, SQL Server, DB2, SSIS, LINUX, MDM, PL/SQL, ETL, Excel, Pivot tables, SAS, SPSS, SSRS.

We'd love your feedback!