We provide IT Staff Augmentation Services!

Sr Data Modeler/data Analyst Resume

0/5 (Submit Your Rating)

Houston, TX

SUMMARY

  • Over 8 years of strong experience in Business and Data Modeling/Data Analysis, Data Profiling, Data Migration, Data Conversion, Data Quality, Data Governance, Data Integration, MDM, NoSQL and Metadata Management Services and Configuration Management.
  • Experienced on data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing.
  • Experienced in Dimensional Data Modeling experience using Data modeling, RelationalData modeling, ER/Studio, Erwin, and Sybase Power Designer, Star Join Schema/Snowflake modeling, FACT & Dimensions tables, Conceptual, Physical & logicaldatamodeling.
  • Experienced in writing Pig Latin scripts, MapReduce jobs and HiveQL.
  • Expertise in theDataAnalysis, Design, Development, Implementation and Testing usingData Conversions, Extraction, Transformation and Loading (ETL) and SQL Server, ORACLE and other relational and non - relational databases.
  • Extensively worked on ERWIN tool with all features like REVERSE Engineering, FORWARD Engineering, SUBJECT AREA, DOMAIN, Naming Standards Document etc.
  • Experienced in importing and exporting the data using Sqoop from HDFS to Relational Database systems/mainframe and vice-versa.
  • Strong experience in using Excel and MS Access to dump thedataand analyze based on business needs.
  • Experienced with Integration Services (SSIS), Reporting Service (SSRS) and AnalysisServices (SSAS).
  • Expertise in Normalization to 3NF/Denormalization techniques for optimum performance in relational and dimensional database environments.
  • Experienced in carrying out Software Development Life Cycle (SDLC) in relational and object methodologies.
  • Extensive experienced on ER Modeling, Dimensional Modeling (Star Schema, Snowflake Schema) andDatawarehousing and OLAP tools.
  • Expertise indatabase programming (SQL, PLSQL) MS Access Oracle 12c/11g/10g/9i, XML, DB2, Informix, Teradata,,Database tuning and Query optimization.
  • Expertise in performingdataanalysis anddataprofiling using complex SQL on various sources systems including Oracle and Teradata.
  • Experienced in logical/physical database design and review sessions to determine and describedataflow anddatamapping from source to target databases coordinating with End Users, Business Analysts, DBAs and Application Architects.
  • Expertise in Visio, Process Flow Diagrams, Activity Diagrams, Cross Functional Diagram, Swim Lane Diagrams, Use Case Diagrams.
  • Expertise in scheduling JAD (Joint Application Development) with End Users, stake Holders, Subject Matter Experts, Developers and Testers.
  • Expertise inDatamodeling (Dimensional & Relational) concepts like Star-Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
  • Expertise in writing Stored Procedures, Functions, Nested Functions, building Packages and developing Public and Private Sub-Programs using PL/SQL and providing Documentation.
  • Expertise in loading data by using the Teradata loader connection, writing Teradata utilities scripts (Fastload, Multiload) and working with loader logs.
  • Experienced in developing TSQL scripts and stored procedures to perform various tasks and multiple DDL, DML, and DCL activities to carry out business requirements
  • Strong RDBMS concepts and well experience in creating database Tables, Views, Sequences, triggers, joins taking the Performance and Reusability into consideration.
  • Efficient in Extraction, Transformation and Loading (ETL) data from spread sheets, database tables using Microsoft data transformation service (DTS)
  • Extensive knowledge in software testing methodology and developing Test Plans, Test Procedures, Test Case Design and Execution, Modification Requests.
  • Strong in conceptualizing and communicating enterprisedataarchitecture frameworks for global enterprises for inter operation ofdatawarehouses, middleware, and web applications.

TECHNICAL SKILLS

DataModeling Tools: Erwin 9x, 8x, 7x, ER Studio and Oracle Designer.

OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9

ETL Tools: SSIS, Pentaho, Informatica.

Programming Languages: Java, Base SAS and SAS/SQL, SQL, T-SQL, HTML, Java Script, CSS, UNIX shells scripting, PL/SQL.

Database Tools: Microsoft SQL Server 2000/2008, Teradata, Oracle 12c/10g/9i, and MS Access, PostgerSQL.

Web technologies: HTML, DHTML, XML, JavaScript

Reporting Tools: Business Objects, Crystal Reports

Packages: Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point Portal Server 2003/2007.

Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX

Quality Assurance Tools: Quick Test Pro, Win Runner, Load Runner, Quality Center.

Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume.

Other Tools: MS-Office suite (Word, Excel, Project and Outlook), BTEQ, Teradata, SQL Assistant, Aginity

PROFESSIONAL EXPERIENCE

Confidential, Houston, TX

Sr Data Modeler/Data Analyst

Responsibilities:

  • Involved with BusinessAnalyststeam in requirements gathering and in preparing functional specifications and changing them into technical specifications.
  • Involved in understanding and creating Logical and Physical Data model using Erwin Tool.
  • Guide the full lifecycle of a Hadoop solution, including requirements analysis, platform selection, technical architecture design, application design and development, testing, and deployment
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
  • Creating logical and physical data models using best practices to ensure high data quality and reduced redundancy.
  • Manage timely flow of business intelligence information to users.
  • Involved in making screen designs, Use Cases and ER diagrams for the project using ERWIN and Visio.
  • Define Big Data strategy, including designing multi-phased implementation roadmaps.
  • Analyze the Business information requirements and research the OLTP source systems to identify the measures, dimensions and facts required for the reports.
  • Performed Data mapping between source systems to Target systems, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data
  • Lead design of high-level conceptual and logical models that facilitate a cross-system/cross functional view of data requirements
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
  • Maintaining conceptual, logical and physical data models along with corresponding metadata.
  • Done data migration from an RDBMS to a NoSQL database, and gives the whole picture for data deployed in various data systems.
  • Designed and developed thedatadictionary and Metadataof the models and maintain them.
  • Involved in Data Warehouse Support - Star Schema and Dimensional modeling to help design data marts and data warehouse
  • Coordinated with DBA ondatabase build and table normalizations and de-normalizations
  • Developed triggers, stored procedures, functions and packages using cursors and ref cursor concepts associated with the project using Pl/SQL
  • Prepared documentation for all entities, attributes,datarelationships, primary and foreign key structures, allowed values, codes, business rules, glossary evolve and change during the project
  • Exported the patterns analyzed back to Teradata using Sqoop.
  • Troubleshoot test scripts, SQL queries, ETL jobs, and data warehouse/data mart/data store models.
  • Responsible for different Data mapping activities from Source systems to Teradata
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS
  • Developed the performance tuning of the database by using EXPLAIN PLAN, TKPROF utilities and also debugging the SQL code.

Environment: Erwin, Oracle, MS-SQL Server, Hive, NoSQL, Teradata, Netezza, PL/SQL, MS-Visio, Informatica, T-SQL, SQL, Crystal Reports 2008, Java, HDFS, PIG.

Confidential, Chicago IL

Sr. Data Analyst/Data Modeler

Responsibilities:

  • Design database, data models, ETL processes, data warehouse applications and business intelligence (BI) reports through the use of best practices and tools, including SQL, SSIS, SSRS and OLAP.
  • Transformed Logical Data Model to Physical Data Model ensuring the Primary Key and Foreign Key relationships in PDM, Consistency of definitions of Data Attributes and Primary Index Considerations.
  • Validated the data of reports by writing SQL queries in PL/SQL Developer against ODS.
  • Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
  • Involved withDataAnalysis primarily IdentifyingDataSets, SourceData, Source MetaData, Data Definitions andDataFormats
  • Developed Star and Snowflake schemas based dimensional model to develop thedatawarehouse.
  • DevelopedDataMapping,DataGovernance, and Transformation and cleansing rules for the MasterData Management Architecture involving OLTP, ODS.
  • Created Hive queries that helped analysts spot emerging trends by comparing fresh data with EDW tables and historical metrics.
  • Designed the ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin
  • Involved with all the phases of Software Development Life Cycle (SDLC) methodologies throughout the project life cycle.
  • Extensively involved in the Physical/logical modeling and development of ReportingData Warehousing System.
  • Performing reverse engineering of physical data models from databases and SQL scripts.
  • Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website.
  • Excellent experience managing the whole testing process, Project Status and monitor workload using HP Quality Centre.
  • Used Normalization (1NF, 2NF&3NF) and De-normalization techniques for effective performance in OLTP and OLAP systems.
  • Created fully fledged Source to Target Mapping documents (S2T), documented business and transformation rules.
  • Managed and reviewed Hadoop log files.
  • Created Complex SQL Queries using Views, Indexes, Triggers, Roles, Stored procedures and User Defined Functions Worked with different methods of logging in SSIS.
  • Worked in importing and cleansing of data from various sources like DB2, Oracle, flat files onto SQL Server with high volume data.
  • Involved in Migrating thedatamodel from one database to Teradata database and prepared a Teradata staging model.
  • Exported the analyzed data to the relational databases using Sqoop for visualization and to generate reports by our BI team.
  • Collecting, aggregating, matching, consolidating, quality-assuring, persisting and distributing such data throughout an organization to ensure consistency trough MDM.
  • Responsible for creating Hive tables, loading data and writing hive queries.
  • Worked on differentdataformats such as Flat files, SQL files, Databases, XML schema, CSV files.

Environment: Erwin, Teradata V14, Teradata SQL Assistant, Informatica, Oracle 11g, Netezza, SQL Server 2008, Mainframes, SQL, PL/SQL, XML, Hive, Hadoop, PIG.

Confidential, Pittsburgh, PA

Sr. Data Analyst/Data Modeler

Responsibilities:

  • Worked with business users to gather requirements and create data flow, process flows and functional specification documents.
  • Designed & Created Test Cases based on the Business requirements (Also referred Source to Target Detailed mapping document & Transformation rules document)
  • Developed, enhanced and maintained Snow Flakes Schemas withindatawarehouse and datamart with conceptual datamodels.
  • Designed 3rd normal form targetdatamodel and mapped to logical model.
  • Involved in extensive Data validation using SQL queries and back-end testing
  • Used SQL for Querying the database in UNIX environment
  • Designed and developed ETL processes using Informatica ETL tool for dimension and fact file creation
  • Involved in data analysis and creating data mapping documents to capture source to target transformation rules.
  • Used Erwin and Visio to create 3NF and dimensional data models and published to the business users and ETL / BI teams.
  • Involved in Data mapping specifications to create and execute detailed system test plans. The data mapping specifies what data will be extracted from an internal data warehouse, transformed and sent to an external entity.
  • Creating or modifying the T-SQL queries as per the business requirements.
  • Worked on creating role playing dimensions, factless Fact, snowflake and star schemas.
  • Using Erwin modeling tool, publishing of adatadictionary, review of the model and dictionary with subject matter experts and generation ofdatadefinition language.
  • Managed full SDLC processes involving requirements management, workflow analysis, source data analysis,datamapping, metadata management,dataquality, testing strategy and maintenance of the model.
  • Wrote complex SQL queries for validating thedataagainst different kinds of reports generated by Business Objects.
  • Analysis of functional and non-functional categorized data elements for data profiling and mapping from source to target data environment. Developed working documents to support findings and assign specific tasks.
  • Involved in fixing invalid mappings, testing of Stored Procedures and Functions, Unit and Integrating testing of Informatica Sessions, Batches and the Target Data.
  • Involved in the validation of the OLAP Unit testing and System Testing of the OLAP Report Functionality and data displayed in the reports

Environment: Erwin, Informatica Power Center 8.1/9.1, Power Connect/ Power exchange, Oracle 11g, Main frames,DB2 MS SQL Server 2008, SQL,PL/SQL, XML, Windows NT 4.0, Unix Shell Scripting.

Confidential, RI

Sr. Data Analyst/Data Modeler

Responsibilities:

  • Created and maintained Logical and Physical models for the data mart. Created partitions and indexes for the tables in the data mart.
  • Performed data profiling and analysis applied various data cleansing rules designed data standards and architecture/designed the relational models.
  • Maintained metadata (data definitions of table structures) and version controlling for the data model.
  • Developed SQL scripts for creating tables, Sequences, Triggers, views and materialized views
  • Worked on query optimization and performance tuning using SQL Profiler and performance monitoring.
  • Assesses and determines governance, stewardship, and frameworks for managing data across the organization.
  • Utilized Erwin's forward/reverse engineering tools and target database schema conversion process.
  • Worked on creating enterprise wide Model EDM for products and services in Teradata Environment based on the data from PDM. Conceived, designed, developed and implemented this model from the scratch.
  • Write SQL scripts to test the mappings and Developed Traceability Matrix of Business Requirements mapped to Test Scripts to ensure any Change Control in requirements leads to test case update.
  • Created the functions and preformed the queries for different application usingPostgreSQL.
  • Involved in extensive DATA validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
  • Created SQL scripts to find data quality issues and to identify keys, data anomalies, and data validation issues.
  • Used Graphical Entity-Relationship Diagramming to create new database design via easy to use, graphical interface.
  • Designed different type of STAR schemas for detailed data marts and plan data marts in the OLAP environment.

Environment: MS SQL Server 2005, DB2, Oracle SQL Developer, PL/SQL, Business Objects, Erwin 7.0.x, MS office suite, Windows xp, TOAD, SQL*PLUS, SQL*LOADER

Confidential

Data Analyst

Responsibilities:

  • Suggested various changes in the physical model to support the business requirements.
  • Developed server jobs to load thedatafrom flat files, text files, tag text files and MS SQL.
  • Utilized shared containers for code reusability and for implementing the predefined business logic.
  • Created and scheduled the job sequences by checking job dependencies.
  • Wrote complex SQL queries using joins, sub queries and correlated sub queries.
  • Wrote PL/SQL stored procedures, functions and packages and triggers to implement business rules into the application.
  • Developed shell scripts to invoke back end SQL and PL/SQL programs
  • Performed unit testing to check the validity of thedataat each stage.
  • UsedDataStage Director to debug the jobs and to view the error log to check for errors.
  • Implemented best practices in the development environment (code standards, code migration).
  • Used Informatica features to implement Type I & II changes in slowly changing dimension tables.
  • Created and ran workflows and Worklets using Workflow Manager to load thedatainto the target database.
  • Performance tuning of SQL Queries, Sources, Targets and sessions

Environment: Oracle 9i, MS SQL Server, PL/SQL, Toad, UNIX Shell Scripting.

We'd love your feedback!