We provide IT Staff Augmentation Services!

Data Analyst /data Modeler Resume

0/5 (Submit Your Rating)

Wilmington, DE

SUMMARY

  • 9 years of excellent experience as a Data Modeler, Data Analyst and ETL/SQL Developer in Finance, Insurance, Healthcare, banking domains. Also experienced in database management, strategic analysis, requirements gathering, application and decision support, which plays a key role in software development life cycle (SDLC).
  • 9 Years of IT experience in System Analysis, Application Design, Dimensional Data Modeling, Implementation and Testing of Enterprise Data Warehousing and Enterprise Database.
  • 9 years of experience in Data Modeling using Erwin 8.0/7.2.9/4.5/4.1.4/3.5.2 , ER/Studio7.1.1/7.5, Oracle Designer 2000 and SQL Designer.
  • Extensive Experience in Relational Modeling, Dimensional Modeling, Conceptual, Logical Modeling, Physical Modeling, Star Schema, Snowflakes Schema, ERD(IDEF1X and IE notation) ER Diagrams, Granularity, Cardinality and Database Reengineering.
  • Experience in developing data models which will serve both OLTP and OLAP functionality.
  • Has a good experience in Pharmacy related Database design.
  • Has a good knowledge on Tax domain.
  • Expertise in working with business users as part of requirements gathering and converting them into technical solution.
  • Designed databases involving complex views, triggers in Oracle 11g/10g/9i/8i/8, Microsoft Access, DB2, and SQL Server 2005/2000.
  • Experience in creating, maintaining and designing database diagrams, setting table relationships and indexes.
  • Designed ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies).
  • Worked closely with ETL teams such as Informatica to extract data from Source system to target systems and created mappings, Data Reports.
  • Deployed naming standard to the Data Model and followed company standard for Project Documentation.
  • Has a good experience on Consolidating data from different data source and performing Data Analysis
  • Expertise in Normalization techniques for data consistency and Flexible Database design.
  • Expertise database normalization method defined by E.F CODD (3NF - Third Normal Form).
  • Expertise in Creating UNIX Shell scripts.
  • Expertise in data design and modeling, system study, design, and development by applying Ralph Kimball methodology of dimensional and relational data modeling.
  • Designed the PL/SQL procedures for getting the data from all systems to Data Warehousing system. The data was standardized to store various Business Units in tables.
  • Expertise in creating Schema objects like Indexes, Views, and Sequences.
  • Worked in a team environment and also coordinated with business/functional team.
  • Excellent interpersonal and oral/written communication skills.
  • Involved in the preparation of Detailed Design Documentation.
  • Excellent analytical and communication skills with clear understanding of business process flow, SDLC and facilitated JAD sessions.
  • Good interpersonal and human relation skills to interact with clients and co-workers

TECHNICAL SKILLS

Analysis and Modeling Tools: Erwin 7.x/8.x/9.x, ER/Studio, MS Visio

ETL Tools: Informatica Power Center 9.x/8.x/7.x, clover ETL

BI Tools: Cognos 8, MicroStrategy, Business Objects, Qlikview, Tableau

Languages: SQL, PL/SQL, XML, HTML, UNIX Shell Scripting, C, C++, Java AWK

Databases: MS SQL Server 2012/2008/2005 , Oracle11g/10g/9i, MS Access 2000, IBM DB2.

Operating Systems: UNIX (Sun), Windows, GNU/Linux, IOS

ProjectExecution Methodologies: Ralph Kimball and Bill Inmon data warehousing methodology, Rational Unified Process (RUP), Agile, Rapid Application Development (RAD), Joint Application Development (JAD)

PROFESSIONAL EXPERIENCE

Confidential, Wilmington, DE

Data Analyst /Data Modeler

Responsibilities:

  • Understanding the specification and analyzed data according to the client requirement
  • Conducting requirement gathering sessions with Mortgage Banking Report Owners/SMEs to understand the current process flow of report generation and identifying the Risk Data elements.
  • Back-tracking of Risk Data elements to its source system and capturing all the calculation/derivation logic involved in underlying SAS/SQL codes.
  • Mapping the Risk Data elements to the Authoritative Data Source and documenting the Schema, Database, Table details for data modeling purpose.
  • Performed Risk Data aggregation i.e. defining, gathering and processing risk data to enable SIFIs (Significantly Important Financial Institutions) to measure performance against internal tolerance criteria. This includes ability to sort, align, merge and analyze multiple sets of data across the enterprise.
  • Participated in DRB Sessions with the reviewer by presenting the designed data models and explained how we followed the standards while designing the models.
  • Prepared Conceptual Model Use Cases for representing the various dimensions and measures involved in recreation of Mortgage Banking reports.
  • Interact with the customer and the SME(s) to identify the relationship between the data element and the business rule to document the data
  • Decode the SAS AND SQL queries to find all the data attributes involved and document for the purpose of development
  • Capturing data elements - definitions and characteristics for information requirements
  • Mapping data to the authoritative source field and table name as part of the data migration analysis activity
  • Worked as a part of ICDW Migration Testing Project where the duties were to writing SQL queries and executive the queries to match with the data in the target. Analysis the logic written by the developer and testing it with our own developed queries to check the migration of data accuracy.
  • If any discrepancy is observed in the data, in depth analysis is done (whether the defect is due to coding issue or due to requirement change) and the defects are logged in ALM.
  • Working on Application lifecycle management tool for uploading the test cases, raising the defects and managing the defects.
  • Prepared Testing Strategies document and documentation of Test Plans and Test Cases.
  • Understanding how the different attributes are categorized in different Servicing system and analyzed how data undergo changes in specific time period.
  • Analyzed how specific category of loans impacted on all the attribute and worked on finding the strategy to resolve this issue by using some filter conditions on the execution query.
  • Each month data load is done and the attributes are tested based on the SQL or Ab-initio logic provided in the BRD document. Provided high level compare results for attributes for all the servicing systems
  • Working especially on case statements and join condition to monitor the performance of the system from an end user perspective

Environment: ORACLE, Teradata SQL assistant 12.0, Windows Server 2012, MS SQL 2012, UNIX, Erwin 7x, Crystal Reports8, Informatica 9x, TOOD, SQL Developer.

Confidential, Hartford, CT

Sr.Data Analyst

Responsibilities:

  • Developed daily and monthly reports for health services using SQL Server.
  • Designeddataprocess flows using Informatica to sourcedatainto Statements database on Oracle platform.
  • Worked on Multiple projects CHRDS, CLAIMS, PHARMACY, FINANCE, and HEALTH CARE.
  • Extensively worked with DTS to load the data from source systems, and run in periodic intervals
  • Worked with data transformations in both normalized and denormalized data environments.
  • Involved in data manipulation using stored procedures and Integration Services.
  • Worked on query optimization, stored procedures, views, and triggers.
  • Assisted in OLAP and Data Warehouse environment when assigned.
  • Created tables, views, triggers, stored procedures and indexes.
  • KIMBALL and INMONN methodologies were used for architecture.
  • Designed and implemented database replication strategies for both internal and Disaster Recovery
  • Created ftp connections, database connections for the sources and targets.
  • Maintained security and data integrity of the database.
  • Performed extensive data analysis and data validation on Teradata.
  • Metadata repository was updated and maintained according to change requests.
  • Developed several forms & reports using Crystal Reports.
  • Experienced in master data management (MDM), Metadata management, data quality.
  • Experienced in building Data marts, data structures, data storages, data warehouses, data archives, and data analysis
  • Worked with the Implementation team to ensure a smooth transition from the design to the implementation phase.

Environment: DB2, SQL, ORACLE, Teradata SQL assistant 12.0, Windows Server 2012, MS SQL 2012, UNIX, Erwin 9x, Crystal Reports8, Informatica 9x, TOOD.

Confidential, Raleigh, NC

Sr. Data Modeler/ App Analyst

Responsibilities:

  • Built Conceptual, Logical and Physical model and implemented them on Care Synergy.
  • Understanding the application and building the database accordingly.
  • Participated in meeting with different teams of CMIS, MDM and CAREVIEW and took the requirements and implemented them on Care Synergy.
  • Interacted with the business Analyst and getting the functional requirements
  • Gathering accurate data by data analysis and functional analysis
  • Provided 24/7 support on Database perspective in case of any issue raises
  • As this project follows Agile Methodology, I will be working with the QA people closely and load the data for Unit Testing and Manual Testing
  • Check to see data model and database are in sync.
  • Data modeling, design & implementation based on established standards
  • Done Data Loading using ETL for QA for testing and also worked closely with the developers and getting the requirement for the database side
  • Forward engineer logical models into physical tables with Confidential naming standard class words.
  • Data Quality, Data Migration while loading the data from one database to another.
  • Has built a good knowledge on the application and gone through the each and every requirement clearly and implemented those on Database
  • Review the data model with functional team and technical team
  • Followed Data Standard Documents while building the database and participated in Database sessions for finalizing the database structure
  • Created a Data Structure document for each table and column that is implemented on database
  • Designed and developed Data Quality Reports, Chain of Trust Mechanisms for File processing.

Environment: SQL Server 2012, Windows Server 2003, Teradata, Windows XP, SQL Profiler, VSS, TFS, Erwin, DBTA, Initiate - IBM Tool for Person MDM, SQL Management Studio, DBTA, SSIS, Informatica

Confidential, Louisville

Data Modeler

Responsibilities:

  • Architected an enterprise solution and led a technical team of 6 to build the Person Master Data Management process and solution for Confidential 's insurance business.
  • Person/Customer MDM Solutions
  • Custom and Automated Maintenance Plans (Optimization, Integrity Checks etc.) for Very Large Databases. The first to be done in Confidential .
  • Architected an enterprise solution that provides data isolation capabilities around co-located data. This ensures that data is provided based on appropriate security context.
  • Data Modeling for OCH DB
  • Performance Tuning of the SDR/OCH DB's
  • Familiar with data architecture including data ingestion pipeline design, Hadoop information architecture, data modeling and data mining, machine learning and advanced data processing. Experience optimizing ETL workflows.
  • To build new servers with an enhanced architecture solution to accommodate ETL, OLTP, Reporting requirements for SDR
  • Has been working on Application Design for Data Isolation, Preferences.
  • Designed and developed Data Quality Reports, Chain of Trust Mechanisms for File processing.
  • Built the system for non-functional activities like backup, restore, recovery, auditing and archival processes for the OCH DB.

Environment: SQL Server 2005, Windows Server 2003, Teradata, Windows XP, SQL Profiler, VSS, TFS, Erwin, Enterprise Architect Sparx 7.5, DBTA, Initiate - IBM Tool for Person MDM, SQL Management Studio, DBTA, SSIS, Informatica

Confidential, Wilmington, DE

Sr. Data Modeler

Responsibilities:

  • Involve in project kickoff meetings to understand the need and scope of the project.
  • Attend cross team meetings with project business Analysts and technical leads to gather basic requirements to create a high level project diagrams.
  • Perform data profiling in Production environment to analyze and collect the statistical information.
  • Design the logical diagrams according to Confidential standards for initial review with Technical leads.
  • Analyze the relationship between the entities based on the requirements and define STAR and Snowflake schema models in Erwin.
  • Maintain metadata of all the new and existing models in Erwin to be readily available for business users.
  • Apply normalization techniques on data set and embed them in the model
  • Forward engineer logical models into physical tables with Confidential naming standard class words.
  • Analyze the table structure based on the size and design indexes to boost the performance.
  • Created new reports based on requirements.
  • Created Report-Models for ad-hoc reporting and analysis.
  • Maintain the script versions in WinCVS to deploy in QA environments.
  • Coordinate with DBAs during deployments and troubleshoot the issues if any
  • Document the work flow in high level and detail level for future reference and enhancements.

Environment: Oracle 10G/11G, TOAD, Teradata, Erwin 9.2, Windows, SQL plus, WinCVS, HP Quality Center 11, SDLC, UNIX, Windows XP

Confidential

Data Modeler

Responsibilities:

  • Gathered business requirements through interviews, surveys, prototyping and observing from account Managers and UI (User Interface) of the existing Broker Portal system.
  • Wrote Requirement and Design document for Master Data Entities. Used Corporate Modeler to design the Process
  • Conducted controlled brain-storming sessions with project focus groups.
  • Prepared High Level Logical Data Models and BRDs (Business Requirement Documents)
  • Supporting documents containing the essential business elements, detailed definitions, and descriptions of the relationships between the actors to analyze and document business data requirements.
  • Analyzed different source systems in order to come up with Source system which was generating the data.
  • Planned and defined system requirements to Use Case, Use Case Scenario and Use Case Narrative using the UML (Unified Modeling Language) methodologies.
  • Created Use Case Diagrams, Activity Diagrams, Sequence Diagrams in Rational Rose.
  • Conducted JAD sessions with management, SME (Subject Matter Expertise), vendors, users and other
  • Stakeholders for open and pending issues to develop specifications.
  • Participated in the Data Modeling to generate target database layout use for the data marts.
  • Generated data mapping documents to define mapping from source to target.
  • Reviewed the business requirements or use cases, do the analysis, and convert them to logical data Models.
  • Participated in high-level and low-level design sessions including writing technical project documentation.
  • Identified/documented data sources and transformation rules required populating and maintaining data Warehouse content.
  • Defined the data definitions, and create the target tables in the database.
  • Mapped the source and target databases by studying the specifications and analyzing the required transforms.
  • Provided data modeling support for numerous strategic application development projects.
  • Documented the Purpose of mapping so as to facilitate the personnel to understand the process and
  • In corporate the changes as and when necessary.

Environment: Windows Server 2003, MS SQL 2000, UNIX, Erwin 5, Crystal Reports8

We'd love your feedback!