Data Architect / Sr. Data Modeler Resume
Malvern, PA
SUMMARY
- 11+ yrs of progressive Oracle development (Data Architect) experience in Development, Support and Maintenance of various Client/Server applications in Finance, Telecom and Sales projects.
- Solid understanding of Data Modeling, Evaluating Data Sources and strong understanding of Data Warehouse/Data Mart Design, ETL, BI, OLAP, Client/Server applications.
- Expertise in writing SQL queries and optimizing the queries in Oracle, SQL Server 2008 and TeradataV2R6/R12/R13.
- Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
- Excellent knowledge in Data Analysis, Data Validation, Data Profiling, Data Cleansing, Data Verification and identifying data mismatch.
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata V2R6/R12/R13.
- Excellent experience on Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and FastExport.
- Design, install, patch and supporting databases and related functionality, providing the highest possible performance, availability, and database security.
- Experienced in Data masking using various tools.
- Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
- Good experience on Greenplum, Hive, Teradata, Oracle, and SQL Server database.
- Excellent experience with IBM Ionosphere utilizing MDM, Data profiling and data Modeling.
- Extensive working experience in Normalization and De - Normalization techniques for both OLTP and OLAP systems in creating Database Objects like tables, Constraints (Primary key, Foreign Key, Unique, Default), Indexes.
- Extensively worked on ERWIN tool with all features like REVERSE Engineering, FORWARD Engineering, SUBJECT AREA, DOMAIN, Naming Standards Document etc.
- Capabilities to provide AWS operations and deployment guidance and best practices throughout the lifecycle of a project.
- Experience with application / data migration to AWS.
- Experienced working with Excel Pivot and VBA macros for various business scenarios.
- Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export through the use of multiple ETL tools such as Ab Initio and Informatica Power Center.
- Experienced in testing and writing SQL and PL/SQL statements - Stored Procedures, Functions, Triggers and packages.
- Excellent knowledge on creating reports on Pentaho Business Intelligence, Tibco Spotfire & Tableau.
- Extensive knowledge and experience in producing tables, reports, graphs and listings using various procedures and handling large databases to perform complex data manipulations.
- Excellent experience in Data mining with querying and mining large datasets to discover transition patterns and examine financial data.
- Excellent in creating various artifacts for projects which include specification documents, data mapping and data analysis documents.
- An excellent team player & technically strong person who has capability to work with business users, project managers, team leads, architects and peers, thus maintaining healthy environment in the project.
TECHNICAL SKILLS
Data Warehousing: Informatica 9.5/9.1/8.6/7.1.2 (Repository Manager, Designer, Pentaho (BI), Workflow Manager, and Workflow Monitor), SSIS, Data Stage 8.x
Reporting Tools: Business Objects6.5, XIR2
Data Modeling: Star-Schema Modeling, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables, Erwin
Testing Tools: Win Runner, Load Runner, Test Director, Mercury Quality Center, Rational Clear Quest
RDBMS: Oracle 12c/11g/10g/9i/8i/7.x, MS SQL Server, UDB DB2 9.x, Teradata V2R6/R12/R13, MS Access 7.0, Greenplum (PostgreSQL)
Programming: SQL, PL/SQL, UNIX Shell Scripting, VB Script
Environment: Windows (95, 98, 2000, NT, XP), UNIX
Other Tools: TOAD, MS-Office suite (Word, Excel, Project and Outlook), BTEQ, Teradata V2R6/R12/R13 SQL Assistant
PROFESSIONAL EXPERIENCE
Confidential, Malvern PA
Data Architect / Sr. Data Modeler
Responsibilities:
- Create logical and physical data models using Erwin to meet the needs of the organization's information systems and business requirements.
- Analysis of functional and non-functional categorized data elements for Data Migration, Data profiling and Mapping from source to target data environment.
- Analyzed the business information requirements and research the OLTP source systems to identify the measures, dimensions and facts required for the reports.
- Demonstrated database technical functionality, such as performance, security and reliability. Enforced naming standards and data dictionary for data models.
- Created reports from Greenplum (Pivotal) which references the positions/transactions for each customer's monthly invoice for all jurisdiction (CFTC, ESMA, CANADA, ASIC & MAS) and make it available on the Portal .
- Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
- Working on Amazon Redshift and AWS and architecting a solution to load data, create data models and run BI on it.
- Involved in Netezza Administration Activities like backup/restore, performance tuning, and Security configuration.
- Create architectural documents such as, the e2e data flow for the IBM Information data warehouse system. Identified security loopholes, established data quality assurance and addressed data governance.
- Designed Physical Data Model (PDM) using IBM Info sphere Data Architect data modeling tool and Oracle PL/SQL.
- Used push down optimization in Informatica to call Greenplum GPLoad functions
- Lead the design and modeling of tactical architectures for development, delivery, and support of projects.
- Called Greenplum Business Rules, Data Rules and Transform Rules functions using Informatica Stored Procedure Transformation.
- Developed and maintained policies, standards, and guidelines to ensure that a consistent framework is applied across the company.
- Promoted the use of a shared infrastructure, application roadmap, and documentation of interfaces to improve information flow and reduce costs.
- Architect and build meta data repository to describe Digital business data, technical data, and processes.
- Designed solutions for multiple large data warehouses with a good understanding of cluster and parallel architecture as well as high-scale or distributed RDBMS and/or knowledge on NoSQL platforms.
- Developed Ad-Hoc Queries, Views and functions in Greenplum in order to make data accessible for Business Analyst and Managers.
- Providing extensive technical, strategic advice and guidance of the highest level to senior managers and technical resources in the creation and implementation of new data standards and databases.
- Created/generated/analyzed reports from SQL database on a scheduled/ad hoc basis for upper management.
- Overseeing the design, evaluation, selection, implementation and support of major databases and metadata structures.
- Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
- Developed data transformation and cleansing rules for migration using ETL tools.
- Coordinated Data Migration from old to new system during Teradata system capacity expansion.
- Moved data from Greenplum to Apache Solr.
- Performed Data Mining on data using very complex SQL queries and discovered pattern.
- Provided Administration and Support for various DTCC Greenplum environments.
- Widely used Normalization methods and have done different normalizations (3NF).
- Worked at Conceptual/Logical/Physical data model level using Erwin according to requirements.
- Worked on performance tuning of queries and was able to reduce the run time 3 hours to 35 minutes.
- Developed the performance tuning of the database by using EXPLAIN PLAN, TKPROF utilities and also debugging the SQL code.
- Extensive experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like Erwin.
- Resolveddataissues and updates for multiple applications using SQL queries/scripts.
- Created dimensional model for the reporting system by identifying required dimensions and facts using Erwin.
Environment: Erwin 9x, DB2, Oracle 12c, Information Analyzer, Informatica, IBM Infosphere, MDM, Quality Centre, Excel, Teradata, PL/SQL, UNIX, Netezza, MS-Word, SQL.
Confidential, Omaha, NB
Data Architect / Sr. Data Modeler
Responsibilities:
- Worked with Data Stewards and subject matter experts to research reported data anomalies, identified root causes, and determined appropriate solutions.
- Worked with data source systems and Client systems to identify data issues, data gaps, identified and recommended solutions.
- Worked on creating SQL queries and performance tuning of queries.
- Prepared Logical Data Models using ERWIN that contains set of diagrams and supporting documents containing the essential business elements, detailed definitions, and descriptions of the relationships between the data elements to analyze and document business data requirements.
- Helped client evaluate Greenplum as a viable solution, which includes Greenplum working with Informatica and MicroStrategy.
- Defined Data Governance process and document plans to address security and operations.
- Uphold architecture responsibilities and collaborated with other senior architects and engineers in ensuring holistic enterprise architecture is established across our environment.
- Executed architecture and code reviews for our core ETL processes, enterprise data warehouse, and reporting tools.
- Created a detailed plan to migrate the EDW from Oracle to Greenplum.
- Collaborated with product teams to understand and translate requirements, ensuring specifications and architecture design meet broader technical strategic goals.
- Collaborated with engineering teams throughout the development process in ensuring best practices and designing scalable analytics infrastructure and data transformation / aggregation processes.
- Implemented Dimensional model for theDataMart and responsible for generating DDL scripts using Erwin.
- Created a high-level industry standard, generalizeddatamodel to convert it into logical and physical model at later stages of the project using Erwin.
- Performed Greenplum database versions upgrades on QA, development and production environments.
- Designed Metadata Repository to storedatadefinitions for entities, attributes & mappings between datawarehouse and source systemdataelements.
- Created native Oracle functions to Greenplum DDLs for Business Objects to use
- Worked on data manipulation and analysis & accessed raw data in varied formats with different methods and analyzing and processing data.
- Performed data modeling and data analysis as required.
- Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
- Created Crystals Reports were converted to run against Greenplum .
- Optimized slow running queries and BO reports on Greenplum.
- Worked on data warehousing, ETL, SQL, scripting and big data (MPP + Hadoop).
- PerformedDatamapping between source systems to Target systems, logicaldatamodeling, created class diagrams and ER diagrams and used SQL queries to filterdata.
- Supported business areas and database platforms to ensure logical data model and database design, creation, and generation follows enterprise standards, processes, and procedures.
- Provided input into database systems optimization for performance efficiency and worked on full lifecycle of data modeling (logical - physical - deployment)
- Involved with datacleansing/scrubbing and validation.
- Performed dicing and slicing ondatausing Pivot tables to acquire the churn rate pattern and prepared reports as required.
- In depth analyses ofdatareport was prepared weekly, biweekly, monthly using MS Excel, SQL & UNIX.
Environment: PL/SQL, Informatica 9.x, Oracle 11G, Netezza, Aginity, ERWIN data modeler, UNIX, SQL, Hadoop, Mainframe, Teradata.
Confidential - Bellevue, WA
Sr. Data Analyst / Data Modeler
Responsibilities:
- Actively involved in creating Physical and Logical models using Erwin.
- Created data masking process to mask the sensitive data between production and test environment.
- Worked on IDQ for Data profiling, cleansing and matching.
- PerformedDatamodeling using Erwin. Identified objects and relationships and how those all fit together as logical entities, these are then translated into physical design using the forward engineering Erwin tool.
- Performed literature searches and ad-hoc data collection based on requests.
- Developed Logical and Physicaldatamodels that capture current state/future statedataelements and dataflows using Erwin / Star Schema.
- Worked with Informatica Data Quality 8.6.1 (IDQ) toolkit, Analysis, data cleansing, data matching, data conversion, exception handling, and reporting and monitoring capabilities of IDQ 8.6.1.
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
- Written several shell scripts using UNIX Korn shell for file transfers, error logging, data archiving, checking the log files and cleanup process.
- Used ERWIN Studio CW Erwin tool for modeling logical models.
- Written SQL scripts to test the mappings and Developed Traceability Matrix of Business Requirements mapped to Test Scripts to ensure any Change Control in requirements leads to test case update.
- Created Logical/PhysicalDatamodels in 3NF in the Warehouse area of EnterpriseData Warehouse.
- Worked on all phases of data warehouse development lifecycle, from gathering requirements to testing, implementation, and support using Pentaho Data Integration
Environment: - Erwin, MS SQL Server, Oracle, MS office, Business Objects Clear Quest, Clear Case, SQL, PL/SQL, Informatica, Teradata.
Confidential, Livonia, MI
Data Analyst/Data Modeler
Responsibilities:
- Performeddataanalysis and profiling of sourcedatato better understand the sources.
- Managed, updated and manipulated report orientation and structures with the use of advanced Excel functions including Pivot Tables and V-Lookups.
- Used data cleansing techniques, Excel pivot tables, formulas, and charts.
- Involved in mapping spreadsheets that will provide theDataWarehouse Development (ETL) team with source to targetdatamapping, inclusive of logical names, physical names,datatypes, domain definitions, and corporate meta-datadefinitions.
- Extensively used SQL forDataAnalysis and to understand thedatabehavior.
- Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Created Schema objects like Indexes, Views, and Sequences, triggers, grants, roles, Snapshots.
- Designed the procedures for getting thedatafrom all systems toDataWarehousing system. The datawas standardized to store various Business Units in tables.
- Responsible for different Data mapping activities from Source systems to Teradata
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, flat files, SQL Server 2005 with high volume data.
- Developed Star Schema and Snowflake Schema in designing the Logical Model into Dimensional Model.
- Tuning the Informatica Mappings for optimum performance and scheduling ETL Sessions.
- Conducted several PhysicalDataModel training sessions with the ETL Developers. Worked with them on day-to-day basis to resolve any questions on Physical Model.
- Extensively worked on documentation ofDataModel, Mapping, Transformations and Scheduling jobs.
- Involved in ETL mapping documents indatawarehouse projects.
Environment: Quality Center, MS Excel 2007, PL/SQL, Business Objects XIR2, ETL Tools Informatica, Oracle 10G, Teradata R12, Teradata SQL Assistant.
Confidential, Boston, MA
Data Analyst/Data Modeler
Responsibilities:
- Used and supported database applications and tools for extraction, transformation and analysis of raw data
- Developed, managed and validated existing data models including logical and physical models of the data warehouse and source systems utilizing a 3NF model
- Developed and programmed test scripts to identify and manage data inconsistencies and testing of ETL processes
- Wrote simple and advanced SQL queries and scripts to create standard and ad hoc reports for senior managers.
- Involved with data profiling for multiple sources and answered complex business questions by providing data to business users.
- Worked on enhancements to theDataWarehouse model using Erwin as per the Business reporting requirements.
- Worked on Fidelity Information Systems (FI Systems) and Maintained security and data integrity of the database.
- Worked withDataQuality Team in defining and configuration of Rules, Monitoring and preparation of DataQuality Analysis and Dashboards using Trillium.
- Analyzed the business requirements by dividing them into subject areas and understood thedataflow within the organization
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle, SQL server and DB2.
- Flexible to work late hours to coordinate with offshore team.
Environment: - Oracle, SQL, Erwin, MS office, Business Objects, Clear Quest, Clear Case, DB2
Confidential, Southfield, MI
Data Modeler/Analyst
Responsibilities:
- Created logical and physical models using best practices to ensure high quality and reduced redundancy.
- Optimized and update logical and physicaldatamodels to support new and existingdatamodels
- Recommend opportunities for reuse ofdatamodels in new environments
- Performed reverse engineering of physicaldatamodels from databases and SQL scripts.
- Validated businessdataobjects for accuracy and completeness
- Wrote SQL code and debug them, unit testing was performed.
- Extensively used star schema methodologies in building and designing the logicaldatamodels into dimensional model.
- Created data masking mappings to mask the sensitive data between production and test environment.
- Worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
- Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Netezza.
- Written several shell scripts using UNIX Korn shell for file transfers, error logging, data archiving, checking the log files and cleanup process.
- Written SQL scripts to test the mappings and Developed Traceability Matrix of Business Requirements mapped to Test Scripts to ensure any Change Control in requirements leads to test case update.
- Wrote stored procedures and triggers extensively and working very closely with developers, business analysts and end users to generate various audit reports and troubleshoot their query problems.
- Provided maintenance and support to customized reports developed in crystal reports.
- Gathered statistics on large reports and redesigned indexes.
ENVIRONMENT: Oracle, DB2, Erwin, Teradata, SQL Server, SQL, Pl/SQL, crystal reports, windows XP, SQL *PLUS.