Sr. Data Architect/data Modeler Resume
Dallas, TX
SUMMARY
- Over 12 + years of working experience as a Data Architect/Modeler and Data Analyst with high proficiency in requirement gathering and data modeling including design and support of various applications in OLTP, Data Warehousing, OLAP and ETL Environment.
- Experienced with data architecting, data mining, data modeling and business requirements gathering/analysis.
- Experienced with Management and implementation of database models, data flow diagrams, database schemas, db scripts, DTD schemas, structures and data standards to support a data management infrastructure.
- Experienced in designing the Conceptual, Logical and Physical data modeling using Erwin, Power Designer and ER Studio Data modeling tools.
- Expertise in system analysis, ER Dimensional Modeling, Database design and implementing RDBMS specific features.
- Experienced in designing solutions for multiple large data warehouses with a good understanding of cluster and parallel architecture as well as high - scale or distributed RDBMS and/or knowledge on platforms.
- Excellent experience with Normalization (1NF, 2NF and 3NF) and De-normalization techniques for improved database performance in OLTP, OLAP, Data Warehouse and Data Mart environments.
- Experienced in writing SQL queries and optimizing the queries in Netezza, Hive, Oracle and SQL Server and Teradata.
- Solid experience with data architecture support tools (e.g., ErWin, Enterprise Data Dictionary Tools - IBM InfoSphere, etc.)
- Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling
- Good understanding of cloud configuration in Amazon web services AWS.
- Experienced in using Excel and MS Access to dump the data and analyze based on business needs.
- Experienced in Multiple Relational Databases primarily like Oracle, SQL Server, MySQL and knowledge of non-relational and NOSQL database HBase
- Excellent experience on Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and FastExport.
- Excellent experience in writing SQL queries to validate data movement between different layers in data warehouse environment.
- Expertise in Importing, cleaning, transforming, validating or modeling data with the purpose of understanding or making conclusions from the data for decision making purposes.
- Experienced with NoSQL databases like Hbase, and MongoDB as well as other ecosystems like ZooKeeper, AWSRedshift etc.
- Excellent oral and written communication skills, analytical skills, problem solving, leadership and management capabilities with quick learning abilities
- Experience in performing root level analysis for data problems by identifying the system processes that lead to inaccurate data
TECHNICAL SKILLS
Programming Languages: SQL, PL/SQL, UNIX shell Scripting, Python, PERL, AWK, SED
Databases: Oracle 10/11g/12c, Teradata, MS SQL Server, MS Access, Netezza, MongoDB, HBase.
Tools: MS-Office suite (Word, Excel, MS Project and Outlook), VSS
Testing and defect tracking Tools: HP/Mercury (Quality Center, Win Runner, Quick Test Professional, Performance Center, Requisite, MS Visio & Visual Source Safe
Operating System: Windows, Unix, Sun Solaris
ETL/Data warehouse Tools: Informatica 9.1/9.5, Talend, SSIS.
Data Modeling: Star-Schema Modeling, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables, Erwin, ER Studio.
OLAP Tools: MS SQL Analysis Manager, DB2 OLAP, Cognos Powerplay, SAP Business Objects XIR3.1/XIR2, Web Intelligence,and Tableau.
Tools: & Software: TOAD, MS Office, BTEQ, Teradata SQL Assistant
Project Execution Methodologies: Ralph Kimball and Bill Inmon data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
PROFESSIONAL EXPERIENCE
Confidential, Dallas TX
Sr. Data Architect/Data Modeler
Responsibilities:
- Responsible for validating the data quality and integration of all Data Architecture Components deemed to be cross-domain or enterprise in scope.Responsible for capture of necessary Data Architecture metrics to support quality of processes and related artifacts.
- Oversee project deliverables thru the whole system development life cycle process and manage the delivery of the project by creating an overall project plan.
- Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data model in model mart, created naming convention files, co-coordinated with DBAs' to apply the data model changes.
- Built and articulated an architecture and delivery roadmap indicating current to future state, focused around clear benefits to both IT and business users.
- Developed, documented and implemented best practices for Big Data solutions and services
- Developed the logical data models and physical data models that confine existing condition/potential status data fundamentals and data flows using Erwin.
- Worked with cloud services like Amazon web services (AWS).
- Involved data analysis, business analysis, user requirement gathering, user requirement analysis, gap analysis, gathering metadata data cleansing, data transformations, data relationships, source systems analysis and reporting analysis
- Working with Model management stewards for model integration and DDL’s publishing.
- Responsible for creating design patterns and best practices for data and the integration of data through the enterprise
- Developed and recommended data management policies, standards, practices and security measures to ensure effective and consistent data management operations.
- Develop suitable AWS-based and Hybrid solutions based on customer requirements.
- Analyzed the business requirements by dividing them into subject areas and understood the data flow within the organization
- Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Used Erwin for reverse engineering to connect to existing database and ODS to create graphical representation in the form of Entity Relationships and elicit more information
- Evaluated system performance and validated NoSQL solutions.
- Used Graphical Entity-Relationship Diagramming to create new database design via easy to use, graphical interface.
- Designing and customizing data models for Data warehouse supporting data from multiple sources on real time
- Defined the Primary Keys (PKs) and Foreign Keys (FKs) for the Entities, created dimensions model star and snowflake schemas using Kimball methodology.
- Worked on AWSredshift data warehouse for columnar data storage.
- Applied data naming standards, created the data dictionary and documented data model translation decisions and also maintained DW metadata.
- Used IBM Information Analyzer data profiling tool to ensure data quality for data requirements.
Environment: Oracle 12c, Microsoft 2008, SQL Developer, SQL Manager, Metadata, Erwin, SQL, Teradata, MongoDB, Microsoft SQL Server 2012, SQL Manager, SAS, Perl, NoSQL, Teradata, Netezza.
Confidential, Chicago, IL
Sr. Data Architect/Data Modeler
Responsibilities:
- Facilitated development, testing and maintenance of quality guidelines and procedures along with necessary documentation
- Developed relational modeling, dimensional modeling, conceptual, logical modeling, physical modeling, star schema, snowflakes schema, ER diagrams, granularity, cardinality and database reengineering.
- Conducted logical data analysis and data modeling JAD sessions, communicated data-related standards.
- Generating AWS migration roadmaps and driving buy-in across complex organization structures.
- Responsible for evaluating and recommending Data Query and Modeling Tools to assist with Data Architecture activities.
- Done rationalizing of various data areas and addressing Data Management Processes (data lineage, data profiling, data quality management, data stewardship and governance).
- Design and develop technical solution to provide business users with functionalities requested.
- Managed and maintained Oracle and NoSQL databases in production domain and designed and developed NoSQL solutions for all users.
- Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement
- Review, approve and ascertain execution of the test scripts and review test results in development environment prior to promotion of code to QA environment.
- Designing standards for modeling, logical and physical data models, including logical and physical naming conventions, logical to physical model transformation rules, data domains.
- Creating indexes and tuning the SQL queries in Hive and Involved in database connection by using Sqoop.
- Conduct code and standard checks, including code review and peer-testing for programming modules developed.
- Used Star Schema and Snow flake schema methodologies in building and designing the logical data model in the dimensional Models.
- Redefined many attributes and relationships in the model and cleansed unwanted tables/columns as part of Data Analysis responsibilities.
- Analyzed the Business information requirements and examined the OLAP source systems to identify the measures, dimensions and facts required for the reports.
- Submitted SAS ETL programs against Mainframe environment for COBOL created data files.
- Documented logical, physical, relational and dimensional data models. Designed the data marts in dimensional data modeling using star and snowflake schemas.
- Redefined attributes and relationships in the model and cleansed unwanted tables/columns as part of data analysis responsibilities.
- Worked on the reporting requirements and involved in generating the reports for the Data Model.
- Initiated and conducted JAD sessions inviting various teams to finalize the required data fields and their formats.
- Conversed with Business Analyst and developers to gather information about the data models (Data Definition) to place the data dictionary in place.
- Worked with Hive/HQL to query data from Hive tables in HDFS.
- Designed different type of STAR schemas for detailed data marts and plan data marts in the OLAP environment.
- Created DMR mapping spreadsheets, Source to Target mapping documents including the transformation rules
Environment: Erwin, Informatica, Teradata, Windows XP/NT/2000, SQL Server 2000/2005, SQL, Oracle RAC 10g, Hive, MS Excel, MS Visio, HBase, Rational Rose, NoSQL, Teradata 6.2, Oracle 11g, TOAD, CSV files, XML files, Cognos, UNIX, Linux, AWK
Confidential, Charlotte, NC
Sr. Data Modeler/Analyst
Responsibilities:
- Created and maintained Database Objects (Tables, Views, Indexes, Partitions, Synonyms, Database triggers, Stored Procedures) in the data model.
- Presented the data scenarios via, ER Studio logical models and excel mockups to visualize the data better.
- Worked on building the data model using ER Studio as per the requirements, discussion and approval of the model from the BA.
- Involved with Data Analysis primarily Identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats
- SAS and SQL - writing SAS programs statements, ETL and UNIX.
- Designed and Developed logical & physical data models and Meta Data to support the requirements using ER Studio.
- Involved with Data Profiling activities for new sources before creating new subject areas in warehouse
- Conducted the Data Analysis and identified the Data quality issues using Data profiling methodologies.
- Created data masking mappings to mask the sensitive data between production and test environment.
- Enforced referential integrity in the OLTP data model for consistent relationship between tables and efficient database design.
- Created a high-level industry standard, generalized data model to convert it into logical and physical model at later stages of the project using ER Studio and Visio
- Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data warehouse, data mart reporting system in accordance with requirements.
- Involved in Teradata SQL Development, Unit Testing and Performance Tuning and to ensure testing issues are resolved on the basis of using defect reports.
- Using SQL in a relational database environment (DB2, Oracle, SQL Server (SSIS), Teradata, and MS Access) to access from SAS Data Integration Studio 4.2.
- Tested the ETL process for both before data validation and after data validation process. Tested the messages published by ETL tool and data loaded into various databases.
- Worked with business to identify the distinct data elements in each report to determine the number of reports needed to satisfy all reporting requirements.
- Taking the data from databases to Data warehousing of DB2, Oracle, SQL Server and MS Access.
- Responsible for analyzing various data sources such as flat files, ASCII Data, EBCDIC Data, Relational Data (Oracle, DB2 UDB, MS SQL Server) from various heterogeneous data sources.
- Tested the database to check field size validation, check constraints, stored procedures and cross verifying the field size defined within the application with metadata.
Environment: Netezza, ER Studio, DB2, Information Analyzer, Informatica, Quality centre, Excel, Toad, Oracle Designer, SQL Server 2008, SAS, TOAD, PL/SQL, Flat Files, Teradata, NIX AIX 5.2, PERL, Shell Scripting
Confidential, Lake Mary, FL
Sr. Data Modeler/Data Analyst
Responsibilities:
- Created conceptual, logical and physical models based on requirements gathered through interviews with the business users.
- Updated existing models to integrate new functionality into an existing application. Conducted one-on-one sessions with business users to gather warehouse requirements.
- Analyzed database requirements in detail with the project stakeholders by conducting joint Requirement Development sessions.
- Developed normalized Logical and Physical database models to design OLTP system.
- Coded using SAS/SQL, PL/SQL to extract data from Teradata and Oracle tables
- Created dimensional model for the reporting system by identifying required dimensions and facts using Erwin.
- Developed module to initiate and run standard database processes in SQL Server, Oracle.
- Used forward engineering to create a Physical Data Model with DDL that best suits the requirements from the Logical Data Model.
- Maintaining and implementing Data Models for Enterprise Data Warehouse using ERWIN.
- Extensive experience in PL SQL programming Stored Procedures, Functions, Packages and Triggers
- Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Eliminated errors in Erwin models through the implementation of Model Mart (a companion tool to Erwin that controls the versioning of models).
- Used SAS/STAT Procedures such as Proc Freq, Proc Means, Proc Summary, Proc Univariate for Analysis
- Used Erwin for reverse engineering to connect to existing database and ODS to create graphical representation in the form of Entity Relationships and elicit more information
- Verified that the correct authoritative sources were being used and that the extract, transform and load (ETL) routines would not compromise the integrity of the source data.
Environment: Erwin7, Windows XP NT 2000, SQL Server 2005, Teradata, Oracle10g, DB2, Informix, MS Excel, Mainframes MS Visio, MYSQL, Shell Scripting, Rational Rose, SAS, Metadata, Requisite Pro
Confidential, San Francisco, CA
Sr. Data Analyst
Responsibilities:
- Mapping and tracing data from system to system in order to solve a given business or system problem.
- Carry out analysis of the given data and draw correct inferences, in keeping the objectives of the analysis
- Focal point for making sound decisions related to data collection, data analysis, data security, methodologies and designs.
- Managed, updated and manipulated report orientation and structures with the use of advanced Excel functions including Pivot Tables and V-Lookups
- Gather data necessary to accomplish analysis from a number of sources, accumulate it together in prearranged format and enter the data in several data analysis software
- Involved with Data profiling for multiple sources and answered complex business questions by providing Data to business users
- Involved in Data mining, transformation and loading from the source systems to the target system
- Facilitated (JAD) Joint Application Development sessions to identify business rules and requirements and documented them in a format that can be reviewed and understood by both business people and technical people
- Provided quality data review for completeness, inconsistencies, erroneous and missing data according to data review plan
- Design and development of ETL processes using Informatica ETL tool for dimension and fact file creation
- Created and executed test cases for ETL jobs to upload master data to repository.
- Create and execute test scripts, cases, and scenarios that will determine optimal system performance according to specifications.
- Modified the automated scripts from time to time to accommodate the changes/upgrades in the application interface.
- Tested the database to check field size validation, check constraints, stored procedures and cross verifying the field size defined within the application with metadata.
- Involved in Teradata SQL Development, Unit Testing and Performance Tuning and to ensure testing issues are resolved on the basis of using defect reports.
- Tested the ETL process for both before data validation and after data validation process. Tested the messages published by ETL tool and data loaded into various databases
Environment: Quality Center 9.2, MS Excel 2007, PL/SQL, Java, Business Objects XIR2, ETL Tools Informatica 8.6/9.1/9.5, SSIS, Oracle 11G, Teradata R13, Teradata SQL Assistant, Load Runner 7.0, Oracle 10g, UNIX AIX 5.2, PERL, Shell Scripting