We provide IT Staff Augmentation Services!

Sr. Data Architect Resume

4.00/5 (Submit Your Rating)

Dallas, TX

SUMMARY:

  • Over 12+ years of experience as a Data Architecture, Data Analyst and Data Modeler with solid understanding of Business Requirements Gathering, Business Process Mapping Evaluating Data Sources and Data mapping, data profiling, Hadoop ecosystem, AWS, Data Analytics, Data Warehousing and ETL. 
  • Experienced in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
  • Solid experience in development of Conceptual, Logical and Physical Models for Online Transaction Processing and Online Analytical Processing (OLTP & OLAP) and Power Design.
  • Experiences with data modeling with NoSQL databases such as MongoDB for document models, AWS Redshift etc.
  • Excellent experience in trouble shooting SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Experienced with distributed  data warehousing and/or  data mining systems, using one or more Big  Data/NoSQL technologies (Hadoop, Hive, HBase, Pig, Cassandra, MongoDB) 
  • Hands on experience with modeling using ERWIN, ER Studio, MS Visio in both forward and reverse engineering cases and skillful in Data Analysis using SQL on Oracle, MS SQL Server, Netezza, DB2 & Teradata
  • Well versed in Normalization / De normalization techniques for optimum performance in relational and dimensional database environments. 
  • Experience in working with Business Intelligence and Enterprise  Data Warehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum and Amazon Redshift (AWS).
  • Good experience and understanding of Teradata SQL Assistant, Teradata Administrator and data load/ export utilities like BTEQ, Fast Load, Multi Load, Fast Export.
  • Hands on experience with various  Data  Architect and ETL  Architect, subsystem and patterns, including Change Date Capture, Slow Change Dimension,  Data Cleansing, auditing and validation, etc.  Solid experience with data governance, data steward, data quality concepts and implementations and expertise in performing User Acceptance Testing (UAT) and conducting end user training sessions. 
  • Strong background in various Data Modeling tools using Erwin, ER Studio and Power Designer. 
  • Experience in integration of various relational and non - relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, and Netezza database.
  • Extensive knowledge and experience in producing tables, reports, graphs and listings using various procedures and handling large databases to perform complex data manipulations.
  • Experienced in data transformation, data mapping from source to target database schemas and data cleansing procedures using Informatica Power Center, Talend and Pentaho. 
  • Good in Data warehouse loads, determining hierarchies, building various logics to handle Slowly Changing Dimensions. 
  • Strong experience in Normalization (1NF, 2NF, 3NF and BCNF) and De-normalization techniques for effective and optimum performance in OLTP and OLAP environments. 
  • Expertise in Informatica Power Center and Data Quality (IDQ) tools. 
  • Excellent experience in trouble shooting SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Expertise in SQL Server Analysis Services (SSAS) and SQL Server Reporting  Services (SSRS) 
  • Trained end users, prepare related documentation (requirements, training materials, process and data flows, use cases, functional design, etc), and addressed critical questions from the user base involving the CRM
  • Good experience with use of Access queries, excel functions V-Lookup, formulas, Pivot Tables, etc. 
TECHNICAL SKILLS:

Operating Systems: Linux, HP UNIX, Sun Solaris, and Windows 2000/XP/Vista MS Office Tools MS Word, MS Excel, MS Access, MS PowerPoint, MS Visio 

Languages: SQL, PL/SQL, JSON, UNIX script, XML, Scala, R

Data warehousing Tools: Business Objects, Cognos, Microstrategy, Tableau, Informatica, Talend, Pentaho, Datastage

Hadoop Ecosystem: Hadoop Framework, HiveQL Queries, Pig, HBase, MongoDB, Sqoop and spark, scala

Modeling Tools: ERWIN r9.6/r 9.5/9.1, Embarcadero, ER/Studio, MSVisio, Sybase Power Designer

Databases: Netezza, Teradata R14, R13, R12, Oracle 12c/11g/10g, DB2, SQL, Oracle, MS Access, SQL Server 2000, MongoDB

AWS: AWS S3, AWS Redshift, AWS EMR.

Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Waterfall Model

ETL Tools: Informatica Power Center 9x/8x, Pentaho, SSIS.

Reporting Tools: Crystal reports XI, Business Intelligence, SSRS, Business Objects 5.x/ 6.x, Tableau.

WORK EXPERIENCE:

Confidential, Dallas TX

Sr. Data Architect

Responsibilities:

  • Provide  data architecture support to enterprise  data management efforts, such as the development of the enterprise  data model, Master and Reference  data.
  • Lead the strategy, architecture and process improvements for  data architecture and  data management, balancing long and short-term needs of the business.
  • Building relationships and trust with key stakeholders to support program delivery and adoption of enterprise architecture. 
  • Defined best practices for  data modeling and extraction and ensure architectural alignment of the designs and development. 
  • Implemented logical and physical relational database and maintained Database Objects in the  data model using Erwin 9.6 
  • Developed a high performance, scalable  data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs. 
  • Working in Hadoop Environment using pig, Sqoop, Hive, HBase, MongoDB, and detailed understanding of map reduce programs 
  • Participated in integration of MDM (Master  Data Management) Hub and  data warehouses. 
  • Responsible of designing the whole  data warehouse system including ODS, DWH and  data marts. 
  • Used ETL methodology for supporting  data extraction, transformations and loading processing, in a complex MDM using Informatica. 
  • Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch  data from Teradata database. 
  • Used Normalization methods up to 3NF and De-normalization techniques for effective performance in OLTP systems. 
  • Extensively using Agile methodology as the Organization Standard to implement the  data Models. 
  • Performed the  Data Mapping,  Data design ( Data Modeling) to integrate the  data across the multiple databases in to EDW. 
  • Developed and configured on Informatica MDM hub supports the Master  Data Management (MDM), Business Intelligence (BI) and  Data Warehousing platforms to meet business needs. 
  • Used Load utilities (Fast Load & Multi Load) with the mainframe interface to load the  data into Teradata and Data reconciliation activities between Source and EDW Teradata databases. 
  • Involved in designing and developing  Data Models and  Data Marts that support the Business Intelligence  Data Warehouse. 
  • Handled importing  data from various  data sources, performed transformations using Hive, Map Reduce, and loaded  data into HDFS. 
  • Involved in Planning, Defining and Designing  data base using Erwin on business requirement and provided documentation. 
  • Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections. 
  • Transforming staging area  data into a STAR schema (hosted on Amazon Redshift) which was then used for developing embedded Tableau dashboards.
  • Maintained metadata ( data definitions of table structures) and version controlling for the  data model. 
  • Created stored procedures, functions, database triggers and packages as per the business needs for developing ad-hoc and robust reports. 
  • Used Spark  Data frames, Spark-SQL, Spark MLLib extensively and developing and designing POC's using Scala, Spark SQL and MLlib libraries. 
  • Involved in integration of various relational and non-relational sources such as Oracle, XML and Flat Files. 
  • Developed SQL scripts for creating tables, Sequences, Triggers, views and materialized views. 
  • Scheduling the SSIS packages execution in SQL Server Agent and tracking the success or failure of the execution of SSIS Packages and configuring the email notifications through SQL Server Agent. 
  • Developed multiple MapReduce jobs in java for  Data Cleaning and pre-processing analyzing  data in PIG. 
  • Analyze existing source system with the help of  Data Profiling and source system  data models thus creating individual  data models for various domains/subject areas for the proposed  data warehouse solution. 
  • Worked in  Data Analysis,  data profiling and  data governance identifying  Data Sets, Source  Data, Source Meta  Data,  Data Definitions and  Data Formats. 
  • Developed  Data Mapping,  Data Governance, and Transformation and cleansing rules for the Master  Data Management Architecture involving OLTP, ODS. 
  • Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, and Oracle) 
  • Coordinating with DBA team to implement physical models & to setup development, test, staging & production environments for DDL & DML Generation & usage. 
  • Used Teradata OLAP functions like RANK, ROW NUMBER, QUALIFY, CSUM and SAMPLE. 

Environment: Erwin 9.6, MDM, AWS, Hadoop, Spark, Sqoop, Hive, HDFS, OLAP, OLTP, ODS, PL/SQL, Tableau, Teradata R15, Teradata SQL Assistant 15.0, Oracle 12c, SQL, PL/SQL, Flat Files, Informaitca, XML, SSIS, SSRS, AWS Redshift, MongoDB, HBase, AWS S3.

Confidential, Burlington NJ

Sr. Data Architect

Responsibilities:

  • Developed a high performance, scalable  data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs. 
  • Participated in the design, development, and support of the corporate operation  data store and enterprise  data warehouse database environment. 
  • Translating high level business requirements into solution, infrastructure architectures, involved enterprise architects to align strategic view. 
  • Designed the Logical  Data Model using ER/Studio with the entities and attributes for each subject areas and Worked on ER Studio for multiple Operations across in both OLAP and OLTP applications.
  • Used the Agile Scrum methodology to build the different phases of Software development life cycle. 
  • Involved in several facets of MDM implementations including  Data Profiling, Metadata acquisition and  data migration. 
  • Independently coded new programs and design Tables to load and test the program effectively for the given POC's using Big  Data/Hadoop. 
  • Involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Oracle, and DB2 database 
  • Developed Map Reduce programs to cleanse the  data in HDFS obtained from heterogeneous  data sources to make it suitable for ingestion into Hive schema for analysis. 
  • Responsible for full  data loads from production to AWS Redshift staging environment and Worked on migrating of EDW to AWS using EMR and various other technologies. 
  • Design of ODS layer, Dimensional modeling using Kimball Methodologies, of the  Data Warehouse sourced from MDM Base tables and other Transactional systems. 
  • Extensively used Aginity Netezza work bench to perform various DML, DDL etc operations on Netezza database. 
  • Designed and produced client reports using Excel, Access, Tableau and SAS. 
  • Created logical and physical  data model using Cassandra's model 
  • Developed  Data Mapping,  Data Governance, Transformation and Cleansing rules for the Master  Data Management Architecture involving OLTP, ODS and OLAP. 
  • Extracting Mega  Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports 
  • Worked with the ETL team to document the transformation rules for  data migration from OLTP to Warehouse environment for reporting purposes. 
  • Worked with Netezza and Oracle databases and implemented various logical and physical  data models for them. 
  • Developed complex SQL scripts for Teradata database for creating BI layer on DW for Tableau reporting. 
  • Worked on debugging and identifying the unexpected real-time issues in the production server SSIS packages. 
  • Involved in mapping the  data elements from the User Interface to the Database and help identify the gaps and Prepared of  Data Dictionary/Meta  Data of the  data elements. 
  • Developed and configured on Informatica MDM hub supports the Master  Data Management (MDM), Business Intelligence (BI) and  Data Warehousing platforms to meet business needs. 
  • Generated preliminary Source to Target mapping by writing a ER/Studio Macro in the ER/Studio VBA platform document and handed over to BAs for further transformation analysis. 
  • Submitted the DDL and Security document by writing an ERStudio Macro to the SVN repository and communicated to the DBA for implementation. 
  • Extracting Mega  Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports.
  • Developed long term  data warehouse roadmap and architectures, designs and builds the  data warehouse framework per the roadmap. 
  • Designed and Developed Oracle PL/SQL and Shell Scripts,  Data Import/Export,  Data Conversions and  Data Cleansing 
  • Extensively worked on documentation of  Data Model, Mapping Transformation and Scheduling jobs. 
  • Worked extensively with Business Objects XI Report Developers in solving critical issues of defining hierarchy, loops and Contexts. 

Environment: ER/Studio, Oracle 12c, Informatica, DB2, MDM AWS Redshift, AWS S3, Hadoop, Hive, Netezza, MongoDB, HBase, Sqoop, Spark, HDFS, MapReduce, OLAP, OLTP, PL/SQL, Flat Files, SQL, XML, Teradata, Tableau, SSRS, SSIS and Cassandra.

Confidential

Data Architect

Responsibilities:

  • Provided a consultative approach with business users, asking questions to understand the business need and deriving the  data flow, logical, and physical  data models based on those needs. 
  • Designed and developed architecture for  data services ecosystem spanning Relational, NoSQL, and Big  Data technologies.
  • Specifies overall  Data Architecture for all areas and domains of the enterprise, including  Data Acquisition, ODS, MDM,  Data Warehouse,  Data Provisioning, ETL, and BI. 
  • Created Physical & logical  data model from the conceptual model and it's conversion into the physical database with the DDL's using forward engineering options in Erwin. 
  • Developed  Data Mapping,  Data Governance, and Transformation and cleansing rules for the Master  Data Management Architecture.
  • Completed enhancement for MDM (Master  data management) and suggested the implementation for hybrid MDM (Master  Data Management).
  • Handled importing  data from various  data sources, performed transformations using Hive, MapReduce, and loaded  data into HDFS. 
  • Worked on designing, implementing and deploying into production an Enterprise  data warehouse 
  • Developed SQl process using SSIS with Various Control Flow,  Data Flow tasks and Store Procedures for Work Oracle Validation process .
  • Worked on importing and exporting  data from Oracle and Teradata into HDFS and HIVE using Sqoop.
  • Designed and Developed Oracle PL/SQL and Shell Scripts,  Data Import/Export,  Data Conversions and  Data Cleansing.
  • Used forward engineering to create a Physical  Data Model with DDL that best suits the requirements from the Logical  Data Model.
  • Involved in capturing  data lineage, table and column  data definitions, valid values and others necessary information in the  data model.
  • Responsible for  data profiling and  data quality checks to suffice the report requirements gathered above and provide an ETL Mapping.
  • Generated periodic reports based on the statistical analysis of the  data using SQL Server Reporting Services (SSRS).
  • Designed both 3NF  data models for ODS, OLTP systems and dimensional  data models using Star and Snow Flake Schemas.
  • Worked in using Teradata tools like Fast Load, Multi Load, Tpump, FastExport, Teradata Parallel Transporter (TPT) and BTEQ. 
  • Used ETL methodology for supporting  data extraction, transformations and loading processing, in a complex MDM using Informatica. 
  • Used SQL for Querying the database in UNIX environment 
  • Worked on the metadata management and part of  data governance team which created the  Data 
  • Creating or modifying the T-SQL queries as per the business requirements. 
  • Developed various QlikView  Data Models by extracting and using the  data from various sources files Excel, Flat Files and Big  data. 
  • Primarily responsible for Tableau customization for statistical dashboard to monitor sales effectiveness and also used Tableau for customer marketing  data visualization. 
  • Developed the performance tuning of the database by using EXPLAIN PLAN, TKPROF utilities and also debugging the SQL code. 

Environment: Erwin 9.1, MapReduce, ODS, MDM, OLAP, OLTP, Pig, Oracle 12, Hadoop, HDFS, Sqoop, Hive, NoSQL, Teradata R13, Netezza, PL/SQL, MS-Visio, T-SQL, SSIS, SSRS, SQL, Unix, Tableau, Informatica, PL/SQL.

Confidential

Sr. Data Modeler

Responsibilities:

  • Performed data analysis and profiling of source data to better understand the sources. 
  • Created logical data model from the conceptual model and it's conversion into the physical database design using Erwin. 
  • Worked with DBA's to create a best-fit Physical Data Model from the logical data model. 
  • Designed Sources to Targets mappings from SQL Server, Excel/Flat files, Xml Files to Teradata using Informatica Power Center Data cleansing, integrating, and matching using Informatica Data Quality IDQ 
  • Redefined many attributes and relationships in the reverse engineered model and cleansed unwanted tables/columns as part of data analysis responsibilities. 
  • Interacted with the database administrators and business analysts for data type and class words. 
  • Conducted design sessions with business analysts and ETL developers to come up with a design that satisfies the organization's requirements. 
  • Worked on enterprise logical data modeling project (in third normal form) to gather data requirements for OLTP enhancements. Converted third normal form ERDs into dimensional ERDs for data warehouse effort. 
  • Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement. 
  • Created ER Diagrams, Data Flow Diagrams, grouped and created the tables, validated the data, identified PK/ FK for lookup tables. 
  • Created 3NF business area data modeling with de-normalized physical implementation data and information requirements analysis using Erwin tool. 
  • Developed Star Schema and Snowflake Schema in designing the Logical Model into Dimensional Model. 
  • Ensured the quality, consistency, and accuracy of data in a timely, effective and reliable manner. 
  • Involved in ETL mapping documents in data warehouse projects. 
  • Created high level ETL design document. 
  • Identified and tracked slowly changing dimensions and determined the hierarchies in dimensions. 
  • Assisted the ETL team to document the transformation rules for data migration from OLTP to Warehouse environment for reporting purposes. 
  • Implement necessary DQ rules in IDQ Analyst while profiling the data. 
  • Involved in extensive data analysis on the Teradata and Oracle systems querying and writing in SQL and TOAD. 
  • Used SQL joins, aggregate functions, analytical functions, group by, order by clauses and interacted with DBA and developers for query optimization and tuning. 
  • Conducted several Physical Data Model training sessions with the ETL Developers. Worked with them on day-to-day basis to resolve any questions on Physical Model. . 

Environment: CA Erwin 9.1, Oracle11g, SQL server 2005, IBM DB2, Informatica Power Center, IDQ, SQL BI 2008, Oracle BI, Visual Studio, SSIS&SSRS, Tibco Spotfire, SQL server management studio 2012.

Confidential

Sr. Data Modeler

Responsibilities:

  • Gathered high level requirements and converted into business requirements. 
  • Created data masking mappings to mask the sensitive data between production and test environment.
  • Strong understanding of data quality assurance process and procedures. 
  • Created and maintained the data dictionaries. 
  • Created Entity Relationship Diagrams (ERD), Functional diagrams, Data flow diagrams and enforced referential integrity constraints. 
  • Created logical and physical models using Erwin. 
  • Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Expertise in the Forward Engineering of the logical models to generate the physical model using Erwin. 
  • Created tables, views, sequences, indexes, constraints and generated SQL scripts for implementing physical data model. 
  • Created Entity Relationships diagrams, data flow diagrams and implemented referential integrity using ERWIN. 
  • Star schema was developed for proposed central model and normalized star schema to snow flake schema. 
  • Worked in data mapping, enabling the extract, transform and loading of the data into target tables. 
  • Data mapping documents between Legacy, Production, and User Interface Systems were developed. 
  • Data sourcing and data staging are performed. 
  • Working along with ETL team for documentation of transformation rules for data migration from OLTP to warehouse for purpose of reporting. 
  • Metadata repository was maintained and updated based on change requests. 
  • Developed operational data store to design data marts and enterprise data warehouses. 
  • Walkthroughs with DBA were conducted to update the changes made to the data model. 
  • Developed statistics and visual analysis for warranty data using MS Excel, MS Access and Tableau Software.
  • Conducted meetings with business and development teams for data validation and end-to-end data mapping. 

Environment: ERWIN, Talend, SQL Server 2008, PL/SQL, Rational Rose, Windows XP, Oracle 10g, Windows XP, XML, Informatica, MS Excel, SSRS, SSIS, SAS.

Confidential

Data Analyst

Responsibilities:

  • Involved in Data mapping specifications to capture source to target mapping rules for ETL and QA teams
  • The data mapping specifies what data will be extracted from an internal data warehouse, transformed and sent to an external entity. 
  • Analyzed business requirements, system requirements, data mapping requirement specifications, and responsible for documenting functional requirements and supplementary requirements in Quality Center. 
  • Reviewed requirements together with QA Manager, ETL leads to enhancing the data warehouse for the originations systems and servicing systems. 
  • Designed an Industry standard data Model, Translated the business requirements into detailed production level using Workflow Diagrams, Sequence Diagrams, Activity Diagrams and Use Case Modeling.
  • Perform semantic data modeling for the ELDM, domain logical data models, and physical data model formats.
  • Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data warehouse, data mart reporting system in accordance with requirements.
  • Responsible for evaluating various RDBMS like OLTP modeling, documentation, and metadata reporting tools including Erwin, Developed logical/ physical data models using Erwin tool across the subject areas based on the specifications and established referential integrity of the system. 
  • Involved in project cycle plan for the data warehouse, source data analysis, data extraction process, transformation and ETL loading strategy designing. 
  • Created snapshots, views, and database indexes for improving the query performance. 
  • Responsible for analyzing various data sources such as flat files, ASCII Data, EBCDIC Data, Relational Data (Oracle, DB2 UDB, MS SQL Server) from various heterogeneous data sources. 
  • Created and maintained Logical Data Model (LDM) / Physical Data Modeling. Includes documentation of all entities, attributes, data relationships, primary and foreign key relationships, allowed values, codes, business rules, glossary terms, etc. 
  • Created dimensional model for the reporting system by identifying required dimensions and facts using Erwin.
  • Developed various transformations like Source Qualifier, Joiner transformations, Update Strategy, Lookup transformation, Stored procedure Transformations, Expressions and Sequence Generator for loading the data into target Data Mart. 
  • Worked with Comparison between Data Model Vs Database and generate difference Reports.

Environment: Erwin, Informatica, Greenplum, Windows XP/NT/2000, SQL Server 2005/2008, SQL, PL,SQL, Oacle, DBVisualizer 6.5, DB2, MS Excel, Pentaho, SQL, Rational Rose, Crystal Reports 9.

We'd love your feedback!