Sr. Data Architect/data Modeler Resume
3.00/5 (Submit Your Rating)
Minneapolis, MN
SUMMARY:
- Over 9+ years of experience in Oracle/Data Architect /Data Modeler & RAC Oracle Database Administration with extensive data architecture, database development and administration experience in full SDLC, designing and building relational databases in OLTP, OLAP, decision support (DSS), client/server, data warehouse and multi - tier web based projects.
- Experience in Agile methodology/Scrum and waterfall models of DW complete life cycle projects
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
- Expertise in developing solutions around NOSQL databases like MongoDB and HBase.
- Extensive expertise in Data Warehousing on different database (s), as well as data modeling, both logical and physical data modeling tools like Erwin, Power Designer and ER Studio.
- Skills in scientific application design and development and scientific data processing related technologies: VBA, Pipeline Pilot, R, Python, or web service integrations.
- Profound experience on SSIS, SSRS, SSAS deployment's and maintenance of L1/L3 support with required service and proxies and granular level permissions of regional security levels.
- Strong Data modeling experience using ER diagram, Dimensional data modeling, Conceptual/ Logical/ Physical Modeling using 3NormalForm (3NF), Star Schema modeling, Snowflake modeling using tools like Erwin, ER-Studio, SAP Power Designer.
- Experienced in integration of various relational and non-relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, COBOL, XML and Flat Files, to Netezza database.
- Business Intelligence: Requirements analysis, Key Performance Indicators (KPI), metrics development, sourcing and gap analysis, OLAP concepts and methods, aggregates / materialized views and performance, rapid prototyping, tool selection, semantic layers Excellent experience in writing SQL queries to validate data movement between different layers in data warehouse environment.
TECHNICAL SKILLS:
- Big Data technologies: MapReduce, HBase, HDFS, Sqoop, Spark, Hadoop, Hive, PIG, Impala.
- Data Modeling Tools: ER/Studio 9.7/9.0, Erwin 9.6/9.5, Power Sybase Designer.
- OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9/7
- Programming Languages: SQL, PL/SQL, UNIX shell Scripting, PERL, AWK, SED, Scala, Python
- Databases: Oracle 12c/11g, Teradata R15/R14, MS SQL Server DB2.
- Testing and defect tracking Tools:: HP/Mercury (Quality Center, Win Runner, Quick Test Professional, Performance Center, Requisite, MS Visio & Visual Source Safe
- Operating System: Windows, UNIX, Sun Solaris
- ETL/Data warehouse Tools: Informatica 9.6/9.1, SAP Business Objects XIR3.1/XIR2, Talend, Tableau, Pentaho.
- Methodologies: Agile, RAD, JAD, RUP, UML, System Development Life Cycle (SDLC), Ralph Kimball and Bill Inmon's, Waterfall Model.
WORK EXPERIENCE:
Confidential - Minneapolis, MN
Sr. Data Architect/Data Modeler
Responsibilities:
- Helped the Client solve critical business needs by modeling and maintaining the business intelligence database environments and researched, evaluated, architect, and deployed new tools, frameworks and patterns to build sustainable Big Data platforms for our clients.
- Designed and architecting AWS Cloud solutions for data and analytical workloads such as warehouses, Big Data, data lakes, real-time streams and advanced analytics.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies and implemented Agile Methodology for building Integrated Data Warehouse, involved in multiple sprints for various tracks throughout the project lifecycle.
- Develop and Implement End to End Big Data Analytic and EDW/BI projects using multiple tools and platform which include.
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Worked on Normalization and De-normalization concepts and design methodologies like Ralph Kimball and Bill Inmon's Data Warehouse methodology.
- Developed prototype solutions to verify capabilities for new systems development, enhancement, and maintenance of MDM.
- Reviewed the Conceptual EDW (Enterprise Data Warehouse) Data Model with Business Users, App Dev. and Information Architects to make sure all the requirements are fully covered.
- Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration and Responsible for Metadata Management, keeping up to date centralized metadata repositories using Erwin modeling tools.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and SnowFlake Schemas and involved in Normalization/De-normalization techniques for optimum performance in relational and dimensional database environments.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop.
- Developed ETL processes extracted data daily and loaded data into a SSIS based Decision Support Warehouse.
- Driven the technical design of AWS solutions by working with customers to understand their needs and conducted numerous POCs (Proof of Concepts) to efficiently import large data sets into the database from AWS S3 Bucket.
- Worked on analyzing source systems and their connectivity, discovery, data profiling and data mapping and generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database
- Created data models for AWS Redshift and Hive from dimensional data models and worked on Data modeling, Advanced SQL with Columnar Databases using AWS and driven the technical design of AWS solutions by working with customers to understand their needs.
- Collected large amounts of log data using Apache Flume and aggregating using PIG in HDFS for further analysis.
- Formulated procedures for migration of source data with Salesforce applications.
- Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
- Interacted with End-users for gathering Business Requirements and Strategizing the Data Warehouse processes
- Write complex Netezza views to improve performance and push down the load to database rather than doing it in the ETL tool.
- Created data migration templates to migrate from multiple sourced into Salesforce.
- Involved in data model reviews with internal data architect, business analysts, and business users with explanation of the data model to make sure it is in-line with business requirements.
- Created DDL scripts using Erwin and source to Confidential mappings to bring the data from source to the warehouse.
- Developed enhancements to MongoDB architecture to improve performance and scalability and worked with MapReduce frameworks such as Hadoop and associated tools (pig, Sqoop, etc)
- Used Power pivot, power view to build dashboards using OLAP cube.
- Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
- Generated the framework model from IBM data Architect for the Cognos reporting team.
Environment: Erwin r9.6/7, Netezza, SQL Server 2016, Informatica 10.2, Taradata15, PowerBI, OLAP, OLTP, UNIX, MDM, Hadoop, Hive, Pig, Salesforce.com, HBase, HDFS, SAP, AWS, Redshift, EMR, S3, Apache Flume, Ralph Kimball and Bill Inmon's, PL/SQL, BTEQ, Python.
Sr. Data Architect/Data Modeler
Confidential Chicago IL
Responsibilities:
- Collaborate in identifying the current problems, constraints and root causes with data sets to identify the descriptive and predictive solution with support of the Hadoop HDFS, MapReduce, Pig, Hive, and Hbase and further to develop reports in Tableau.
- Massively involved in Data Architect role to review business requirement and compose source to Confidential data mapping documents and involved in relational and dimensional Data Modeling for creating Logical and Physical design of the database and ER diagrams using data modeling like Erwin.
- Created the data model for the Subject Area in the Enterprise Data Warehouse (EDW) and applied Data Governance rules (primary qualifier, class words and valid abbreviation in Table name and Column names)
- Worked on AWS Redshift and RDS for implementing models and data on RDS and Redshift.
- Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
- Worked on NoSQL databases including HBase, Mongo DB, and Cassandra. Implemented multidata center and multi-rack Cassandra cluster.
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture
- Worked closely with Product Owner and Solution Architects to find out the optimal solution for the Database design from and ETL and Data Governance perspective.
- Worked on Normalization and De-Normalization techniques for both OLTP and OLAP systems and worked with data investigation, discovery and mapping tools to scan every single data record from many sources.
- Performed Data mapping between source systems to Confidential systems, logical data modeling, created class diagrams and ER diagrams and used SQL queries to filter data.
- Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL access on Hadoop data
- Automation of SSIS Packages for production deployment with xml configurations and developed Historical/Incremental of SSIS Packages with SCD2 concept of Star Schema.
- Involved in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi-dimensional models such as Star Schemas and Snowflake Schema
- Involved in Teradata utilities (BTEQ, Fast Load, Fast Export, Multiload, and Tpump) in both Windows and Mainframe platforms.
- Designed both 3NF Data models for DB2 systems and dimensional Data models using Star and Snowflake Schemas.
- Administered and managed data management and reporting services related to salesforce.com.
- Extensively used Metadata & Data Dictionary Management, Data Profiling and Data Mapping and involved in Performance Tuning in SSIS packages by using Row, Block and Unblock Transformations.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snowflake Schemas.
- Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data warehouse, data mart reporting system in accordance with requirements.
- Conducted design walk through sessions with Business Intelligence team to ensure that reporting requirements are met for the business.
- Designed and developed Power BI graphical and visualization solutions with business requirement documents and plans for creating interactive dashboards.
- Performed Extracting, Transforming and Loading (ETL) data from Flat file, MS SQL Server by using SSIS services and involved in debugging and Tuning the PL/SQL code, tuning queries, optimization for the Sql database.
- Completed the data migration into Salesforce and guide clients to perform the data validation and tracks.
- Worked on Tableau for insight reporting and data visualization and extracted data from IBM Cognos to create automated visualization reports and dashboards on Tableau.
- Developed and implemented data cleansing, data security, data profiling and data monitoring processes and generated DDL and created the tables and views in the corresponding architectural layers.
- Created PL/SQL Tables, Collections, and Records, Partitions and Used Dynamic SQL and triggers for faster data access and to in corporate business logic and facilitated in developing testing procedures, test cases and User Acceptance Testing (UAT)
- Extensively worked with Developers and ETL team in enhancing the models and co-ordinate the same with DBA in implementing those changes into Applications or Databases.
Environment: ERWIN r9.5, Cognos, SQL Server2016, DB2, SSIS, OLAP, OLTP, LINUX, MDM, Hadoop, Hive, Pig, HBase, SAP, AWS, Redshift, PL/SQL, ETL, MondoDB, AWS S3, Informatica 9.6, Salesforce.com, Power BI, AWS EMR, SQL, Teradata, Netezza, Oracle and SSRS.