Sr. Data Architect/ Data Modeler Resume
Minnetonka, MN
SUMMARY:
- Above 9+ years of experience as Data Architect/Data Modeler and Data Analyst in Data Analysis, Data Modeling, Data Architecture, designing, developing, and implementing data models for enterprise - level applications and systems.
- Proficient in gathering business requirements and handling requirements management.
- Experience in Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data.
- Strong experience with different project methodologies including Agile Scrum Methodology and Waterfall methodology.
- Working knowledge of IBM's Banking and Financial Data Warehouse Industry Models, IBM's Insurance Information Warehouse Models
- Experienced in JIRA software for Plan, Track, and Report and Release management.
- Experience in writing SQL queries and optimizing the queries in Oracle, SQL Server, and Netezza, Teradata.
- Working Knowledge of RDBMS (MySQL, PostgreSQL, SQL Server) and NoSQL databases such as HBase, CouchBase, Vertica, MongoDB, Cassandra.
- Experienced in design, development, deployment, management of complex solutions on AWS for clients.
- Experience in Statistical or Econometrics hands-on work (can include Graduate School Research work) using SAS
- Experience with ETL/ELT & workflow scheduling with Airflow, AWS Data Pipelines, Azure, SSIS, etc.
- Strong virtualization skills and production experience with VMWare VSphere.
- Proficient with MySQL, Informix, DB2, OWB, ER/Studio, SQL Server, OBIEE, OID, OAS, and Portal.
- Experience in Adobe Analytics and Dashboard Creation Tableau/Qlik View will be highly regarded.
- Experience in SAP R/3 ABAP, BW, Business Objects, HANA, BO, BODS and SAP Industry solution IS-PS.
- Solid understanding of statistical analysis (e.g., hypothesis testing, different types of regressions).
- Working knowledge of data services offerings on other Cloud platforms, e.g. Amazon Web Services (AWS), Google, Pivotal Cloud Foundry, etc.
- Experience in designing, building and implementing complete Hadoop ecosystem comprising of Map Reduce, HDFS, Hive, Impala, Pig, Sqoop, Oozie, Cassandra, HBase, MongoDB, and Spark.
- Knowledge and experience in Big Data suite of tools such as Greenplum, Gemfire, Kafka, SpringCloud, Spark, and HAWQ
- Extensive knowledge of big data, Hadoop, Map-Reduce, Hive, NoSQL Databases and other emerging technologies.
- Extensive experience in Normalization (1NF, 2NF, 3NF and BCNF) and De-normalization techniques for improved database performance in OLTP, OLAP and Data Warehouse/Data Mart environments.
- Have a good experience using SAS and R.
- Experience in Batch processes, Import, Export, Backup, Database Monitoring tools and Application support.
- Experience in Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and Fast Export.
- Experience in using the databases like DB2, Teradata and its utilities, Netezza, Oracle, SQL Server Integration Services (SSIS)
- Experience in ER/Studio and Dimensional Models using Erwin advanced features, Conceptual, logical and physical data models using Erwin.
- Experience in development and support knowledge on Oracle, SQL, PL/SQL, T-SQL queries.
- Excellent Knowledge of Ralph Kimball and Bill Inmon's approaches to Data Warehousing.
- Extensive experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like ERWIN, ER/Studio.
- Working knowledge of SAS or other advanced statistical/econometric analysis software
- Very good knowledge and experience on AWS, Redshift, S3 and EMR.
- Expertise in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
- Experience with Adobe Analytics, Tealeaf, and SiteSpect.
- Experience in SQL and good knowledge in PL/SQL programming and developed Stored Procedures and Triggers and Data Stage, DB2, UNIX, Cognos, MDM, UNIX, Hadoop, Pig.
- Heavy use of Access queries, V-Lookup, formulas, Pivot Tables, etc. Working knowledge of CRM Automation Salesforce.com, SAP.
- Extensive experience on usage of ETL & Reporting tools like SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS)
- Experience in metadata design, real time BI Architecture including Data Governance for greater ROI.
- Implementation of Big data solutions on the AWS Cloud platform and Azure cloud platform
- Expertise in Data Migration, Data Profiling, Data Cleansing, transformation, integration, Data Import, data mapping and Data Export through the use of multiple ETL tools such as Informatica Power Centre.
- Extensively experience on Excel Pivot tables to run and analyze the result data set and perform UNIX scripting.
- Experience with Object Oriented Analysis and Design (OOAD) using UML, Rational Unified Process (RUP), Rational Rose and MS Visio.
TECHNICAL SKILLS:
Big Data Tools: Apache Hadoop 3.0, MapReduce, Sqoop 1.4, Pig, Hive 2.3, NoSql, Cassandra 3.11, MongoDB 3.6, Spark 2.2, HBase 1.2, Scala 2.12, Apache Storm, Elastic search, and R-Programming
Cloud Platform: Amazon AWS, EC2, EC3, Elastic Search, Elastic Load balancing, Redshift, AMI
Languages: SQL, PL/SQL, T-SQL, XML, HTML, UNIX, Shell Scripting and basic C & C++
OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9
Data Modeling Tools: ER/Studio 9.7/9.0, Erwin 9.7/9.5, Power Sybase Designer.
Database Tools: Microsoft SQL Server12.0, Teradata 15.0, Oracle 12c and MS Access Version Tool VSS, SVN, GIT
Project Execution Methodologies: Agile, Ralph Kimball and BillInmon s data warehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
Operating Systems: Microsoft Windows 10, Linux, and UNIX.
WORK EXPERIENCE:
Confidential, Minnetonka, MN
Sr. Data Architect/ Data Modeler
Responsibilities:
- Worked as a Data Architect / Modeler to generate Data Models using Erwin and developed relational database system.
- Worked on BODS for extracting data from SAP and non-SAP systems and loaded data into new SAP system using AIO methodology.
- Developed new queries and Alteryx flows for extracting, transforming and analyzing data to address project requirements.
- Used Python to place data into JSON files for testing Django Websites.
- Generate appropriate DDL for the target database, such as Informix Dynamic Server.
- Massively involved in Data Architect role to review business requirement and compose source to target data mapping documents.
- Used R statistical software for effective analysis by hypothesis testing to validate data and interpretations.
- Experience with Document - oriented NoSQL databases - MongoDB and CouchDB.
- Designed, implemented and maintained all AWS infrastructure and services within a managed service environment.
- Experience using Azure cloud, data lake storage, installation, security and infrastructure (IaaS) set up.
- Conducted ad hoc quantitative analyses, modeling, or programming using SAS, SQL, R, or Python
- Coordinated with the project management team in establishing business analytics systems that ensured timely and orderly completion of reports.
- Gathered, cleanses and analyzes data from multiple sources including but not limited to; Adobe Analytics, Tealeaf, Sitespect and Enterprise Databases
- Experience designing utilizing Data Federation and Data Virtualization technologies.
- Good experience in pivotal role in advancing the company’s aggressive growth and innovation strategy utilizing multifaceted data and our own custom-built platform to discover new opportunities in the parking industry.
- Implemented New Features Roadshow twice/year to educate Delphix Clients with new capabilities.
- Worked on managing and analyzing large data sets using tools such as MS Access, SQL Server or SAS preferred
- Developed and maintained an Enterprise Data Model (EDM) to serve as both the strategic and tactical planning vehicles to manage the enterprise data warehouse.
- Assigned tasks among development team monitored and tracked progress of project following Agile methodology.
- Worked with sales team to develop and implement account penetration strategies and drive the adoption of Pivotal solutions within strategic accounts.
- Worked with delivery of Data & Analytics applications involving structured and un-structured data on Hadoop based platforms on AWS EMR
- Designed SAP Data Services (BODS ) to load data from SAP BW to external systems
- Used Cassandra as NoSQL database.
- Used statistical techniques for hypothesis testing to validate data and interpretations
- Familiar with retargeting, Adobe Marketing Cloud, and SEO fundamentals
- Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from NoSQL and a variety of portfolios.
- Performed data migration from on premises environments into AWS.
- Worked with multiple Microsoft SME to define requirements for Azure POC including project plan and execution with (ADGILE/Waterfall) approaches.
- Designed and developed architecture for data services ecosystem spanning Relational, NoSQL, and Big Data technologies.
- Developed datasets with tools such as Alteryx and SQL, document requirements, and work closely with the build team to fully automate into the ASDW
- Implemented multi-data center and multi-rack Cassandra cluster.
- Generated a physical design for a particular data source from a logical design; IDA supports heterogeneous data environments across DB2, Informix, Oracle, Microsoft, Sybase, and mySQL.
- Provided expert technical pre-sales support for opportunities involving Pivotal Cloud Foundry.
- Experience in Information Integration using leading ETL and/or Data Virtualization tool
- Specified overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
- Gathered and analyzed existing physical data models for in scope applications and proposed the changes to the data models according to the requirements.
- Designed the Logical Data Model using Erwin 9.7 with the entities and attributes for each subject areas.
- Used Python scripts to update content in the database and manipulate files.
- Working directly with large volumes of data using SQL, Spark and other Hadoop technologies), Alteryx, R, Adobe Analytics, Google Big Query, Python
- Experience troubleshooting Cloudera production environments
- Worked on Amazon Redshift and AWS and architecting a solution to load data, create data models.
- Develop and Implement Hadoop, Spark, Big Data Analytics and Integrations, Microsoft Azure cloud data solutions, AWS Cloud Big Data solutions.
- Worked for map reduce and query optimization for Hadoop hive and HBase architecture
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for the Master Data Management Architecture.
- Extensively used Erwin r9.7 for Data modeling also created Staging and Target Models for the Enterprise Data Warehouse.
- Involved in Normalization/Denormalization techniques for optimum performance in relational and dimensional database environments.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using star and snow flake Schemas.
- Used IDA with Informix Dynamic Server
- Forward Engineering the Data Models, Reverse Engineering on the existing Data Models and updates the data models.
- Creation of BTEQ, Fast export, Multiload, Tpump, Fast load scripts for extracting data from various production systems.
- Involved in Data loading using PL\SQL Scripts and SQL Server Integration Services packages.
- Developed data Mart for the base data in Star Schema, Snow-Flake Schema involved in developing the data warehouse for the database.
- Created Erwin reports in HTML, RTF format depending upon the requirement, Published Data model in model mart.
- Developed SAS macros for data cleaning and Reporting and to support routing processing.
- Used Erwin model mart for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Designed ER diagrams (Physical and Logical using Erwin) and mapping the data into database objects and identified the Facts and Dimensions from the business requirements.
- Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
- Testing responsibilities included unit testing, integration testing and business acceptance testing.
- Applied appropriate level of abstraction in designs and confirmed that Data designs support the integration of data and information flow across systems and platforms.
Environment: Erwin 9.7, OLTP, Informix, Hadoop,Cloudera, Informatica, MDM, BTEQ, PL\SQL, SQL, Amazon Redshift, NoSQL, Teradata 15, CouchDB, HBase 1.2, MongoDB 3.6, Alteryx, Cassandra 3.11
Confidential, Merrimack, NH
Sr. Data Architect/ Data Modeler
Responsibilities:
- Responsible for the data architecture design delivery, data model development, review, approval and Data warehouse implementation.
- Developed technical expertise around Pivotal Cloud Foundry and the ecosystem
- Responsible for the data architecture design delivery, data model development, content creation, review, approval and used Agile Methodology for Data Warehouse development
- Working with Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data and conventional RDBMS.
- Worked extensively on Data conversion projects using BODS tool for SAP IDOC and LSMW loading method.
- Used the Django Framework to develop the application.
- Created and maintained SAS Datasets that are extracted from an Oracle Database.
- Led development team in designing/migrating AWS cloud based solutions.
- Collaborated with power users and IT management teams to streamline business reports and identify all end users.
- Utilized Azure cloud to correct storage and performance bottlenecks.
- Experience with infrastructure components such as LDAP, Single Sign On, SQL and NoSQL database technology, and virtualization technology used to manage IaaS and PaaS infrastructures.
- Developed normalized Logical and Physical database models to design OLTP system for insurance applications.
- Production experience with the Cloudera distribution of Hadoop
- Provided suggestion to implement multitasking for existing Hive Architecture in Hadoop. Also suggested UI customization in Hadoop
- Developed and automated multiple departmental Reports using Tableau and MS Excel.
- Responsible for all metadata relating to the EDW's overall data architecture, descriptions of data objects, access methods and security requirements.
- Involved in relational and dimensional Data Modeling for creating Logical and Physical design of the database and ER diagrams using data modeling like Erwin.
- Designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using ERWIN.
- Experience with related data management disciplines and technologies, such as master data management, data quality, data integration, metadata, data warehousing, and data virtualization.
- Generated Python Django Forms to record data of online users.
- Created and execute a strategy to build mindshare and broad use of AWS within a wide range of customers and partners.
- Designed both 3NF data models for ODS, OLTP systems and dimensional data models using Star and Snow Flake Schemas.
- Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.
- Designed both OLTP and ODS databases for high performance using ERWIN modeling tool.
- Worked on Normalization and De-Normalization techniques for both OLTP and OLAP systems.
- Managed the comprehensive assessment of Azure, Hadoop, DW and BI environments focusing on related data warehouse logical and physical design and implementation
- Worked on multiple projects with different business units including Insurance actuaries
- Involved in OLAP model based on Dimension and FACTS for efficient loads of data based on Star Schema structure on levels of reports using multi-dimensional models such as Star Schemas and Snowflake Schema
- Established uniform Master Data Dictionary and Mapping rules for metadata, data mapping and lineage.
- Developed MapReduce programs to parse the raw data, populate staging tables and store the refined data in partitioned tables in the EDW.
- Handled importing of data from various data sources, performed transformations using Hive, MapReduce, loaded data into HDFS and Extracted the data from Oracle into HDFS using Sqoop.
- Worked with delivery of Data & Analytics applications involving structured and un-structured data on Hadoop based platforms on AWS EMR
- Designed and implemented Oracle PL/SQL store procedures, functions and packages for data manipulation and validation.
- Involved in all the steps and scope of the project reference data approach to MDM and Created Data Dictionary and Data Mapping from Sources to the Target in MDM Data Model.
- Participated in the creation of Business Objects Universes using complex and advanced database features.
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS.
- Worked on building Aptitude Operational Data Store (ODS) model in an Oracle Ex-data database.
- Set up of environments to be used for testing and the range of functionalities to be tested as per technical specifications.
- Reviewed Complex ETL Mappings and Sessions based on business user requirements and business rules to load data from source flat files and RDBMS tables to target tables.
- Created Complex SQL Queries using Views, Indexes, Triggers, Roles, Stored procedures and User Defined Functions worked with different methods of logging in SSIS.
- Automation of SSIS Packages for production deployment with xml configurations.
- Developed Historical/Incremental of SSIS Packages with SCD2 concept of Star Schema.
Environment: Erwin 9.7, Oracle PL/SQL, SSIS, ODS, Hadoop,Cloudera,OLTP, Hadoop 3.2, HDFS, Oracle 12c, Sqoop 1.4, AWS, Agile, Sql, ETL & MDM.
Confidential - Juno Beach, FL
Sr. Data Modeler/Data Analyst
Responsibilities:
- Worked with Business users for requirements gathering, business analysis and project coordination.
- Coordinated with business unit teams in planning strategies to improve the performance of management processes and tools.
- Understood and translate business needs into data models supporting underwriting workstation services.
- Worked with the Application Development team to implement data strategies, build data flows and develop data models.
- Designed and developed Use Cases, Activity Diagrams, Sequence Diagrams, OOD (Object oriented Design) using UML and Visio.
- Involved in the entire data Migration process from analyzing the existing data, cleansing, validating, translating tables, converting and subsequent upload into new platform.
- Generated and DDL (Data Definition Language) scripts using Erwin and assisted DBA in Physical Implementation of data Models.
- Developed Conceptual, Logical and Physical data models for central model consolidation.
- Used Normalization (1NF, 2NF & 3NF) and de-normalization techniques for effective performance in OLTP and OLAP systems.
- Designed Star and Snowflake Data Models for Enterprise Data Warehouse using Erwin.
- Developed and maintained data Dictionary to create Metadata Reports for technical and business purpose.
- Worked on Performance Tuning of the database which includes indexes, optimizing SQL Statements.
- Conducted data modeling JAD sessions and communicated data-related standards.
- Developed SQL Queries to fetch complex data from different tables in remote databases using joins, database links and Bulk collects.
- Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
- Implemented Forward engineering to create tables, views and SQL scripts and mapping documents.
- Worked on PL/SQL programming Stored Procedures, Functions, Packages and Triggers.
- Wrote DDL and DML statements for creating, altering tables and converting characters into numeric values.
- Involved in development and implementation of SSIS, SSRS and SSAS application solutions for various business units across the organization.
- Worked with business analyst to design weekly reports using combination of Crystal Reports.
Environment: Erwin, OLTP, ETL, Sql, PL/SQL, DB2, SSIS, SSRS
Confidential - Houston, TX
Data Modeler/Data Analyst
Responsibilities:
- Gathered and translated business requirements into detailed, production-level technical specifications, new features, and enhancements to existing technical business functionality.
- Part of team conducting logical data analysis and data modeling JAD sessions, communicated data-related standards.
- Designed ER diagrams (Physical and Logical using Erwin) and mapping the data into database objects.
- Performed Reverse Engineering of the current application using Erwin, and developed Logical and Physical data models for Central Model consolidation.
- Translated logical data models into physical database models, generated DDLs for DBAs
- Performed Data Analysis and Data Profiling and worked on data transformations and data quality rules.
- Involved in extensive data validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
- Collected, analyze and interpret complex data for reporting and/or performance trend analysis
- Wrote and executed unit, system, integration and UAT scripts in a data warehouse projects.
- Extensively used ETL methodology for supporting data extraction, transformations and loading processing, in a complex DW using Informatica.
- Developed and maintain sales reporting using in MS Excel queries, SQL in Teradata, and MS Access.
- Involved in writing T-SQL working on SSIS, SSRS, SSAS, Data Cleansing, Data Scrubbing and Data Migration.
- Redefined many attributes and relationships in the reverse engineered model and cleansed unwanted tables/columns as part of Data Analysis responsibilities.
- Designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using Erwin.
- Written complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, flat files, with high volume data
- Written SQL scripts to test the mappings and Developed Traceability Matrix of Business
- Involved in extensive data validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
- Created SQL tables with referential integrity, constraints and developed queries using SQL, SQL*PLUS and PL/SQL.
- Performed GAP analysis of current state to desired state and document requirements to control the gaps identified.
- Developed the batch program in PL/SQL for the OLTP processing and used Unix Shell scripts to run in corn tab.
- Used Microsoft Excel tools like pivot tables, graphs, charts, solver to perform quantitative analysis.
- Wrote and executed unit, system, integration and UAT scripts in a data warehouse projects.
- Identified & record defects with required information for issue to be reproduced by development team.
- Worked on the reporting requirements and involved in generating the reports for the Data Model using crystal reports
Environment: Erwin 8.5, PL/SQL, Business Objects XIR2, Informatica 8.6, Oracle 10g, Teradata R13, Teradata SQL Assistant 12.0, PL/SQL, Flat Files
Confidential
Data Analyst
Responsibilities:
- Responsible for the analysis, design, development, coding, generation of reports using SQL, testing and documentation.
- Resolved the data type inconsistencies between the source systems and the target system using the mapping documents and analyzing the database using SQL queries.
- Created Database Maintenance Plans for the performance of SQL Server which covers Database Integrity checks, update database Statistics and Re-indexing.
- Used Teradata SQL Assistant to process ad hoc and business requests and to extract data from various data sources.
- Extracted data from production database (Oracle) and prepared financial reports using Oracle, SQL, PL/SQL, and Bteq.
- Developed and optimized stored procedures for use as a data window source for complex reporting purpose.
- Developed a Conceptual model using Erwin based on requirements analysis
- Created tables, views, SQL Scripts.
- Used Joins like Inner join; Outer joins while creating tables from multiple tables.
- Used MS-Excel, SQL and Unix for weekly and monthly.
- Responsible for developing and creating Tables, Views using DDL, DCL, and DML.
- Modified/developed Unix shell scripts to automate pre-session and post session- tasks and Bteq scripts.
- Used Excel Pivot Tables to represent data and presentation
- Monitored the existing code performance and to change the code for better performance.
- Used Inner Join and Outer join to retrieve data from multiple tables.
- Have used analytical skills and quantitative knowledge for problem solving.
- Wrote SQL scripts to run ad-hoc queries, PL/SQL scripts, Stored Procedures & Triggers and prepare reports to the management.
- Designed automated reports through MySQL and Excel to reduce manual work.
- Developed PL/SQL programs, stored procedures for data loading and data validations.
- Developed Oracle and Teradata queries to replace current data warehouse reports.
- Generated comprehensive analytical reports by running SQL queries against current databases to conduct data analysis.
- Created or modifying the T-SQL queries as per the business requirements.
- Designed and Developed Oracle database Tables, Views, Indexes with proper privileges and Maintained and updated the database by deleting and removing old data
Environment: Teradata, Oracle 10g, SQL, PL/SQL, Erwin, MS-Excel, MySQL, Teradata, T-SQL