We provide IT Staff Augmentation Services!

Sr Data Modeler / Lead Data Architect Resume

5.00/5 (Submit Your Rating)

Atlanta, GA

SUMMARY

  • Having 13+ yrs. of Data warehousing, Big Data and Cloud experience using Hadoop, Salesforce CRM, AWS, Azure, GCP, Snowflake, Informatica Power Center 9.1/8.6, Informatica Cloud, SAP ERP, ETL, Business Intelligence, Data Profiling, Data Mining, Metadata Management, Dimensional Modeling, ER Modeling and Data Governance.
  • Well - versed with all stages of Software Development Life Cycle (SDLC) and Software Testing Life Cycle (STLC).
  • Played lead role in all stages of software development, including planning, implementation, testing, documentation, and release.
  • Experience in software development methodologies, test-driven development, data analysis, and reporting.
  • Strong understanding of Cloud, Big Data, Data warehouse and BI Analytics concepts.
  • Experience in Metadata design, real time BI Architecture including Data Governance for greater ROI.
  • Strong knowledge on Cloud, Big Data, Data warehouse and BI Analytics concepts.
  • Hands-on experience on implementing Cloud Solutions using various AWS Services including S3, Glacier, RDS, Athena, DynamoDB, Redshift etc.
  • Hands-on experience on creating datalake logical and physical data models on Azure Cloud.
  • Good Experience on Data migrations from On-premise to AWS/ Azure / GCP /Snowflake Cloud.
  • Worked on GCP Cloud platform and its features which includesGCP Storage, Bigquery, Dataprep, Dataflow, Cloud SQL etc.
  • Have very good experience on Snowflake Cloud Data Warehouse Migration from On-Prem DWH.
  • Created Conceptual, Logical Data model and Physical data models Using ERWIN and SAP Power designer tools.
  • Expert with versioning, change management, problem management, and troubleshooting.
  • Designed, set up and maintained Salesforce standard objects, custom objects and junction objects, while also structuring user roles, security profiles and workflow rules.
  • Experience in preparing Test Strategy, developing Test Plan, Detailed Test Cases, writing Test Scripts by decomposing Business Requirements, and developing Test Scenarios to support quality deliverables.

TECHNICAL SKILLS

Big Data/Cloud Technologies: Hortonworks, Cloudera, HDFS, Sqoop, Hive, HBase, Impala, Spark, Pig, Oozie, HUE, Salesforce, AWS, S3, Redshift, Athena, RDS, Dynamodb, Salesforce Sales Cloud, Snowflake Cloud, Informatica Cloud, GCP BigQuery, GCS, DataPrep, DataFlow, MSD 365

ETL Tool: Informatica Power Center 9.6.1, Informatica MDM, Informatica IDQ, Informatica Axon, Enterprise Information Catalog, Microsoft SSIS

Reporting Tool: SAP Business Objects 14.2.3, POWER BI, Tableau

Scheduling Tool: Control-M 8v, Autosys

Versioning Tools: GIT, SVN

RDBMS: Oracle 10g/11g/12C, Teradata 13v, SQL Server, Mainframe DB2, SAP HANA, Netezza

Programming Languages: SQL, Core Java, PL/SQL

Business/ ERP Applications: MS-Office, MS Visio, SAP ERP, Zuora

Scripting Languages: HTML, XML, Shell Scripting, Python

Modelling Tools: ERStudio, SAP PowerDesigner, ERWIN

Methods/ Frameworks: Change Management, Problem Management, Incident Management, TOGAF, Metadata Management, Data Governance, Agile, Waterfall, Dimensional Modeling, Data Vault, OSDU, DDMS, Industry Compliance policies

Testing Methodologies: Big Data / ETL Testing Methodologies

Operating Systems: Windows 9x/2000, UNIX, LINUX

PROFESSIONAL EXPERIENCE

Confidential

Environment: Oracle, SQL Server, Teradata, Mongodb, Informatica Power Center, Informatica MDM, Informatica IDQ, Informatica Axon, Informatica ICS, Tableau, SAP Hana, SAP Sales & Distribution, SAP Supply Chain, SAP FICO, Salesforce CRM,ERP, ZUORA, AWS S3, RDS, SNS, Redshift, Dynamodb, MSD 365, GCP Storage, Cloud SQL, BigQuery,DataPrep, DataFlow, Snowflake Cloud, Snowpipe, Snowflake Streams, Snow Tasks, REST API’S, Hortonworks, HDFS, HIVE, Impala, Hbase,Sqoop, Spark, Kafka, Oozie, UNIX, Tableau, Collibra, Agile Methodology, SOAP, SAP PowerDesigner, ERWIN

Sr Data Modeler / Lead Data Architect

Responsibilities:

  • Lead, plan and build quality Conceptual, Logical, Physical and Semantic Data models for OLTP, Operational Data Stores, Data Warehouses and Data Marts using tools like PowerDesigner.
  • Developed data modeling standards including version control, best practices & guidelines and ensure data models comply.
  • Lead, define and implement end-to-end modern data platforms in support of analytics.
  • Documented business data rules and translated them into usable database designs.
  • Document and review requirements, technical specifications, and customer expectations. Participated in design of Informatica MDM solutions, Informatica Data Quality and Data Integration based on customer needs.
  • Analyzing and documenting client's business requirements and processes; communicate these requirements to technical personnel by constructing basic conceptual data and process models, including data dictionaries and volume estimates.
  • Analyzed and Designed physical data models for TXT, CSV, XML, JSON files landed in HDFS storage layer.
  • Created semantic data models on Redshift for Lead to Cash Enterprise BI reporting.
  • Created Redshift Dimension tables with Dist Key and Sort keys for better performance.
  • Worked on Athena extensively to create external tables on S3 Storage objects and done data profiling on 3rd party vendor source systems.
  • Used GLUE Crawler for Metadata analysis.
  • Implemented IAM policies on AWS S3 Storage buckets to control unauthorized access.
  • Extracted data from Zuora Billing and Subscription system and created data pipelines to Snowflake Cloud environment.
  • Extracted data from Microsoft Dynamics 365 CRM application for Accounts, Leads, Quotes, Opportunities, Order and Subscription data to GCP Bigquery and Cloudsql databases.
  • Done Data Profiling on 3rd party source systems using Trifacta Dataprep tool on GCP.
  • Created Logical and Physical Data Mart design for Account, Product, Sales, Marketing, Subscription and Finance Data Domains.
  • Analyzed Accounts, Billing and Subscription data elements and done extensive data profiling to identify the data integration issues.
  • Done data extraction from Zuora application using REST API’s.
  • Extensively worked on Zuora Central Platform for product, order, billing and subscription data analysis for Order to Cash process.
  • Worked on Zuora CPQ has an integration with Salesforce for 360 synch for Product Catalog, Quotes, Accounts and Related objects.
  • Analyzed the data sets for Subscription life cycle quoting, Real-Time subscription metrics, One-click quote to cash.
  • Have good understanding on Zuora CPQ Object data model, on how the orders and related objects are mapped between Zuora and Zuora quotes.
  • Integrated Financial Accounting data elements like Chart of Account, Company Code, General Ledger, Account Receivable and Accounts Payable to the Data Lake / Data Warehouse environment.
  • Configured the accounting periods based on company financial calendar.
  • Assigned Accounting codes to transaction types, to summarize accounting transactions for accounting purposes.
  • Have good understanding on Invoice Items, Invoice payments, Tax Items, Credit Balance Payment, Payment Refund and Invoice adjustment financial processes.
  • Created physical data models for Big data applications using Hive and Hbase.
  • Created application specific nosql data models on Amazon Dynamodb database by determining all the access patterns.
  • Defined partition key and sort key on Dynamodb tables for better performance.
  • Defined Dynamodb tables with proper data types.
  • Created Physical Data models for Snowflake Cloud DW Tables using Dimensional Modeling approach.
  • Created Data Architecture diagram and ETL Process flow diagram for Snowflake Cloud DWH implementation.
  • Have very good understanding on Snowflake Architecture and Services.
  • Created database schemas and tables for both DWH Stage and DWH tables in Snowflake Cloud DWH.
  • Migrated SAP, Oracle, Sql Server, Teradata database applications to Snowflake cloud environment.
  • Created data pipelines from Salesforce CRM to Snowflake Cloud for Accounts, Leads, Quotes, Opportunities and Sales Order data sets.
  • Created Clustered tables on Snowflake for performance tuning and cost optimization.
  • Created Database clones for DEV and TEST purpose on Snowflake environment.
  • Implemented Role based security on Snowflake DWH.
  • Reverse Engineered the Physical data models using SAP Power Designer for any changes to the existing data models.
  • Published data models on SAP Power Designer web portal to access internal and external stakeholders.
  • Exported model objects/entities/attributes/references into an excel sheet.
  • From a high-level data governance perspective, drives data quality across systems by recommending and instituting data ownership with responsibilities, by monitoring\auditing master data across systems, implements data cleansing procedures, and recommends integration solutions and\or tools that minimizes data inconsistencies.
  • Written SQL scripts extensively to identify data quality and integration issues in existing data warehouse and Data Mart tables.
  • Created scorecards to monitor the data quality metrics and set up the email notifications to the Data stewards and Governance team if any changes in the threshold.
  • Map data sources to data models including required transformation logic, e.g. business rules, data cleansing rules, etc.
  • Conducts data cleaning to rid the system of old, unused data, or duplicate data for better management and quicker access.
  • Do extensive data profiling on source systems data to identify data anomalies.
  • Do extensive data analysis on finding data integration issues on different interrelated data domains.
  • Maintain a complete understanding of and adheres to all IT policies and processes
  • Develops and maintains comprehensive documentation on all standards and ensures that the information is appropriately published and distributed.

Confidential, Atlanta GA

Environment: Oracle, PL/SQL,SQL Server, Teradata, Informatica Power Center, Informatica MDM, Informatica IDQ, Informatica Axon, Informatica ICS, Tableau, SAP Hana, Salesforce Sales Cloud, SAP Financials, SAP Sales & Distribution, SAP Supply Chain, SAP Hana, Salesforce, Azure storage, Databricks, Azure SQL Server, Azure Synapse SQL, Snowflake, Snowpipe, Snowflake Streams, Snow Tasks, SAP PowerDesigner, Hortonworks, HDFS, HIVE, Hbase,Sqoop, Spark, Dremio, UNIX, Tableau, Collibra, Agile Methodology, SOAP

Data Architect / Sr Data Modeler /Data Governance

Responsibilities:

  • Lead, plan and build quality Conceptual, Logical, Physical and Semantic Data models for OLTP, Operational Data Stores, Data Warehouses and Data Marts using tools like PowerDesigner.
  • Developed data modeling standards including version control, best practices & guidelines and ensure data models comply.
  • Documented business data rules and translated them into usable database designs.
  • Document and review requirements, technical specifications, and customer expectations. Participated in design of Informatica MDM solutions, Informatica Data Quality and Data Integration based on customer needs.
  • Analyzing and documenting client's business requirements and processes; communicate these requirements to technical personnel by constructing basic conceptual data and process models, including data dictionaries and volume estimates.
  • Created physical data models for Big data applications using Hive and Hbase.
  • Created physical data models for Transactional systems using RDS SQL Server.
  • Created Physical Data models for Snowflake Cloud DW Tables using Data Vault approach.
  • Created Data Architecture diagram and ETL Process flow diagram for Snowflake Cloud DWH implementation.
  • Have very good understanding on Snowflake Architecture and Services.
  • Created Clustered tables on Snowflake for performance tuning and cost optimization.
  • Created Database clones for DEV and TEST purpose on Snowflake environment.
  • Implemented Role based security on Snowflake DWH.
  • Created database schemas and tables for both DWH Stage and DWH tables in Snowflake Cloud DWH.
  • Created physical data models for both structured and semi structured data.
  • Migrated SAP, Oracle, Sql Server database applications to Snowflake cloud environment.
  • Involved in Migrating Objects from Teradata to Snowflake.
  • Involved in testing Snowflake to understand best possible way to use the cloud resources.
  • Developed data warehouse model in snowflake for over 20 datasets.
  • Reverse Engineered the Physical data models using SAP Power Designer for any changes to the existing data models.
  • Published data models on SAP Power Designer web portal to access internal and external stakeholders.
  • Exported model objects/entities/attributes/references into an excel sheet.
  • Involved in Customer MDM and Product MDM data model designs and interacted with all the internal and external stakeholders to understand the data cleansing and extraction rules.
  • From a high-level data governance perspective, drives data quality across systems by recommending and instituting data ownership with responsibilities, by monitoring\auditing master data across systems, implements data cleansing procedures, and recommends integration solutions and\or tools that minimizes data inconsistencies.
  • Documented Informatica MDM Match rules for Customer and Product master data.
  • Written SQL scripts extensively to identify data quality and integration issues in existing data warehouse and Data Mart tables.
  • Created scorecards to monitor the data quality metrics and set up the email notifications to the Data stewards and Governance team if any changes in the threshold.
  • Created tableau reports and dashboards for data visualization and mining to identify the trends and patterns in the data.
  • Map data sources to data models including required transformation logic, e.g. business rules, data cleansing rules, etc.
  • Implemented data blending to blend related data from different data sources using Tableau Desktop.
  • Created various reports for sales & marketing Business team using calculations, pie chart, bar graph, Geographic maps etc.
  • Conducts data cleaning to rid the system of old, unused data, or duplicate data for better management and quicker access.
  • Do extensive data profiling on source systems data to identify data anomalies.
  • Do extensive data analysis on finding data integration issues on different interrelated data domains.
  • Maintain a complete understanding of and adheres to all IT policies and processes
  • Develops and maintains comprehensive documentation on all standards and ensures that the information is appropriately published and distributed.

Confidential, Atlanta GA

Environment: Hortonworks, HDFS, HIVE, HUE, YARN, PIG, SQOOP, HBase, Cassandra, Teradata, Mongodb,DB2, Oracle, PL/SQL, SAP Hana, SQL Server, Netezza, PL/SQL, UNIX, Informatica 9.6.1, Tableau, SAP Power Designer, Guidewire Policy Center, Guidewire Claims Center, Microsoft Visio, Agile Methodology

Data Architect / Data Modeler /Sr Business Data Analyst

Responsibilities:

  • Provide architect responsibilities including roadmaps, leadership, planning, technical innovation, security, IT governance, etc
  • Develop solutions for key business initiatives ensuring alignment with future state analytics vision.
  • Collaborate with clients to ensure information and approvals are accurate and documented in alignment with Firmwide Data Management Standards for new data requests.
  • Assist with the definition of how the data will be sourced, managed and consumed within and across business processes.
  • Evaluated on new technologies and executed proof-of-concepts on Amazon Cloud implementations.
  • Work alongside agile development teams to help improve completeness, accuracy and timeliness of datasets.
  • Worked with Database Administrators, Business Analysts and Content Developers to conduct design reviews and validate the developed models.
  • Define and govern data modeling and design standards, naming standards, tools, best practices, and related development methodologies for the organization.
  • Participate in Data Governance procedures and policy management.
  • Perform exploratory data analysis, generate and test working hypotheses, and uncover interesting trends and relationships
  • Analysed Mongodb source system data and document structures and created an application specific data models using normalized approach.
  • Analysed Mongodb Health Innovation Program data sets in Mongodb Json document formats.
  • Serve as a Subject Matter Expert for Big Data analytics, to partner and strategize with stakeholders in developing analytical routines to achieve their objectives. Clearly articulate and present design solutions, rationale issues and tradeoffs
  • Implement automation, audit and alert strategies throughout ETL processes.
  • Developed aConceptual modelusing Power Designer based on requirements analysis.
  • Developed normalizedLogicalandPhysicaldatabase models to design OLTP system for insurance applications.
  • Created dimensional model for the reporting system by identifying required dimensions and facts usingPower Designer.
  • Used forward engineering to create a Physical Data Model withDDLthat best suits the requirements from the Logical Data Model.
  • UsedPower Designerfor reverse engineering to connect to existing database and ODS to create graphical representation in the form of Entity Relationships and elicit more information.
  • Exhaustively collected business and technical metadata and maintained naming standards.
  • Extracted data from the databases (Oracle and SQL Server, DB2, FLAT FILES) using Big Data Spark/Sqoop/Informatica to load it into data lake/Datawarehouse.
  • Experienced with using different kind ofcompression techniquesto save data and optimize data transfer over network usingSnappy.

Confidential, Schaumburg IL

Environment: Cloudera, HDFS, HIVE, HUE, YARN, PIG, SQOOP, Map Reduce, HBase, SPARK,, DB2, Oracle, PL/SQL, SAP Financials, SAP Hana, SQL Server, SAP HANA, Netezza, UNIX, HP ALM, Autosys, Informatica 9.6.1, Power BI Desktop, Guidewire Policy Center, Guidewire Claims Center, Power BI Services, Salesforce Service Cloud, ERStudio, Microsoft Visio

ETL Data Architect / Data Modeler

Responsibilities:

  • Creating the Conceptual Data Model as per the Business process flow in Microsoft Visio.
  • Involved in the Analysis of the Business requirements and Design the Logical, Physical data Model using ERWIN.
  • Analyzed database requirements in detail with the project stakeholders by conducting Joint Requirements Development sessions.
  • Created data models for Reinsurance data domain.
  • Worked on Commercial lines LOB dimensional models design (Workers Compensation, Commercial General Liability)
  • Integrated crop insurance database objects into existing Confidential Atomic data warehouse.
  • Updated Policy, Claims and Product data models with new data/functional requirements to support BI and analytical usecases.
  • Creating Proof of Concepts from scratch illustrating how these data integration techniques can meet specific business requirements reducing cost and time to market.
  • Working with BA engineers to answer questions about data quality and integrity as well as identifying opportunities for presenting existing data in novel ways.
  • Used forward engineering to create a Physical Data Model withDDLthat best suits the requirements from the Logical Data Model.
  • Generated DDL (DataDefinition Language) scripts using Erwin 8 and supported the DBA in Physical Implementation ofdataModels.
  • Recognized Use Cases from the requirements. Created UML Diagrams including Use Cases Diagrams, Sequence Diagrams using MS Visio.
  • Designed and Deployed Big data solutions on Cloudera Platform.
  • Development of database objects like Hive External tables and HBase table DDL’s.
  • Performed the Data Mapping, Data design (Data Modeling) to integrate the data across the multiple databases in to EDW.
  • Understanding the data model, Data Extract, data profiling, de-duplication of data, data mapping, Source to Target mapping, look ups and transformations in excel, working with the development team to meet business requirements and logic for the Datawarehouse.
  • Created HBase tables to load large sets of structured, semi-structured and unstructured data coming from UNIX, NoSQL and a variety of portfolios.
  • Developed dimensional model forDataWarehouse/OLAP applications by identifying required facts and dimensions.
  • Worked Extensively with DBA and Reporting team for improving the Report Performance with the Use of appropriate indexes and Partitioning.

Confidential, Pheonix AZ

Environment: Hortonworks, HDFS, HIVE, PIG, HUE, HBASE, SQOOP, Flume, YARN, Oozie, Oracle, SQL Server, DB2, UNIX, HP ALM, Guidewire Policy Center, Guidewire Claims Center, ERWIN, JIRA, Microsoft Visio, Autosys, Informatica

Big Data Analyst

Responsibilities:

  • Have sound knowledge in analyzing the Policy, Member, Premium and Claims source data sets to identify the data integration issues between dependent source systems.
  • Have good insight into the Life Insurance domain to identify the gaps in requirements and data loss from source systems.
  • Review daily Database Change Requests for data model structure and clients standards.
  • Created Business Process Models and DFD's using ERWIN tool.
  • Have good technical knowledge to understand the Project Architecture and to identify the design issues.
  • Involved in software development experience with Big Data technologies such as Hadoop, HDFS, SQOOP, Hive, HBASE, Flume, Spark, Kafka, Pig and Oozie.
  • Prepared Test Plan/Approach, Test Scenario’s and Test Cases Design and Review.
  • Validated the data loaded in HDFS against the RDBMS tables using Hive and SQL Queries.
  • Written Hive/Sql queries for data validation to meet the business requirements
  • Written Test Cases for Data Correctness, Data Transformation, Metadata, Data Integrity, Data Quality, Data Security, Negative Scenario Tests.
  • Written HIVE queries and HDFS commands to validate the data between HDFS Files and HIVE External tables to validate the data loaded in hive external tables.
  • Have sound knowledge on Environment Setup and END to END Data loads in QA and UAT Env for data refresh.
  • Ensure the MapReduce jobs are running at peak performance.
  • Experienced in reviewing Hive Query log files.
  • Validated data in different file formats like tab delimited, fixed width and Json files.
  • Experienced in writing HDFS commands to validate the data loaded in HADOOP File system.
  • Have knowledge to verify available resource utilization in YARN server
  • Experienced in importing data from different RDBMS source systems into Hive tables using Hadoop echo system SQOOP.

Confidential, Dublin OH

Environment: Informatica 9.1, Abinitio, Teradata, Oracle, Control-M, Unix, Business Objects, MS Office, Erwin, Microsoft Visio, SharePoint, Jira

System Analyst / Data Modeler

Responsibilities:

  • Worked with Business, Architects and System Analysts to fully understand the business requirements and translate them into Conceptual, Logical and Physical models.
  • Prepared low-level design documents and S-T Mapping sheets to convert business requirements into functional specifications.
  • Participate in the discovery, documentation and presentation of rules, issues, metadata, and business/technical decision for the Datawarehouse.
  • Analyzed all errors generated by Informatica and fixed those errors in data collection.
  • Perform data profiling on elements, entities, and entity relationship, data reconciliation, Data Extract, identify elements for metadata standardization.
  • Extensively involved in analyzing variousdataformats using industry standard tools and effectively communicate them with business users and SME's.
  • DevelopedDataMigration and Cleansing rules for the Integration Architecture (OLTP, ODS, DW)
  • Involved in the Analysis of the Business requirements and Design the Logical, Physical data Model usingm ERWIN.
  • Worked with Business team to get the metadata approved for the newdataelements that are added for this project.
  • Used forward engineering to create a Physical Data Model withDDLthat best suits the requirements from the Logical Data Model.
  • Reverse engineered the reports and identified the Data Elements (in the source systems),
  • Performed Performance tuning of the SQL queries to increase the speediness in running query and generate the data accordingly by imposing proper Indexes and Creating Subqueries in the Datawarehouse.
  • Maintain Comprehensive documentation of business processes, analytics design, measure definitions, data integration and usability to assist with the development of enterprise level data concepts and metadata.

Confidential

Environment: Informatica 8.6, Oracle 10g, PL/SQL, Control-M, ERWIN, UNIX

ETL Developer

Responsibilities:

  • Discussions with the Business Analyst and Work Streams (Source Systems) about the Business Requirements and Pre-requisites.
  • Perform data analysis and data profiling to discover and understand data relationships in a system.
  • Created Physical Data Models, Entity Relationship Diagrams, DFD's and Mapping Documents for Policy and Claims data domains.
  • Involved in analyzing Schema's, Tables and Indexes as part of OPTIMIZATION.
  • Involved in the creation of Partitioned Tables and Indexes.
  • Define and implement best practices across database design and ETL.
  • Worked on Informatica - Source Analyzer, Warehouse Designer, Mapping Designer & Mapplet, and Transformation Developer.
  • Developed Informatica workflows/worklets/sessions associated with the mappings across various sources likeXML, COBOL, flat files, Webservices, Salesforce.
  • Worked with SCD Type1, Type 2 and Type 3 to maintain history in Dimension tables.
  • Created complex Informatica mappings using transformations Unconnected Lookup, joiner, Rank, Source Qualifier, Sorter, Aggregator, Lookup and Router transformations to extract, transform and loaded data mart area.
  • Createdpre-session, post session, pre-sql, post sql commandsin Informatica.
  • Extensively usedWorkflow Managerto createconnections,sessions, tasks and workflows
  • Attended production implementation calls and coordinated with the build engineers during migration of the code and was highly acknowledged.

We'd love your feedback!