We provide IT Staff Augmentation Services!

Sr. Data Architect/data Modeler Resume

4.00/5 (Submit Your Rating)

Long Island, NY

SUMMARY:

  • Over 7 years of Industry experienced in IT with solid understanding of Data Architecture, Data Modeling, Data Analysis, Evaluating Data Sources and strong understanding of Data Warehouse/Data Mart Design, BI, OLAP, OLTP, Client/Server applications.
  • Experience as Architect UML models and leverage the advanced executable code generators to target different domains.
  • Strong Knowledge of Data Warehouse Architecture and Star Schema, Snow flake Schema, FACT and Dimensional Tables.
  • Strong Experience in Big Data HadoopEcosystem in ingestion, storage, querying, processing and analysis of big data.
  • Experience in analyzing data using Hadoop Ecosystem including Map Reduce, HDFS, Hive, Spark, Spark Streaming, Elastic Search, Kibana, Kafka, HBase, Zookeeper, PIG, Sqoop and Flume.
  • Experienced in designing Star Schema, Snowflake schema for Data Warehouse, by using tools like Erwin data modeler, Power Designer and Embarcadero E - R Studio.
  • Experience as a solution architect for business information systems, focusing on Data Architecture, Data Stores and Data Mart/Data Warehouse concepts.
  • Experienced in big data analysis and developing data models using Hive, PIG , and Map reduce , SQL with strong data architecting skills designing data-centric solutions.
  • Very Good experienced in Amazon platform (AWS ), such as AWS S3, AWS EMR, AWS Redshift and AWS EC2 instances.
  • Experienced in Data modeling for Data Mart/Data Warehouse development including conceptual, logical and physical model design, developing Entity Relationship Diagram (ERD), reverse/forward engineer (ERD) with CA ERwin data modeler.
  • Extensive experience in shell scripting, Python , Perl or any other scripting language
  • Experienced in Netezza tools and Utilities NzLoad, NzSql, NzPL/SQL, Sqltoolkits , Analytical functions etc.
  • Extensive experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like Erwin, ER Studio .
  • Experience in different Azure services like Storage, Data Factory, Data Lake, App Services, Security Center, Multi Factor Authentication, Certification management, Networking components etc..
  • Excellent Knowledge of Ralph Kimball and Bill Inmon's approaches to Data Warehousing and experienced in development and support knowledge on Oracle, SQL, PL/SQL,T-SQL queries.
  • Experienced in Logical Data Model (LDM) and Physical Data Models (PDM) using Erwin data modeling tool.
  • Experienced in migration of Data from Excel, Flat file, Oracle to MS SQL Server by using SQL Server SSIS.
  • Experienced in process improvement, Normalization/De-normalization, data extraction, data cleansing, and data manipulation.
  • Experienced in ETL design, development and maintenance using Oracle SQL, PL/SQL, TOAD SQL Loader, and Database Management System (RDBMS).
  • Experienced in Designed and developed Data models for Database ( OLTP ), the Operational Data Store ( ODS ), Data warehouse ( OLAP ), and federated databases to support client enterprise Information Management Strategy.
  • Working experience with Kimball Methodology and Data Vault Modeling
  • Experienced in data from various sources like Oracle Database, Flat Files, and CSV files and loaded to target warehouse.
  • Experienced in Transform, and Load data from heterogeneous data sources to SQL Server using SQL Server Integration Services (SSIS) Packages.
  • Good knowledge Developing Informatica Mappings, Mapplets, Sessions, Workflows and Work lets for data loads from various sources such as Oracle, Flat Files, DB2, SQL Server etc.
  • Excellent understanding and working experience of industry standard methodologies like System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE Methodologies.
  • Experienced in Business Intelligence (SSIS, SSRS), Data Warehousing and Dashboards.
  • Expertise in source to target mapping in Enterprise and Corporate Data Warehouse environments.
  • Experienced in MDM (Master Data management) in removing duplicates, standardizing data, and to eliminate incorrect data.

TECHNICAL SKILLS:

Analysis and Modeling Tools: : Erwin r9.6/r9.5/r9.1/r8.x, Sybase Power Designer, Oracle Designer, BP win ER/Studio, .1, MS Access 2000, Star-Schema, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables.

OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9.

Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Spark, MongDB, Apache Cassandra.

Cloud: Azure Data Factory, Azure Data Storage, AWS, Redshift, S3, EC2, EMR

Oracle: Oracle12c/11g/10g/9i/8.x, R2 database servers with RAC, ASM, Data Guard, Grid Control and Oracle Golden Gate(Oracle Enterprise Manager),Oracle Data Guard, SQL, Net, SQL Loader and SQL PLUS, AWR,ASH, ADDM, Explain Plan.

ETL Tools: SSIS, Pentaho, Informatica Power Center 9.7/9.6/9.5/9.1 , Data stage.

Programming Languages: Java, Base SAS, SSIS and SAS/SQL, SQL, T-SQL, HTML/ XHTML/ HTML4.0.1/ HTML3.2, Java Script, CSS3/CSS2/CSS1, UNIX shells scripting, PL/SQL.

Database Tools: Microsoft SQL Server 2014/2012/2008/2005 , Teradata, and MS Access, Poster SQL, Netezza, SQL Server, Oracle.

Web technologies: Python, HTML, XHTML, DHTML, XML, JavaScript

Reporting Tools: Business Objects, SSRS, Tableau, Crystal Reports

Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX Windows 95, 98, Windows NT, Windows XP, 7.

Tools: & Software: TOAD, MS Office, BTEQ, Teradata 15/14.1/14/13.1/13 , SQL Assistant

Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, C++, UNIX, PL/SQL etc.

PROFESSIONAL EXPERIENCE:

Sr. Data Architect/Data modeler

Confidential, Long Island, NY

  • Massively involved in Data Architect role to review business requirement and compose source to target data mapping documents.
  • Owned and managed all changes to the data models. Created data models, solution designs and data architecture documentation for complex information systems.
  • Developed long term data warehouse roadmap and architectures, designs and builds the data warehouse framework per the roadmap.
  • Specifies overall Data Architecture for all areas and domains of the enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
  • Responsible for leading the architectural development of project/metadata and reporting framework.
  • Worked on designing Canonical, Conceptual, Logical and Physical data modeling using Erwin r9.6 , Power Designer in both OLAP and OLTP applications.
  • Defined facts, dimensions and designed the data marts using the Ralph Kimball's Dimensional Data Mart modeling methodology using Erwin.
  • Responsible for data profiling, data analysis, data specification, data flow mappings, and business logic documentation associated with new or modified product data capture requirements.
  • Create conceptual, logical and physical models for OLTP , Data Warehouse, Data Vault and Data Mart Star/Snowflake schema implementations
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS
  • Responsible for full data loads from production to AWS Redshift staging environment and Worked on migrating of EDW to AWS using EMR and various other technologies.
  • Used Data Stage to Extract, Transform and Load data into Netezza Data Warehouse from various sources like Oracle12c and flat files.
  • Creating Functional design document, developed detailed mapping document to provide reporting and Data Warehouse team with Source to Target mapping includes logical name, Physical name, data types, corporate Metadata definitions and translation rules.
  • Extensively worked across various DBMS like SQL Server 2014 , Oracle, Teradata 15 and IBM DB2 and worked on Hadoop ecosystem, Hive queries, MongoDB , Cassandra , Pig , and Apache Strom .
  • Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
  • Used Informatica Power Center for extraction, loading and transformation (ETL) of data in the data warehouse.
  • Worked on data pre-processing and cleaning the data to perform feature engineering and performed data imputation techniques for the missing values in the dataset using Python.
  • Enabled speedy reviews and first mover advantages by using Oozie to automate data loading into the Hadoop Distributed File System and PIG to pre-process the data.
  • Worked with cloud based technology like Redshift, S3, AWS, EC2 Machine , etc. and extracting the data from the Oracle financials and the Redshift database and worked with various RDBMS like Oracle 12c, SQL Server, DB2 UDB , and Teradata 14.1, Netezza.
  • Used SQL tools like Teradata SQL Assistant and TOAD to run SQL queries and validate the data in warehouse.
  • Used Data Vault technique and achieved many advantages of Data Vault approach some of them are simplified the data ingestion process, removed the cleansing requirement of a star schema and easily allowed for the addition of new data sources without disruption to existing schema
  • Extensively worked in Oracle environment using Oracle, PL/SQL, SQL developer, SQL Plus, SQL loader, Netezza, UNIX, Toad and Oracle Forms/Reports.
  • Created Source to Target mappings and Transformations. Mapped data between Source and Targets and perform data profiling and data analysis to enable identify data gaps and familiarize with new source system data
  • Worked with DB2 group to create Best-Fit Physical Data Model from the Logical Data Model using Forward engineering using Erwin and involved in Normalization and De-Normalization of existing tables for faster query retrieval.
  • Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS
  • Developed ETL's for Data Extraction, Data Mapping and data Conversion using SQL, PL/SQL and various shell scripts in Data stage.
  • Developed Data Mapping, Data Governance, and transformation and cleansing rules for the Master Data Management Architecture involving OLTP, ODS .
  • Created Dashboards on Tableau from different sources using data blending from Oracle, SQL Server, MS Access and CSV at single instance.
  • Designed Physical Data Model (PDM) using IBM Info sphere Data Architect data modeling tool and Oracle, PL/SQL and Data Vault used in both a data loading technique and methodology which accommodates historical data, auditing, and tracking of data.
  • Extensively using Agile methodology as the Organization Standard to implement the data Models.
  • Processed the data using HQL (like SQL) on top of Map-reduce.
  • Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS) .

Environment: ERWIN r9.6, Netezza Aginity, Oracle12c, Taradata15, T-SQL, SQL Server 2014, DB2, SSIS, SSRS, R, SAS, HTML, EMR, EHR, Agile, Python, UNIX, Tableau, MySQL, Hadoop, Hive, Pig, Map Reduce, Spark, Mongodb, MDM, PL/SQL, ETL, facet etc.

Sr. Data Analyst/Data modeler

Confidential, Austin, TX

  • Delivered complex and ambiguous business requirements while elevating data architecture to the next level and developed strategies for data acquisitions, archive recovery, and implementation of databases.
  • Responsible for developing and supporting a data model and architecture that supports and enables the overall strategy of expanded data deliverables, services, process optimization and advanced business intelligence.
  • Drove requirement sessions with the client to the degree ER/Studio analytical logical and physical star schema data models could be designed from three source systems and developed Logical and Physical Data models by using ER Studio and its converted into physical database design.
  • Migration of on premise MongoDB to CosmosDB using Azure Data Factory .
  • Worked on the development of Data Warehouse, Business Intelligence architecture that involves data integration and the conversion of data from multiple sources and platforms.
  • Developed automated data pipelines from various external data sources (web pages, API etc) to internal data warehouse (SQL server) then export to reporting tools by Python .
  • Analyzed the Data from different sourcing using Big Data Solution Hadoop by implementing Azure Data Factory, Azure Data Lake, PIG, Hive, Sqoop.
  • Designed and developed SSIS Packages to import and export data from MS Excel, SQL Server 2012 and Flat files.
  • Strong knowledge in Data Warehousing Concepts like Ralph Kimball Methodology, Bill Inmon Methodology, OLAP, OLTP, Star Schema , Snow Flake Schema , Fact Table and Dimension Table.
  • Analyzed the web log data using the HiveQL to extract number of unique visitors per day, page views, visit duration, most purchased product on website.
  • Developed Azure Data Factory Pipelines / SSIS packages to automate cleansing implementation and importing the cleansed data into Azure SQLServer Master Data Services.
  • Used Data Governance of RAW, Staging, Curated and Presentation Layers in Azure Data Lake Store
  • Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings.
  • Build and maintain scalable data pipelines using the Hadoop ecosystem and other open source components like Hive and HBase .
  • Involved in Automating and Scheduling the Teradata SQL Scripts in UNIX using Korn Shell scripting and developed Initio graphs to fetch data from Oracle, Teradata, Flat Files and mainframe Files.
  • Deployed and scheduled Reports using SSRS to generate all daily, weekly, monthly and quarterly Reports including current status.
  • Involved in Normalization and de-normalization OLAP and OLTP systems, process including relational database, table, constraints (Primary key, foreign key, Unique and check) and Indexes.
  • Involved in Teradata utilities (BTEQ, Fast Load, Fast Export, Multiload, and Tpump) in both Windows and Mainframe platforms.
  • Work on SQL Data warehouse using Azure for designing services to handle computational and data intensive queries in database.
  • Worked with ETL Developers in creating External Batches to execute mappings, Mapplets using Informatica workflow designer to integrate Shire's data from varied sources like Oracle, DB2, flat files and SQL databases and loaded into landing tables of Informatica MDM Hub.
  • Performed data profiling of data vault hubs, links and satellites using Erwin generated SQL scripts and designed Physical Data Model (PDM) using ER Studio data modeling tool and PL/SQL and T-SQL Managed Meta-data for data models.
  • Extensively Used Sqoop to import/export data between RDBMS and hive tables, incremental imports and created Sqoop jobs for last saved value
  • Worked with the UNIX team and installed TIDAL job scheduler on QA and Production Netezza environment.
  • Prepared Tableau reports and dashboards with calculated fields, parameters, sets, groups or bins and publish on the server.

Environment: ER Studio, Metadata, Netezza, Oracle12c, Taradata14.1, T-SQL, SQL Server 2012, DB2, SSIS, R, Python, Azure, Hadoop, Spark, Map Reduce, UNIX, HTML, SQL, Informcatica, Tableau, Aginity, MySQL, Hive, Pig, MDM, PL/SQL, SPSS, ETL, Data stage etc.

Data modeler/Data Analyst

Confidential, Tampa, FL

  • Worked on designing Conceptual, Logical and Physical data models and performed data design reviews with the Project team members Erwin r9.1 .
  • Worked in complex SQLs needed for ETL and analyzing data, and is proficient and worked with databases like Oracle 9i,SQL Server2008 , Teradata 13.1, DB2, MS SQL,Excel sheets, Flat Files, Sybase , and Netezza .
  • Extensively used Erwin for developing data model using star schema and Snowflake Schema methodologies.
  • Involved in data analysis and creating datamapping documents to capture source to target transformation rules.
  • Used Netezza SQL to maintain ETL frameworks and methodologies in use for the company and also accessed Netezza environment for implementation of ETL solutions.
  • Involved in Logical modeling using the Dimensional Modeling techniques such as Star Schema and Snow Flake Schema .
  • Worked on the Netezza Admin Console when the issues were not solved at the session/workflow level.
  • Worked in importing and cleansing of data from various sources like Teradata, Oracle, Netezza flat files, SQL Server with high volume data.
  • Created SSIS Reusable Packages to extract data from Multi formatted Flat files, Excel, XML files into Database and Billing Systems
  • Involved in PL/SQL programming and developed Stored Procedures and Triggers and Data Stage, DB2, UNIX, Cognos, MDM, UNIX, Hadoop, Pig .
  • Extensively worked on Teradata tools & utilities like Fast Load, Multi Load, T Pump, Fast Export, Teradata Parallel Transporter (TPT) and BTEQ.
  • Involved in Netezza Administration Activities like backup/restore, performance tuning, and Security configuration.
  • Utilized SDLC and Agile methodologies such as SCRUM .
  • Generated periodic reports based on the statistical analysis of the data using SQL Server Reporting Services (SSRS) .
  • Worked in data from Flat files, SQL Server, Oracle, DB2 , and Sybase and load the data into flat files, Oracle and SQLServer using Data stage.
  • Created various transformation procedures by using SAS ETL and SAS Enterprise guide.

Environment: Netezza, Agile, Oracle10g, Taradata13.1, R, SAS, T-SQL, SQL Server, DB2, SSIS, ERWIN, Aginity, SSRS, MDM, PL/SQL, ETL, Data stage etc.

Data Analyst

Confidential

  • Designed different type of STAR schemas and Snowflake Schema using ERWIN 8.x with various Dimensions like time, services, customers and FACT tables.
  • Involved in data from different databases like Oracle8.x, Neteza, Abnitio and external source systems like flat files using ETL tool.
  • Designed Star schema (identification of facts, measures and dimensions) and Snowflake schema .
  • Analysis of functional and non-functional categorized data elements for data profiling and mapping from source to target data environment.
  • Provided technical guidance for re-engineering functions of Teradata13 warehouse operations into Netezza .
  • Designed and Developed Oracle PL/SQL and Shell Scripts , Data Import/Export , Data Conversions and Data Cleansing .
  • Involved in data from different sources like Excel, CRM, SAP, Netezza , and others integrate these and built Dashboard.
  • Involved in Integration of various data sources like DB2, Oracle, Netezza, SQL Server, SAP BW, XML Files .
  • Assisted in Batch processes using Fast Load, BTEQ, MDM, UNIX Shell and LINUX, Teradata SQL to transfer cleanup and summarize data.
  • Worked with requirements management, workflow analysis, source data analysis, data mapping, Metadata management, data quality, testing strategy and maintenance of the model.
  • Involved in data from various sources like Oracle Database, XML, Flat Files, CSV files and loaded to target warehouse.
  • Designed the ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata as per business requirements using Erwin .
  • Extensively worked in SQL, PL/SQL, SQL Plus, SQL Loader, Query performance tuning, DDL scripts, database objects like Tables, Views Indexes, Synonyms and Sequences.

Environment: ERWIN8.x, Data Modeler, Netezza, Oracle8.x, SQL, PL/SQL, Taradata13, T-SQL, SQL Server, DB2, SSIS, SSRS, LINUX, MDM, PL/SQL, ETL, etc.

We'd love your feedback!