Senior Data Architect/modeler Resume
Newark, NJ
SUMMARY
- 9+ years of experience as Data Architect/Modeler and Data Analyst wif high proficiency in requirement gathering and data modeling including design and support of various applications in OLTP, Data Warehousing, OLAP and ETL Environment.
- Skillful in Data Analysis using SQL on Oracle, MS SQL Server, DB2 & Teradata.
- Proficient in System Analysis, ER/Dimensional Data Modeling, Database design and implementing RDBMS specific features.
- Practical understanding of teh Data modeling (Dimensional & Relational) concepts like Star - Schema Modeling, Snowflake Schema Modeling, Fact and Dimension tables.
- Experienced in Technical consulting and end-to-end delivery wif architecture, data modeling, data governance and design - development - implementation of solutions.
- Heavy use of Access queries, V-Lookup, formulas, Pivot Tables, etc. Working knowledge of CRM Automation Salesforce.com, SAP.
- Extensive experience on usage of ETL& Reporting tools like SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS)
- Experienced in integration of various relational and non-relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, COBOL, XML and Flat Files, to Netezza database.
- Responsible for detail architectural design and data wrangling, data profiling to ensure data quality of vendor data, Source to target mapping
- Experience in SQL and good knowledge in PL/SQL programming and developed Stored Procedures and Triggers and Data Stage, DB2, UNIX, Cognos, MDM, UNIX, Hadoop, Pig.
- Logical and physical database designing like Tables, Constraints, Index, etc. using Erwin, ER Studio, TOAD Modeler and SQL Modeler.
- Experience in Big Data, NoSQL Database like Cassandra and technical expertise in Hadoop.
- Data Warehousing: Full life-cycle project leadership, business-driven requirements, capacity planning, gathering, feasibility analysis, enterprise and solution architecture, design, construction, data quality, profiling and cleansing, source-target mapping, gap analysis, data integration/ETL, SOA, ODA, data marts, Inman/Kimball methodology, Data Modeling for OLTP, canonical modeling, Dimension Modeling for data ware house star/snowflake design.
- Good understanding of AWS, big data concepts and Hadoop ecosystem.
- Extensive ETL testing experience using Informatica 9x/8x, Talend, Pentaho.
- Work on Background process in oracle Architecture. Also drill down to teh lowest levels of systems design and construction.
- Experience in designing Architecture for Modeling a Data warehouse by using tools like Erwin r9.6/r9.5, Sybase Power Designer and E-R Studio.
- Experience in BI/DW solution (ETL,OLAP, Data mart), Informatica, BI Reporting tool like Tableau and Qlikview and also experienced leading teh team of application, ETL, BI developers, Testing team
- Experience in working wifMicro StrategySecurity model including Users, Groups, Security Roles and Security filters.
- Experience in Big Data Hadoop Ecosystem in ingestion, storage, querying, processing and analysis of big data
- Experience in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
- Good experience inSFDCrelated technologies such as Apex, Visualforce, Apex triggers
- Expertise on Relational Data modeling (3NF) and Dimensional data modeling.
- Worked on Informatica Power Center tools-Designer, Repository Manager, Workflow Manager.
- Business Intelligence: Requirements analysis, Key Performance Indicators (KPI), metrics development, sourcing and gap analysis, OLAP concepts and methods, aggregates / materialized views and performance, rapid prototyping, tool selection, semantic layers Excellent experience in writing SQL queries to validatedatamovement between different layers in data warehouse environment.
TECHNICAL SKILLS
Database Tools: Microsoft SQL Server12.0, Teradata 15.0, Oracle 11g/9i/12c and MS Access
Version Tool: VSS, SVN, GIT
BI Tools: Tableau 7.0/8.2, Tableau server 8.2, Tableau Reader 8.1,SAP Business Objects, Crystal Reports Packages Microsoft Office 2010, Microsoft Project 2010, SAP and Microsoft Visio, Share point Portal Server
Tools: OBIE 10g/11g/12c, SAP ECC6 EHP5, Go to meeting, Docusign, Insidesales.com, Share point, Mat-lab
ETL/Datawarehouse Tools: Informatica 9.6/9.1/8.6.1/8.1 , SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau 8.2
Quality Assurance Tools: Win Runner, Load Runner, Test Director, Quick Test Pro, Quality Center, Rational Functional Tester.
Testing and defect tracking Tools: HP/Mercury (Quality Center), Quick Test Professional, Performance Center, Requisite, MS Visio.
AWS: AWS, EC2, S3
Project Execution Methodologies: Agile, Ralph Kimball and BillInmondatawarehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
Operating System: Windows, Unix, Sun Solaris
Data Modeling Tools: Erwin, Rational System Architect, IBM Infosphere Data Architect, ER Studio and Oracle Designer.
PROFESSIONAL EXPERIENCE:
Confidential, Newark, NJ
Senior Data Architect/Modeler
Responsibilities:
- Designed and implemented a Data Lake to consolidate data from multiple sources, using Hadoop stack technologies like SQOOP, HIVE/HQL.
- TalendAdministrative tasks like - Upgrades, create and manage user profiles and projects, manage access, monitoring, setup TAC notification.
- Analyzed metrics surrounding teh activities associated wif theacquisition, documentation, review, cleaning and processing ofdata.
- Specifies overall Data Architecture for all areas and domains of teh enterprise, including Data Acquisition, ODS, MDM, Data Warehouse, Data Provisioning, ETL, and BI.
- Reverse engineered some of teh databases using Erwin.
- Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, SQL Server and Oracle).
- Worked wifMicro StrategyObject Manager to move objects among teh Development, Test, and Prodenvironments.
- Specializes in technology platform migrations for Windows Server, Windows Desktop, Hyper-V, Office 365 andAzure
- Developed Data Mapping, Data Governance, Transformation, Acquisition and cleansing rules for teh Master Data Management Architecture involving OLTP, ODS.
- Working as a Data Modeler/Architect to generate Data Models using Erwin and developed relational database system.
- Also involved in Data Architect role to review business requirement and compose source to target data mapping documents.
- Researched, evaluated, architect, and deployed new tools, frameworks, and patterns to build sustainable Big Data platforms for our clients
- Designed and developed architecture fordataservices ecosystem spanning Relational, NoSQL, and BigDatatechnologies.
- Collected large amounts of log data using Apache Flume and aggregating using PIG/HIVE in HDFS for further analysis
- Designed teh Logical Data Model using ERWIN 9.64 wif teh entities and attributes for each subject areas.
- Designed Common Information Model (CIM) using IBM Infosphere Data Architect data modeling tool
- Worked extensively in creatingMicro StrategyApplication Objects such as Compound Metrics, Level Metrics, complex Filters, Custom Groups, Consolidations, and nested Prompts.
- Worked on multipleSFDCprojects as Administrator andDeveloper.
- Worked wif theSparkfor improving performance and optimization of teh existing algorithms in Hadoop usingSparkContext,Spark-SQL,SparkMLlib, Data Frame, Pair RDD,SparkYARN.
- Developed U-SQL Scripts for schematizing teh data inAzureData Lake Analytics.
- Experienced in using Talend Data Fabric tools (Talend DI, Talend MDM, Talend DQ, Talend Data Preparation, ESB, TAC)
- InstalledHortonworksHadoop clusters and supporting packages.
- Gathered and analyzed existing physical data models for in scope applications and proposed teh changes to teh data models according to teh requirements.
- Advises on and enforces data governance to improve teh quality/integrity of data and oversight on teh collection and management of operational data.
- Designed Physical Data Model (PDM) using IBM Infosphere Data Architect data modeling tool and ORACLE PL/SQL
- Used theSparkData Stax Cassandra Connector to load data to and from Cassandra.
- Design Architecture for API development & deployment as Microservice including Python code in Docker container andAzure.
- Design reference data and data quality rules usingIDQand involved in cleaning teh data usingIDQin Informatica Data Quality 9.1 environment.
- Integrated crystal reports using Erwin Data Modeler.
- Us Erwin to support for TeradataV15 and SSL.
- Designed and developed Reference Integrity, Technical and Business Data
- Quality rules using IDQ and involved in cleaning teh data using IDQ in Informatica
- Data Quality Data modeling, Design, implement, and deploy high-performance, custom applications at scale on Hadoop /Spark.
- Loaded data into Hive Tables from Hadoop Distributed File System (HDFS) to provide SQL-like access on Hadoop data.
- Developed and implemented data cleansing, data security, data profiling and data monitoring processes.
- Applied data analysis, data mining and data engineering to present data clearly.
- Developed long term data warehouse roadmap and architectures, designs and builds teh data warehouse framework per teh roadmap.
- Involved in designing Logical and Physical data models for different database applications using teh Erwin.
- Migrated data from MS Excel / CSV files toSFDCusing Data Loader.
- Worked wif Cloudera, AWS, andHortonworks
- Experience wif AWS ecosystem (EC2, S3, RDS, Redshift).
- WrittenPythonScripts, mappers to run on Hadoop distributed file system (HDFS).
- Advises/leads projects involving teh ETL related activities and teh migration or conversion of data between enterprise data systems. Coordinates interactions between central IT, business units, and data stewards to achieve desired organizational outcomes.
Environment: Oracle 12c, MS-Office, SQL Architect, TOAD Benchmark Factory, Teradatav15, SQL Loader, SFDC, Big Data, SharePoint, ERwin r 9.64,Micro strategy,DB2, MS-Office, SQL Server 2008/2012, AWS.
Confidential, Bronx, NY
Sr. Data Analyst/Modeler
Responsibilities:
- Worked in importing and cleansing ofdatafrom various sources like Teradata, Oracle, flat files, MS SQL Server wif high volumedata
- Designed Logical & Physical Data Model /Metadata/ data dictionary usingErwinfor both OLTP and OLAP based systems.
- Reverse Engineered DB2 databases and then forward engineered them to Teradata using ER Studio.
- Part of team conducting logical data analysis and data modeling JAD sessions, communicated data-related standards .
- Involved in meetings wif SME (subject matter experts) for analyzing teh multiple sources.
- Developed teh logical data models and physical data models that capture current state/future state data elements and data flows using ER Studio.
- Delivered dimensional data models using ER/Studio to bring in teh Employee and Facilities domain data into teh oracle data warehouse.
- Performed troubleshooting, fixed and deployed manyPythonbug fixes of teh two main applications that were a main source of data for both customers and internal customer service team.
- Worked on linux system (Red Hat) to deploy theTalendcode.
- Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data warehouse, data mart reporting system in accordance wif requirements.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, flat files, SQL Server 2005 wif high volume data
- Worked extensively on ER Studio for multiple Operations across Atlas Copco in both OLAP and OLTP applications.
- Generated comprehensive analytical reports by running SQL queries against current databases to conductdataanalysis.
- Produced PL/SQL statement and stored procedures in DB2 for extracting as well as writingdata.
- Co-ordinate all teams to centralize Meta-data management updates and follow teh standard Naming Standards and Attributes Standards for DATA &ETL Jobs.
- Finalize teh naming Standards for Data Elements and ETL Jobs and create a Data Dictionary for Meta Data Management.
- Developed teh design & Process flow to ensure that teh process is repeatable.
- Performed analysis of teh existing source systems (Transaction database)
- Involved in maintaining and updating Metadata Repository wif details on teh nature and use of applications/datatransformations to facilitate impact analysis.
- Created DDL scripts using ER Studio and source to target mappings to bring teh data from source to teh warehouse.
- Designed teh ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical database (capacity planning, object creation and aggregation strategies) for Oracle and Teradata .
- Involved in SQL queries and optimizing teh queries in Teradata.
- Created DDL scripts using ER Studio and source to target mappings to bring teh data from source to teh warehouse.
- Identify, assess and intimate potential risks associated to testing scope, quality of teh product and schedule.
- Wrote and executed SQL queries to verify that data has been moved from transactional system to DSS, Data warehouse, data mart reporting system in accordance wif requirements.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, flat files, SQL Server 2005 wif high volume data.
Environment: ER Studio, Business Objects XI, Rational Rose,Datastage, MS Office, MS Visio, SQL, SQL Server 2000/2005, Rational Rose, Crystal Reports 9, SQL Server 2008, SQL Server Analysis Services, SSIS, Oracle 10g
Confidential, Miami Lakes, Florida
Data Analyst/Modeler
Responsibilities:
- Worked on data mapping process from source system to target system. Created dimensional model for teh reporting system by identifying required facts and dimensions using Erwin
- Developed enhancements toMongo DBarchitecture to improve performance and scalability.
- Forward Engineering teh Data models, Reverse Engineering on teh existing Data Models and Updates teh Data models.
- Performeddatacleaning anddatamanipulation activities using NZSQL utility.
- Analyzed teh business requirements by dividing them into subject areas and understood teh data flow wifin teh organization
- Generated a separate MRM document wif each assignment and shared it on SharePoint along wif teh PDF of updated data models.
- Created a Data Mapping document after each assignment and wrote teh transformation rules for each field as applicable
- Worked on Unit Testing for three reports and created SQL Test Scripts for each report as required
- Extensively used Erwin as teh main tool for modeling along wif Visio
- Established and maintained comprehensive data model documentation including detailed descriptions of business entities, attributes, and data relationships.
- Worked on Metadata Repository (MRM) for maintaining teh definitions and mapping rules up to mark.
- Developed data Mart for teh base data in Star Schema, Snow-Flake Schema involved in developing teh data warehouse for teh database.
- Designed Logical Data Models and Physical Data Models using Erwin.
- Developed teh Conceptual Data Models, Logical Data models and transformed them to creating schema using ERWIN.
- Created a list of domains in Erwin and worked on building up teh data dictionary for teh company
- Created DDL scripts for implementing Data Modeling changes. Created ERWIN reports in HTML, RTF format depending upon teh requirement, Published Data model in model mart, created naming convention files, co-coordinated wif DBAs' to apply teh data model changes.
- Analyzed teh physicaldatamodel to understand teh relationship between existing tables. Cleansed teh unwanted tables and columns as per teh requirements as part of teh duty being aDataAnalyst.
- Worked very close wifDataArchitectures and DBAteam to implementdatamodel changes in database in all environments.
Environment: OracleDataModeler, Teradata 12, SSIS, Business Objects, Erwin r8.2, Oracle SQL Developer, SQL Server 2008, ER/Studio Windows XP, MS Excel.
Confidential
Data Analyst/Modeler
Responsibilities:
- Designed Star and Snowflake Data Models for Enterprise Data Warehouse using ERWIN
- Created and maintained Logical Data Model (LDM) for teh project. Includes documentation of all entities, attributes, data relationships, primary and foreign key structures, allowed values, codes, business rules, glossary terms, etc.
- Improved performance on SQL queries used Explain plan / hints /indexes for tuning created DDL scripts for database. Created PL/SQL Procedures and Triggers.
- Validated and updated teh appropriate LDM's to process mappings, screen designs, use cases, business object model, and system object model as they evolve and change.
- Worked wif Business users during requirements gathering and prepared Conceptual, Logical and PhysicalDataModels.
- Created conceptual, logical and physical data models using best practices and company standards to ensure high data quality and reduced redundancy.
- Wrote PL/SQL statement, stored procedures and Triggers in DB2 for extracting as well as writing data.
- Attended and participated in information and requirements gathering sessions
- Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
- Performed extensive Data Analysis and Data Validation on Teradata.
- Responsible for teh development and maintenance of Logical and Physical data models, along wif corresponding metadata, to support Applications.
- Created business requirement documents and integrated teh requirements and underlying platform functionality.
- Excellent knowledge and experience in Technical Design and Documentation.
- Used forward engineering to create a physical data model wif DDL that best suits teh requirements from teh Logical Data Model.
- Worked wif teh DBA to convert logical Data models to physical Data models for implementation.
- Involved in preparing teh design flow for theDatastageobjects to pull thedatafrom various upstream applications and do teh required transformations and load thedatainto various downstream applications
- Performed logicaldatamodeling, physicaldatamodeling (including reverse engineering) using teh ErwinDataModeling tool.
- Experience in developing dashboards and client specific tools in Microsoft Excel and Power Point.
Environment: ER- Studio6.0/6.5, Toad 8.6, Informatica 8.0, IBM OS 390(V6.0), DB2 V7.1, Oracle9i, PL/SQL, Solaris 9/10, Windows Server 2003 & 2008. NZSQL,