We provide IT Staff Augmentation Services!

Sr. Data Architect/data Modeler Resume

0/5 (Submit Your Rating)

Chevy Chase, MD

SUMMARY

  • Over 9+ years of experience as a Data Architecture, Data Analyst and Data Modeler with solid understanding of Business Requirements Gathering, Business Process Mapping Evaluating Data Sources and Data mapping, data profiling, Hadoop ecosystem, AWS, Data Analytics, Data Warehousing and ETL.
  • Experienced in Dimensional Data Modeling, Star/Snowflake schema, FACT & Dimension tables.
  • Solid experience in development of Conceptual, Logical and Physical Models for Online Transaction Processing and Online Analytical Processing (OLTP & OLAP) and Power Design.
  • Experiences with data modeling with NoSQL databases such as MongoDB for document models, AWS Redshift etc.
  • Excellent experience in trouble shooting SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Experienced with distributeddatawarehousing and/ordatamining systems, using one or more BigData/NoSQL technologies (Hadoop, Hive, HBase, Pig, Cassandra, MongoDB)
  • Hands on experience with modeling using ERWIN, ER Studio, MS Visio in both forward and reverse engineering cases and skillful in Data Analysis using SQL on Oracle, MS SQL Server, Netezza, DB2 & Teradata
  • Well versed in Normalization / De normalization techniques for optimum performance in relational and dimensional database environments.
  • Experience in working with Business Intelligence and EnterpriseDataWarehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum and Amazon Redshift (AWS).
  • Good experience and understanding of Teradata SQL Assistant, Teradata Administrator anddataload/ export utilities like BTEQ, Fast Load, Multi Load, Fast Export.
  • Hands on experience with variousDataArchitectand ETLArchitect, subsystem and patterns, including Change Date Capture, Slow Change Dimension,DataCleansing, auditing and validation, etc.Solid experience with data governance, data steward, data quality concepts and implementations and expertise in performing User Acceptance Testing (UAT) and conducting end user training sessions.
  • Strong background in variousDataModeling tools using Erwin, ER Studio and Power Designer.
  • Experience in integration of various relational and non - relational sources such as DB2, Teradata, Oracle, Netezza, SQL Server, NoSQL, and Netezza database.
  • Extensive knowledge and experience in producing tables, reports, graphs and listings using various procedures and handling large databases to perform complex data manipulations.
  • Experienced indatatransformation,datamapping from source to target database schemas anddata cleansing procedures using Informatica Power Center, Talend and Pentaho.
  • Good inDatawarehouse loads, determining hierarchies, building various logics to handle Slowly Changing Dimensions.
  • Strong experience in Normalization (1NF, 2NF, 3NF and BCNF) and De-normalization techniques for effective and optimum performance in OLTP and OLAP environments.
  • Expertise in Informatica Power Center and Data Quality (IDQ) tools.
  • Excellent experience in trouble shooting SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Expertise in SQL Server Analysis Services (SSAS) and SQL Server Reporting Services (SSRS)
  • Trained end users, prepare related documentation (requirements, training materials, process and data flows, use cases, functional design, etc), and addressed critical questions from the user base involving the CRM
  • Good experience with use of Access queries, excel functions V-Lookup, formulas, Pivot Tables, etc.

TECHNICAL SKILLS

Operating Systems: Linux, HP UNIX, Sun Solaris, and Windows 2000/XP/Vista

MS Office Tools: MS Word, MS Excel, MS Access, MS PowerPoint, MS Visio

Languages: SQL, PL/SQL, JSON, UNIX script, XML, Scala, R

Datawarehousing Tools: Business Objects, Cognos, Microstrategy, Tableau, Informatica, TalendPentaho, Datastage

Hadoop Ecosystem: Hadoop Framework, HiveQL Queries, Pig, HBase, MongoDB, Sqoop and spark, scala

Modeling Tools: ERWIN r9.6/r 9.5/9.1, Embarcadero, ER/Studio, MSVisio, Sybase Power Designer

Databases: Netezza, Teradata R14, R13, R12, Oracle 12c/11g/10g, DB2, SQL, Oracle, MS Access, SQL Server 2000, MongoDB

AWS: AWS S3, AWS Redshift, AWS EMR.

Methodologies: RAD, JAD, RUP, UML, System Development Life Cycle (SDLC)Waterfall Model

ETL Tools: Informatica Power Center 9x/8x, Pentaho, SSIS.

Reporting Tools: Crystal reports XI, Business Intelligence, SSRS, Business Objects 5.x / 6.x, Tableau.

PROFESSIONAL EXPERIENCE

Confidential, Chevy Chase, MD

Sr. Data Architect/Data Modeler

Responsibilities:

  • Provided a consultative approach with business users, asking questions to understand the business need and deriving the data flow, logical, and physical data models based on those needs.
  • Designed and developedarchitecturefor data services ecosystem spanning Relational,NoSQL, andBigDatatechnologies
  • Specifies overallDataArchitecturefor all areas and domains of the enterprise, includingDataAcquisition,ODS,MDM,DataWarehouse, Data Provisioning,ETL, and BI.
  • CreatedPhysical & logicaldatamodelfrom theconceptualmodeland it's conversion into the physical database with theDDL’susingforwardengineeringoptions inErwin.
  • DevelopedData Mapping, Data Governance, and Transformation and cleansing rules for theMaster Data Management Architecture
  • Completed enhancement forMDM (Master data management)and suggested the implementation forhybridMDM(Master Data Management)
  • Handled importing data from various data sources, performed transformations usingHive,MapReduce, and loaded data intoHDFS.
  • Worked on designing, implementing and deploying into production anEnterprisedatawarehouse
  • DevelopedSQlprocess usingSSISwith VariousControl Flow, Data Flow tasksandStore Proceduresfor WorkOracleValidationprocess
  • Worked on importing and exporting data fromOracleandTeradataintoHDFSandHIVEusingSqoop.
  • UsedSQLon the newAWSDatabases likeRedshiftand Relation Data Services.
  • Designed and DevelopedOraclePL/SQLandShellScripts,Data Import/Export,Data ConversionsandDataCleansing.
  • Usedforwardengineeringto create aPhysical Data ModelwithDDLthat best suits the requirements from the Logical Data Model
  • Involved in capturingdatalineage, table and column data definitions, valid values and others necessary information in the data model
  • ExtractingMegaDatafromAmazonRedshift,AWS, and Elastic Search engine usingSQLQueriesto create reports
  • Responsible fordataprofilingand data quality checks to suffice the report requirements gathered above and provide anETLMapping
  • Generated periodic reports based on the statistical analysis of the data usingSQL Server Reporting Services (SSRS).
  • Designed both3NFdata models forODS,OLTPsystems and dimensional data models usingStarandSnowFlakeSchemas
  • Worked in using Teradata tools likeFast Load, Multi Load, Tpump, FastExport,TeradataParallelTransporter(TPT)andBTEQ.
  • UsedETLmethodology for supporting data extraction, transformations and loading processing, in a complexMDMusingInformatica.
  • Worked in usingTeradatatools likeFastLoad,MultiLoad,Tpump,FastExport,TeradataParallelTransporter(TPT) andBTEQ.
  • UsedSQLfor Querying the database inUNIXenvironment
  • Worked on themetadatamanagementand part ofdatagovernanceteam which created theData
  • Responsible for full data loads from production toAWSRedshiftstaging environment.
  • Creating or modifying theT-SQLqueriesas per the business requirements.
  • Developed variousQlikViewDataModelsby extracting and using the data from various sources files Excel, Flat Files andBigdata.
  • Primarily responsible forTableaucustomization for statistical dashboard to monitor sales effectiveness and also usedTableaufor customer marketingdatavisualization.
  • Developed the performance tuning of the database by usingEXPLAINPLAN,TKPROFutilities and also debugging theSQLcode.

Environment: Erwin 9.6, AWS Redshift, MapReduce, ODS, MDM, OLAP, OLTP, Pig, Oracle 12, Hadoop, HDFS, Sqoop, Hive, NoSQL, Teradata R15, Netezza, PL/SQL, MS-Visio, T-SQL, SSIS, SSRS, SQL, Unix, Tableau.

Sr. Data Architect/Data Modeler

Confidential, Dearborn, MI

Responsibilities:

  • Extensively usedAgilemethodology as the Organization Standard to implement thedata models.
  • Excellent experience and knowledge onDataWarehouseconcepts and dimensional datamodelingusingRalphKimballmethodology.
  • Developed long term data warehouse roadmap and architectures, designs and builds thedatawarehouseframework per theroadmap.
  • Involved in designingLogicalandPhysicaldatamodelsfor different database applications using theErwin 9.6.
  • Translate business and data requirements intoLogicaldatamodelsin support of EnterpriseData models,ODS, OLAP, OLTP, Operational Data Structures and Analytical systems.
  • Developed Extraction, Transformation and Loading(ETL)processes to acquire and load data from internal and external sources.
  • Created several MasterData Models(MDM)those unify data for critical concepts across applications.
  • DevelopedData Mapping, Data Governance, andTransformationand cleansing rules for the Master Data Management Architecture involvingOLTP, ODS.
  • Designed both3NFdata modelsforODS, OLTPsystems and dimensionaldata modelsusingStarandSnow flake Schemas.
  • Responsible for Meta data Management, keeping up to date centralized metadata repositories usingErwinmodeling tools.
  • Involved in the Analysis, design, testing and Implementation of Business Intelligence solutions usingData Warehouse, ETL, OLAP, Client/Server applications.
  • Involved inAWSArchitecture design and modification, gap analysis
  • Handled importing data from various data sources, performed transformations usingHive, Map Reduce,and loaded data intoHDFS.
  • PerformedHiveprogramming for applications that were migrated tobig datausingHadoop
  • Loaded data intoHiveTables from Hadoop Distributed File System(HDFS)to provideSQLaccess on Hadoop data
  • Designed and developed aDataLakeusingHadoopfor processing raw and processed claims viaHiveandInformatica.
  • Created data model and imported data using mongo import.
  • Backed up databases usingMongoDBbackup facility in OPS manager.
  • Performance tuning and stress-testing ofNoSQLdatabase environments in order to ensure acceptable database performance in production mode.
  • Implemented strong referential integrity and auditing by the use oftriggersandSQLScripts.
  • Produced and enforced data standards and maintain a repository of data architecture artifacts and procedures.
  • Provides architectures, patterns, tooling choices and standards for master data and hierarchy life cycle management.

Environment: ERWIN 9.6, MS SQLServer2016,AWS,Oracle12c, SQL, Hive, MapReduce, HDFS, Hadoop, Teradata, Netezza, PL/SQL, Informatica, SSIS, SSRS.

Sr. Data Modeler/Data Analyst

Confidential - San Francisco, CA

Responsibilities:

  • Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
  • Participated in the design, development, and support of the corporate operationdatastore and enterprisedatawarehouse database environment.
  • Designed the LogicalDataModel using ER/Studio with the entities and attributes for each subject areas and Worked on ER Studio for multiple Operations across in both OLAP and OLTP applications.
  • Used the Agile Scrum methodology to build the different phases of Software development life cycle.
  • Involved in several facets of MDM implementations includingDataProfiling, Metadata acquisition anddatamigration.
  • Independently coded new programs and design Tables to load and test the program effectively for the given POC's using BigData/Hadoop.
  • Developed Map Reduce programs to cleanse thedatain HDFS obtained from heterogeneousdata sources to make it suitable for ingestion into Hive schema for analysis.
  • Responsible for fulldataloads from production to AWS Redshift staging environment and Worked on migrating of EDW to AWS using EMR and various other technologies.
  • Design of ODS layer, Dimensional modeling using Kimball Methodologies, of theDataWarehouse sourced from MDM Base tables and other Transactional systems.
  • Extensively used Aginity Netezza work bench to perform various DML, DDL etc operations on Netezza database.
  • DevelopedDataMapping,DataGovernance, Transformation and Cleansing rules for the MasterDataManagement Architecture involving OLTP, ODS and OLAP.
  • Extracting MegaDatafrom Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports
  • Worked with the ETL team to document the transformation rules fordatamigration from OLTP to Warehouse environment for reporting purposes.
  • Worked with Netezza and Oracle databases and implemented various logical and physicaldata models for them.
  • Developed complex SQL scripts for Teradata database for creating BI layer on DW for Tableau reporting.
  • Involved in mapping thedataelements from the User Interface to the Database and help identify the gaps and Prepared ofDataDictionary/MetaDataof thedataelements.
  • Developed and configured on Informatica MDM hub supports the MasterDataManagement (MDM), Business Intelligence (BI) andDataWarehousing platforms to meet business needs.
  • Generated preliminary Source to Target mapping by writing a ER/Studio Macro in the ER/Studio VBA platform document and handed over to BAs for further transformation analysis.
  • Submitted the DDL and Security document by writing an ERStudio Macro to the SVN repository and communicated to the DBA for implementation.
  • Extracting MegaDatafrom Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports.
  • Developed long termdatawarehouse roadmap and architectures, designs and builds thedata warehouse framework per the roadmap.
  • Designed and Developed Oracle PL/SQL and Shell Scripts,DataImport/Export,DataConversions andDataCleansing
  • Extensively worked on documentation ofDataModel, Mapping Transformation and Scheduling jobs.
  • Worked extensively with Business Objects XI Report Developers in solving critical issues of defining hierarchy, loops and Contexts.

Environment: ER/Studio, Oracle 12c, Informatica, DB2, MDM AWS Redshift, AWS S3, Hadoop, Hive, Netezza, MongoDB, HBase, Sqoop, Spark, HDFS, MapReduce, OLAP, OLTP, PL/SQL, Flat Files, SQL, XML, Teradata, Tableau, SSRS, SSIS and Cassandra.

Sr. Data Modeler/Data Analyst

Confidential - New York, NY

Responsibilities:

  • Provided a consultative approach with business users, asking questions to understand the business need and deriving thedataflow, logical, and physicaldatamodels based on those needs.
  • Designed and developed architecture fordataservices ecosystem spanning Relational, NoSQL, and BigDatatechnologies.
  • Specifies overallDataArchitecture for all areas (Healthcare payer domain) and domains of the enterprise, includingData Acquisition, ODS, MDM,DataWarehouse,DataProvisioning, ETL, and BI.
  • Created Physical & logicaldatamodel from the conceptual model and it's conversion into the physical database with the DDL's using forward engineering options in Erwin.
  • DevelopedDataMapping,DataGovernance, and Transformation and cleansing rules for the MasterDataManagement Architecture.
  • Completed enhancement for MDM (Masterdatamanagement) and suggested the implementation for hybrid MDM (MasterDataManagement).
  • Handled importingdatafrom variousdatasources, performed transformations using Hive, MapReduce, and loadeddatainto HDFS.
  • Worked on designing, implementing and deploying into production an Enterprisedatawarehouse
  • Developed SQl process using SSIS with Various Control Flow,DataFlow tasks and Store Procedures for Work Oracle Validation process.
  • Worked on importing and exportingdatafrom Oracle and Teradata into HDFS and HIVE using Sqoop.
  • Designed and Developed Oracle PL/SQL and Shell Scripts,DataImport/Export,DataConversions andDataCleansing.
  • Used forward engineering to create a PhysicalDataModel with DDL that best suits the requirements from the LogicalDataModel.
  • Responsible fordataprofiling anddataquality checks to suffice the report requirements gathered above and provide an ETL Mapping.
  • Generated periodic reports based on the statistical analysis of thedatausing SQL Server Reporting Services (SSRS).
  • Designed both 3NFdatamodels for ODS, OLTP systems and dimensionaldatamodels using Star and Snow Flake Schemas.
  • Worked in using Teradata tools like Fast Load, Multi Load, Tpump, FastExport, Teradata Parallel Transporter (TPT) and BTEQ.
  • Used ETL methodology for supportingdataextraction, transformations and loading processing, in a complex MDM using Informatica.
  • Used SQL for Querying the database in UNIX environment
  • Worked on the metadata management and part ofdatagovernance team which created theData
  • Creating or modifying the T-SQL queries as per the business requirements.
  • Developed various QlikViewDataModels by extracting and using thedatafrom various sources files Excel, Flat Files and Bigdata.
  • Primarily responsible for Tableau customization for statistical dashboard to monitor sales effectiveness and also used Tableau for customer marketingdatavisualization.
  • Developed the performance tuning of the database by using EXPLAIN PLAN, TKPROF utilities and also debugging the SQL code.

Environment: Erwin 9.1, MapReduce, ODS, MDM, OLAP, OLTP, Pig, Oracle 12, Hadoop, HDFS, Sqoop, Hive, NoSQL, Teradata R13, Netezza, PL/SQL, MS-Visio, T-SQL, SSIS, SSRS, SQL, Unix, Tableau, Informatica, PL/SQL.

Data Analyst

Confidential - Pleasanton, CA

Responsibilities:

  • Performeddataanalysis and profiling of sourcedatato better understand the sources.
  • Created logicaldatamodel from the conceptual model and it's conversion into the physical database design using Erwin.
  • Worked with DBA's to create a best-fit PhysicalDataModel from the logicaldatamodel.
  • Designed Sources to Targets mappings from SQL Server, Excel/Flat files, Xml Files to Teradata using Informatica Power Center Data cleansing, integrating, and matching using Informatica Data Quality IDQ
  • Redefined many attributes and relationships in the reverse engineered model and cleansed unwanted tables/columns as part ofdataanalysis responsibilities.
  • Interacted with the database administrators and business analysts for data type and class words.
  • Conducted design sessions with business analysts and ETL developers to come up with a design that satisfies the organization's requirements.
  • Worked on enterprise logicaldatamodeling project (in third normal form) to gatherdatarequirements for OLTP enhancements. Converted third normal form ERDs into dimensional ERDs fordatawarehouse effort.
  • Used Model Mart of Erwin for effective model management of sharing, dividing and reusing model information and design for productivity improvement.
  • Created ER Diagrams,DataFlow Diagrams, grouped and created the tables, validated thedata, identified PK/ FK for lookup tables.
  • Created 3NF business areadatamodeling with de-normalized physical implementationdataand information requirements analysis using Erwin tool.
  • Developed Star Schema and Snowflake Schema in designing the Logical Model into Dimensional Model.
  • Assisted the ETL team to document the transformation rules fordatamigration from OLTP to Warehouse environment for reporting purposes.
  • Implement necessary DQ rules inIDQAnalyst while profiling the data.
  • Involved in extensivedataanalysis on the Teradata and Oracle systems querying and writing in SQL and TOAD.
  • Used SQL joins, aggregate functions, analytical functions, group by, order by clauses and interacted with DBA and developers for query optimization and tuning.
  • Conducted several PhysicalDataModel training sessions with the ETL Developers. Worked with them on day-to-day basis to resolve any questions on Physical Model..

Environment: CA Erwin 9.1, Oracle11g, SQL server 2005, IBM DB2, Informatica Power Center, IDQ, SQL BI 2008, Oracle BI, Visual Studio, SSIS&SSRS, Tibco Spotfire, SQL server management studio 2012.

We'd love your feedback!