Sr. Data Architect/data Modeler Resume
Dallas, TX
SUMMARY
- 10+ years of Industry experienced in IT with solid understanding of Data Modeling, Data Analysis, Data Architecture, Evaluating Data Sources and strong understanding of Data Warehouse/Data Mart Design, ETL, BI, OLAP,OLTP, Client/Server applications.
- Experienced in writing SQL queries and optimizing the queries in Oracle, SQL Server,and Netezza, Teradata.
- Experienced in Dimensional Data Modeling using Data Modeling, Relational Data modeling, ER/Studio, ERwin, and Sybase Power Designer, Star Join Schema/Snowflake modeling, FACT & Dimensions tables, Conceptual, Physical & logical Data Modeling.
- Experienced in Management and implementation of database models, data flow diagrams, database schemas, db scripts, DTD schemas, structures and data standards to support a robust data management infrastructure.
- Experienced in Data Analysis and Data Profiling using complexSQL on various sources systems including Oracle and Teradata.
- Very good knowledge and experience on AWS, Redshift, S3 and EMR.
- Extensive experience in Normalization (1NF, 2NF, 3NF and BCNF) and De - normalization techniques for improved database performance in OLTP, OLAP and Data Warehouse/Data Mart environments.
- Experienced on Metadata definition, implementation and maintenance, new business rules identification and implementation to data rules, transformation program library maintenance, XML file generation and data quality.
- Experienced using MapReduce and Bigdata work on Hadoop and other NO SQL platforms
- Experienced in designing standards for using normalized data structures, de-normalized structures and dimensional structures. Defines common design patterns for modeling various types of relationships.
- Experienced in Batch processes, Import, Export, Backup, Database Monitoring tools and Application support.
- Experienced in big data analysis and developing data models using Hive, PIG, and Map reduce, SQL with strong data architecting skills designing data-centric solutions.
- Experienced in Teradata SQL queries, Teradata Indexes, Utilities such as Mload, Tpump, Fast load and Fast Export.
- Experienced in using the databases like DB2, Teradata and its utilities, Netezza, Oracle, SQL Server Integration Services (SSIS).
- Experienced in data from various data sources/business systems including MS Excel, MS Access, Flat Files etc to SQL Server using SSIS using various features like data conversion etc.
- Experienced in Oracle,Netezza, and Teradata, SQL Server, and DB2 database architecture.
- Expertise in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
- Extensive experience in development of T-SQL, DTS, OLAP, PL/SQL, Stored Procedures, Triggers, Functions, Packages, performance tuning and optimization for business logic implementation.
- Good knowledge of Data Marts, Operational Data Store (ODS), OLAP, Dimensional Data Modeling with Ralph Kimball Methodology (StarSchemaModeling,Snow-FlakeModeling for FACT and Dimensions Tables) using Analysis Services.
- Excellent in performing data transfer activities between SAS and various databases and data file formats like XLS,CSV,DBF,MDB etc.
- Experienced in ER Studio and Dimensional Models using ERwin advanced features, Conceptual, logical and physical data models using ERwin.
- Experienced in development and support knowledge on Oracle, SQL, PL/SQL, T-SQL queries.
- Experienced in testingintegration solutions for Data import,export and Migration using EIM(Enterprise Integration Manager).
- Experienced in Data Scrubbing/Cleansing, Data Quality, Data Mapping, Data Profiling, Data Validation in ETL
- Excellent Knowledge of Ralph Kimball and Bill Inmon's approaches toDataWarehousing.
- Excellent knowledge in developing Informatica Mappings, Mapplets, Sessions, Workflows and Worklets for data loads from various sources such as Oracle, Flat Files, DB2, SQL Server etc.
- Experienced in writing UNIX shell scripting and hands on experienced with scheduling of shell scripts using Control-M.
- Extensive experience in Relational and Dimensional Data modeling for creating Logical and Physical Design of Database and ER Diagrams using multiple data modeling tools like ERWIN, ER Studio.
TECHNICAL SKILLS
Analysis and Modeling Tools: Erwin r9.6/r9.5/9.1, Sybase Power Designer, Oracle Designer, BP win ER/Studio, MS Access 2000, Star-Schema, Snowflake-Schema Modeling, and FACT and dimension tables, Pivot Tables.
OLAP Tools: Tableau, SAP BO, SSAS, Business Objects, and Crystal Reports 9.
Oracle: Oracle12c/11g/10g/9i/8.x, R2 database servers with RAC, ASM, Data Guard, Grid Control and Oracle Golden Gate(Oracle Enterprise Manager),Oracle Data Guard, SQL, Net, SQL Loader and SQL PLUS, AWR,ASH, ADDM, Explain Plan.
ETL Tools: SSIS, Pentaho, Informatica Power Center 9.7/9.6/9.5/9.1 etc.
Programming Languages: Java, Base SAS, SSIS and SAS/SQL, SQL, T-SQL, HTML/ XHTML/ HTML4.0.1/ HTML3.2, Java Script, CSS3/CSS2/CSS1, UNIX shells scripting, PL/SQL.
Database Tools: Microsoft SQL Server 2014/2012/2008/2005 , Teradata, and MS Access, Poster SQL, Netezza, SQL Server, Oracle.
Reporting Tools: Business Objects, Crystal Reports, and SSRS.
Operating Systems: Microsoft Windows 9x / NT / 2000/XP / Vista/7 and UNIX Windows 95, 98, Windows NT, Windows XP, 7.
Tools: & Software: TOAD, MS Office, BTEQ, Teradata 15/14.1/14/13.1/13 , SQL Assistant
Big Data: Hadoop, HDFS 2, Hive, Pig, HBase, Sqoop, Flume.
Other tools: TOAD, SQL PLUS, SQL LOADER, MS Project, MS Visio and MS Office, Have worked on C++, UNIX, PL/SQL, Microsoft Team Foundation Server etc.
PROFESSIONAL EXPERIENCE
Sr. Data Architect/Data Modeler
Confidential, Dallas, TX
Responsibilities:
- Involved in architecting, designing, modeling, developing, maintaining and deploying complex data models, and data integrations utilizing SQL server (SSIS) based ETL processes for new and existing systems.
- Designed ER diagrams, logical model (relationship, cardinality, attributes, and, candidate keys) and physical data models (capacity planning, object creation and aggregation strategies) as per business requirements.
- Involved in documentation ofDataModeler/Data Analyst and ETL specifications forDatawarehouse Erwin r9.6.
- Coordinated new data development ensuring consistency and integration with existing data warehouse structure.
- Worked on setting upAWSDMS and SNS fordatatransfer and replication
- Created and maintained Logical Data Model (LDM) for the project includes documentation of all entities, attributes, data relationships, primary and foreign key structures, allowed values, codes, business rules, glossary terms, etc.
- Analysis of functional and non-functional categorized data elements for data profiling and mapping from source to target data environment.
- Worked on updating the Designer Workflows and InfoPath forms inSharePoint2010.
- Used SQL on the new AWS Databases likeRedShiftand RelationDataServices.
- Involved in OLAP model based on Dimension and FACTS for efficient loads of data based onStarSchema structure on levels of reports using multi-dimensional models such asStarSchemasandSnow Flake Schema for developing Cubes using MDM.
- Used forward engineering to create a PhysicalDataModel with DDL that best suits the requirements from the LogicalDataModel
- Migrating existing on-premise applications and services to AWS.
- Predominantly using Python and AWS (Amazon web services), and MySQL along with NoSQL (mongodb) databases for meeting end requirements and building scalable real time system.
- Executed Hivequeries on Parquet tables stored in Hive to performdataanalysis to meet the business requirements and worked on importing and exportingdatafrom Oracle and DB2 into HDFS and HIVE using Sqoop.
- Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata14.1, Oracle, and Database, analysis, development, testing, implementation and deployment.
- Created Mappings for Initial load from MS SQL server 2012 to Netezza while performing data cleansing.
- Migrated Oracle Database of size 100TB over toAWScloud fromDatacenters.
- UsedInformatica mappingparameters, variables and different tasks such as command task, decision task and timer.
- Loaded multi format data from various sources like flat-file, Excel, MS Access, and Oracle using Netezza and performing file system operation.
- Developed ETL processes extracted data daily from Mainframe DB2 and ORACLE Databases, and loaded data into a SSIS based Decision Support Warehouse.
- Performed Extracting, Transforming and Loading (ETL) data from Excel, Flat file, Oracle to MS SQL Server by using SSIS services
- Involved in extensive DATA validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
- Performed ExploratoryDataAnalysis using Python and Hive on Hadoop HDFS.
- Designed the schema, configured and deployed AWSRedshiftfor optimal storage and fast retrieval ofdata and used Spark Data frames, Spark-SQL, SparkMLLibextensively and developing and designing POC's using Scala, Spark SQL andMLliblibraries.
- Gathered runtime metrics to monitor the activities of your data warehouse using AWS Glue.
- Worked onNormalizationandDe-Normalizationtechniques for both OLTP and OLAP systems.
- Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
- Created PL/SQLTables, Collections, and Records, Partitions and Used DynamicSQLand triggers for faster data access and to incorporate business logic.
- Generates ETL scripts to transform, flatten, and enrich the data from source to target using AWS Glue and created event-driven ETL pipelines with AWS Glue.
- Responsible for creating Hive tables, loading data and writing hive queries.
- Performed cluster analysis and page-rank on newswire articles using Hadoop framework in python to interpret the importance of keywords connecting documents.
- Worked with various performance tuning techniques using tools provided by Oracle like EXPLAIN PLAN,SQLTrace, TKPROF, Oracle hints, DBMS PROFILER, DBMS STATS.
- Responsible for determining the bottlenecks and fixing the bottlenecks with performance tuning using Netezza Database.
- Using SQL to extractdatafrom different databases such as Oracle and AmazonRedshift.
Environment: Erwin r9.6, Netezza, Oracle11g, Taradata14.1, T-SQL, SQL Server 2012, SSRS, SSIS, LINUX, MDM, PL/SQL,SPSS, ETL, Informatica MDM, OLTP and OLAP, EIM, Redshift, AWS Glue, AWS Redshift Spectrum, Aginity, Sharepoint, Python, Power view, SQL Assistance, Metadata, Hadoop, DB2, AWS, UNIX, Microsoft Team Foundation Server, Test Manager, Excel, SAS, Hive, Spark, Scala, Pig, Mapreduce, Mongodb, HBase etc.
Sr. Data Architect/Data Modeler
Confidential, Mentor, OH
Responsibilities:
- Analyzed across multiple data domains and define strategies to reduce data redundancy, improve data availability, and accessibility by partnering with the technical teams and deliver a maintainable and reusable data architecture
- Maintained the end-to-end vision of the data flow diagram and develop logical data models into one or more physical data repositories.
- Documented logical data integration (ETL) strategies for data flows between disparate source/target systems for structured and unstructured data into common data lake and the enterprise information repositories
- Involved in designing the data warehouses and data lakes on regular (Oracle, SQL Server) high performance (Netezzaand Teradata) and big data (Hadoop - MongoDB, Hive, Cassandra and HBase) databases.
- Worked with cloud based technology likeRedshift, S3, AWS, EC2 Machine, etc.
- Implemented dimension model (logical and physical data modeling) in the existing architecture using ER Studio.
- Used External Loaders like Multi Load, T Pump and Fast Load to load data into Teradata14 Database.
- Involved in Data modeling using ER Studio identified objects and relationships and how those all fit together as logical entities, these are then translated into physical design using forward engineering ER Studio tool.
- Created ERD diagrams using ER Studio and implemented concepts likeStar-Schema Modeling, SnowflakeSchemaModeling, Fact and Dimension tables.
- Provided technical guidance for re-engineering functions of Teradata warehouse operations into Netezza.
- Designed and Developed Oracle PL/SQL Procedures LINUX and UNIX Shell Scripts for Data Import/Export and Data Conversions.
- Involved in SQL Server and T-SQL in constructing Tables,Normalizationand De-normalization techniques on database Tables.
- Worked with cloud based technology likeRedshift, S3, AWS, EC2 Machine, etc. and extracting the data from the Oracle financials and theRedshiftdatabase.
- Worked on Cube structure optimization for MDM query performance in Analysis Services (SSAS) and worked in generating and documenting Metadata while designing OLTP and OLAP systems environment.
- PerformedSQLand PL/SQLPerformance tuning and Application Tuning using various tools like TKPROF, AUTOTRACE and DBMS SQLTUNE.
- Handled data loading operations from flat files to tables using NZLOAD utility and extensively used NZSQL and NZLOAD utilities.
- Performed Data Analysis, Data Validation and Data verification using InformaticaDVO(Data Validation Option) from raw data to user acceptance.
- Developed complex TeradataSQL code in BTEQ script using OLAP and Aggregate functions to name few and reviewed and implemented the naming standards for the entities, attributes, alternate keys, and primary keys for the logical model.
- Responsible for managing and reviewingHadooplog files. Designed and developeddata management system using MySQL.
- Handled importing data from various data sources, performed transformations using Hive, Map Reduce, and loaded data into HDFS.
- Worked on SQL Server concepts SSIS (SQL Server Integration Services), SSAS (Analysis Services) and SSRS (Reporting Services).
- Used the Agile Scrum methodology to build the different phases of Software development life cycle.
- Manipulating, cleansing & processingdatausingExcel, and SQL. Responsible for loading, extracting and validation of clientdata.
- Performed Forward Engineeringdatamodels for Schema generation, Reverse Engineered on the ExistingDataModels to accommodate new requirements.
- Designed and developedstarschemamodel for target database using ER Studio Data modeling.
Environment: ER Studio, Netezza, Oracle11g, Taradata14.1, T-SQL, SQL Server 2012, SSAS, SSRS, SSIS, LINUX, MDM, PL/SQL,SPSS, ETL, OLTP and OLAP, Informatica MDM, Pivot, EHR, EMR, Sharepoint, Aginity, SQL Assistance, Metadata, DB2, UNIX, Excel, SAS, Hive, Pig, Power view, DVO, Microsoft Team Foundation Server, Test Manager, Mapreduce, Mongodb, HBase etc.
Sr. Data Modeler/ Data Analyst
Confidential, Greensboro, NC
Responsibilities:
- Extensively used Erwin r9.5 to design Logical/Physical Data Models, relational database design, forward/reverse engineering, publishing Data Model to acrobat files, created ERWIN reports in HTML, RTF format depending upon the requirement, Published Data Model in model mart, created naming convention files, co-coordinated with DBAs to apply the Data Model changes.
- Involved in requirement analysis, ETL design and development for extracting data from the source systems like Taradata13.1, DB2, Sybase, Oracle 9i, flat files and loading into Netezza.
- Involved in Designing Star Schema (identification of facts, measures and dimensions), Snowflake Schema for Data Warehouse, ODS architecture by using tools like Data Model, Erwin.
- Verify the data quality or data validation through Informatica data validation option tool (DVO).
- Involved inNormalizationandDe-Normalizationof existing tables for faster query retrieval.
- Performed Verification, Validation, and Transformations on the Input data (Text files, XML files) before loading into target database.
- Involved in Ralph Kimball and Bill Inman Methodologies (StarSchema,SnowFlakeSchema).
- Coded using Teradata Analytical functions, BTEQ SQL of TERADATA, write UNIX scripts to validate, format and execute the SQLs on LINUX environment.
- Worked in importing and cleansing of data from various sources like Teradata, Oracle, Netezza flat files, SQL Server with high volume data.
- Created SSIS Reusable Packages to extract data from Multi formatted Flat files, Excel, XML files into Database and Billing Systems.
- Creating various type of reports such as drill down & drill through reports, Matrix reports, Sub reports and Charts using SQL Server Reporting Services (SSRS).
- Implemented naming Standards and Warehouse Metadata for fact and dimension of Logical and Physical Data Model.
- Created Position and Transaction OLAP Cubes for mortgages Loans for various measure groups and dimensions (Loan Position, Commitment Schedule, Loan Journal Entry, Loan journal Entry Period Zero, Time Period, Loan, Account, Portfolio, Property, Business Unit, Borrower, Journal Entry Demographic, Journal Entry Line Item etc.) to address the business needs using SSAS.
- Created new database objects like Procedures, Functions, Packages, Triggers, Indexes and Views using T-SQL in SQL Server.
- Created complex Stored Procedures and PL/SQLblocks with optimum performance using Bulk Binds (BULK COLLECT & FORALL), Inline views, Reference cursors, cursor variables, dynamicSQL, v-arrays, external tables, nested tables, etc.
- DesignedInformaticamappingfor Error handling and involved in preparation of the low level design (LLD) documents forInformaticaMappings.
- CreatedSSISPackages usingSSISDesigner for exporting heterogeneous data from OLE DB Source, Excel Spreadsheets to SQL Server.
- Designing and developing SQL Server Database, Tables, and Indexes, Stored procedures, Views, User Defined Functions and other T- SQL statements.
Environment: Netezza, Oracle9i, Taradata13.1, T-SQL, SQL Server, DB2, SSIS, LINUX, ERWIN r9.5, MDM, PL/SQL, ETL, SSRS, SSIS, Cognos, SAS, SSAS, SPSS, Sharepoint, Excel, Pivot tables, Shell scripting, Pivot, Power view, UNIX etc.