Sr. Data Engineer Resume
Chicago, IL
SUMMARY
- Over 12+ years of extensive experience in Data Architecture and Data Engineering wif Development, Testing and Deployment of business applications.
- Experienced in Data Modeling &Data Analysis experience using Dimensional Data Modeling and Relational Data Modeling, Star Schema/Snowflake Modeling, FACT & Dimensions tables, Physical & Logical Data Modeling.
- Experience wif distributed data warehousing and/or data mining systems, using one or more Big Data/NoSQL technologies (Hadoop, Hive, HBase, Pig, Cassandra, MongoDB)
- Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
- Experience in Technical consulting and end - to-end delivery wif architecture, data modeling, data governance and design - development - implementation of solutions.
- Extensive experience in developing and driving strategic direction of SAP operating system (SAPECC) and SAP business intelligence (SAPBI) system
- Working experience of Databricks or similar
- Experience in designing, building and implementing complete Hadoop ecosystem comprising of MapReduce, HDFS, Hive, Impala, Pig, Sqoop, Oozie, HBase, MongoDB, and Spark.
- Experience in various Teradata utilities like Fastload, Multiload, BTEQ, and Teradata SQL Assistant.
- Experience in working wif Business Intelligence and Enterprise Data Warehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum, Amazon Redshift and Azure Data Warehouse
- Excellent Software Development Life Cycle (SDLC) wif good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
- Experienced in Data Scrubbing/Cleansing, Data Quality, Data Mapping, Data Profiling, Data Validation in ETL
- Experienced in creating and documenting Metadata for OLTP and OLAP when designing a systems.
- Performeddataanalysis anddataprofiling using complex SQL on various sources systems including Oracle and Teradata.
- Experience in building CI/CD methodology in Azure using technologies like VS online, Jenkins, TFS etc
- Excellent Knowledge of Ralph Kimball and Bill Inmon's approaches toDataWarehousing.
- Excellent in performing data transfer activities between SAS and various databases and data file formats like XLS,CSV, DBF,MDB etc.
- Experience in architecting hybrid solutions, cloud integration wif on premise environment (Azure app service environment, express route)
- Experience in working wif Excel Pivot and VBA macros for various business scenarios.
- Experience in Data modeling using ER diagram, Dimensional data modeling, Conceptual/Logical/Physical Modeling using 3NormalForm (3NF), Star Schema modeling, Snowflake modeling using tools like ER/Studio, Erwin, Sybase PowerDesigner for both forward and reverse engineering.
- Expert in BI reporting and Data Reporting tools like Pentaho and SAP BI.
- Excellent experience in troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
- Proficient wifDataAnalysis, mapping source and target systems fordatamigration efforts and resolving issues relating todatamigration.
- Excellent understanding and working experience of industry standard methodologies like System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE Methodologies.
- Advanced expertise in visualization design and development, SQL and Oracle/SQL Server.
- Expert in building Enterprise Data Warehouse or Data warehouse appliances from Scratch using both Kimball and Inmon Approach.
- Experience in various databases Design of development and Production environment involving Oracle, SQL server, Netezza, MYSQL, DB2, MS Access and Teradata etc.
- Hands on experience in Normalization &De-normalization techniques design considerations upto3NF Form for OLTP Databases and Models.
- Strong experience in writing SQL and PL/SQL, Transact SQL programs for Stored Procedures, Triggers and Functions.
- Expertise in analyzing and documenting business requirement documents (BRD) and functional requirement documents (FRD) along wif Use Case Modeling and UML.
- Experience in UNIX shell scripting, Perl scripting and automation of ETL Processes.
- Expertise in designing complex Mappings and has expertise in performance tuning and slowly-changing Dimension Tables and Fact tables
- Extensively worked wif Teradata utilities BTEQ, Fast Export and Multi Load to export and load data to/from different source systems including flat files.
TECHNICAL SKILLS
Data Modeling Tools: Erwin 9.6/9.5, ER/Studio 9.7/9.0, Sybase Power DesignerBig Data Technologies: Hadoop, Hive, HDFS, HBase, Flume, Sqoop, Spark, Pig, Impala, MapReduce.
Programming Languages: SQL, PL/SQL, UNIX shell Scripting, PERL, AWK, SED
Databases: Oracle 12c/11g, Teradata R15/R14, MS SQL Server 2014/2016Testing and defect tracking Tools: HP/Mercury (Quality Center, Win Runner, Quick Test Professional, Performance Center, Requisite, MS Visio & Visual Source Safe
Operating System: Windows, Unix, Sun Solaris
ETL/Datawarehouse Tools: Informatica 9.6/9.1, SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau, Pentaho
Tools: & Software: TOAD, MS Office, BTEQ, Teradata SQL Assistant
Other Tools: Teradata SQL Assistant, Toad 9.7/8.0, DB Visualizer 6.0, Microsoft Office, Microsoft Visio, Microsoft Excel, Microsoft Project
Project Execution Methodologies: Ralph Kimball and Bill Inmondatawarehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)
PROFESSIONAL EXPERIENCE
Confidential - Chicago, IL
Sr. Data Engineer
Responsibilities:
- As a Data Engineer, assisted in leading teh plan, building, and running states wifin teh Enterprise Analytics Team.
- Assisted in leading teh plan, building, and running states wifin teh Enterprise Analytics Team.
- Understand current Production state of application and determine teh impact of new implementation on existing business processes.
- Involved in migrating teh client data warehouse architecture from on-premises into Azure cloud.
- Worked wif team on making decisions on how to migrate teh data from on prem to cloud, which tools can be used for ETL or ELT on cloud.
- Involved in migrating teh client data warehouse architecture from on-premises into Azure cloud.
- Designed and Configured Azure Cloud relational servers and databases analyzing current and future business requirements.
- Engaged in solving and supporting real business issues wif you're Hadoop distributed File systems and Open Source framework knowledge.
- Responsible for data governance rules and standards to maintain teh consistency of teh business element names in teh different data layers.
- Involved in various phases of development analyzed and developed teh system going through Agile Scrum methodology.
- Performed detailed analysis of business problems and technical environments and use this data in designing teh solution and maintaining data architecture.
- Built teh data pipelines that will enable faster, better, data-informed decision-making wifin teh business.
- Developed Spark scripts by using python and bash Shell commands as per teh requirement.
- Implemented Azure Data Factory operations and deployment into Azure for moving data from on-premise into cloud.
- Used Azure Data Factory extensively for ingesting data from disparate source systems.
- Analyze, design and build Modern data solutions using Azure PaaS service to support visualization of data.
- To meet specific business requirements wrote UDF’s in Scala and Pyspark.
- Designed and developed ETL pipeline in Azure cloud which gets customer data from API and process it to Azure SQL DB.
- Orchestrated all Data pipelines using Azure Data Factory (V2) and built a custom alerts platform for monitoring.
- Created custom alerts queries in Log Analytics and used Web hook actions to automate custom alerts.
- Created Azure Databricks Job workflows which extract data from SQL server and upload teh files to SFTP using Pyspark and python.
- Used Azure Key vault as central repository for maintaining secrets and referenced teh secrets in Azure Data Factory and also in Databricks notebooks.
- Designed of process oriented UNIX script and ETL processes for loading data into data warehouse.
- Working wif Azure BLOB and Data lake storage and loading data into Azure SQL Synapse analytics (DW).
- Developed JSON Scripts for deploying teh Pipeline in Azure Data Factory (ADF) that process teh data using teh Sql Activity.
- Working wif complex SQL, Stored Procedures, Triggers, and packages in large databases from various servers.
- Successfully Generated consumer group lags from Kafka using their API.
- Created workflows to automate teh Sqoop jobs and data ingestion into teh HDFS.
- Worked on Azure Power BI Embedded to integrate teh reports to application.
- Used Azure reporting services to upload and download reports.
- Helping team member to resolve any technical issue, Troubleshooting, Project Risk & Issue identification and management
- Addressing resource issue, Monthly one on one, Weekly meeting.
Environment: Hadoop 3.1, Spark 3.0, Azure cloud, ADF V2, Azure Sql DB, Power BI, ETL, ELT, PySpark 2.4, Python 2.7, JOSN, Kafka, Sqoop, HDFS and Agile methodology.
Confidential - Juno Beach, FL
Lead Data Architect
Responsibilities:
- As a Data Architect designed infrastructure architecture based on hub and spoke model for Big Data requirements.
- Developed detailed source to target mapping from client data to our data specifications and explain teh mapping to internal technical team members.
- Implemented multi-factor authentication solutions that ca help to secure Office 365 by providing an account password and another form of authentication.
- Key member of MDM implementation, responsible for detailed data analysis, Data profiling, data standardization define Master data domain.
- Responsible for estimating teh cluster size, monitoring and troubleshooting of teh Spark Databricks cluster.
- Managed users access onto environments based on roles, access policy and resource groups.
- Provided high availability for IAAS and PAAS role instances for access from other services in teh VNet wif Azure Internal Load Balancer.
- Performed data profiling and transformation on teh raw data usingPython.
- Developed long term data warehouse roadmap and architectures, designs and builds teh data warehouse framework per teh roadmap.
- Created Azure Data Factory pipeline to insert teh flat file, ORC file data into Azure SQL.
- Architect and design data integration interface into Azure DataLake and SQLDW.
- Used U-SQL to interact multiple source streams wif in Azure Data Lake.
- Implemented high availability wif Azure Classic and Azure Resource Manager deployment models.
- Worked wif security team for implementing security control over all teh environments of teh project.
- Designed Microservices based architecture on Azure Service Fabric cluster for lead management system to replace existing sales tool.
- Worked on migrating of EDW to AWS using EMR and various other technologies
- Performed ETLSQL optimization designed OLTP system environment and maintained documentation of Metadata.
- Create rich web user interfaces using HTML, CSS, and JavaScript
- Defined data model for each confidential service for managing information.
- Configure Azure Active Directory (Azure AD) Application Proxy connectors to work wif outbound proxy servers.
- Extract Transform and Load data from Sources Systems to Azure Data Storage services.
- Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Azure SQL, Azure DW, Azure APP) and processing teh data in InAzure Databricks.
- Actively participate in agile iteration planning, scrum meeting, sprint review in iterative &incremental agile delivery framework
- Developed DB tables and Sql script for displaying required information.
- Transformed data using Hadoop Streaming activity in Azure Data Factory
- Developed Pipelines to load data from on prem to AZURE cloud database.
- Developed Pipelines in Azure data factory using copy activity, Notebook, Hive, U-SQL to load data.
- Developed Pipelines in Azure data factory to call Notebooks to transform data for reporting and analytics.
- Reports are developed on Power BI on top of Views in Azure SQL.
- Scheduled Pipelines in Azure pipeline.
- Designed various Jenkins jobs to continuously integrate teh processes and executed CI/CD pipeline using Jenkins,Ansible Playbooks and Ansible Tower.
- Designed and developed a Data Lake using Hadoop for processing raw and processed data via Hive.
- Implemented cluster security using certificates and configured client access based on RBAC (Role Based Access Control).
- Deployed Azure IaaS virtual machines (VMs) and Cloud services (PaaS role instances) into secure VNets and subnets.
- Created Web API for fetching teh data from database and pass to UI Layer.
- Assigned Projects, Deadline Creation, Meetings wif Clients, Code creation and Test Plans.
Environment: IAAS, PAAS, Azure Data Lake, Azure Data Factory, AWS, SQL, Python, Power BI, Microservices, U-Sql, Hive, Hadoop
Confidential, Chicago, IL
Lead Data Architect
Responsibilities:
- Managed technical team of 6 members, overall data/application architecture & application performance and technical/audit items delivery.
- Developed Data Mapping, Data profiling, Data Governance, transformation and cleansing rules for teh Master Data Management Architecture involving OLTP, ODS.
- Data Governance of RAW, Staging and Presentation Layers in Azure Data Lake Store.
- Migrated AIM Data Warehouse applications to cloud environment using IaaS.
- Performed Proof of concept (POC) for AIM Data warehouse application by comparing performance wif new multiplexing features and wifout multiplexing on different versions.
- Design Setup maintain Administrator teh Azure SQL Database, Azure Analysis Service, Azure SQL Data warehouse, Azure Data Factory, Azure SQL Data warehouse
- Written SQL scripts to test teh mappings and Developed Traceability Matrix of Business Requirements.
- Involved in extensive Data validation by writing several complex SQL queries and Involved in back-end testing and worked wif data quality issues
- Designing and implementing data pipelines on Hadoop and NoSQL platforms to enable rapid prototyping and accelerated production
- Created Mappings, Tech Design, loading strategies for ETL to load newly created or existing tables.
- Implemented Hadoop ecosystem for scale and performance.
- Ingested relational data to HDFS using Sqoop and benchmarked performance to notice 75% performance gains and potential cost saving by economies of scale wif future expected exponential data growth.
- Architected IaaS solutions and Migrated MongoDB servers to cloud to provide infrastructure savings.
- Built Data Lake in Azure using Hadoop (HDInsight clusters) and migrated Data using Azure Data Factory pipeline.
- Designed Lambda architecture to process streaming data using Spark. Data was ingested using Sqoop for structured data and Kafka for unstructured data.
- Creation Azure Event Hubs, Azure Service Bus, Azure Service Analysis, Azure APP service, Power BI for handling IOT Messages.
- Responsible to develop advance PL/SQL packages, procedures, triggers, functions, Collections to implement business logic.
- Created packages, created tables, indexes, constraints, PL/SQL tables, sequences, synonyms and views.
- Created jobs, alerts to run SSIS, SSRS packages periodically.
- Performed data analysis, statistical analysis, generated reports, listings and graphs using SAS Integration Studio, SAS/Graph and SAS/SQL.
- Generated parameterized queries for generating tabular reports using global variables, expressions, functions, and stored procedures using SSRS.
- Used SSRS to create reports, customized Reports, on-demand reports, ad-hoc reports and involved in analyzing multi-dimensional reports in SSRS.
Environment: OLTP, IaaS, SQL, HDFS, Sqoop 1.4, MongoDB, MS SQL Server 2014, Azure, Hadoop 3.0, Spark 2.4, Kafka 2.2, Power BI, PL/SQL, SSIS, SSRS, SAS
Confidential - Philadelphia, PA
Sr. Data Architect/Data Modeler
Responsibilities:
- Lead teh strategy, architecture and process improvements for data architecture and data management, balancing long and short-term needs of teh business.
- Provide data architecture support to enterprise data management efforts, such as teh development of teh enterprise data model, Master and Reference data.
- Implemented logical and physical relational database and maintained Database Objects in teh data model using Erwin9.6
- Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
- Participated in integration of MDM (MasterDataManagement) Hub anddatawarehouses.
- Responsible of designing teh whole data warehouse system including ODS, DWH and data marts.
- Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
- Responsible for full data loads from production to AWS Redshift staging environment.
- Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
- Used Normalization methods up to 3NF and De-normalization techniques for TEMPeffective performance in OLTP systems.
- Extensively using Agile methodology as teh Organization Standard to implement thedataModels.
- Recreating existing application logic and functionality in teh Azure Data Lake, Data Factory, SQL Database and SQL data warehouse environment. experience in DWH/BI project implementation using Azure DF
- Performed theDataMapping,Datadesign (DataModeling) to integrate thedataacross teh multiple databases in to EDW.
- Developed and configured on Informatica MDM hub supports teh Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
- Used Load utilities (Fast Load &Multi Load) wif teh mainframe interface to load teh data into Teradata.
- Data reconciliation activities between Source and EDW Teradata databases.
- Involved in designing and developing Data Models and Data Marts that support teh Business Intelligence Data Warehouse.
- Handled importing data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
- Involved in Planning, Defining and Designing data base using Erwin on business requirement and provided documentation.
- Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections.
- Designed and created queries to extract teh desired data and runs teh report. Formatted teh data and validated teh results.
- Transforming staging area data into a STAR schema (hosted on AmazonRedshift) which was then used for developing embedded Tableau dashboards
- Translating business requirements into SAS code for use wifin internal systems and models
- Building relationships and trust wif key stakeholders to support program delivery and adoption of enterprise architecture.
- Maintained metadata (datadefinitions of table structures) and version controlling for thedatamodel.
- Created stored procedures, functions, database triggers and packages as per teh business needs for developing ad-hoc and robust reports.
- Defined best practices fordatamodeling and extraction and ensure architectural alignment of teh designs and development.
- Used Spark Data frames, Spark-SQL, Spark MLLib extensively and developing and designing POC's using Scala, SparkSQL and MLlib libraries.
- Working in Hadoop Environment using pig, Sqoop, Hive, HBase and detailed understanding of map reduce programs
- Involved in integration of various relational and non-relational sources such as Oracle, XML and Flat Files.
- Developed SQL scripts for creating tables, Sequences, Triggers, views and materialized views.
- Scheduling theSSISpackages execution in SQL Server Agent and tracking teh success or failure of teh execution ofSSISPackages and configuring teh email notifications through SQL Server Agent.
- Developed multiple MapReduce jobs in java for Data Cleaning and pre-processing analyzing data in PIG.
- Worked in Data Analysis, data profiling and data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
- Developed Data Mapping, Data Governance, and Transformation and cleansing rules for teh Master Data Management Architecture involving OLTP, ODS.
- Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, and Oracle)
- Coordinating wif DBA team to implement physical models & to setup development, test, staging & production environments for DDL&DML Generation & usage.
Environment: Erwin 9.6, MDM, AWS, Hadoop, Agile, Spark, Sqoop, Hive, HDFS, OLAP, OLTP, ODS, PL/SQL, Tableau, Teradata R15, Teradata SQL Assistant 15.0,Oracle 12c, SQL, PL/SQL, Flat Files.
Confidential - Union, NJ
Sr. Data Architect/Data Modeler
Responsibilities:
- Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
- Participated in teh design, development, and support of teh corporate operation data store and enterprise data warehouse database environment.
- Designed teh Logical Data Model using ER/Studio9.0 wif teh entities and attributes for each subject area.
- Used teh Agile Scrum methodology to build teh different phases of Software development life cycle.
- Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
- Migration of on premise data (Oracle/MongoDB) to Azure Data Lake Store(ADLS) using Azure Data Factory
- Independently coded new programs and design Tables to load and test teh program TEMPeffectively for teh given POC's using Big Data/Hadoop.
- Translating high level business requirements into solution, infrastructure architectures, involved enterprise architects to align strategic view.
- Worked on ER Studio for multiple Operations across in both OLAP and OLTP applications.
- Involved in debugging and Tuning teh PL/SQL code, tuning queries, optimization for teh Oracle, and DB2 database
- Developed MapReduce programs to cleanse teh data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
- Designed teh Fact and Dimension table for Data Marts using ER Studio 9.0.
- Design of ODS layer, Dimensional modeling using Kimball Methodologies, of teh Data Warehouse sourced from MDM Base tables and other Transactional systems.
- Designed and produced client reports using Excel, Access, Tableau and SAS.
- Created logical and physical data model using Cassandra’s model
- Developed Data Mapping, Data Governance, Transformation and Cleansing rules for teh Master Data Management Architecture involving OLTP, ODS and OLAP.
- Extracting Mega Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports
- Worked wif teh ETL team to document teh transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
- Worked wif Netezza and Oracle databases and implemented various logical and physical data models for them.
- Generated comprehensive analytical reports by running SQL queries against current databases to conductdataanalysis.
- Developed complex SQL scripts for Teradata database for creating BI layer on DW for Tableau reporting.
- Worked on debugging and identifying teh unexpected real-time issues in teh production serverSSIS packages.
- Involved in mapping teh data elements from teh User Interface to teh Database and help identify teh gaps.
- Prepared ofDataDictionary/MetaDataof thedataelements.
- Developed and configured on Informatica MDM hub supports teh Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
- Generated preliminary Source to Target mapping by writing a ER/Studio Macro in teh ER/Studio VBA platform document and handed over to BAs for further transformation analysis.
- Submitted teh DDL and Security document by writing a ER Studio Macro to teh SVN repository and communicated to teh DBA for implementation.
- Designed and Developed Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing
- Worked extensively wif Business Objects XI Report Developers in solving critical issues of defining hierarchy, loops and Contexts.
Environment: ER/Studio 9.0, Oracle 11g, DB2, Agile, MDM AWS, Hadoop, Hive, Netezza, Sqoop, Spark, HDFS, MapReduce, OLAP, OLTP, PL/SQL, Flat Files.
Confidential - Chicago, IL
Sr. Data Analyst/Data Modeler
Responsibilities:
- Created Physical Data Analyst from teh Logical Data Analyst using Compare and Merge Utility in ER Studio and worked wif teh naming standards utility.
- Developed normalized Logical and Physical database models for designing an OLTP application.
- Extensively used Star Schema methodologies in building and designing teh logicaldatamodel into Dimensional Models
- Creation of database objects like tables, views, Materialized views, procedures, packages using Oracle tools like PL/SQL, SQL*Loader and Handled Exceptions.
- Enforced referential integrity in teh OLTP data model for consistent relationship between tables and efficient database design.
- Worked wif data investigation, discovery and mapping tools to scan every single data record from many sources.
- Utilized SDLC and Agile methodologies such as SCRUM.
- Involved in administrative tasks, including creation of database objects such as database, tables, and views, using SQL, DDL, and DML requests.
- Worked on Data Analysis, Data profiling, and Data Modeling, data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
- Loaded multi format data from various sources like flat-file, Excel, MS Access and performing file system operation.
- Used T-SQL stored procedures to transfer data from OLTP databases to staging area and finally transfer into data marts.
- Worked on Physical design for both SMP and MPPRDBMS, wif understanding of RDMBS scaling features.
- Wrote SQL Queries, Dynamic-queries, sub-queries and complex joins for generating Complex Stored Procedures, Triggers, User-defined Functions, Views and Cursors.
- Wrote simple and advanced SQL queries and scripts to create standard and ad hoc reports for senior managers.
- Involved wifDataAnalysis primarily IdentifyingDataSets, SourceData, Source MetaData, Data Definitions andDataFormats
- Worked wif developers on data Normalization and De-normalization, performance tuning issues, and provided assistance in stored procedures as needed.
- Used Teradata for OLTP systems by generating models to support Revenue Management Applications that connect to SAS.
- Created SSIS Packages for import and export of data between Oracle database and others like MS Excel and Flat Files.
- Worked in teh capacity of ETL Developer (Oracle Data Integrator (ODI) / PL/SQL) to migrate data from different sources in to target Oracle Data Warehouse.
- Designed and Developed PL/SQL procedures, functions and packages to create Summary tables.
- Involved in creating tasks to pull and push data from Salesforce to Oracle Staging/Data Mart.
- Created VBA Macros to convert teh Excel Input files in to correct format and loaded them to SQL Server.
- Helped teh BI, ETL Developers in understanding teh Data Model, data flow and teh expected output for each model created
Environment: ER/Studio 8.0, Oracle 10g Application Server, Agile, Oracle Developer Suite, PL/SQL, T-SQL, SQL Plus, SSIS, Teradata 13, OLAP, OLTP, SAS, MS Excel.
Confidential -Trenton, NJ
Data Analyst/Data Modeler
Responsibilities:
- Gatheird Business requirements by organizing and managing meetings wif business stake holders, Application architects, Technical architects and IT analysts on a scheduled basis
- Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
- Performed Normalization, Indexes Tuned and Optimized teh existing database design.
- Performed reverse engineering using teh Erwin Data Modeling tool.
- Performed various tasks that included data massaging, data extraction from flat files, data transformations into various file formats, validations, loading data and checking for any data integrity errors.
- Responsible for teh development and maintenance of Logical and Physical data models, along wif corresponding metadata, to support Applications.
- Conducted meetings wif teh business and technical team to gather necessary analytical data requirements in JAD sessions.
- Analyzed all existing SSIS packages, SQL Server objects & new functional specs.
- Identified and tracked teh slowly changing dimensions, heterogeneous sources and determined teh hierarchies in dimensions.
- Worked extensively wif MicroStrategy Report developers in creating data marts and develop reports
- Created databases for OLAP Metadata catalog tables using forward engineering of models in Erwin.
- Enforced referential integrity in teh OLTPdatamodel for consistent relationship between tables and efficient database design.
- Involved in importing/exporting large amounts ofdatafrom files to Teradata and vice versa.
- Designed Mapping Documents and Mapping Templates for Data Stage ETL developer.
- Designed, Build teh Dimensions, cubes wif Star schema and Snow Flake Schema using SQL Server Analysis Services (SSAS).
- Wrote PL/SQL statement, stored procedures and Triggers in DB2 for extracting as well as writing data.
- Created a Data Mapping document after each assignment and wrote teh transformation rules for each field as applicable
- Analyzed OLTP source systems and Operational Data Store and research teh tables/entities required for teh project. Designing teh measures, dimensions and facts matrix document for teh ease while designing.
- Tuning all database via indexing of tables, MSSQLServer2005 configuration parameters and stored procedures SQLcode optimization.
- Used forward engineering to create a physical data model wif DDL that best suits teh requirements from teh Logical Data Model.
Environment: Erwin 7.3, SSIS, Metadata, OLAP, OLTP, Data Mart, PL/SQL, T-SQL, DB2, SQL Plus, Microsoft SQL Server 2005, Flat files.