We provide IT Staff Augmentation Services!

Sr. Data Engineer Resume

3.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY

  • Over 12+ years of extensive experience in Data Architecture and Data Engineering wif Development, Testing and Deployment of business applications.
  • Experienced in Data Modeling &Data Analysis experience using Dimensional Data Modeling and Relational Data Modeling, Star Schema/Snowflake Modeling, FACT & Dimensions tables, Physical & Logical Data Modeling.
  • Experience wif distributed data warehousing and/or data mining systems, using one or more Big Data/NoSQL technologies (Hadoop, Hive, HBase, Pig, Cassandra, MongoDB)
  • Experience in importing and exporting data using Sqoop from HDFS to Relational Database Systems (RDBMS) and from RDBMS to HDFS.
  • Experience in Technical consulting and end - to-end delivery wif architecture, data modeling, data governance and design - development - implementation of solutions.
  • Extensive experience in developing and driving strategic direction of SAP operating system (SAPECC) and SAP business intelligence (SAPBI) system
  • Working experience of Databricks or similar
  • Experience in designing, building and implementing complete Hadoop ecosystem comprising of MapReduce, HDFS, Hive, Impala, Pig, Sqoop, Oozie, HBase, MongoDB, and Spark.
  • Experience in various Teradata utilities like Fastload, Multiload, BTEQ, and Teradata SQL Assistant.
  • Experience in working wif Business Intelligence and Enterprise Data Warehouse(EDW) including SSAS, Pentaho, Cognos, OBIEE, QlikView, Greenplum, Amazon Redshift and Azure Data Warehouse
  • Excellent Software Development Life Cycle (SDLC) wif good working knowledge of testing methodologies, disciplines, tasks, resources and scheduling.
  • Experienced in Data Scrubbing/Cleansing, Data Quality, Data Mapping, Data Profiling, Data Validation in ETL
  • Experienced in creating and documenting Metadata for OLTP and OLAP when designing a systems.
  • Performeddataanalysis anddataprofiling using complex SQL on various sources systems including Oracle and Teradata.
  • Experience in building CI/CD methodology in Azure using technologies like VS online, Jenkins, TFS etc
  • Excellent Knowledge of Ralph Kimball and Bill Inmon's approaches toDataWarehousing.
  • Excellent in performing data transfer activities between SAS and various databases and data file formats like XLS,CSV, DBF,MDB etc.
  • Experience in architecting hybrid solutions, cloud integration wif on premise environment (Azure app service environment, express route)
  • Experience in working wif Excel Pivot and VBA macros for various business scenarios.
  • Experience in Data modeling using ER diagram, Dimensional data modeling, Conceptual/Logical/Physical Modeling using 3NormalForm (3NF), Star Schema modeling, Snowflake modeling using tools like ER/Studio, Erwin, Sybase PowerDesigner for both forward and reverse engineering.
  • Expert in BI reporting and Data Reporting tools like Pentaho and SAP BI.
  • Excellent experience in troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Proficient wifDataAnalysis, mapping source and target systems fordatamigration efforts and resolving issues relating todatamigration.
  • Excellent understanding and working experience of industry standard methodologies like System Development Life Cycle (SDLC), as per Rational Unified Process (RUP), AGILE Methodologies.
  • Advanced expertise in visualization design and development, SQL and Oracle/SQL Server.
  • Expert in building Enterprise Data Warehouse or Data warehouse appliances from Scratch using both Kimball and Inmon Approach.
  • Experience in various databases Design of development and Production environment involving Oracle, SQL server, Netezza, MYSQL, DB2, MS Access and Teradata etc.
  • Hands on experience in Normalization &De-normalization techniques design considerations upto3NF Form for OLTP Databases and Models.
  • Strong experience in writing SQL and PL/SQL, Transact SQL programs for Stored Procedures, Triggers and Functions.
  • Expertise in analyzing and documenting business requirement documents (BRD) and functional requirement documents (FRD) along wif Use Case Modeling and UML.
  • Experience in UNIX shell scripting, Perl scripting and automation of ETL Processes.
  • Expertise in designing complex Mappings and has expertise in performance tuning and slowly-changing Dimension Tables and Fact tables
  • Extensively worked wif Teradata utilities BTEQ, Fast Export and Multi Load to export and load data to/from different source systems including flat files.

TECHNICAL SKILLS

Data Modeling Tools: Erwin 9.6/9.5, ER/Studio 9.7/9.0, Sybase Power DesignerBig Data Technologies: Hadoop, Hive, HDFS, HBase, Flume, Sqoop, Spark, Pig, Impala, MapReduce.

Programming Languages: SQL, PL/SQL, UNIX shell Scripting, PERL, AWK, SED

Databases: Oracle 12c/11g, Teradata R15/R14, MS SQL Server 2014/2016Testing and defect tracking Tools: HP/Mercury (Quality Center, Win Runner, Quick Test Professional, Performance Center, Requisite, MS Visio & Visual Source Safe

Operating System: Windows, Unix, Sun Solaris

ETL/Datawarehouse Tools: Informatica 9.6/9.1, SAP Business Objects XIR3.1/XIR2, Web Intelligence, Talend, Tableau, Pentaho

Tools: & Software: TOAD, MS Office, BTEQ, Teradata SQL Assistant

Other Tools: Teradata SQL Assistant, Toad 9.7/8.0, DB Visualizer 6.0, Microsoft Office, Microsoft Visio, Microsoft Excel, Microsoft Project

Project Execution Methodologies: Ralph Kimball and Bill Inmondatawarehousing methodology, Rational Unified Process (RUP), Rapid Application Development (RAD), Joint Application Development (JAD)

PROFESSIONAL EXPERIENCE

Confidential - Chicago, IL

Sr. Data Engineer

Responsibilities:

  • As a Data Engineer, assisted in leading teh plan, building, and running states wifin teh Enterprise Analytics Team.
  • Assisted in leading teh plan, building, and running states wifin teh Enterprise Analytics Team.
  • Understand current Production state of application and determine teh impact of new implementation on existing business processes.
  • Involved in migrating teh client data warehouse architecture from on-premises into Azure cloud.
  • Worked wif team on making decisions on how to migrate teh data from on prem to cloud, which tools can be used for ETL or ELT on cloud.
  • Involved in migrating teh client data warehouse architecture from on-premises into Azure cloud.
  • Designed and Configured Azure Cloud relational servers and databases analyzing current and future business requirements.
  • Engaged in solving and supporting real business issues wif you're Hadoop distributed File systems and Open Source framework knowledge.
  • Responsible for data governance rules and standards to maintain teh consistency of teh business element names in teh different data layers.
  • Involved in various phases of development analyzed and developed teh system going through Agile Scrum methodology.
  • Performed detailed analysis of business problems and technical environments and use this data in designing teh solution and maintaining data architecture.
  • Built teh data pipelines that will enable faster, better, data-informed decision-making wifin teh business.
  • Developed Spark scripts by using python and bash Shell commands as per teh requirement.
  • Implemented Azure Data Factory operations and deployment into Azure for moving data from on-premise into cloud.
  • Used Azure Data Factory extensively for ingesting data from disparate source systems.
  • Analyze, design and build Modern data solutions using Azure PaaS service to support visualization of data.
  • To meet specific business requirements wrote UDF’s in Scala and Pyspark.
  • Designed and developed ETL pipeline in Azure cloud which gets customer data from API and process it to Azure SQL DB.
  • Orchestrated all Data pipelines using Azure Data Factory (V2) and built a custom alerts platform for monitoring.
  • Created custom alerts queries in Log Analytics and used Web hook actions to automate custom alerts.
  • Created Azure Databricks Job workflows which extract data from SQL server and upload teh files to SFTP using Pyspark and python.
  • Used Azure Key vault as central repository for maintaining secrets and referenced teh secrets in Azure Data Factory and also in Databricks notebooks.
  • Designed of process oriented UNIX script and ETL processes for loading data into data warehouse.
  • Working wif Azure BLOB and Data lake storage and loading data into Azure SQL Synapse analytics (DW).
  • Developed JSON Scripts for deploying teh Pipeline in Azure Data Factory (ADF) that process teh data using teh Sql Activity.
  • Working wif complex SQL, Stored Procedures, Triggers, and packages in large databases from various servers.
  • Successfully Generated consumer group lags from Kafka using their API.
  • Created workflows to automate teh Sqoop jobs and data ingestion into teh HDFS.
  • Worked on Azure Power BI Embedded to integrate teh reports to application.
  • Used Azure reporting services to upload and download reports.
  • Helping team member to resolve any technical issue, Troubleshooting, Project Risk & Issue identification and management
  • Addressing resource issue, Monthly one on one, Weekly meeting.

Environment: Hadoop 3.1, Spark 3.0, Azure cloud, ADF V2, Azure Sql DB, Power BI, ETL, ELT, PySpark 2.4, Python 2.7, JOSN, Kafka, Sqoop, HDFS and Agile methodology.

Confidential - Juno Beach, FL

Lead Data Architect

Responsibilities:

  • As a Data Architect designed infrastructure architecture based on hub and spoke model for Big Data requirements.
  • Developed detailed source to target mapping from client data to our data specifications and explain teh mapping to internal technical team members.
  • Implemented multi-factor authentication solutions that ca help to secure Office 365 by providing an account password and another form of authentication.
  • Key member of MDM implementation, responsible for detailed data analysis, Data profiling, data standardization define Master data domain.
  • Responsible for estimating teh cluster size, monitoring and troubleshooting of teh Spark Databricks cluster.
  • Managed users access onto environments based on roles, access policy and resource groups.
  • Provided high availability for IAAS and PAAS role instances for access from other services in teh VNet wif Azure Internal Load Balancer.
  • Performed data profiling and transformation on teh raw data usingPython.
  • Developed long term data warehouse roadmap and architectures, designs and builds teh data warehouse framework per teh roadmap.
  • Created Azure Data Factory pipeline to insert teh flat file, ORC file data into Azure SQL.
  • Architect and design data integration interface into Azure DataLake and SQLDW.
  • Used U-SQL to interact multiple source streams wif in Azure Data Lake.
  • Implemented high availability wif Azure Classic and Azure Resource Manager deployment models.
  • Worked wif security team for implementing security control over all teh environments of teh project.
  • Designed Microservices based architecture on Azure Service Fabric cluster for lead management system to replace existing sales tool.
  • Worked on migrating of EDW to AWS using EMR and various other technologies
  • Performed ETLSQL optimization designed OLTP system environment and maintained documentation of Metadata.
  • Create rich web user interfaces using HTML, CSS, and JavaScript
  • Defined data model for each confidential service for managing information.
  • Configure Azure Active Directory (Azure AD) Application Proxy connectors to work wif outbound proxy servers.
  • Extract Transform and Load data from Sources Systems to Azure Data Storage services.
  • Data Ingestion to one or more Azure Services - (Azure Data Lake, Azure Storage, Azure SQL, Azure DW, Azure APP) and processing teh data in InAzure Databricks.
  • Actively participate in agile iteration planning, scrum meeting, sprint review in iterative &incremental agile delivery framework
  • Developed DB tables and Sql script for displaying required information.
  • Transformed data using Hadoop Streaming activity in Azure Data Factory
  • Developed Pipelines to load data from on prem to AZURE cloud database.
  • Developed Pipelines in Azure data factory using copy activity, Notebook, Hive, U-SQL to load data.
  • Developed Pipelines in Azure data factory to call Notebooks to transform data for reporting and analytics.
  • Reports are developed on Power BI on top of Views in Azure SQL.
  • Scheduled Pipelines in Azure pipeline.
  • Designed various Jenkins jobs to continuously integrate teh processes and executed CI/CD pipeline using Jenkins,Ansible Playbooks and Ansible Tower.
  • Designed and developed a Data Lake using Hadoop for processing raw and processed data via Hive.
  • Implemented cluster security using certificates and configured client access based on RBAC (Role Based Access Control).
  • Deployed Azure IaaS virtual machines (VMs) and Cloud services (PaaS role instances) into secure VNets and subnets.
  • Created Web API for fetching teh data from database and pass to UI Layer.
  • Assigned Projects, Deadline Creation, Meetings wif Clients, Code creation and Test Plans.

Environment: IAAS, PAAS, Azure Data Lake, Azure Data Factory, AWS, SQL, Python, Power BI, Microservices, U-Sql, Hive, Hadoop

Confidential, Chicago, IL

Lead Data Architect

Responsibilities:

  • Managed technical team of 6 members, overall data/application architecture & application performance and technical/audit items delivery.
  • Developed Data Mapping, Data profiling, Data Governance, transformation and cleansing rules for teh Master Data Management Architecture involving OLTP, ODS.
  • Data Governance of RAW, Staging and Presentation Layers in Azure Data Lake Store.
  • Migrated AIM Data Warehouse applications to cloud environment using IaaS.
  • Performed Proof of concept (POC) for AIM Data warehouse application by comparing performance wif new multiplexing features and wifout multiplexing on different versions.
  • Design Setup maintain Administrator teh Azure SQL Database, Azure Analysis Service, Azure SQL Data warehouse, Azure Data Factory, Azure SQL Data warehouse
  • Written SQL scripts to test teh mappings and Developed Traceability Matrix of Business Requirements.
  • Involved in extensive Data validation by writing several complex SQL queries and Involved in back-end testing and worked wif data quality issues
  • Designing and implementing data pipelines on Hadoop and NoSQL platforms to enable rapid prototyping and accelerated production
  • Created Mappings, Tech Design, loading strategies for ETL to load newly created or existing tables.
  • Implemented Hadoop ecosystem for scale and performance.
  • Ingested relational data to HDFS using Sqoop and benchmarked performance to notice 75% performance gains and potential cost saving by economies of scale wif future expected exponential data growth.
  • Architected IaaS solutions and Migrated MongoDB servers to cloud to provide infrastructure savings.
  • Built Data Lake in Azure using Hadoop (HDInsight clusters) and migrated Data using Azure Data Factory pipeline.
  • Designed Lambda architecture to process streaming data using Spark. Data was ingested using Sqoop for structured data and Kafka for unstructured data.
  • Creation Azure Event Hubs, Azure Service Bus, Azure Service Analysis, Azure APP service, Power BI for handling IOT Messages.
  • Responsible to develop advance PL/SQL packages, procedures, triggers, functions, Collections to implement business logic.
  • Created packages, created tables, indexes, constraints, PL/SQL tables, sequences, synonyms and views.
  • Created jobs, alerts to run SSIS, SSRS packages periodically.
  • Performed data analysis, statistical analysis, generated reports, listings and graphs using SAS Integration Studio, SAS/Graph and SAS/SQL.
  • Generated parameterized queries for generating tabular reports using global variables, expressions, functions, and stored procedures using SSRS.
  • Used SSRS to create reports, customized Reports, on-demand reports, ad-hoc reports and involved in analyzing multi-dimensional reports in SSRS.

Environment: OLTP, IaaS, SQL, HDFS, Sqoop 1.4, MongoDB, MS SQL Server 2014, Azure, Hadoop 3.0, Spark 2.4, Kafka 2.2, Power BI, PL/SQL, SSIS, SSRS, SAS

Confidential - Philadelphia, PA

Sr. Data Architect/Data Modeler

Responsibilities:

  • Lead teh strategy, architecture and process improvements for data architecture and data management, balancing long and short-term needs of teh business.
  • Provide data architecture support to enterprise data management efforts, such as teh development of teh enterprise data model, Master and Reference data.
  • Implemented logical and physical relational database and maintained Database Objects in teh data model using Erwin9.6
  • Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
  • Participated in integration of MDM (MasterDataManagement) Hub anddatawarehouses.
  • Responsible of designing teh whole data warehouse system including ODS, DWH and data marts.
  • Used ETL methodology for supporting data extraction, transformations and loading processing, in a complex MDM using Informatica.
  • Responsible for full data loads from production to AWS Redshift staging environment.
  • Generated ad-hoc SQL queries using joins, database connections and transformation rules to fetch data from Teradata database.
  • Used Normalization methods up to 3NF and De-normalization techniques for TEMPeffective performance in OLTP systems.
  • Extensively using Agile methodology as teh Organization Standard to implement thedataModels.
  • Recreating existing application logic and functionality in teh Azure Data Lake, Data Factory, SQL Database and SQL data warehouse environment. experience in DWH/BI project implementation using Azure DF
  • Performed theDataMapping,Datadesign (DataModeling) to integrate thedataacross teh multiple databases in to EDW.
  • Developed and configured on Informatica MDM hub supports teh Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Used Load utilities (Fast Load &Multi Load) wif teh mainframe interface to load teh data into Teradata.
  • Data reconciliation activities between Source and EDW Teradata databases.
  • Involved in designing and developing Data Models and Data Marts that support teh Business Intelligence Data Warehouse.
  • Handled importing data from various data sources, performed transformations using Hive, MapReduce, and loaded data into HDFS.
  • Involved in Planning, Defining and Designing data base using Erwin on business requirement and provided documentation.
  • Involved in database development by creating Oracle PL/SQL Functions, Procedures and Collections.
  • Designed and created queries to extract teh desired data and runs teh report. Formatted teh data and validated teh results.
  • Transforming staging area data into a STAR schema (hosted on AmazonRedshift) which was then used for developing embedded Tableau dashboards
  • Translating business requirements into SAS code for use wifin internal systems and models
  • Building relationships and trust wif key stakeholders to support program delivery and adoption of enterprise architecture.
  • Maintained metadata (datadefinitions of table structures) and version controlling for thedatamodel.
  • Created stored procedures, functions, database triggers and packages as per teh business needs for developing ad-hoc and robust reports.
  • Defined best practices fordatamodeling and extraction and ensure architectural alignment of teh designs and development.
  • Used Spark Data frames, Spark-SQL, Spark MLLib extensively and developing and designing POC's using Scala, SparkSQL and MLlib libraries.
  • Working in Hadoop Environment using pig, Sqoop, Hive, HBase and detailed understanding of map reduce programs
  • Involved in integration of various relational and non-relational sources such as Oracle, XML and Flat Files.
  • Developed SQL scripts for creating tables, Sequences, Triggers, views and materialized views.
  • Scheduling theSSISpackages execution in SQL Server Agent and tracking teh success or failure of teh execution ofSSISPackages and configuring teh email notifications through SQL Server Agent.
  • Developed multiple MapReduce jobs in java for Data Cleaning and pre-processing analyzing data in PIG.
  • Worked in Data Analysis, data profiling and data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
  • Developed Data Mapping, Data Governance, and Transformation and cleansing rules for teh Master Data Management Architecture involving OLTP, ODS.
  • Proficiency in SQL across a number of dialects (we commonly write MySQL, PostgreSQL, Redshift, and Oracle)
  • Coordinating wif DBA team to implement physical models & to setup development, test, staging & production environments for DDL&DML Generation & usage.

Environment: Erwin 9.6, MDM, AWS, Hadoop, Agile, Spark, Sqoop, Hive, HDFS, OLAP, OLTP, ODS, PL/SQL, Tableau, Teradata R15, Teradata SQL Assistant 15.0,Oracle 12c, SQL, PL/SQL, Flat Files.

Confidential - Union, NJ

Sr. Data Architect/Data Modeler

Responsibilities:

  • Developed a high performance, scalable data architecture solution that incorporates a matrix of technology to relate architectural decision to business needs.
  • Participated in teh design, development, and support of teh corporate operation data store and enterprise data warehouse database environment.
  • Designed teh Logical Data Model using ER/Studio9.0 wif teh entities and attributes for each subject area.
  • Used teh Agile Scrum methodology to build teh different phases of Software development life cycle.
  • Involved in several facets of MDM implementations including Data Profiling, Metadata acquisition and data migration.
  • Migration of on premise data (Oracle/MongoDB) to Azure Data Lake Store(ADLS) using Azure Data Factory
  • Independently coded new programs and design Tables to load and test teh program TEMPeffectively for teh given POC's using Big Data/Hadoop.
  • Translating high level business requirements into solution, infrastructure architectures, involved enterprise architects to align strategic view.
  • Worked on ER Studio for multiple Operations across in both OLAP and OLTP applications.
  • Involved in debugging and Tuning teh PL/SQL code, tuning queries, optimization for teh Oracle, and DB2 database
  • Developed MapReduce programs to cleanse teh data in HDFS obtained from heterogeneous data sources to make it suitable for ingestion into Hive schema for analysis.
  • Designed teh Fact and Dimension table for Data Marts using ER Studio 9.0.
  • Design of ODS layer, Dimensional modeling using Kimball Methodologies, of teh Data Warehouse sourced from MDM Base tables and other Transactional systems.
  • Designed and produced client reports using Excel, Access, Tableau and SAS.
  • Created logical and physical data model using Cassandra’s model
  • Developed Data Mapping, Data Governance, Transformation and Cleansing rules for teh Master Data Management Architecture involving OLTP, ODS and OLAP.
  • Extracting Mega Data from Amazon Redshift, AWS, and Elastic Search engine using SQL Queries to create reports
  • Worked wif teh ETL team to document teh transformation rules for data migration from OLTP to Warehouse environment for reporting purposes.
  • Worked wif Netezza and Oracle databases and implemented various logical and physical data models for them.
  • Generated comprehensive analytical reports by running SQL queries against current databases to conductdataanalysis.
  • Developed complex SQL scripts for Teradata database for creating BI layer on DW for Tableau reporting.
  • Worked on debugging and identifying teh unexpected real-time issues in teh production serverSSIS packages.
  • Involved in mapping teh data elements from teh User Interface to teh Database and help identify teh gaps.
  • Prepared ofDataDictionary/MetaDataof thedataelements.
  • Developed and configured on Informatica MDM hub supports teh Master Data Management (MDM), Business Intelligence (BI) and Data Warehousing platforms to meet business needs.
  • Generated preliminary Source to Target mapping by writing a ER/Studio Macro in teh ER/Studio VBA platform document and handed over to BAs for further transformation analysis.
  • Submitted teh DDL and Security document by writing a ER Studio Macro to teh SVN repository and communicated to teh DBA for implementation.
  • Designed and Developed Oracle PL/SQL and Shell Scripts, Data Import/Export, Data Conversions and Data Cleansing
  • Worked extensively wif Business Objects XI Report Developers in solving critical issues of defining hierarchy, loops and Contexts.

Environment: ER/Studio 9.0, Oracle 11g, DB2, Agile, MDM AWS, Hadoop, Hive, Netezza, Sqoop, Spark, HDFS, MapReduce, OLAP, OLTP, PL/SQL, Flat Files.

Confidential - Chicago, IL

Sr. Data Analyst/Data Modeler

Responsibilities:

  • Created Physical Data Analyst from teh Logical Data Analyst using Compare and Merge Utility in ER Studio and worked wif teh naming standards utility.
  • Developed normalized Logical and Physical database models for designing an OLTP application.
  • Extensively used Star Schema methodologies in building and designing teh logicaldatamodel into Dimensional Models
  • Creation of database objects like tables, views, Materialized views, procedures, packages using Oracle tools like PL/SQL, SQL*Loader and Handled Exceptions.
  • Enforced referential integrity in teh OLTP data model for consistent relationship between tables and efficient database design.
  • Worked wif data investigation, discovery and mapping tools to scan every single data record from many sources.
  • Utilized SDLC and Agile methodologies such as SCRUM.
  • Involved in administrative tasks, including creation of database objects such as database, tables, and views, using SQL, DDL, and DML requests.
  • Worked on Data Analysis, Data profiling, and Data Modeling, data governance identifying Data Sets, Source Data, Source Meta Data, Data Definitions and Data Formats.
  • Loaded multi format data from various sources like flat-file, Excel, MS Access and performing file system operation.
  • Used T-SQL stored procedures to transfer data from OLTP databases to staging area and finally transfer into data marts.
  • Worked on Physical design for both SMP and MPPRDBMS, wif understanding of RDMBS scaling features.
  • Wrote SQL Queries, Dynamic-queries, sub-queries and complex joins for generating Complex Stored Procedures, Triggers, User-defined Functions, Views and Cursors.
  • Wrote simple and advanced SQL queries and scripts to create standard and ad hoc reports for senior managers.
  • Involved wifDataAnalysis primarily IdentifyingDataSets, SourceData, Source MetaData, Data Definitions andDataFormats
  • Worked wif developers on data Normalization and De-normalization, performance tuning issues, and provided assistance in stored procedures as needed.
  • Used Teradata for OLTP systems by generating models to support Revenue Management Applications that connect to SAS.
  • Created SSIS Packages for import and export of data between Oracle database and others like MS Excel and Flat Files.
  • Worked in teh capacity of ETL Developer (Oracle Data Integrator (ODI) / PL/SQL) to migrate data from different sources in to target Oracle Data Warehouse.
  • Designed and Developed PL/SQL procedures, functions and packages to create Summary tables.
  • Involved in creating tasks to pull and push data from Salesforce to Oracle Staging/Data Mart.
  • Created VBA Macros to convert teh Excel Input files in to correct format and loaded them to SQL Server.
  • Helped teh BI, ETL Developers in understanding teh Data Model, data flow and teh expected output for each model created

Environment: ER/Studio 8.0, Oracle 10g Application Server, Agile, Oracle Developer Suite, PL/SQL, T-SQL, SQL Plus, SSIS, Teradata 13, OLAP, OLTP, SAS, MS Excel.

Confidential -Trenton, NJ

Data Analyst/Data Modeler

Responsibilities:

  • Gatheird Business requirements by organizing and managing meetings wif business stake holders, Application architects, Technical architects and IT analysts on a scheduled basis
  • Translated business requirements into working logical and physical data models for Data warehouse, Data marts and OLAP applications.
  • Performed Normalization, Indexes Tuned and Optimized teh existing database design.
  • Performed reverse engineering using teh Erwin Data Modeling tool.
  • Performed various tasks that included data massaging, data extraction from flat files, data transformations into various file formats, validations, loading data and checking for any data integrity errors.
  • Responsible for teh development and maintenance of Logical and Physical data models, along wif corresponding metadata, to support Applications.
  • Conducted meetings wif teh business and technical team to gather necessary analytical data requirements in JAD sessions.
  • Analyzed all existing SSIS packages, SQL Server objects & new functional specs.
  • Identified and tracked teh slowly changing dimensions, heterogeneous sources and determined teh hierarchies in dimensions.
  • Worked extensively wif MicroStrategy Report developers in creating data marts and develop reports
  • Created databases for OLAP Metadata catalog tables using forward engineering of models in Erwin.
  • Enforced referential integrity in teh OLTPdatamodel for consistent relationship between tables and efficient database design.
  • Involved in importing/exporting large amounts ofdatafrom files to Teradata and vice versa.
  • Designed Mapping Documents and Mapping Templates for Data Stage ETL developer.
  • Designed, Build teh Dimensions, cubes wif Star schema and Snow Flake Schema using SQL Server Analysis Services (SSAS).
  • Wrote PL/SQL statement, stored procedures and Triggers in DB2 for extracting as well as writing data.
  • Created a Data Mapping document after each assignment and wrote teh transformation rules for each field as applicable
  • Analyzed OLTP source systems and Operational Data Store and research teh tables/entities required for teh project. Designing teh measures, dimensions and facts matrix document for teh ease while designing.
  • Tuning all database via indexing of tables, MSSQLServer2005 configuration parameters and stored procedures SQLcode optimization.
  • Used forward engineering to create a physical data model wif DDL that best suits teh requirements from teh Logical Data Model.

Environment: Erwin 7.3, SSIS, Metadata, OLAP, OLTP, Data Mart, PL/SQL, T-SQL, DB2, SQL Plus, Microsoft SQL Server 2005, Flat files.

We'd love your feedback!