Senior Azure Data Engineer Resume
Pittsburgh, PA
SUMMARY
- 18+ years of experience in the analysis, design, development, testing, performance and documentation of Database and Client Server applications.
- Experience in data architecture designs.
- Expert in Data Warehouse development starting from inception to implementation and ongoing support, Strong
- Understanding of BI application design and development principles.
- Excellent command with MS SQL Server 2012/2014/2017 ; worked on Integration services (SSIS), Reporting services (SSRS) and Analysis services (SSAS).
- Experience in developingOLAP Cubesby using SQL Server Analysis Services (SSAS), and defined data sources, data source views, Dimensions, Measures, Hierarchies, Attributes.
- Hands on experience on Unified Data Analytics with Databricks, Databricks workspace user interface, and Managing Databricks Notebooks.
- Experienced in managing Azure Data Lakes (ADLS) and Data Lake Analytics and deep understanding in the Azure Services integration.
- Experience in data ingestion using Azure Data Factory from multiple sources like on - premises, Azure Data Lake, Azure SQL database, Azure Databricks and Azure Synapse Analytics.
- Good experience in creating data pipelines in the Azure Data Factory (ADF) using Linked Services, Datasets, Azure key vault.
- Deep understanding on the functioning of Azure data factory (ADF) objects- ADF activities, Data Flows, Pipelines, parameters, variables, Integration Runtime services, and Triggers.
- Implemented SCD1 and SCD2 using merge functionality and created a dimension framework using config files in JSON format.
- Experienced in orchestrating data pipelines using Apache Airflow to interact with services like Azure Databricks, Azure Data Factory, Azure Data Lake, and Azure Synapse Analytics
- Automated jobs using different triggers (Event, Scheduled and Tumbling) in ADF
- Good understanding of Spark Architecture including Spark Core, Spark SQL, Data Frames, Spark Streaming, Driver Node, Worker Node, Stages, Executors and Tasks.
- Expertise in developing spark application using Spark-SQL and PySpark in Databricks for data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.
- Expertise in establishing connectivity from Databricks to Azure Blob Storage and Data Lake Gen2.
- Experience with different file formats like Avro, parquet, ORC, JSON, XML, tsv and compressions like snappy, gzip etc.,
- Experience in writing ETL scripts using Databricks in Azure for various data transformations.
- Created several Databricks Spark jobs with PySpark to perform several table-to-table operations
- Used Databricks utilities called widgets to pass runtime parameters in notebooks.
- In-depth understanding with real-time data streaming technologies such as Kafka and Spark streaming.
- Good understanding in Delta Lake Architecture features like time-travel, schema enforcement and merge functionalities.
- Experience in DTS Migration and Metadata Management, Migrating DTS Packages to SSIS, Package Migration Wizard, BULK INSERT and Storage Management.
- Analyzed the source data, Coordinated with Data Warehouse team in developing Relational Model. Designed and developed logical and physical models to store data retrieved from other sources including legacy systems.
- Experienced in developing Data Mapping, Performance Tuning and Identifying Bottlenecks of sources, mappings, targets and sessions.
- Expert in creating SQL objects like Tables, Complex Stored Procedures, Triggers, Views, Indexes, and User Defined Functions to facilitate efficient data manipulation and consistent data storage.
- Experience in Database design, Development, Data Modeling, Administration, ETL packages using DTS / SSIS, Performance Tuning, Stored Procedures for Database applications using MS SQL Server 2017/2014 Proficient in implementing all transformations available on the SSIS toolbox.
- Experience in Database design, Development, Data Modeling, Administration, ETL packages using DTS / SSIS, Performance Tuning, Stored Procedures for Database applications using MS SQL Server 2017/2014 Proficient in implementing all transformations available on the SSIS toolbox.
- Well experienced in SQL Server Management Studio, SQL Server Business Intelligence Development Studio, SQL Server Enterprise Manager, and SQL Query Analyzer.
- Experienced in SQL Server Management Studio, SQL Server Business Intelligence Development Studio, SQL Server Enterprise Manager, and SQL Query Analyzer.
- Experienced in building Data Integration and Workflow Solutions and Extract, Transform, and Load (ETL) solutions for data warehousing using tools like SSIS import and export and SSIS packages.
TECHNICAL SKILLS
Programming Languages: C, C++, C#, VB.Net, Python, PySpark, SparkSQL, PL/SQL, DAX, VB scripting, VBA
Cloud Platform: Microsoft Azure
ETL Tools: Azure Data Factory, Azure Databricks, SSIS, Power Query
Databases: Azure SQL DB, MySQL, SQL Server, Oracle
NoSQL Databases: Azure Cosmos DB
Reporting Tools: Power BI, SSRSss
Programming Tools: Jupiter Notebook, Azure Data Studio, Visual Studio, SSMS, SSDT
Version Control: TFS, GitHub, Azure DevOps
Operating System: Windows Server 2008/2012/2016 , Linux, Windows
Methodologies: Agile
PROFESSIONAL EXPERIENCE
Confidential, Pittsburgh, PA
Senior Azure Data Engineer
Responsibilities:
- Worked on Full life cycle development (SDLC) involving in all stages of development.
- Worked in Building Tabular Models using SSAS
- Involved in gathering Business Requirement after interacting with the Manager and development team to build up a solution to implement in SQL Server 2017.
- Creating and managing schema objects such as tables, views, indexes, stored procedures, and triggers & maintaining Referential Integrity.
- Used stored procedures, wrote new stored procedures, modified existing ones, and tuned them such that theyperform well.
- Used advanced features ofT-SQLin order to design and tune T-SQL to interface with the Database and other applications in the most efficient manner and createdstored Procedures for the business logic using T-SQL.
- Usedindexestoenhance the performanceof individual queries and enhance the Stored Procedures.
- Continuous review of Business critical Database to proactively identify space issues, performance-tuning issues.
- Created SSIS packages to populate data from various data sources.
- Developed SSIS packages using various transformations such as Conditional Splits, Fuzzy Look Ups, Aggregations, and Multicasting.
- Created packages using SSIS for data extraction from Flat Files, Excel Files, and OLEDB to SQL Server.
- Experience in providing Logging, Error handling by using Event Handler, and Custom Logging for SSIS Packages.
- Expert in Migration 2017 of data from various OLTP sources Such as Oracle, MS Excel, flat filesand CSV files.
- Worked on migration for large amount of data fromOLTPto OLAPby usingETLPackages.
- Worked on Azure SQL Data Factory to load Azure SQL Data Warehouses for analytics and reporting.
- Developed PySpark, Python for regular expression (regex) in DataBricks.
- Data integrated from multiple internal sources into an AzureSQLServer with Azure Data factory and Azure Automation.
- Used SQL Azure Databases, Data Lakes, SSIS to capture features used to train risk attribute models used to score and identify high-risk deals.
- Large Volume of data from blob storage to Azure SQL Databases using ADF.
- Architect and implement ETL and data movement solutions using Azure Data Factory, SSIS create and run SSIS Package ADF V2 Azure-SSIS IR
- Integrated Custom Visuals based on business requirements using POWER BI Desktop.
- Created DAXQueries to generated computed columns in Power BI.
- Wrote calculated columns, Measures query’s in POWER BI desktop to show good data analysis techniques.
- Worked on Power Query in Power BI to pivot and un-pivot the data model for data cleansing and data massaging.
- Created Grouping, Sorting columns and writing expressions according to the business requirements.
Environment: MS SQL Server 2016, SSMS, Azure Data Factory (ADF), Azure SQL, Azure Blob, Data Bricks, PySpark, Data Lake Gen2, T-SQL, SSIS, SSRS, Power BI.
Confidential
Lead Azure Data Engineer
Responsibilities:
- Worked on Full life cycle development (SDLC) involving in all stages of development.
- Involved in designing technical specifications for data migration from on premises to Azure cloud.
- Created Database Objects - Schemas, Tables, Indexes, Views, User defined functions, Stored Procedure, Constraints and Roles.
- Created Pipeline for movement of data from On-Prime ODS to Azure cloud MI-ODS. Loading the data incrementally.
- Created Logical Apps for email notification for Pipelines
- Used SQL Azure Databases, Data Lakes, SSIS to capture features used to train risk attribute models used to score and identify high-risk deals.
- Large Volume of data from blob storage to Azure SQL Databases using ADF.
- Architect and implement ETL and data movement solutions using Azure Data Factory, SSIS create and run SSIS Package ADF V2 Azure-SSIS IR
- Integrated Custom Visuals based on business requirements using POWER BI Desktop.
- Created DAX Queries to generated computed columns in Power BI.
- Wrote calculated columns, Measures queries in POWER BI desktop to show good data analysis techniques.
- Worked on Power Query in Power BI to pivot and un-pivot the data model for data cleansing and data massaging.
- Created Grouping, sorting columns and writing expressions according to the business requirements.
Environment: SQL Server 2014/2012 Management Studio, SQL Server Business Intelligence Studio (BIDS) SQL Server Integration Services (SSIS), SQL Server Reporting Services (SSRS), ADF, Azure SQL DB, ADW, POWER BI
Confidential
ADF/SSIS/POWER BI/ Azure SQL DB/ Oracle
Responsibilities:
- Involved in creating Azure Data Factory pipelines that move, transform, and analyze data from a wide variety of sources.
- Created Triggers to run pipelines as per schedule.
- Configured ADF pipeline parameters and variables.
- Worked on connecting to a SQL instance running on a VM, an Azure Managed Instance, or an on-premises server.
- Development level experience in Microsoft Azure providing data movement and scheduling functionality to cloud-based technologies such as Azure Blob Storage and Azure SQL Database.
- Design, build and deploy ETL/SSIS packages with different data sources (SQL Server, Flat Files, Excel source files, Share Point etc.) and then loaded the data into destination tables by performing complex transformations using SSIS packages.
- Developed ETL Packages Involving transformations like derived column, conditional split, lookup, data conversion to load data into SQL Data Warehouse.
- Developed reports in Power BI Desktop and deployed to Power BI Service.
- Developed complex Data Analysis Expression Language (DAX) calculations, measures, report parameters, and report performance optimization.
- Designed complex reports and dashboards using Filters, Drill-down Reports and Sub-reports reports in Power BI Desktop.
- Provide support and fix issues to all ETL/SSIS schedule jobs and maintain compliance to same and develop and maintain various standards to perform ETL codes and maintain an effective project life cycle on all ETL processes.
- Developed and deployed SSAS cube (Tabular model).
- Migrated On-Premise databases to Azure SQL Server.
- Managing Scheduled tasks for the servers and automating all maintenance activities.
- Recovering the databases from backup in case of disasters.
Environment: Azure SQL Database, SQL Server 2017,2016 and 2014, Azure Data Factory, PowerBI
Confidential
Responsibilities:
- Getting customer/bookings data from various sources like - Hadoop,Teradata,NRT servers, Responsible for refreshing customer/subscriber/Search/Cancellation data daily basis
- Design and development of new ETM campaigns. Modifying existing campaigns based on business needs.
- Migration of Tailwind on-premise to AWS - TCS
- Design, build and deploy ETL/SSIS packages with different data sources (SQL Server, Flat Files, Excel source files, Share Point etc.) and then loaded the data into destination tables by performing complex transformations using SSIS packages.
- Developed ETL Packages Involving transformations like derived column, conditional split, lookup, data conversion to load data into SQL Data Warehouse.
- Developed reports in Power BI Desktop and deployed to Power BI Service.
- Developed complex Data Analysis Expression Language (DAX) calculations, measures, report parameters, and report performance optimization.
- Designed complex reports and dashboards using Filters, Drill-down Reports and Sub-reports reports in Power BI Desktop.
- Provide support and fix issues to all ETL/SSIS schedule jobs and maintain compliance to same and develop and maintain various standards to perform ETL codes and maintain an effective project life cycle on all ETL processes.
- Developed and deployed SSAS cube (Tabular model).
- Migrated On-Premise databases to Azure SQL Server.
- Managing Scheduled tasks for the servers and automating all maintenance activities.
- Recovering the databases from backup in case of disasters.
Environment: SQL Server 2008R2, SQL Server 2016, SSIS, SQL Scripts, Hadoop, pig, Hive, MapReduce.
Confidential
Responsibilities:
- Interact with customer to understand the business requirements and for any clarifications required.
- Reported task status with Client/Onshore team on a daily basis.
- Creation of mappings, sessions and workflows according to the specifications.
- Scheduling workflows based on the dependency/requirements.
- Coordinating with the team, helping them out in functional and technical aspects.
- Reviewing all project related documents and giving suggestions, if required.
Confidential
Responsibilities:
- Implementing end to end DW/BI system
- Preparing IDD-Interface design documents, STDM- Source target data mapping and providing ETL framework to the new developers
- Review the packages developed in the team and providing necessary best practice sessions and docs to the developers
- Support existing reporting platform and applications to ensure operational availability and scalability
- Designing data marts and different cubes to satisfy the senior management requirement, deployed the cube into different environments and created KPI’s and dashboard reports and extensively written MDX queries in report generation
- Working directly with functional analysts and business users in understanding the information needs of the business and developing new enhancements
- Creating data marts, cubes, deploying and maintaining to meet the business requirements
- Designing, developing, deploying and Maintaining SSRS reports to meet the business requirements
- Managing the team and providing required information to the team
- Creating technical specs, technical review meeting with client team and ensure entire offshore delivery progress smoothly.
Confidential
Responsibilities:
- Preparing IDD-Interface design documents, STDM- Source target data mapping and providing ETL framework to the new developers
- Review the packages developed in the team and providing necessary best practice sessions and docs to the developers
- Support existing reporting platform and applications to ensure operational availability and scalability
- Designed data marts and different cubes to satisfy the senior management requirement, deployed the cube into different environments and created KPI’s and dashboard reports and extensively written MDX queries in report generation
- Work directly with functional analysts and business users in understanding the information needs of the business and developing new enhancements
- Creating data marts, cubes, deploying and maintaining to meet the business requirements
- Designing, developing, deploying and Maintaining SSRS reports to meet the business requirements
- Managing the team and providing required information to the team
- Creating technical specs, technical review meeting with client team and ensure entire offshore delivery progress smoothly.
Confidential
Responsibilities:
- Preparing IDD-Interface design documents, STDM- Source target data mapping and providing ETL framework to the new developers
- Review the packages developed in the team and providing necessary best practice sessions and docs to the developers
- Support existing reporting platform and applications to ensure operational availability and scalability
- Designed data marts and different cubes to satisfy the senior management requirement, deployed the cube into different environments and created KPI’s and dashboard reports and extensively written MDX queries in report generation
- Work directly with functional analysts and business users in understanding the information needs of the business and developing new enhancements
- Creating data marts, cubes, deploying and maintaining to meet the business requirements
- Designing, developing, deploying and Maintaining SSRS reports to meet the business requirements
- Providing Status reports to different stake holders
- Managing the team and providing required information to the team
- Creating technical specs, technical review meeting with client team and ensure entire offshore delivery progress smoothly.
Environment: SQL Server 2008, SSIS, SSRS, SSAS 2008 and MDX