We provide IT Staff Augmentation Services!

Azure Data Engineer Resume

4.00/5 (Submit Your Rating)

Minneapolis, MN

SUMMARY

  • Having 13+ years of professional experience in ETL development, End to End DWH development.
  • Experienced in designing and deploying solutions on Microsoft Azure platform using Data Factory, Data Lake store, Data Lake Analytics, Azure Data Bricks, and Integration solutions.
  • Experienced in using Spark SQL with various data sources like JSON, Parquet and Hive.
  • Experienced in extracting and loading the data from different relational data bases such as Oracle, Teradata, DB2 into Azure Datalake storage using Azure data factory.
  • Created Pipelines in ADF using Linked Services/Datasets/Pipeline/ to Extract, Transform, and load data from different sources like Azure SQL, Blob storage, Azure SQL Data warehouse, write - back tool and backwards.
  • Developed Spark applications usingPysparkandSpark-SQLfor data extraction, transformation, and aggregation from multiple file formats for analyzing & transforming the data to uncover insights into the customer usage patterns.
  • Responsible for estimating the cluster size, monitoring, and troubleshooting of the Spark data bricks cluster.
  • Experienced in performance tuning of Spark Applications for setting right Batch Interval time, correct level of Parallelism and memory tuning.
  • To meet specific business requirements wrote UDF’s inScalaandPyspark.
  • Developed JSON Scripts for deploying the Pipeline in Azure Data Factory (ADF) that process the data using the SQL Activity.
  • Hands-on experience on developing SQL Scripts for automation purpose.
  • Extensive experience using Azure cloud components such as Blob storage account, Key Vault, ADLSv2, Azure Databricks.
  • Experienced in developing external tables in SQL DWH for data analysis and reporting.
  • Experienced in troubleshooting failures in Production and resolve them as per the given SLA.
  • Hands on experience in developing Teradata scripts data processing and to apply transformations.
  • Experienced in extracting and transforming the data using Informatica PowerCenter ETL tool.
  • Proficient in Software Development Life cycle processes - Agile and Waterfall.
  • Excellent ability to grasp technologies and interact effectively with peers and clients and ability to work with multiple technologies and to take responsibilities at typical deliveries.

TECHNICAL SKILLS

Technologies: Azure Data Factory (ADF), Azure Datalake (ADLS), Apache Spark2.4.0, Informatica Power Center 10.x/9.1/8.x

Databases: Oracle 11g/10g/9i, MS SQL Server 2008/2005/2000 , IBM-DB2, IBM Netezza 6.0, Azure SQL DB, DB2

Languages: Java, C, C++, SQL, UNIX Shell Scripting, HTML, XML, Python3X, Scala2.11

Tools: Autosys, Control-M, Toad, SQL Developer, JIRA, HPQC, WinSCP, Microsoft Azure Storage explorer, Jenkins, Bitbucket, GitHub, Git.

Operating Systems: Windows NT/95/98/XP, MS-DOS, UNIX

PROFESSIONAL EXPERIENCE

Confidential, Minneapolis, MN

Azure Data Engineer

Responsibilities:

  • Extracted the data from On-premises shared location and ingested into Azure Blob storage using Azure data factory.
  • Setup Key vaults to store secrets securely and Pyspark package installation, IR, Linked service to manage the ADF and cluster.
  • Worked on designing dataset and mapping of dataset whiling performing ingestion and various trans formation using business driven logic like removing of duplicate records, eliminating, or analyzing missing value from file-based ingestion of unwanted characters and making union of data frame apart from making joins.
  • Extracted the data from relational and non-relational sources using Azure data factory.
  • Involved in developing the Logic Apps for email notifications and to develop the custom business transformations.
  • Experienced in developing ADB notebooks for data profiling and analysis and to generate ad hoc reports for business teams.
  • Developed external tables pointing to ADLS in Azure synapse analytics for data validation and reporting needs.
  • Developed complex stored procedures in Azure SQL DWH to achieve business functionality and optimized long running stored procedures.
  • Experienced in migrating the existing legacy applications to Azure cloud using azure echo system tools.
  • Participated in SIT testing and resolved defects raised by SIT team and business teams.
  • Participated in code review meeting and made code changes as per the feedback received.
  • Experienced in both agile and waterfall methodology.

Confidential, Charlotte, NC

Azure Data Engineer

Responsibilities:

  • Extracted the data from Teradata and ingested into Azure Blob storage using Azure data factory.
  • Setup Databricks creation and involved in cluster management.
  • Setup Key vaults to store secrets securely and Pyspark package installation, IR, Linked service to manage the ADF and cluster.
  • Worked on designing dataset and mapping of dataset whiling performing ingestion and various transformation using business driven logic like removing of duplicate records, eliminating, or analyzing missing value from file-based ingestion of unwanted characters and making union of data frame apart from making joins.
  • Extracted the data from relational and non-relational sources using Azure data factory.
  • Involved in developing the Logic Apps for email notifications and to develop the custom business transformations.
  • Experienced in developing ADB notebooks for data profiling and analysisand to generate ad hoc reports for business teams.
  • Developed external tables pointing to ADLS in Azure synapse analytics for data validation and reporting needs.
  • Developed complex stored procedures in Azure SQL DWH to achieve business functionality and optimized long running stored procedures.
  • Experienced in migrating the existing legacy applications to Azure cloud using azure echo system tools.
  • Participated in SIT testing and resolved defects raised by SIT team and business teams.
  • Participated in code review meeting and made code changes as per the feedback received.
  • Experienced in both agile and waterfall methodology.

Environment: ADF, Azure SQL DB, Azure Synapse analytics, Azure Logic Apps, Azure Databricks, ADLS, ADO, Azure BLOB Storage, Informatica Power Center 10.x, Oracle, SQL Server

Confidential, Charlotte, NC

Azure Data Engineer

Responsibilities:

  • Helped the team in the Design, Development and Execution of the ETL code.
  • Extracted the data from relational and non-relational sources using Azure data factory.
  • Involved in developing the Logic Apps for email notifications and to develop the custom business transformations.
  • Having good experience Azure Data Factory, Azure Data bricks Azure Delta Lake.
  • Experience in Managing and storing confidential credential in Azure Key vault.
  • Expertise in implement of end-to-end of right from requirement gathering/analysis, designing, coding, testing and deployment.
  • Expertise in Design the Pipeline in Azure Data Factory using Activities, Data Flows and ADB.
  • Performed data factory pipeline optimizations.
  • Performed Data validation checks after trigger activity on every pipeline.
  • In ADB design data frame on transient folder, add ETL Date, Apply Encryption/Decryption on sensitive columns and merge to Refined tables.
  • Knowledge on streaming data using Event hubs.
  • Get the metadata through using Azure function API call.
  • Implemented the CI-CD process through Azure DevOps Organization.
  • Implemented the Log Analytics and exception handling (eMail).
  • Experience in On-Prem to Cloud through PaaS Migration.
  • Good Knowledge on Azure Cosmos DB and Azure Synapse Analytics.
  • Design Queries, functions, Stored Procedures for data Preparation in SQL Server.
  • Collaborating with cross cluster teams (Onshore and offshore) and Partners to improve data quality and data relevance.
  • Extensive experience working in an Agile development environment.

Environment: Azure Data Factory, ADLS, Azure BLOB storage, Azure SQL DB, Informatica Power Center 9.x/10.x, Oracle 11g, MS SQL Server 2008 R2, EventHub, Log analytics, Git.

Confidential, Oklahoma, OK

ETL Senior Informatica Developer

Responsibilities:

  • Experienced in Interacting with customers and business analyst daily to understand business requirements.
  • Developed and tested extraction, transformation, and load (ETL) processes.
  • Designed and developed Informatica's Mappings and Sessions based on business user requirements and business rules to load data from source flat files and oracle tables to target tables.
  • Designed and implemented stored procedures, views, and other application database code objects to aid complex mappings.
  • Interacted with third party vendors and identified different external and internal homogenous and heterogeneous sources and extracted and integrated data from flat files, Oracle, SQL Server sources and loaded to staging area and database/Datamart tables.
  • Experienced in developing Unix shell scripting for processing the files
  • Optimized the long running ETL Informatica processes and implemented the best practices.
  • Experienced in developing Mapplets to reuse the business functionality in multiple mappings.
  • Experienced in writing unit test cases and executed them and captured the results in HPQC.
  • Experienced in working on Onsite and offshore development model.

Environment: Informatica Power Center 9.x, Oracle 11g, UNIX, FileZilla, Control -M, Fast load, Multiload, TPump, BTEQ, SQL Server, DB2, XML

Confidential

Associate Consultant

Responsibilities:

  • Daily Interacting with customers and business analyst daily to understand business requirements.
  • Participated in Team meetings to ensure a mutual understanding with business, development, and test teams.
  • Encapsulated frequently executed SQL statements into stored procedures to reduce the query execution times.
  • CreatedSSIS packagesto implement error/failure handling with event handlers, row redirects, and loggings.
  • Managed packages the inSSISDBcatalog with environments; automated deployment and execution with SQL agent jobs.
  • Involved in the design of Data-warehouse usingStar-Schemamethodology and converted data from various sources to Sql tables.
  • Designed complex data intensive reports inPower BIutilizing various graph features such as gauge, funnel.
  • Experienced in working with Admin team on environment setup and getting the required access to start the development activities.
  • Optimized the long running ETL Informatica processes and implemented the best practices.
  • Experienced in developing Mapplets to reuse the business functionality in multiple mappings.

Environment: Informatica Power Center 9.0/9.5. Oracle 11g, Sql Server, Microsoft Office Tools, UNIX, FileZilla.

Confidential

Software Engineer

Responsibilities:

  • Interacted with third party vendors and identified different external and internal homogenous and heterogeneous sources and extracted and integrated data from flat files, Oracle, SQL Server sources and loaded to staging area and database/Datamart tables.
  • Created Informatica Data quality plans according to the business requirements to standardize, cleanse data and validate addresses.
  • Integrated data quality plans as a part of ETL processes.
  • Debugged existing ETL processes and did performance tuning to fix bugs.
  • Experienced in Creating the technical documents and get them review by business stakeholders.
  • Optimized the long running ETL Informatica processes and implemented the best practices.
  • Experienced in developing Mapplets to reuse the business functionality in multiple mappings.
  • Resolved production issues on time to meet the SLA.
  • Designed and developed Informatica's Mappings and Sessions based on business user requirements and business rules to load data from source flat files and oracle tables to target tables.
  • Experienced in analyzing the data and provide ad hoc reports to business.
  • Used different ETL transformations to transform the data using Informatica Power Center.

Environment: Informatica Power Center 8.x. Oracle 11g, Toad 10.5, UNIX, WinSCP.

We'd love your feedback!