We provide IT Staff Augmentation Services!

Data Engineer Resume

4.00/5 (Submit Your Rating)

Plano -, TX

SUMMARY

  • 8+ years of professional experience in information technology as Data Engineer with an expert hand in the areas of Database Development, ETL Development, Data modelling Technologies.
  • Experience in Data Integration and Data Warehousing using various ETL tools like IBM WebSphere Data Stage PX, Apache Nifi, DBT.
  • Extensively used IBM WebSphere Data Stage for extracting, transforming, loading, and cleansing data from various source data inputs to various targets, in batch and real time.
  • Experience working on Snowflake with Amazon Web Services (AWS) cloud and its services.
  • Strong expertise in Relational Data Base systems Netezza, DB2 design and database development using SQL, AQT, TOAD, COGINITY. Highly proficient in writing, testing and implementation of triggers, stored procedures, functions, packages, Cursors using PL/SQL.
  • Hands on Experience on Snowflake with AWS Snowflake cloud data warehouse and AWS S3 bucket for integrating data from multiple source system which include loading nested JSON formatted data into Snowflake table.
  • Worked in building multiple Data pipelines, end to end ETL and ELT process for Data ingestion and transformation in Snowflake.
  • Worked with different scheduling tools like Talon, Control - M and TWS (Tivoli Workload Scheduler).
  • Experienced in design, development, Unit testing, integration, debugging and implementation and production support, client interaction and understanding business application, business data flow and data relations.
  • Worked on extracting the data using API calls.
  • Worked on AWS Data Pipeline to configure data loads from Snowflake.
  • Worked on Data Migration from Data Stage to AWS Snowflake Environment using DBT.
  • Developed Automated scripts to do the migration using Unix shell scripting, Python.
  • Good Knowledge on No SQL database like HBase.
  • Expert-level mastery in designing and developing complex mappings to extract data from diverse sources including flat files, RDBMS tables, legacy system files, XML files, Applications, COBOL Sources.
  • Worked on JIRA for defect/issues logging & tracking and documented all my work using CONFLUENCE.
  • Experience in identifying Bottlenecks in ETL Processes and Performance tuning of the production applications using ETL Tuning, Partitioning, Index Usage, Session partitioning, Load strategies and transformation tuning.
  • Worked on performance tuning of user queries by analysing the explain plans, recreating the user driver tables by right Primary Index, scheduled collection of statistics, secondary or various join indexes.
  • Experience with scripting languages like PowerShell, Perl, Shell, etc.
  • Expert knowledge and experience in fact dimensional modelling (Star schema, Snowflake schema), transactional modelling and SCD (Slowly changing dimension).
  • Excellent interpersonal and communication skills, experienced in working with senior level managers, businesspeople, and developers across multiple disciplines.
  • Strong problem solving, analytical and can work both independently and as a team. Highly enthusiastic, self-motivated, and rapidly assimilate with new concepts and technologies.
  • Very good Knowledge on Agile Methodologies.

TECHNICAL SKILLS

ETL: IBM WebSphere Data Stage 9.X/11.x, DBT, Apache Nifi

DATABASES & TOOLS: Netezza & DB2

CLOUD ENVIRONMENT: SNOWFLAKE cloud data warehousing & AWS

SCHEDULER: Control-M & Tivoli & Talon

BIG DATA ECOSYSTEM: Hive, Sqoop, HBase, Kafka, HDFS, Yarn

PROGRAMMING LANGUAGES: Unix Shell Scripting, SQL, PL/SQL, Python

VERSIONING & DEPLOYMENT TOOL: GIT LAB & UCD(urban code deploy)

AGILE TOOL: Jira, IBM Rational Team Concert

PROFESSIONAL EXPERIENCE

Confidential, PLANO - TX

Data Engineer

Responsibilities:

  • Involved in gathering requirement specifications by having Interactive Sessions with the business.
  • Very good Knowledge on Agile Methodologies.
  • Analysis and Reverse Engineering of Data Stage code.
  • Develop ELT data pipeline to migrate applications usingDBTand Snowflake framework.
  • Responsible for generation of DDL statements which are executed for database creation.
  • Implement ETL Solution fulfilling stringent performance requirements.
  • Designed and developed the ELT jobs using DBT to achieve the best performance.
  • Very good Knowledge on Agile Methodologies.
  • Scheduling and dependency management of data pipeline with proper auditing
  • Register and Build ingestion pipeline with test automation of code.
  • Implement test automation in the data pipeline
  • Performed data ingestion from the source S3 bucket, transform and load to Snowflake Data Marts and load the processed files into target S3 bucket.
  • Implemented SCD Type 2 using python & DBT, on the claims system and incremental load on claims subject area.
  • Experience in Building snow pipe and Data sharing in Snowflake.
  • Experience in using snowflake clone and Time Travel.
  • Involved in all steps of development and deployment processes.
  • Design, development, debugging and Unit Testing of theDBT jobs
  • Involved in implementing different logics like incremental loading, change capturing and slowly changing dimensions.
  • Doing Unit and Integration testing doc.
  • Used Talon & Control-M to schedule the jobs.

Environment: Nifi, Snowflake, Python, Hadoop and DBT and Qtest.

Confidential, PLANO - TX

Senior Developer

Responsibilities:

  • Working closely with business intelligence team to understand requirements and develop ETL jobs and procedures using IBM Data Stage and tools.
  • Create technical design specification documentation for data Extraction, Transformation, and loading into EDW Data Warehouse.
  • Prepare flow charts and diagrams to illustrate the sequence of steps the program must follow describing logical operations
  • Actively participate in weekly team meetings and solving the Technical Issues.
  • Involved in gathering requirement specifications by having Interactive Sessions with the business.
  • Develop Solution in highly demanding environment and provide hands on guidance to the offshore team members.
  • Implement ETL Solution USING NIFI to extract data from Kafka.
  • Create External tables in Hadoop.
  • Create Framework for SFTP the Files to ETL Server from Hadoop.
  • Guiding The Offshore team by providing the solutions.
  • Involved in resolving the user queries by looking into their issues.
  • Designed and customized data models for Data warehouse supporting data from multiple sources on real time.
  • Worked on various Stages likeTransformer,Aggregator,Lookup, Join, Merge, Remove duplicates, Funnel Filter, Pivot, hierarchical data, Unstructured data, and complex flat file stagesetc.
  • Provided production support and performed enhancement on existing multiple jobs.

Environment: Nifi, Hadoop, IBM Data Stage PX 11.5, Netezza, Snowflake.

Confidential, PLANO - TX

Software developer

Responsibilities:

  • Develop project plans to implement ETL Solution and identify resource requirements.
  • Involved in gathering requirement specifications by having Interactive Sessions with the business.
  • Designed the ETL jobs using IBM Info Sphere Data Stage to Extract, Transform and load the data into staging, ODS and EDW.
  • Designed and developed the ETL jobs using Parallel Edition which distributed the incoming data concurrently across all the processors, to achieve the best performance.
  • Designed parallel jobs using stages such as Join, Merge, Lookup, Remove Duplicates, Copy, Filter, Funnel, Dataset, Lookup, Pivot, and Sort, Surrogate key Generator, Change Data Capture (CDC), Modify, Row Generator and Aggregator.
  • Responsible for generation of DDL statements which are executed for database creation.
  • Implement ETL Solution fulfilling stringent performance requirements.
  • Used DataStage Parallel Extender stages namely Datasets, Sort, Lookup, Change Capture, Funnel, Peek, SCD, and Row Generator.
  • Extensively worked with Netezza connector. Loaded high volume of data into Netezza Database.
  • Used Control-M to schedule the jobs.
  • Implement ETL Solution fulfilling performance requirements.
  • Developed job sequencer with proper job dependencies, job control stages, triggers.
  • Involved in Unit testing, System testing to check whether the data is loading into target, which was extracted from different source systems according to the user requirements.
  • Involved in production support working with various mitigation tickets.
  • Experience in Building snow pipe and Data sharing in Snowflake.
  • Experience in using snowflake clone and Time Travel.
  • Involved in all steps of development and deployment processes.
  • Worked on all the standards & guidelines Best Practices Documentation, and the functional & Technical design specification documentation.

Environment: Nifi, Snowflake, Hadoop, Hive, Netezza, UNIX, Data Stage.

Confidential, PLANO - TX

Software developer

Responsibilities:

  • Involved in gathering requirement specifications by having Interactive Sessions with the business.
  • Develop Solution in highly demanding environment and provide hands on guidance to other team members.
  • Implement ETL Solution fulfilling stringent performance requirements.
  • Identify, recommend, and implement ETL process and architecture improvements.
  • Developed Data Migration methodologies, specifically around DataStage to SAP BODS migration.
  • Connection setup for SAP ECC and SAP BW systems in ETL.
  • Worked on Batch/ real time jobs to extract the data from the SAP BW/ECC system, and load into the staging database for the multiple projects.
  • Creating Jobs, Workflows, data flows and scripts in BODS Designer to extract data from all the Relational Databases including oracle, Microsoft SQL server and loaded in different environments which include BW, ORACLE, SQL SERVER, and .txt & .CSV files.
  • Created Global variables to read the files from the shared location from the BODS Designer Also worked on the FTP process to put & get the files on the data-exchange servers.
  • Extracted data from different source systems and loaded into target SAP BW system.
  • Extracted the data from the client system through the FTP then using these flat file data source, created ETL to load into target system.
  • Created JOBs, Workflows, and complex Data flows.
  • Involved in resolving the user queries by looking into their issues.
  • Used SAP BODS transformers like query transformer, pivot, case, validation, while, Conditional, Merge.
  • Administered on QlikView Server and Publisher, to manage user access, schedule refreshing of documents, mount completed applications.
  • Applied security to document (section access) and QMC server level (reduction, distribution, and loop)
  • Enhance the existing QVD files created by ETL logic to feedQlikViewapplications
  • Monitoring dashboard refresh in QlikView publisher and NPrinting.

Confidential - OH

ETL Developer

Responsibilities:

  • Working closely with business intelligence team to understand requirements and develop ETL jobs and procedures using IBM DataStage and tools.
  • Create technical design specification documentation for data Extraction, Transformation, and loading into EDW Data Warehouse.
  • Prepare flow charts and diagrams to illustrate the sequence of steps the program must follow describing logical operations
  • Actively participate in weekly team meetings and solving the Technical Issues.
  • Involved in gathering requirement specifications by having Interactive Sessions with the business.
  • Develop Solution in highly demanding environment and provide hands on guidance to the offshore team members.
  • Implement ETL Solution fulfilling stringent performance requirements.
  • Involved in resolving the user queries by looking into their issues.
  • Designed and customized data models for Data warehouse supporting data from multiple sources on real time.
  • Implement ETL Solution fulfilling performance requirements.
  • Developed job sequencer with proper job dependencies, job control stages, triggers.
  • Worked on various Stages likeTransformer,Aggregator,Lookup, Join, Merge, Remove duplicates, Funnel Filter, Pivot, hierarchical data, Unstructured data, and complex flat file stagesetc.
  • Provided production support and performed enhancement on existing multiple jobs.
  • Involved in all steps of development and deployment processes.
  • Extensively worked with Netezza connector. Loaded high volume of data into Netezza Database.
  • Used Control-M to schedule the jobs.

Environment: IBM Data Stage PX 9.1, DB2, UNIX.

Confidential - OH

ETL Developer

Responsibilities:

  • Monitoring jobs and job sequencers.
  • Preparing Unit and Assembly Test Plans.
  • Analysing and resolving the Tickets.
  • Debugging the jobs.
  • Analysing change requests and preparing detail designing of technical documents.
  • Performing impact analysis for new CRs.
  • Designed and Prepared a KB DOCS for the Users.
  • Updating the resolutions with Solution
  • Monitoring the ETL jobs by the time of moving into production and maintain the report of the database transactions.
  • Preparing Unit and integration Test Plans.
  • Provided design Solution for issues caused during support.
  • Implement ETL Solution fulfilling performance requirements.
  • Extensively worked with Netezza connector. Loaded high volume of data into Netezza Database.
  • Used Tivoli to schedule the jobs.

Environment: IBM Data Stage PX 9.1, DB2, UNIX.

We'd love your feedback!