We provide IT Staff Augmentation Services!

Data Engineer Resume

3.00/5 (Submit Your Rating)

Nashville, TN

SUMMARY

  • 8+ years of IT experience in Data Warehousing, ETL, BI and Big Data
  • Skilled in full life cycle including project scoping, requirements gathering, systems analysis, design, development, testing, implementation and maintenance
  • Proven skills in AWS Redshift, Oracle PL/SQL, PostgreSQL including writing queries, stored procedures, functions and triggers
  • Expertise in Data Extraction, Cleansing, Transformation, Integration, Data Analysis, Logical/Physical Relational/Dimensional Database Modeling & Designing
  • Experience in designing and developing dashboards and reports by extracting data from different sources like oracle, flat files, and excel by using Qlik View, Tableau, Power BI
  • Active team player and self - motivated with passion for technology and data

TECHNICAL SKILLS

  • ETL Tools: Pentaho, Informatica, Data Pipeline, SAP, SSIS
  • Reporting Tools: Qlikview, Tableau, Qlik Sense, PBI
  • Databases: Redshift, Oracle, SQL Server, Postgresql
  • Reporting Tools: Tableau, Pentaho, SSRS, QlikView, SAP BO
  • Languages: Java, T - SQL, PL/SQL, MySQL, C++, Shell
  • Operating System: Windows, Mac, Linux

PROFESSIONAL EXPERIENCE

Confidential - Nashville, TN

Data Engineer

Responsibilities:

  • Develop quality framework to ensure delivery of high quality data and analyses to stakeholders
  • ETL jobs to transfer data from external applications to AWS cloud on a periodic basis, data modeling to ensure standardization across systems, deploying (and possibly coding) data processing scripts in a distributed manner, creating databases and orchestrate pipelines (event based/time based).
  • Advanced SQL with Redshift, Oracle, and Columnar Databases
  • Experience with AWS services including S3, Data-pipeline and other big data technologies
  • Hands-on experience with big data technologies such as S3, Redshift, Aurora, Athena, Glue, EMR, DMS and can explain their varied use cases.
  • Can code using any of the following languages: JavaScript, Python, and Java.
  • Developed PL/SQL triggers and Master tables for automatic creation of primary keys.
  • Created PL/SQL stored procedure, functions and packages for moving data from staging area to data mart.
  • Used different types of input and output steps for various data sources including Tables, Access, Text File, Excel and CSV files
  • Used dimension lookup/update step to populate data into SCDs and other various steps as per requirements.
  • Identify and analyze data discrepancies and data quality issues and works to ensure data consistency and integrity.
  • Created Data Flow Mappings to extract data from source system and Loading to Target .
  • Created Database logging for each Transformation/Jobs into ETL PDI TRANS and setup Email notification failure on each Component level

Environment: Pentaho BI Server, Pentaho Data Integration (PDI/Kettle), Pentaho Report, Qlik view Report, Designer, Pentaho Dashboard Designer, Pentaho Business Analytics, Java, MySQL, Oracle 10g, Oracle SQL Developer, Linux, C++

Confidential - Plano, TX

Sr. Pentaho ETL/ Report Developer

Responsibilities:

  • Interacted with the Business Analysts to understand the process flow and the business.
  • Actively participated in team to gather requirements to develop this BI project and also participated in designing Physical and Logical of Data warehouse
  • Responsible for designing, developing, automating, monitoring and maintaining ETL jobs, batches, processes and metadata associated with the transfer of data to and from internal and external locations
  • Developed PL/SQL triggers and Master tables for automatic creation of primary keys.
  • Created PL/SQL stored procedure, functions and packages for moving data from staging area to data mart.
  • Used different types of input and output steps for various data sources including Tables, Access, Text File, Excel and CSV files
  • Used dimension lookup/update step to populate data into SCDs and other various steps as per requirements.
  • Identify and analyze data discrepancies and data quality issues and works to ensure data consistency and integrity.
  • Created Data Flow Mappings to extract data from source system and Loading to Target.
  • Created Database logging for each Transformation/Jobs into ETL PDI TRANS and setup Email notification failure on each Component level

Environment: Pentaho BI Server, Pentaho Data Integration (PDI/Kettle), Pentaho Report, Qlik view Report, Designer, Pentaho Dashboard Designer, Pentaho Business Analytics, Java, MySQL, Oracle 10g, Oracle SQL Developer, Linux, C++

Confidential - Austin, TX

Data Consultant

Responsibilities:

  • Configured AWS environment to extract data from various sources and loaded the data in Redshift using distribution and sorting.
  • Utilized SQL to pull data from various databases and created various dashboards, scorecards and reports to support business decision making.
  • Actively interacted with business stakeholders, sponsors and client partners for Business Analysis & requirement collection.
  • Exported the analyzed data into relational databases using data pipeline for visualization and to generate reports for the BI team.
  • Developed ETL process using Pentaho PDI to extract the data from various data sources and populated it in to our BI data warehouse which includes SQL Server, PostgreSQL, and Redshift .
  • Worked with analysts to understand and build complex business logic, cleaned up and standardized data to create meta tables, fact tables, dimension tables and detailed data marts, applied business rules during the process to derive valuable business information.
  • Made use of different transformations in PDI designer to cleanse and manipulate source data e.g. Database Lookup & Join, Stream lookup, merge join, select values, UDJE, UDJC, generate rows, Calculator, Row normalizer & demoralizers, Java Script, add constant, Add Sequence, Get & Set Variables etc.
  • Used Kettle to build the ETL process from data sources to data mart, as well as scheduled the transformations and jobs.
  • Performance tuning of slowly running transformations and jobs, Redshift SQL Queries & Stored procedures .
  • Designed project through the SDLC process, with best practices and trends in the field of BI and Data Warehousing to help shape company’s BI strategy and solution designs.
  • Created deployment documentation with the help of other users and operational team to deploy the new release in Production environment

Environment: Amazon Redshift, Oracle, Pentaho, Tableau

Confidential - Austin, TX

ETL Developer

Responsibilities:

  • Installed and Configured Pentaho BI Suite along with Enterprise Repository in Pentaho BI server.
  • Configured Pentaho BI Server for report deployment by creating database connections in Pentaho enterprise console for central usage by the reports deployed in the repository.
  • Coded several database Stored Procedures, Functions, Packages and Triggers in PL/SQL, and maintained Integrity Constraints.
  • Designed automation system using Pentaho Kettle for ETL enhancements and Pentaho BI Server for publishing the Pentaho Reports on user funds transfer measures.
  • Identify, document and communicate BI and ETL best practices and industry accepted development methodologies and techniques.
  • Used various types of inputs and outputs in Pentaho Kettle including Database Tables, MS Access, Text Files, Excel files and CSV files.
  • Implemented miscellaneous transformations in Kettle Spoon Designer including Database Lookup, Database Join, Calculator, Generate Rows, Mapping Transformation, Filter Rows, Dimension Lookup/Update, Add Sequence, Add Constants, Row Normalizer and Demoralizer.
  • Used different components in Pentaho like Database Lookup & Join, generate rows, Calculator, Row normalizer & demoralizers, JavaScript, add constant, Add Sequence.
  • Worked on all phases of data warehouse development lifecycle, from gathering requirements to testing, implementation, and support using Pentaho Data Integration.

Environment: Pentaho BI Server, Pentaho Data Integration (PDI/Kettle), Pentaho Report Designer, Pentaho Dashboard Designer, MySQL, Postgres

Confidential - Salt Lake City, UT

Sr. Pentaho/ ETL Engineer

Responsibilities:

  • Participated in cost and schedule estimations of software, database, tools and platforms.
  • Influence the BI tool decisions and process design. Assist creation of conceptual, logical, and physical models for data analysis and modeling to help develop a training strategy and plan for transition management.
  • Participated in design of Redshift Staging Databases and Data Warehouse/Data mart.
  • Automated data transfer processes and mail notifications by using FTP Task and send mail task in Transformations.
  • Performed advanced-level business system analysis. Worked closely with product owners and analysts to understand and translate requirements into design.
  • Helped company refine the process for intake, prioritization, routing, and estimation of BI development requests from the business.
  • Using the data Integration tool Pentaho for designing ETL jobs in the process of building Redshift Data warehouses.
  • Used various types of inputs and outputs in Pentaho Kettle including Database Tables, MS Access, Text Files, Excel files and CSV files.
  • Responsible for creating database objects like table, views, Functions etc. using Redshift SQL to provide structure to store data and to maintain database efficiently.
  • Involved in optimizing code and improving efficiency in databases by re-indexing, updating statistics, recompiling stored procedures and performing other maintenance tasks.
  • Created web based reports and dashboards using Report Designer. Presented weekly/monthly reports and dashboards to management team and cross-functional stakeholders.
  • Involved in administration tasks such as creating and managing Sites, setting permissions, managing ownerships and providing access to the users using user filters and adding them to the specific group.

Environment: AWS S3, EC2, EMR, Redshift, PostgreSQL, NetBeans-Java, Shell, SQL Workbench, Pentaho, Tableau, QlikView

Confidential - San Fransisco, CA

BI / ETL Developer

Responsibilities:

  • Interaction with business and technical teams across geographic boundaries, with clients, technical architects, project managers, requirements analysts and testing teams.
  • Provided production support for the production database and also to the code deployed into the production environment.
  • Provide analysis and development support for ETL and data integration initiatives
  • Worked with business users/analytics team, data architect to identify the business requirements and designed ETL flow, data architecture and created various complex Pentaho Transformation and Jobs using PDI Spoon.
  • Worked on all phases of data warehouse development lifecycle, from gathering requirements to testing, implementation, and support using Pentaho Data Integration.
  • Identify, document and communicate Redshift and ETL best practices and industry accepted development methodologies and techniques.
  • Worked very closely with Project Manager to understand the requirement of reporting solutions to be built.
  • Implemented Logic with Database lookup table to maintain Parent- Child relationship and maintain hierarchy.
  • Applied Configuration, Logging, Error reporting to all packages to make package deployment easy and troubleshoot package on run time.
  • Worked on Performance manager to create various metrics, Analytics and used Dashboard Manager to build corporate dashboards.
  • Working closely with DBA team to regularly monitor system for bottlenecks and implement appropriate solutions.

Environment: Pentaho BI Server, Pentaho Data Integration, SSIS, Redshift, SQL Server, Shell, Unix, SQL profiler, XML, SSRS, Tableau

We'd love your feedback!