We provide IT Staff Augmentation Services!

Big Data Engineer Resume

4.00/5 (Submit Your Rating)

Watertown, MA

SUMMARY

  • Overall 7 years of Industry experience as a Big Data Engineer with solid understanding of Data Modeling, Data Validation, Evaluating Data Sources, and strong understanding of Data Warehouse/Data Mart Design, ETL, BI, OLAP, Client/Server applications.
  • Expert in writing Complex SQL queries and optimizing the queries in Oracle, SQL Server, and Teradata. Excellent Software Development Life Cycle (SDLC) with good working knowledge of testing methodologies, disciplines, tasks, resources, and scheduling.
  • Worked with packages like ggplot2 and shiny in R to understand data and developing applications.
  • Worked on Tableau, Power BI to create dashboards and visualizations.
  • Excellent knowledge in Data Analysis, Data Validation, Data Cleansing, Data Verification and identifying data mismatch.
  • Performed data analysis and data profiling using complex SQL on various sources systems including Oracle and Teradata.
  • Excellent experience on Teradata SQL queries, Teradata Indexes, Utilities such as Multiload, Tpump, Fast load and Fast Export.
  • Strong experience in using Excel and MS Access to dump the data and analyze based on business needs.
  • Excellent knowledge in Alteryx.
  • Excellent knowledge on Perl & UNIX.
  • Experienced working with Excel Pivot and VBA macros for various business scenarios.
  • Strong experience in Data Analysis, Data Migration, Data Cleansing, Transformation, Integration, Data Import, and Data Export using multiple ETL tools such as Ab Initio and Informatica Power Center Experience in testing and writing SQL and PL/SQL statements - Stored Procedures, Functions, Triggers, and packages.
  • Excellent experience on using Teradata SQL Assistant, Teradata Administrator, PMON and data load/export utilities like BTEQ, Fast Load, Multi Load, Fast Export, and Exposure to T pump on UNIX/Windows environments and running the batch process for Teradata.
  • Excellent experience in writing SQL queries to validate data movement between different layers in data warehouse environment.
  • Excellent experience in troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Excellent knowledge in preparing required project documentation and tracking and reporting regularly on the status of projects to all project stakeholders.
  • Extensive knowledge and experience in producing tables, reports, graphs, and listings using various procedures and handling large databases to perform complex data manipulations.
  • Experience in testing Business Intelligence reports generated by various BI Tools like Cognos and Tableau.
  • Have good exposure on working in offshore/onsite model with ability to understand and/or create functional requirements working with client also have Good experience in requirement analysis and generating test artifacts from requirements docs.
  • Proficient in Microsoft business intelligence stack SSRS.

TECHNICAL SKILLS

Databases: Microsoft Access, Oracle 11g, MySQL, Amazon Dynamo DB

ETL Tools: Informatica, Alteryx, Talend, Data Stage, Pentaho

Database Programming: SQL, PL/SQL, HIVE, PostgreSQL

BI and Visualization tools: Tableau, Cognos, Qlikview, Microsoft Power BI

Programming Languages: Python, R, SQL, Perl

Cloud: AWS, Microsoft Azure, Google Cloud

Microsoft Tools: SSIS, SSAS, SSRS, Microsoft Excel

PROFESSIONAL EXPERIENCE

Confidential, Watertown MA

Big Data Engineer

Responsibilities:

  • Analyzed client products data and ingested onto Master Data Management (MDM) with compliance oversight into data governance standards.
  • Investigated market sizing, competitive analysis, and positioning for product feasibility.
  • Wrote SQL Scripts for various MDM tables which links all the customer’s demographic details along their associated products together and mapped them to Persistent Id’s which uniquely identifies each client.
  • Automation of mastering Customers daily transactions and ingesting into MDM.
  • Performed data management projects and fulfilling ad-hoc requests per user specifications by utilizing data management software programs and tools like Perl, Toad, MS Access, Excel, and SQL. Written SQL scripts to test the mappings and Developed Traceability Matrix of Business Requirements mapped to Test Scripts to ensure any Change Control in requirements leads to test case update.
  • Generated graphs and reports using ggplot package in R-Studio for analytical models.
  • Developed and implemented R statistical analysis for business forecasting.
  • Performed time series analysis using Tableau.
  • Worked with AWS S3, AWS Glue, Amazon DynamoDB for extracting, transforming, data from various data sources and ingesting into MDM.
  • Developed various workbooks in Tableau from multiple data sources.
  • Created dashboards and visualizations using Tableau desktop.
  • Created dashboards in Power BI to visualize data.
  • Later used Alteryx designer to blend the data and to validate data lineage.
  • Performed analysis using JMP SAS.
  • Written connectors to extract data from databases.
  • Analysis on Mainframe data to generate reports for business users.
  • Identified & recorded defects with required information for issue to be reproduced by development team.
  • Worked on SAS for data analytics and data quality checks.
  • Worked with the business in gathering requirements from the existing reporting tool.
  • Designing, developing, testing, and maintaining Tableau reports and dashboards based on user requirements.
  • Working closely with business in creating dashboards based on the data from various Data sources.
  • Worked on Tableau in migrating the Excel reports and Business objects to Tableau Dashboards.
  • Used Tableau for data visualization to create reports, dashboards for insights and business process improvement.
  • Extensively used Python’s multiple data science packages like Pandas, NumPy, Matplotlib, SciPy, Scikit-learn and NLTK.
  • Used TensorFlow Python API’s to perform TensorFlow graphs.
  • Worked on Spark Python modules for machine learning and predictive analytics in Spark on AWS.
  • Worked on end to end pipe line in Spark.
  • Explored and analyzed the customer specific features by using SparkSQL.
  • Performed data imputation using Scikit-learn package in Python.
  • Created the dashboards and reports in tableau for visualizing thedata in required format.
  • Collaborated with team members and translated functional requirements to technical requirements for development.
  • Conducted Code Review for the Fit Gap done by the team members.

Environment: Tableau, SQL, Business Objects XIR2, ETL Tools Informatica 8.6/9.1, Oracle 11G, Teradata V2R12/ R13.10, Teradata SQL Assistant 12.

Confidential, Evansville IN

Big Data Engineer

Responsibilities:

  • Analysis of functional and non-functional categorical data elements for data profiling and mapping from source to target data using SSRS.
  • Involved with data profiling/validation for multiple sources using Aginity Pro, AWS Redshift, MySQL.
  • Worked with data investigation, discovery, and mapping tools to validate the data across various environments.
  • Wrote and executed unit, system, integration and UAT scripts in a data warehousing applications like Ataccama, EDW.
  • Worked on Troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Extensively used ETL for supporting data extraction, transformations and loading processing, in a complex EDW using Talend/Data stage.
  • Metrics reporting, data mining and trends in helpdesk environment using Access.
  • Written complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2
  • Extensively used MS Access to pull the data from various data bases and integrate the data.
  • Worked on SAS for data analytics and data quality checks.
  • Worked with the business in gathering requirements from the existing reporting tool.
  • Designing, developing, testing, and maintaining Tableau reports and dashboards based on user requirements.
  • Working closely with business in creating dashboards based on the data from various Data sources.
  • Worked on Tableau in migrating the Excel reports and Business objects to Tableau Dashboards.
  • Used Tableau for data visualization to create reports, dashboards for insights and business process improvement.

Environment: Aginity Pro-Netezza, Oracle 11G, Teradata SQL Assistant 12.0, AWS Redshift, Tableau, MS-Excel

Confidential, Bridgewater NJ

Big Data Engineer

Responsibilities:

  • Gathered and documented MDM application, conversion and integration requirements.
  • Interacting with Business Analysts and Developers in identifying the requirements, designing and implementing the Database Schema.
  • Performing codebase maintenance and quality checks for Microsoft Azure.
  • Documenting and maintaining database system specifications, diagrams, and connectivity charts.
  • Participating in T-SQL code reviews and technical quality standards reviews with the development teams.
  • Involved with Query Optimization to increase the performance.
  • Supporting solution architects in problem analysis and solution design.
  • Developing and optimizing Stored Procedures, Views, and User-Defined Functions for the Application.
  • Support Data and Analytics/Transformation Architecture teams who are building a data strategy aligned with global strategic direction - develop canonical and other models as required, implementing data architecture platforms and solutions and data services, develop MDM foundation, participate in the design and implementation of unified data warehouse.
  • Developing physical data models and creating DML scripts to create database schema and database objects.
  • Created Clustered and Non-Clustered Indexes to improve data access performance.
  • Identified Relationships between tables and enforce referential integrity using foreign key constraints.
  • Created Functional Design Documents and Transaction Definition Documents.
  • Implemented metadata standards, data governance and stewardship, master data management, ETL, ODS, data warehouse, data marts, reporting, dashboard, analytics, segmentation, and predictive modeling
  • Designing dashboards and reports, parameterized reports, predictive analysis in Power BI.
  • Creating dashboards with Combination Charts, Custom Charts based on the requirement.
  • Deploying and managing user permissions for reports and dashboards on Power BI web portal.
  • Creating DAX Queries to generated computed columns in Power BI.
  • Evaluated data profiling, cleansing, integration and extraction tools (e.g. Informatica).
  • Responsible for the Database backup and Restoration using SQL native tool.
  • Partnering closely with business and IT teams in meeting the deadlines pertaining to design and development deliverables and maintaining audit and compliance needs.

Environment: SQL, Business Objects XIR2, ETL Tools Informatica 8.6/9.1, 11G, Enterprise BI in Azure with Azure Data slake/Synapse, Microsoft Power BI

Confidential

Big Data Engineer

Responsibilities:

  • Worked on providing analytics insights and decision support tools for executives for accurate decision making for banking and e-commerce clients.
  • Applied highly advanced data access routines to extract data from source systems for monitoring operations compliance to banking Laws, Rules and Regulations using Visual Basic Apps (VBA), SQL Server SSIS, SAS and SQL.
  • Identified, measured, and recommended improvement strategies for KPIs across all business areas.
  • Analysis of functional and non-functional categorical data elements for data profiling and mapping from source to target data using SSRS.
  • Involved with data profiling/validation for multiple sources using Agility Pro, AWS Redshift, MySQL.
  • Worked with data investigation, discovery, and mapping tools to validate the data across various environments.
  • Wrote and executed unit, system, integration and UAT scripts in a data warehousing applications like Ataccama, EDW.
  • Worked on Troubleshooting test scripts, SQL queries, ETL jobs, data warehouse/data mart/data store models.
  • Extensively used ETL for supporting data extraction, transformations and loading processing, in a complex EDW using Talend/Data stage.
  • Metrics reporting, data mining and trends in helpdesk environment using Access.
  • Written complex SQL queries for validating the data against different kinds of reports generated by Business Objects XIR2
  • Extensively used MS Access to pull the data from various data bases and integrate the data.
  • Worked on SAS for data analytics and data quality checks.
  • Worked with the business in gathering requirements from the existing reporting tool.
  • Designing, developing, testing, and maintaining Tableau reports and dashboards based on user requirements.
  • Working closely with business in creating dashboards based on the data from various Data sources.
  • Created and Monitored workflows using workflow designer and workflow monitor.
  • Involved in extensive DATA validation by writing several complex SQL queries and Involved in back-end testing and worked with data quality issues.
  • Developed regression test scripts for the application and Involved in metrics gathering, analysis and reporting to concerned team and tested the testing programs.

Environment: Aginity Pro-Netezza, Oracle 11G, Teradata SQL Assistant 12.0, AWS Redshift, Tableau, MS-Excel

Confidential 

Big Data Engineer

Responsibilities:

  • Assisted in defining, implementing, and utilizing business metrics calculations and methodologies.
  • Designed and provided complex excel reports including summaries, charts, and graphs to interpret findings to team and stakeholders.
  • Assisted the team for standardization of reports using SAS macros and SQL.
  • Responsible for creation of Credit data related warehouse to help with Risk Assessment for Commercial loans via Microsoft Access data dumps.
  • Performed competitor and customer analysis, risk and pricing analysis and forecasted results for credit card holders on demographical basis.
  • Created macros and used existing macros to develop SAS programs for data analysis.
  • Created and manipulated various management reports in MS Excel for sales metrics using VLOOKUP and Pivot tables.
  • Developed transformation logic for BI tools (Informatica) for data transformation into various layers in Data warehouse.
  • Utilized SQL to develop stored procedures, views to create result sets to meet varying reporting requirements.
  • Used advanced excel formulas (lookup functions, pivot table, If Statements etc.) for analyzing data.
  • Identified process improvements that significantly reduce workloads or improve quality.
  • Worked for BI Analytics team to conduct A/B testing, data extraction and exploratory analysis.
  • Generated dashboards and presented the analysis to researchers explaining insights on the data.

Environment: Microsoft Access, Microsoft Excel, R studio, Informatica Power Center 9.0, MS SQL Server 200.

We'd love your feedback!