We provide IT Staff Augmentation Services!

Senior Data Engineer/architect Resume

5.00/5 (Submit Your Rating)

Burlington, MA

SUMMARY:

  • Developed complex visualizations and machine learning models to determine analytical approaches and modeling techniques. Created prototypes.
  • Expertise in building logical & physical data models for Proof - Of- Concept analytics & reporting requirements using various analytical engines like Tableau, Si Sense, Power BI.
  • Expertise & good understanding of statistical hypothesis testing and experimental design, data visualization.
  • Excellent experience in building machine learning models & exploratory data analysis with Python IDE with Seaborn/Matplotlib libraries.
  • Translate business problems to an analytics problem, recommending and applying the most appropriate methods to yield insights and results. Worked on building data models/ETL/business rules for meeting data needs for numerous studies between clients & publications.
  • Expertise in applying analytical and statistical methods to analyze large amounts of data, using advanced statistical techniques.
  • Built technical plan and strategy to Lead change to data analysis and reporting processes for R&D Medical device algorithmic enhancement and other data science initiatives.
  • Expertise in building tools/dashboards to automate reporting and generate insights that will allow QI Managers/Directors to quickly and accurately see how our programs and processes are performing.
  • Ensure data integrity of the supported financial reports and dashboards. Ensures required audit documentation and validation reports for applications are maintained.
  • Built data governance initiative for company-wide data reporting for Outcomes & Patient Health metrics through representation on internal initiatives including master data, data lake and other projects.
  • Collaborate with business partners to understand their problems and goals, develop predictive modeling, statistical analysis, data reports and performance metrics.
  • Advised management & business stakeholders on use of quality business analytics, tools, and market information for approaches & opportunities.
  • Collected, Interpreted and analyzed data across to support management decisions on projects, improve inefficiencies, and enhance business performance.
  • Strong experience integrating loosely-coupled systems, both internally and externally, utilizing API’s and file-based integration methods.
  • Developed code and prepared data for diagnostic and predictive analytics.
  • Identified, measured and recommended improvement strategies for KPIs across all business areas.
  • Designing and developing solutions for ingestion, curation, organization, and dissemination of data in support of Data Engineering and Analytics projects.
  • Great experience working with building data lakes on Azure.
  • Provided subject matter expertise in addressing projects and issues that encompass a wide range of internal and external systems (Core banking, Wealth, Data Warehousing), components, and processes
  • Work closely with business stakeholders and other engineering teams to research technical requirements and resolve issues
  • Develop, test, and maintain user-facing semantic layers, i.e. cubes, models. Provide production support and troubleshooting as needed.

PROFESSIONAL EXPERIENCE:

Confidential, Burlington, MA

Senior Data Engineer/Architect

  • Created visualizations for risk adjustments/Visit performance with Qlik Sense and Tableau showing how each provider and medical facility are performing with respect to HEDIS measures for utilization and cost calculation.
  • Extensively used Postgres SQL to Critically evaluate information gathered from multiple sources, reconcile conflicts, and decompose high-level information into detailed reports & dashboards.
  • Deep understanding of data modeling and the use of data modeling tools.
  • Strong experience using development tools including Microsoft Visual Studio, SSDT, and SQL Server Management Studio (SSMS).
  • Solid understanding of software engineering and processes (i.e. design patterns, algorithms, data structures, schemas and queries, system design, unit testing, code reviews, Agile, and DevOps).
  • Strong written, verbal, and interpersonal communication skills.
  • History of successful software design and development and utilization of latest technologies.
  • Designing and developing solutions for ingestion, curation, organization, and dissemination of data in support of Data Engineering and Analytics projects.
  • Build data pipelines using Hadoop, Spark, Hive, in specific built scripts in Sqoop to import, export and update the data to HDFS.
  • Build data exploratory analysis with Python (3.x) with understanding of concurrency, performance, optimization and implementation of algorithms in devOps environment.
  • Good knowledge of Hadoop HDFS, Hive, Pig, Flume and Sqoop. Understanding of data loading tools such as Flume, Sqoop.
  • Good knowledge of database principles, practices, structures, and theories
  • Implement code changes, Conduct SQL Code Reviews and Maintain and improve existing database objects
  • Leverages industry best practices in ETL, data models, query optimizations & analytics; share same with team.
  • Develop, test, and maintain user-facing semantic layers, i.e. cubes, models. Provide production support and troubleshooting as needed
  • Lead new development efforts for adding new structures to existing data model or maintenance of existing operational processes.
  • Worked extensively to store code in GitHub & deploy code using Jenkins OR Integrated Development Environment using Microsoft Team Foundation services.
  • Managed Agile Development cycle/Kanban board and apply the methodologies to story points; allocate time with Sprint cycle to complete development for any of data integration/architecture work.

Environment: QLIK Sense, Python, Postgres, AWS, PySpark

Confidential, Beverly, MA

Senior BI/ETL Developer

  • Built Dashboards in QLIK View to assist business owners with Self-Serve analytics. Created Dashboards with different style using QlikView components like List Box, Slider, Buttons, charts and Bookmarks
  • Built ETL processes in SSIS & data model to Load & extract data from Salesforce.com to existing CRM warehouse for supporting the sales & marketing strategy for GE Synchrony Project.
  • Worked extensively in extracting data from multiple sources, integrate disparate data into a common data model, and integrate data into a target database, application, or file using efficient ETL processes.
  • Worked extensively on VLDB & Implementation of complex business logic for matching, linking multiple data sources and deriving new attributes from the data sources.
  • Good Knowledge of Sqoop to write scripts to import, export and update the data to HDFS. Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Good knowledge of Hadoop HDFS, Hive, Pig, Flume and Sqoop. Understanding of data loading tools such as Flume, Sqoop.
  • Good knowledge of database principles, practices, structures, and theories
  • Implement code changes, Conduct SQL Code Reviews and Maintain and improve existing database objects
  • Leverages industry best practices in ETL, data models, query optimizations & analytics; share same with team.
  • Developed complex T-SQL queries and designed SSIS packages to load the data into Marketing/Sales warehouse.
  • Experience with of all aspects of data systems (both Big data and traditional) including database design, ETL/ELT, aggregation strategy, performance optimization.
  • Very strong Product knowledge in business rules and building interfaces in Life/Annuity/Mutual Fund

Environment: Salesforce.com, QLIKVIEW, SSIS, Python, Hadoop, SQL

Confidential, Philadelphia, PA

IT MDM Lead

  • Worked on resolving data quality issues by profiling the data and building data quality metrics for Data Management group.
  • Developed custom reports to help clients in building data feeds for State & Federal compliance reporting.
  • Provided guidelines and Subject matter expertise with relation to Data collection and storage to Internal & external source data providers and resolve data quality issues between sources & IBM Solution team for all ETL Inbound Spec and file format related questions.
  • Worked on automating the ETL process through scheduling and exception-handling routines as well as source to target mapping development, support, and maintenance in IPM.
  • Imported Source/Target Tables from the respective databases and created reusable transformations (Joiner, Routers, Lookups, Rank, Filter, Expression, and Aggregator) in a Mapplet and created new mappings using Designer. Build Data quality processes using Informatica IDQ 8.6.1 to Analyze, troubleshoot, diagnose Data Quality workflow issues & recommend and implement appropriate solutions.
  • Built ETL processes in SSIS/Talend accessing various data sources like flat files/XMLfiles/Oracle DB/PDF documents for creating a consolidated view of data to review and clean the data prior to data submissions.
  • Very strong background in data management/analysis, project management, and the ability to translate business needs into requirements and assist throughout the development cycle.
  • Great management expertise with monitoring, disaster recovery, backup, automated testing, automated schema migration, and continuous deployment

Environment: T-SQL/PLSQL, Informatica 9.0, IDQ, Oracle 10g, SQL Server 2012

Confidential, Philadelphia, PA

Senior DW/ETL Developer

  • Build processes to assist Actuarial/Rating team in building SQL procedures for pulling and loading data with relation to coverage's, rating factors, premium & exposures to Person Lines Warehouse.
  • Working extensively on data, data issues, data profiling, actuarial Formulae errors and SQL Optimization.
  • Designing SSIS Packages for data conversion from Flat files/XML files to SQL server tables using several transformations to perform Data profiling, Data Cleansing and Data Transformation
  • Developed ETL solution using SQL procedures/Informatica mappings to load data for Claims/Loss DataMart

Environment: Microsoft SQL Server 2016, T-SQL, Excel/XML, SSIS,SSAS,SSRS

Confidential, Philadelphia, PA

Senior ETL Developer

  • Mentor & train new employees or contractors in ETL concepts/DW processes and provide standard operating procedures for deployment of code.
  • Worked on automating the ETL process through scheduling and exception-handling routines as well as source to target mapping development, support, and maintenance in autopsies.
  • Worked on Performance tuning of long running exception handling jobs in production and tuning Oracle SQL and redeveloping the code to bring the load time from 3.5 hrs. to 28 secs.
  • Worked on data conversion & data integration initiatives at Barclay's capital for conversion of data from legacy/mainframe files to client server and integrating the load process with product catalog system.
  • Built ETL processes to load data from complex legacy VSAM & Sequential files for APB into
  • Outbound files for Micro Focus or database tables.
  • Aided in automating ETL testing by creating ETL/SQL scripts & test data required for validating business rules and data checks of data & ETL load process.
  • Worked with operations on reviewing JIRA tickets, opening Change Requests, scheduling ETL jobs on production and providing required scheduling documentation.
  • Worked on Upgrade and Migration of code from Informatica 8.1.1 to 8.6.1.

Environment: Informatic 8.6.1, Trillium, Netezza

Confidential, Wayne, PA

Senior Data Integration Analyst

  • Worked to provide a single consistent view of data to all Commercial Business users across North America & UK which enabled faster data remediation and greater data certainty for International Commercial Business Insight Project.
  • Built Data Migration processes to convert & migrate data from MySQL/Oracle/SAP by building data model and ETL code to Load to Netezza DB using NZSQL by keeping the business logic same and validating the results across the disparate systems.
  • Assisted in building BO Universes and setting up relationships between source tables and defining dimensions/metrics & setting up of security access level.in Universe
  • Worked on Trillium Software- Parser for Data parsing (elementizing) - context and destination of each component of each field & Data standardization to forma that are standard throughout the DW.
  • Worked on building back-end end to end solution using Power Center 8.6 for International commercial business Insight project.
  • Worked on providing complete load schedule for execution of load with focus on points of failure and recovery.
  • Provide ETL design, development and testing estimation based on business requirements and research into source data.

Environment: Erwin, Informatica 8.6.1, Business Objects, Oracle 10G, Power Exchange for SAP/NETEZZA

Confidential, Conshohocken, PA

Senior BI/DW Consultant, Idea Integration

  • Worked as data architect for Financial & Expense reporting data by formulating business rules, working on data model & data migration projects.
  • During the requirements gathering phase assisted the business in identifying potential data quality and system needs that would improve the effectiveness of data development and QA/QC testing.
  • Assisted in building the IDQ environment and providing help with setup and implementation of data cleansing routines for Address standardization and matching algorithms
  • Worked as data architect for Financial & Expense reporting data by formulating business rules, working on data model & moving data from DB2 to Oracle 10G.
  • Worked on building ETL processes for moving data in & out of SAP R3.
  • Worked on Reviewing interfaces and assessing for criticality in the production process and documenting and potentially upgrading the environments.

TECHNICAL SKILLS:

BI/Data Visualization tools: QLIK Sense/ Tableau

Data Integration Tools: Talend. SSIS, Informatica 9.5

Programming: T-SQL, SSIS, Python

Databases: Oracle 11g, SQL Server 2016, PostgreSQL

Cloud: Azure DW/Data Lake/AWS

We'd love your feedback!