We provide IT Staff Augmentation Services!

Datawarehouse Engineer Resume

4.00/5 (Submit Your Rating)

Seattle, WashingtoN

PROFESSIONAL SUMMARY:

  • Overall 13 years of experience in Analysis, Design, Development and maintenance in IT. Data Gathering, Data Analytics, Data visualization, and analyzing large volume transactional data and historical data in Banking, Travel E - commerce, Automotive Manufacturing and Insurance domain.
  • Highly analytical and process-oriented data analyst with in-depth knowledge of database types; research methodologies; and big data capture includes cloud infrastructure, curation, manipulation and visualization. Furnish insights, analytics and business intelligence used to advance opportunity identification, process reengineering and corporate growth
  • Business Intelligence, Data Analytics and Data Warehouse architecture skills.
  • Hands on experience in installing, configuring and using ecosystem components like Hadoop MapReduce, Apache Storm, HDFS, HBase, Zoo Keeper, Oozie, Hive, Cassandra, Sqoop, Pig, Flume, Avro and Informatica Powercenter - ETL tool.
  • Hands on Experience in working with ecosystems like Hive, Pig, Sqoop, Map Reduce.
  • Hands on experience on Terraform cloud and Terraform Enterprise installations and managing versioned infrastructure.
  • Strong Knowledge of Hadoop and Hive and Hive's analytical functions.
  • Implemented Dockers - Containerization. Good understanding on containers vs VMs.
  • Knowledge on enabling Kubernetes, Swarn using Docker.
  • Efficient in building Hive, pig and map Reduce scripts.
  • Integrated Message Queuing and Mango DB using Apache Storm 1.2.2. Re-partitioned the data streams between each stage of computation.
  • Implemented Full Backups, Incremental Back ups, Differential Back ups, Synthetic Backups.
  • Implemented on Hadoop stack and different big data analytic tools, migration from different databases SQL Server2008 R2, Oracle, MYSQL to Hadoop.
  • Successfully loaded files to Hive and HDFS from MYSQL.
  • Loaded the dataset into Hive for ETL Operation.
  • Good knowledge on Hadoop Cluster architecture and monitoring the cluster.
  • Experience in using Zoo keeper and Horton works Hue and HDP.
  • In-depth understanding of Data Structure and Algorithms.
  • Created reports and charts to visualize BigQuery data using Google Data Storage
  • Real time log analysis using Fluentd and BigQuery
  • Performed realtime analysis and implementation on MUFG financial time series data using BigQuery
  • Developed ETL pipelines in and out of data warehouse using combination of Python and Snowflake s SnowSQL and GCP Cloud Fusion.
  • Understanding of snowflake data life cycle - organizing data, storing data, querying data, working with data, Removing Data.
  • Used Column Re ordering, Column Omission, Casts, Truncating Texts, Bulk Loading.
  • Knowledge on automation of snowpipe for S3 Bucket.
  • Handling Cloud parameters URL,Storage Integration, Encryption for snowpipe configuration.
  • Hands on experience on SQL queries against Snowflake.
  • Developed scripts (Unix, Python) to do Extract, Load and Transform data
  • Production support for Data Warehouse issues such data load problems, transformation/translation problems etc.
  • Developed solutions leveraging the Informatica Intelligent Cloud Services (IICS) Platform
  • Demonstrated Informatica Cloud Data Integration (CDI) and Informatica Cloud Application Integration (CAI) solutions at Confidential and MUFG
  • Developed taskflows and linear taskflows
  • Understanding on generation of CDI swagger file
  • Knowledge of various mapping task configuration Defining a mapping task, configuring sources, configuring parameters, view mapping task details, task synchronization, masking tasks and powercenter tasks.
  • Translate requirements for BI and Reporting to Database design and reporting design
  • In Depth understanding of data transformation and translation requirements and which tools to leverage to get the job done
  • Ability to understand data pipelines and modern ways of automating data pipeline using cloud based and on premise technologies - GCP and SnowFlake
  • Actively test and clearly document implementations, so others can easily understand the requirements, implementation, and test conditions..
  • Strong understanding of various data formats such as CSV, XML, JSON etc.
  • Partner with the Data architects, Product managers and Scrum Masters to deliver data integrations and BI solutions required for MUFG Enterprise data Hub.
  • Enable Continuous Delivery (CD) to production for all data warehousing and BI builds. Collaborate with DevOps team to align with CI/CD requirements for assigned projects.
  • Ability to understand end to end data integration requirements and response time SLA s to build data driven solutions that provide best in class customer experience1.
  • Strong understanding of incident management and change management process to support day to day production issues2. Experience working directly with technical and business teams.
  • Implemented DIH Publication Process with batch workflow, real-time workflow, informatica cloud task.
  • Handled Developer user roles across powercenter, big data management mapping, data quality mapping, informatica cloud task.
  • Managed source and target connections for publications and subscriptions in the operation console.
  • Good understanding on DX Start Publication and DX Notification transformations in informatica DIH.
  • Realtime workflows are published through JMS and WSP.
  • Good understanding on Publication repository.
  • Mapping configuration in run time environment and application creation.
  • Created DQ mappings for DIH subscriptions.
  • Understanding on informatica Big data management RPM package and Data integration hub big data management.
  • Configured HBase, Hive, HDFS connection properties using Informatica BDM
  • Understanding on Address validator, Aggregator, Case converter, consolidation, data processor, data Masking and filter transformations in Informatica BDM.
  • Understanding of Hive limitations using SQl transformation in BDM.
  • Understanding on configuring a mapping to run in Hadoop environment (Hive on MapReduce and Blasé engines).
  • Understanding of truncating partitions on HIVE target using BDM.
  • Configured Hadoop cluster to enable comparison on Landing Tables.
  • Performs a key Design Or Development role in the area of MDM using any of Informatica PIM.
  • Understanding requirements performing vendor and product evaluations mentoring team deliver PoCs or solutions and documentation
  • Knowledge of basic Data Architecture principles Data Management Data quality
  • Review and integrate the technical architecture requirements. Provide input into final decisions regarding hardware, network products, system software and security
  • Used Oracle Data Integrator Designer ODI to develop processes for extracting, cleansing, transforming, integrating, and loading data into data warehouse database.
  • Worked on slowly changing dimensions SCD, Change data capture CDC as part of data warehousing concepts.
  • Oracle Data Integrator ODI reduce the dependency on Excel and other proprietary tools for data entry and reporting, and provide property and department level budgeting/forecasting so as to produce a consolidated budget and forecast.
  • Primary responsibility include but not limited to installing and configuring the Oracle Data Integrator ODI software tool in a three-tier environment and performing periodic upgrades, performing source-to-target mappings, storage capacity planning, developing ETL.
  • Specialized in developing stored procedures, functions, packages and database triggers using Oracle SQL and PL/SQL to improve query performance.
  • Established and maintained productive working relationship with Management, Clients and Team Members. Self-starter, entrepreneurial, highly motivated team- player with excellent communication / problem-solving skills. Quick learning ability vis- -vis new technology. Easy adaptability to new environments.
  • PIM installation and upgrades
  • Understanding on PIM repository Design and maintaining Master catalog
  • Experience working on Informatica MDM to design, develop, test and review & optimize Informatica MDM.
  • Familiar with Active VOS features such as Active VOS Designer, Active VOS central, screenflow and server.
  • Expertise in creating Mappings, Trust and Validation rules, Match Path, Match Column, Match rules,
  • Merge properties and Batch Group creation.
  • Experience in creation and maintenance of entity objects, hierarchies, entity types, relationship objects
  • Relationship types using Hierarchy tool to enable Hierarchy Manager (HM) in MDM HUB implementation.
  • Hands on experience in design and configuration of landing tables, staging tables, base objects, hierarchies, foreign - key relationships, lookups, query groups, queries/custom queries and packages.
  • Designed, Installed, Configured core Informatica MDM Hub components such as Informatica MDM Hub Console, Hub Store, Hub Server, Cleanse Match Server, Cleanse Adapter, Data Modeling.
  • Solid expertise in Data Extraction, Data Migration, Data Transformation and Data Loading using ETL process in Informatica Power Center 9.x/8.x/7.x.
  • Experience in designing Reusable Transformations in Informatica such as Joiner, Sorter, Aggregator, Expression, Lookup, Router, Filter, Update Strategy, Sequence Generator, Normalizer and Rank) and Mappings using Informatica Designer and processing tasks using Workflow Manager to move data from multiple sources into targets.
  • Worked on Data Profiling using IDE-Informatica Data Explorer and IDQ-Informatica Data Quality to examine different patterns of source data. Proficient in developing Informatica IDQ transformations like Parser, Classifier, Standardizer and Decision.
  • 11 years of extensive experience in Data Warehouse applications using
  • Informatica Power Center, Tableau, VB Scripts, Advanced Excel, Informatica IDQ, Oracle, DB2, MS SQL server on Windows, IBM an d UNIX/Linux platforms.
  • 3 years of experience in data analytics - R Programming.
  • Prepared scripts to ensure proper data access, manipulation and reporting functions with R programming languages
  • Developed tactical and strategic plans to implement technology solutions and effectively manage client expectations.
  • Road mapping technologies, Project Design implementation and review, Project audit to ensure deliverables are in accordance with the architecture standards.
  • Developed effective working relationships with client team to understand support requirements.
  • Experienced to work in Development team, Production support teams in handling critical situations to meet the deadlines for successful completion of the tasks/projects.
  • Excellent interpersonal and communication skills, technically competent and result- oriented with problem solving skills and ability to work independently and use sound judgment.
  • Strong expertise in designing and developing Business Intelligence solutions in staging, populating Operational Data Store (ODS), Enterprise Data Warehouse (EDW), Data Marts Decision Support Systems using Informatica Power Center 9.x/8.x/7.x/6.x ETL tool.
  • Expertise in Data Modeling using Star Schema/Snowflake Schema, OLAP/ROLAP tools, Fact and Dimensions tables, Physical and logical data modeling using ERWIN 4.x/3.x
  • Experience in documenting High Level Design, Low level Design, STM's, Unit test plan, Unit test cases and Deployment documents.
  • Experienced in Repository Configuration/using Transformations, creating Informatica Mappings, Mapplets, Sessions, Worklets, Workflows, Processing tasks using Informatica Designer / Workflow Manager to move data from multiple source systems into targets.
  • Experienced in Installation, Configuration, and Administration of Informatica Power Center 9.x/8.x/7.x/6.x.
  • Experienced in Performance tuning of Informatica (sources, mappings, targets and sessions) and tuning the SQL queries.
  • Experienced in integration and transforming of various data sources from Databases like MS Access, Oracle, DB2, SQL Server and formats like flat-files, COBOL files, XML, etc.
  • Experienced in using ETL tools like Informatica (Power Center) Designer, Repository Manager, Administration console and Workflow Manager.
  • Experience in Oracle and MS SQL Server environments using Triggers, functions, SQL, T-SQL and PL/SQL.
  • Expertise in scheduling Informatica jobs using Informatica, Windows scheduler and with Unix.
  • Expertise in creating Unix shell scripts.
  • Experienced in working for the post development cycle and applications in Production Support.
  • Strong Communication skills of written, oral, interpersonal and presentation
  • Ability to perform at a high level, meet deadlines, adaptable to ever changing priorities
  • Implemented established Software Development Lifecycles and methodology to support the success of the larger IT organization, including waterfall, iterative, agile and scrum
  • Ability to produce realistic software schedules, Project Management, Strategic/ Tactical Planning, Business Analysis and Development, Budgeting/Forecasting, and New Product Development.

CORE COMPETENCIES:

  • Leadership Skills Customer Service Client Relationship Management Sales Team Management & Development Organizing Skills
  • 3Organized and goal-oriented, with more than 10 years’ experience in project management and a strong customer focus.
  • Analytical problem- solver, able to anticipate issues and create new systems that streamline operations, resolve concerns and improve efficiency. skilled in building excellent rapport with clients and team members.
  • Able to see the big picture, delegate effectively and motivate team members to achieve on-time project completion

TECHNICAL SKILLS:

Computer skills include: MS Word, MS Excel, MS Access, MS Project Publisher, MS Powerpoint *MS - Microsoft tools

Data Integration: Informatica Data Integration Hub 10.2, Informatica IICS 10.XPaaS Cloud Infrastructure: SnowFlake Elastic Cloud Datawarehouse, Google Cloud Platform

Modeling Tools: Erwin 4.0 data modeler, ER studio 7.5, MS Visio 2007

ETL tool: GCP Cloud Fusion, GCP Cloud Data Flow, Informatica Power center 10.x

Data Backup and Recovery Tool: Veritas Net Backup 8.1

Streaming Application: Apache Storm

Elastic Container: Docker Platform

Database: Oracle, DB2, SQL server

Data Mining: SAS

Data Analytics: R Programming and Python scripting

Bigdata: Hadoop Architecture and Big Data tools, Informatica BDM 10.0

Scripting: Unix Shell/Perl Scripting, VB scripting, Advanced Excel

OS: Lunix /Windows

Methodologies: Agile, waterfall, UML, Design Patterns

Project Tracking/ Defect Tracking/ Production Issue Tracking: JIRA

Code Repository: GitHub

Scheduler: AirFlow, Google Composer, AutoSYS, Crontab

PROFESSIONAL EXPERIENCE:

Confidential, Seattle, Washington

Datawarehouse Engineer

Environment: Google Cloud Platform, Snowflake DWAS, Google Cloud Data Flow, Google Cloud Data Fusion, Oracle 10g, SQL server 2005, SQL, T-SQL, PL/SQL, Toad, Erwin4.x, Unix, Flat files, Tableau, SQL, T-SQL, PL/SQL, Toad, Erwin4.x, GItHub

Confidential, Monterey Park, California

Applications Consultant (Cloud Datawarehouse)

Environment: Google BigQuery, Snowflake DWAS, Informatica IICS, Informatica Data Integration Hub 10.2, Informatica 10.x, Oracle 10g, SQL server 2005, SQL, T-SQL, PL/SQL, Toad, Erwin4.x, Unix, Tortoise SVN, Flat files, Mainframes, CICS, AS400, DB2, Oracle, Apache Storm, Marketo, Veritas Net Backup 8.1, Terraform, ECS, Tableau, SQL, T-SQL, PL/SQL, Toad, Erwin4.x, Unix, Tortoise SVN, Flat files, Hadoop Architecture and Big Data tools - MapReduce, HDFS, HBase, Zoo Keeper, Oozie, Hive, Cassandra, Sqoop, Pig, Flume, Avro, Unix Shell Scripting, Perl Scripting, R Programming

Confidential, CA

Datawarehouse Developer

Environment: Informatica 10.x, Oracle 10g, Oracle Data Integrator 11g, Informatica PIMSQL server 2005, SQL, T-SQL, PL/SQL, Toad, Erwin4.x, Unix, Tortoise SVN, Flat files, Mainframes, CICS, AS400, DB2, Oracle, Apache Storm, Marketo, Veritas Net Backup 8,Oracle 10g, SQL server 2005, SQL, T-SQL,PL/SQL, Toad, Erwin4.x, Unix, Tortoise SVN, Flat files

Confidential

BI & Datawarehouse Developer

Environment: Informatica Power center 8.6.1, Oracle 10g, Windows XP, Unix ShellScripts, SQL, PL/SQL, Flat files

Confidential

Datawarehouse Developer

Environment: Informatica Power center 6.1/7.1, Oracle 9i,ODI Informatica PIM, SQL Server 2000, SQL,PL/SQL, TOAD, Windows NT, Unix- Shell Scripting

Responsibilities:

  • Received, utilizing various reports to determine needs within the department.
  • Assisted in the preparation of documentation.
  • Performed web based online and other effective and advanced methods available.
  • Maintained key facility staff informed of programs, schedules, employee participation and performance level.
  • Collaborated with Patient Access Management in targeting key areas of performance for the facilities.
  • Delivered project needs on time and within the agreed acceptance criteria in a hybrid methodology environment as they attempted to transition to an Agile Methodology.
  • Reported progress to all stakeholders through Sprint Burn-Down Reports, Iteration Burn-Down Reports, and velocity target updates.
  • Wrote User Stories based on the requirements gathered from all the stake holders.
  • Organized and facilitated Agile and Scrum meetings, which included Sprint Planning, Daily Scrums or Standups, Sprint Check-In, Sprint Review & Retrospective.
  • Coordinated with systems partners to finalize designs and formalize requirements Utilized Story. Sizing and Planning Poker techniques as needed based on the length of the backlog and priorities.
  • Operated in 2 week sprints with flexibility on length based on immediate functionality concerns.
  • Ensured application availability and data integrity through preventative maintenance and upgrades.

We'd love your feedback!