We provide IT Staff Augmentation Services!

Etl, Bi & Dw Lead Engineer Resume

5.00/5 (Submit Your Rating)

SUMMARY

  • Over 8 years of Exclusive Experience in Ab Initio (1.9.1 to 3.3.5). Expertise in ETL tools using Ab Initio GDE, Co>operating system, EME TR, Web Services and Metadata Hub.
  • Over 4 plus years of experience in Python scripting.
  • Over 2 plus years of experience in No SQL databases like MongoDB.
  • Over 4 plus years of experience in Cloud Technologies like AWS S3, AWS ECS, AWS Cloud watch.
  • Over 3 plus years of experience in Apache Spark.
  • 8 plus years of professional IT experience in Analysis, Design, Development, Testing and Implementation of various Data Warehousing, Reporting and Software Applications.
  • Over 6 plus years of Exclusive Experience in Informatica power center 10/9.1/8.6/8.5.
  • Over 7 years of hands on experience in developing Business Intelligence reports like (Tableau, Power BI, SAP BO, AWS Quick sight & OBIEE.
  • Having good knowledge with Java and Big Data Technologies.
  • Over 6 plus years of experience in handling the Offshore team of 15.
  • Experience in preparing code documentation in support of application development, including high level and detailed design documents, unit test specifications, interface specifications, etc.
  • Worked with different source systems ODS (Operational Data Source), DW (Data Warehouse) built on different RDBMS and built Data validation and cleansing process using Ab Initio.
  • Experience in Ab Initio EME/Sandbox to implement version control and impact analysis to manage various project across the organization.
  • Ability to demonstrate command and use of industry accepted Ab Initio best practices.
  • Strong development skills including the ability to work through the entire development life (SDLC) cycle from gathering requirements through implementation, development, production, support and documentation of the complete project.
  • Expertise in creating generic code objects where possible for graphs, plans, feeds and rules to increase the code reusability and decrease the future work.
  • Proficient in various Ab Initio Parallelism, Transformation and Multi file System techniques.
  • Thorough knowledge of DML (Data Manipulation Language), Unix Shell scripting.
  • Developed various UNIX shell wrappers to run Ab Initio and Data base jobs. Practical experience with working on multiple environments like production, development, testing.
  • Experience in working with Data Profiler and Business Rule Engine (BRE).
  • Very Good Programming skills using UNIX/Shell Scripts, Python and PL/SQL for Automation of ETL process.
  • Worked with different source systems like Oracle, DB2, Teradata, MS SQL.
  • Extensive knowledge in Dimensional Data Modeling, Star schema, Snow - Flake schema, creation of Fact and Dimension Tables, OLAP, OLTP and thorough understanding of Data-Warehousing concepts.
  • Excellent Knowledge of Data Warehousing Concepts and Metadata Concepts.

TECHNICAL SKILLS

CRM: Siebel CRM (Loyalty/Marketing), Siebel Tools, EIM, eScript, Configuration

Databases: Oracle, SQL Server, Redshift, Cassandra, Mongo DB, Teradata

Operating Systems: Windows XP/2007/8/10, Linux, Unix

Languages: Scala, Python, Java

Reporting Tools: OBIEE, Power BI, Tableau, AWS Quick sight, SAP BO

Project Management Tools: Confluence, Jira, DASH

ETL Tools: Informatica, Ab-initio, SSIS, Oracle ODI(ELT), Pentaho

Big Data Technologies: Apache Spark, Apache Kafka, Hadoop

SDLC: Agile, Waterfall

Project Management: Jira, Confluence, Rally, Test Rails

Cloud Technologies: AWS (S3, EC2, ECS, CloudWatch, EC

Container Technologies: Docker, Docker Compose

Scripting Languages: Unix Shell Script, PL /SQL, SQL

Testing Tools: HP- Quality Center, Mantis, HP ALM, Test Rails

Scheduling/Monitoring Tools: Control-M, Crontab, DAC, Tivoli

Software Tools: GitHub, Jenkins, Toad for Oracle, SQL Workbench, Microsoft Office, SQL Developer, FTP, SFTP, SCP, Putty, Upload, Atlas, Robo-3T, SQL Management Studio

PROFESSIONAL EXPERIENCE

Confidential

ETL, BI & DW Lead Engineer

Responsibilities:

  • Involved in architecture, flow and the database model of the application.
  • Developed the API connectors using Python to parse the Web based data for Website teams.
  • Developed and used the Python Spark to build the transformation to read data from MongoDB.
  • Developed the data tracing log analyzer to track the data flows for the ED&A Analytics Live Dashboards.
  • Developed the ETL jobs as per the requirements to update the data into the database (Teradata) from various data sources and REST API’s.
  • Developed the various data operations in python using Numpy and Pandas.
  • Developed an Analytical and process monitoring dashboards using plotly and matplotlib.
  • Developed Analytical queries in Teradata, SQL-Server, and Oracle.
  • Developed a grid to store and reutilize the data files to multiple buckets of AWS S3.
  • Developed a Web service on the Teradata database using python Flask framework which was served as a backend for the real-time dashboard.
  • Participated in building the Predictive Analytics using the algorithms like Linear Regression, K- Means.
  • Developed reverse engineering practice of ETL loads with date and file based using the python and shell scripts.
  • Wrote Unit and Integration Tests for all the ETL services.
  • Worked on optimizing and memory management of the ETL services

Tools: & Technologies: Ab Initio 3.1.5/3.2.5/3.3.5 , Metadata Hub 3.2.5, Teradata (14.0,15.0,16.0), Oracle 11.2, Oracle 12c, Python, Linux, Flat Files, XML, Tivoli Scheduler, Salesforce, SQL, Shell Scripts, SAP BO, Power BI, Tableau, JIRA, DASH and Windows XP/7.

Confidential

ETL, BI & DW Lead Engineer

Responsibilities:

  • Interacted with senior management and stake holders on various aspects of the project, guided BSA in gathering requirements; provide guidance to finalize the file layouts to match the target data model, interacting with various teams-data modelers, dbas and stake holders.
  • Provided architecture for a customizable, extendable, scalable and performance tuned solution.
  • Provided leadership for resolving various technical issues on the project which comprised of 10+ member team both on-site and offshore.
  • Worked in creating the data pipelines for Business reporting and Technology costing group.
  • Handled development for major critical components, developed many common graphs for standardized and expandable solution. Worked in all stages of the development lifecycle.
  • Developed custom components for various ETL functions. Used PDL and meta-programming extensively.
  • Interacted with various testing teams.
  • Coordinated on site and offshore resources.
  • Provided research / analysis skills for root cause analysis for defects and for accommodating new requirements.
  • Supported the team in deploying the code base in different test environments.
  • Guided metadata team in loading Technical Metadata from Ab Initio TR.

Confidential

Sr. Software Developer

Responsibilities:

  • Installations of the Databases, Informatica & OBIEE in servers of AWS. (For Century21, Pacsun & Ampersand)
  • Creating the data models for the Retail business and getting it approved with Solution Architects.
  • Creating the services for the installations for auto start and stop for server restarts.
  • Requirements Gathering and estimate the effort required.
  • Design the ETL flow to process data and transform into format that facilitates reporting.
  • Develop Ab Initio graphs to read the source data and apply transformation logic as per the business rules and prepare data to load into the target tables.
  • Build and schedule Unix Shell Scripts to invoke the Ab Initio Graphs.
  • Involved in code review and performance tuning of ETL graphs.
  • Involved in Data profiling that helps in redesigning the database.
  • Coordinate and support Integration Testing & User Acceptance Testing.
  • Analyzing and fixing defects during testing cycles. Analyzing Productions issues, TRs and fixes either the code or creating ad-hoc graphs to cater the requirement.
  • Migration and deployment of code in development, IST, UAT and Production.
  • Responsible for managing deliverables and meeting project schedules/deliverables.
  • Responsible for providing effective system solutions for technical issues identified during post deployment phase.
  • Worked on Informatica Power Center tools- Designer, Repository Manager, Workflow Manager, and Workflow Monitor.
  • Parsed high-level design specification to simple ETL coding and mapping standards.
  • Designed and customized data models for Data warehouse supporting data from multiple sources on real time.
  • Involved in building the ETL architecture and Source to Target mapping to load data into Data warehouse.
  • Created mapping documents to outline data flow from sources to targets.
  • Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.
  • Extracted the data from the flat files and other RDBMS databases into staging area and populated onto Data warehouse.
  • Maintained stored definitions, transformation rules and targets definitions using Informatica repository Manager.
  • Used various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings in the Informatica Designer.
  • Developed mapping parameters and variables to support SQL override.
  • Created mapplets to use them in different mappings.
  • Developed mappings to load into staging tables and then to Dimensions and Facts.
  • Used existing ETL standards to develop these mappings.
  • Worked on different tasks in Workflows like sessions, events raise, event wait, decision, e-mail, command, worklets, Assignment, Timer and scheduling of the workflow.
  • Created sessions, configured workflows to extract data from various sources, transformed data, and loading into data warehouse.
  • Used Type 1 SCD and Type 2 SCD mappings to update slowly Changing Dimension Tables.
  • Extensively used SQL* loader to load data from flat files to the database tables in Oracle.
  • Modified existing mappings for enhancements of new business requirements.
  • Used Debugger to test the mappings and fixed the bugs.
  • Wrote UNIX shell Scripts & PMCMD commands for FTP of files from remote server and backup of repository and folder.
  • Involved in Performance tuning at source, target, mappings, sessions, and system levels.
  • Prepared migration document to move the mappings from development to testing and then to production repositories.
  • Scheduled the Mapping’s in Data Administration Console for Hourly and Daily updates.
  • Analyzing the Reporting Requirements and participated in the documentation of Business Requirements.
  • Developing Oracle BI Metadata repository using BI Administration tool.
  • Developing interactive Dashboards and Reports in Oracle BI Answers as per the reporting specifications.
  • Defining and creating Initialization Blocks, Repository and Session Variables for generating filters and defining security.
  • Creating Hierarchies, Levels, and implemented Business Logic by creating level based measures in OBIEE business model & mapping layer.
  • Developing measure for time comparison using time functions.
  • Using Filters and provided Prompts appropriately to make interactive reports and to avoid excessive data showing up on Reports.
  • Customizing/Creating presentation layer catalog folders, to show different views of the business model to different users and assigned group permissions to the catalog folders.
  • Performing the Performance Tuning of Dashboards and Reports and made required changes to repository and database in consultation with DBA.
  • Implementing security based on LDAP Authentication.
  • Configuring Measures in the Business Model Layer and mapped to multiple logical table sources.
  • Creating schedulers and developed iBots to deliver reports and dashboards to the subscribers as per the schedules and alerts.
  • Creating report prototypes and presented them to users to get approval.
  • Attending weekly meetings with BA and Stakeholders for project status updates.
  • Performing unit and integration testing of reports and dashboards.
  • Creating Requirements Traceability Matrix and Mapping Document to trace requirements in BI implementation.
  • Conducting Functional Requirements review and walk through with BI Architect, Designers, and Stakeholders.
  • Analyzing the data and developed Logical Star Schema model with consultation from Data Modelers and Business Analysts.
  • Developing metadata repository using OBIEE Administration tool in Physical, Business Model and Mapping, and Presentation Layer.
  • Creating new logical columns, dimensional hierarchy, calculated measures, and aggregate mappings in the BMM layer as per the business requirements.

We'd love your feedback!