Etl, Bi & Dw Lead Engineer Resume
SUMMARY
- Over 8 years of Exclusive Experience in Ab Initio (1.9.1 to 3.3.5). Expertise in ETL tools using Ab Initio GDE, Co>operating system, EME TR, Web Services and Metadata Hub.
- Over 4 plus years of experience in Python scripting.
- Over 2 plus years of experience in No SQL databases like MongoDB.
- Over 4 plus years of experience in Cloud Technologies like AWS S3, AWS ECS, AWS Cloud watch.
- Over 3 plus years of experience in Apache Spark.
- 8 plus years of professional IT experience in Analysis, Design, Development, Testing and Implementation of various Data Warehousing, Reporting and Software Applications.
- Over 6 plus years of Exclusive Experience in Informatica power center 10/9.1/8.6/8.5.
- Over 7 years of hands on experience in developing Business Intelligence reports like (Tableau, Power BI, SAP BO, AWS Quick sight & OBIEE.
- Having good knowledge with Java and Big Data Technologies.
- Over 6 plus years of experience in handling the Offshore team of 15.
- Experience in preparing code documentation in support of application development, including high level and detailed design documents, unit test specifications, interface specifications, etc.
- Worked with different source systems ODS (Operational Data Source), DW (Data Warehouse) built on different RDBMS and built Data validation and cleansing process using Ab Initio.
- Experience in Ab Initio EME/Sandbox to implement version control and impact analysis to manage various project across the organization.
- Ability to demonstrate command and use of industry accepted Ab Initio best practices.
- Strong development skills including the ability to work through the entire development life (SDLC) cycle from gathering requirements through implementation, development, production, support and documentation of the complete project.
- Expertise in creating generic code objects where possible for graphs, plans, feeds and rules to increase the code reusability and decrease the future work.
- Proficient in various Ab Initio Parallelism, Transformation and Multi file System techniques.
- Thorough knowledge of DML (Data Manipulation Language), Unix Shell scripting.
- Developed various UNIX shell wrappers to run Ab Initio and Data base jobs. Practical experience with working on multiple environments like production, development, testing.
- Experience in working with Data Profiler and Business Rule Engine (BRE).
- Very Good Programming skills using UNIX/Shell Scripts, Python and PL/SQL for Automation of ETL process.
- Worked with different source systems like Oracle, DB2, Teradata, MS SQL.
- Extensive knowledge in Dimensional Data Modeling, Star schema, Snow - Flake schema, creation of Fact and Dimension Tables, OLAP, OLTP and thorough understanding of Data-Warehousing concepts.
- Excellent Knowledge of Data Warehousing Concepts and Metadata Concepts.
TECHNICAL SKILLS
CRM: Siebel CRM (Loyalty/Marketing), Siebel Tools, EIM, eScript, Configuration
Databases: Oracle, SQL Server, Redshift, Cassandra, Mongo DB, Teradata
Operating Systems: Windows XP/2007/8/10, Linux, Unix
Languages: Scala, Python, Java
Reporting Tools: OBIEE, Power BI, Tableau, AWS Quick sight, SAP BO
Project Management Tools: Confluence, Jira, DASH
ETL Tools: Informatica, Ab-initio, SSIS, Oracle ODI(ELT), Pentaho
Big Data Technologies: Apache Spark, Apache Kafka, Hadoop
SDLC: Agile, Waterfall
Project Management: Jira, Confluence, Rally, Test Rails
Cloud Technologies: AWS (S3, EC2, ECS, CloudWatch, EC
Container Technologies: Docker, Docker Compose
Scripting Languages: Unix Shell Script, PL /SQL, SQL
Testing Tools: HP- Quality Center, Mantis, HP ALM, Test Rails
Scheduling/Monitoring Tools: Control-M, Crontab, DAC, Tivoli
Software Tools: GitHub, Jenkins, Toad for Oracle, SQL Workbench, Microsoft Office, SQL Developer, FTP, SFTP, SCP, Putty, Upload, Atlas, Robo-3T, SQL Management Studio
PROFESSIONAL EXPERIENCE
Confidential
ETL, BI & DW Lead Engineer
Responsibilities:
- Involved in architecture, flow and the database model of the application.
- Developed the API connectors using Python to parse the Web based data for Website teams.
- Developed and used the Python Spark to build the transformation to read data from MongoDB.
- Developed the data tracing log analyzer to track the data flows for the ED&A Analytics Live Dashboards.
- Developed the ETL jobs as per the requirements to update the data into the database (Teradata) from various data sources and REST API’s.
- Developed the various data operations in python using Numpy and Pandas.
- Developed an Analytical and process monitoring dashboards using plotly and matplotlib.
- Developed Analytical queries in Teradata, SQL-Server, and Oracle.
- Developed a grid to store and reutilize the data files to multiple buckets of AWS S3.
- Developed a Web service on the Teradata database using python Flask framework which was served as a backend for the real-time dashboard.
- Participated in building the Predictive Analytics using the algorithms like Linear Regression, K- Means.
- Developed reverse engineering practice of ETL loads with date and file based using the python and shell scripts.
- Wrote Unit and Integration Tests for all the ETL services.
- Worked on optimizing and memory management of the ETL services
Tools: & Technologies: Ab Initio 3.1.5/3.2.5/3.3.5 , Metadata Hub 3.2.5, Teradata (14.0,15.0,16.0), Oracle 11.2, Oracle 12c, Python, Linux, Flat Files, XML, Tivoli Scheduler, Salesforce, SQL, Shell Scripts, SAP BO, Power BI, Tableau, JIRA, DASH and Windows XP/7.
Confidential
ETL, BI & DW Lead Engineer
Responsibilities:
- Interacted with senior management and stake holders on various aspects of the project, guided BSA in gathering requirements; provide guidance to finalize the file layouts to match the target data model, interacting with various teams-data modelers, dbas and stake holders.
- Provided architecture for a customizable, extendable, scalable and performance tuned solution.
- Provided leadership for resolving various technical issues on the project which comprised of 10+ member team both on-site and offshore.
- Worked in creating the data pipelines for Business reporting and Technology costing group.
- Handled development for major critical components, developed many common graphs for standardized and expandable solution. Worked in all stages of the development lifecycle.
- Developed custom components for various ETL functions. Used PDL and meta-programming extensively.
- Interacted with various testing teams.
- Coordinated on site and offshore resources.
- Provided research / analysis skills for root cause analysis for defects and for accommodating new requirements.
- Supported the team in deploying the code base in different test environments.
- Guided metadata team in loading Technical Metadata from Ab Initio TR.
Confidential
Sr. Software Developer
Responsibilities:
- Installations of the Databases, Informatica & OBIEE in servers of AWS. (For Century21, Pacsun & Ampersand)
- Creating the data models for the Retail business and getting it approved with Solution Architects.
- Creating the services for the installations for auto start and stop for server restarts.
- Requirements Gathering and estimate the effort required.
- Design the ETL flow to process data and transform into format that facilitates reporting.
- Develop Ab Initio graphs to read the source data and apply transformation logic as per the business rules and prepare data to load into the target tables.
- Build and schedule Unix Shell Scripts to invoke the Ab Initio Graphs.
- Involved in code review and performance tuning of ETL graphs.
- Involved in Data profiling that helps in redesigning the database.
- Coordinate and support Integration Testing & User Acceptance Testing.
- Analyzing and fixing defects during testing cycles. Analyzing Productions issues, TRs and fixes either the code or creating ad-hoc graphs to cater the requirement.
- Migration and deployment of code in development, IST, UAT and Production.
- Responsible for managing deliverables and meeting project schedules/deliverables.
- Responsible for providing effective system solutions for technical issues identified during post deployment phase.
- Worked on Informatica Power Center tools- Designer, Repository Manager, Workflow Manager, and Workflow Monitor.
- Parsed high-level design specification to simple ETL coding and mapping standards.
- Designed and customized data models for Data warehouse supporting data from multiple sources on real time.
- Involved in building the ETL architecture and Source to Target mapping to load data into Data warehouse.
- Created mapping documents to outline data flow from sources to targets.
- Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.
- Extracted the data from the flat files and other RDBMS databases into staging area and populated onto Data warehouse.
- Maintained stored definitions, transformation rules and targets definitions using Informatica repository Manager.
- Used various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings in the Informatica Designer.
- Developed mapping parameters and variables to support SQL override.
- Created mapplets to use them in different mappings.
- Developed mappings to load into staging tables and then to Dimensions and Facts.
- Used existing ETL standards to develop these mappings.
- Worked on different tasks in Workflows like sessions, events raise, event wait, decision, e-mail, command, worklets, Assignment, Timer and scheduling of the workflow.
- Created sessions, configured workflows to extract data from various sources, transformed data, and loading into data warehouse.
- Used Type 1 SCD and Type 2 SCD mappings to update slowly Changing Dimension Tables.
- Extensively used SQL* loader to load data from flat files to the database tables in Oracle.
- Modified existing mappings for enhancements of new business requirements.
- Used Debugger to test the mappings and fixed the bugs.
- Wrote UNIX shell Scripts & PMCMD commands for FTP of files from remote server and backup of repository and folder.
- Involved in Performance tuning at source, target, mappings, sessions, and system levels.
- Prepared migration document to move the mappings from development to testing and then to production repositories.
- Scheduled the Mapping’s in Data Administration Console for Hourly and Daily updates.
- Analyzing the Reporting Requirements and participated in the documentation of Business Requirements.
- Developing Oracle BI Metadata repository using BI Administration tool.
- Developing interactive Dashboards and Reports in Oracle BI Answers as per the reporting specifications.
- Defining and creating Initialization Blocks, Repository and Session Variables for generating filters and defining security.
- Creating Hierarchies, Levels, and implemented Business Logic by creating level based measures in OBIEE business model & mapping layer.
- Developing measure for time comparison using time functions.
- Using Filters and provided Prompts appropriately to make interactive reports and to avoid excessive data showing up on Reports.
- Customizing/Creating presentation layer catalog folders, to show different views of the business model to different users and assigned group permissions to the catalog folders.
- Performing the Performance Tuning of Dashboards and Reports and made required changes to repository and database in consultation with DBA.
- Implementing security based on LDAP Authentication.
- Configuring Measures in the Business Model Layer and mapped to multiple logical table sources.
- Creating schedulers and developed iBots to deliver reports and dashboards to the subscribers as per the schedules and alerts.
- Creating report prototypes and presented them to users to get approval.
- Attending weekly meetings with BA and Stakeholders for project status updates.
- Performing unit and integration testing of reports and dashboards.
- Creating Requirements Traceability Matrix and Mapping Document to trace requirements in BI implementation.
- Conducting Functional Requirements review and walk through with BI Architect, Designers, and Stakeholders.
- Analyzing the data and developed Logical Star Schema model with consultation from Data Modelers and Business Analysts.
- Developing metadata repository using OBIEE Administration tool in Physical, Business Model and Mapping, and Presentation Layer.
- Creating new logical columns, dimensional hierarchy, calculated measures, and aggregate mappings in the BMM layer as per the business requirements.