Big Data Infrastructure, Development, & Reporting Resume
Columbus, OH
SUMMARY:
- I am senior data engineer with 11+ years of experience in data analytics, ETL & leading/managing end to end technical delivery.
- Experience with Master Data Management, Data Governance, Data Quality, Data Profiling, and Meta Data management.
- Advanced experience with data modeling and data architecture.
- Theoretical understanding of Machine Learning algorithms for text mining, predictions, and sentiment analysis.
- Extensive experience working with Teradata sql scripting using analytical functions, complex joins, indices, volatile/global temporary tables, collect stats etc. while developing presentation layer for various cards subject areas at Confidential including inventory, payments, collections, recoveries etc. Have extensive performance tuning skills for Teradata sqls.
- Strong ability to understand and convert the logical data modeling into ETL data flow and draw a proper mapping to physical data model.
- Well versed with DWH concepts and data modelling ways i.e. relational, dimensional, and data vaults.
- Have good understanding of Hadoop architecture and framework.
- Hands on experience with HDFS, PIG Latin, HIVE, and SQOOP on MAPR cluster.
- Very strong UNIX shell scripting experience. Beginner level python exposure.
- 10+ years of relevant hands - on experience on Ab Initio ETL, Data Mapping, Transformation and Loading in complex and high-volume data processing applications. Experience with other ETL tools i.e. Informatica etc.
- Have experience working with PDL and DML Metaprogramming in Ab initio.
- Have good experience working with Continuous Flows using Ab initio queues and MQ series queues.
- Have data visualization experience using Tableau. Attended Tableau fundamentals & advanced s from Tableau. Have created few dashboards as POCs for few projects.
- Gained knowledge about business intelligence / presentation perspective of multiple projects while working closely with reporting teams.
- Ability to design the ETL considering factors like high data volume, partitioned data, join effectiveness, parallelism, generalization etc.
- Have solid experience of financial domain with good understanding of areas like cards specially Credit Cards, Prepaid Cards, Account Statements, Account Composite, Collection, and Recoveries, Bank, Personal Loans, Home Equity etc.
- Strong experience of agile project methodology and other rapid development models. Drives dev through TDD (Test driven development approach).
- A Team Player with excellent communication, analytical, verbal and writingskills.
- Strong articulation and presentation skill. Take bottom line responsibilities for assigned tasks. Proven track record of delivering results in Fast Track development environment within timeframe.
- Have effective communication across third parties, technical, and business product managers on solution design.
TECHNICAL SKILLS:
ETL Knowledge: Ab initio Co>Op 3.1.1, 3.0.4, 2.15/2.14, GDE 3.1.1, 3.0.4, 3.0, 1.15, 1.14, EME, PDL (Parameter Definition Language), DML Metaprogramming, generic graphs and psets, Continuous Flows, Conduct>IT, XML data processing, BRE, ACE, and Query>IT/Express>IT theoretical knowledge, Very good understanding of data lineage and dependency analysis, performance tuning of Ab initio batch and continuous graphs.Teradata Bteq, Fastload, Fast Export and Multi Load utilities through Ab initio. Trillium standardization, enrichment & matching via Ab initio. Working with load/unload utilities of Netezza.
Big Data Technologies: Hadoop & Map Reduce architecture, HDFS, PIG Latin, Hive, SQOOP, & OOZIE
Business Intelligence / Data Visualization: Tableau 9.0
Data modeling tools: ERWIN, ER diagrams
DBMS: Teradata, Netezza, Oracle8, Exadata, Oracle9i, DB2 iseries via Abinitio
Concepts: Data Warehousing and data modeling concepts including dimensional, relational and data vault concepts.
Database tools: Teradata sql Assistant, SQL*Plus, Toad, DB Visualizer 7.1.3, DB2 data studio.
Scripting/Languages: UNIX shell scripting, SQL/PL SQL scripting, Teradata Bteq, Core Java, Object Oriented programming language concepts.
Scheduling: Control-m, CA7, Autosys, Cron
Data analysis / mining: SQL, HQL
PROFESSIONAL EXPERIENCE:
Confidential
Big data Infrastructure, development, & reporting
Responsibilities:
- Led end to end technical delivery of various use cases in Amex’s brand new Big data/Cornerstone environment. Managed team of multiple engineers and vendor partners to deliver good quality data transformation, complex analytics, and reporting outcome. I extended my role from tech lead to manager and to solution/design architect.
- Architected and designed SPOT-ML metadata driven, reusable platform which exposed various classification ML models (including sentiment analysis models), and one organic unsupervised clustering ML model. Platform integrated model outputs with various big data use case outputs and provided topic detection, sentiments analysis, and identified data clusters. Use case output and ML output together then were used in reporting/dash boarding.
- In addition worked extensively as data analyst and mined complaints, collections, voice to text, global corporate payments, digital marketing and customer data to create one views for reporting. Created various Tableau dashboards as part of these exercise as well.
Technologies involved are - Abinitio, Hadoop, Hive, Pig, Java, Python, Machine Learning algorithms and libraries, Unix shell scripting, Kognitio, Jethro, Data Meer & Tableau.
Confidential
Sr. Engineer
Responsibilities:
- End to end ETL application architecture, design, development, testing, documentation, support and delivery.
- Document architectural high level process flows, design documents.
- Review codes and help team members resolves Abinitio related queries.
- Design and Dev of ETL processes using Abinitio GDE 3.1.4
- Worked on performance tuning and optimization of existing graphs using best practices for effective usage of dynamic lookups, phasing, check pointing, avoiding sorts, using db utilities etc.
- Worked with Tableau 9.0 to create portfolio dashboard with links to navigate to different worksheets on being clicked/actioned.
Confidential
Sr. Ab initio consultant
Responsibilities:
- End to end ETL application architecture, design, development, testing, documentation, support and delivery.
- Document architectural high level process flows, design documents.
- Review codes and help team members resolves Abinitio related queries.
- Design and Dev of ETL processes using Abinitio GDE 3.1.4
- Worked on performance tuning and optimization of existing graphs using best practices for effective usage of dynamic lookups, phasing, check pointing, avoiding sorts, using db utilities etc.
- Addresses defects in existing Abinitio graphs.
- Worked on migration using Abinitio air tag, save and load commands.
- Written complex db2 sql scripts using aggregation functions, multiple joins etc.
Confidential
Sr. Ab initio consultant
Responsibilities:
- End to end ETL application architecture, design, development, testing, documentation, support and delivery.
- Created Continuous application using JMS, MQ queues, Abinitio queues, XML data parsing etc.
- Effectively used and configured compute points and checkpoints.
- Worked on migration using air load, save and tag commands.
- Developed shell scripts to provide additional functionality and wrap Abinitio graphs.
- Wrote complex SQL and PL/SQL scripts to successfully source and update data.
Confidential
Sr Abinitio consultant
Responsibilities:
- End to end ETL application design, development, testing, documentation and delivery.
- Developed Abinitio graphs using GDE 3.1.x
- Worked to resolve DA and DL issues.
- Worked with Control-m to schedule jobs and define dependencies.
- Worked with xml parsing components i.e. xml split and combine.
Confidential, Columbus, OH
Lead
Responsibilities:
- As Abinitio Consultant I am responsible to create POCs while doing performance tuning of existing Ab initio graphs and presenting the same to client as improvement proposals. Idea is to reduce number of existing processes, reduce elapsed time and CPU utilization and improve environmental efficiency.
Confidential, Omaha NE
Responsibilities:
- Lead ETL (Abinitio/Netezza/Unix/Data Vault) Designer, developer, onsite coordinator.
- Technologies used - Abinitio V3-0, UNIX shell scripting, Netezza.
- Participated in all stages of project grouped into various stories, project being agile.
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping.
Confidential, Richmond VA
Responsibilities:
- Development of efficient, easily maintainable Abinitio graphs, shell scripts, & Teradata sql scripts.
- Participated in all stages of project grouped into various stories, project being agile.
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping.
- Developed and reviewed test plans, test scenarios, test cases as part of Unit testing.
- Responsible for setting up Production Jobs and created Production Support Document.
Confidential, Durham NC
ETL Designer and Developer.
Responsibilities:
- Development of high performing tuned Abinitio graphs.
- Responsible for Technical Design, Development, Unit Testing, Code Review, Code migration, Supporting System testing, Job scheduling.
- Developed and reviewed test plans, test scenarios, test cases as part of Unit testing.
- Responsible for setting up Production Jobs and created Production Support Document.
Confidential, Richmond VA
ETL Designer
Responsibilities:
- Development of high performing, easily maintainable Abinitio graphs and shell scripts.
- Participated in all stages of project grouped into various stories, project being agile.
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping.
- Develop & review several Korn scripts to integrate workflows on Control-M.
- Developed and reviewed test plans, test scenarios, test cases as part of Unit testing.
- Responsible for setting up Production Jobs and created Production Support Document.
Confidential, Richmond VA
ETL Designer and developer
Responsibilities:
- Development of high performing, easily maintainable Abinitio graphs and shell scripts.
- Participated in all stages of project grouped into various stories, project being agile.
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping.
- Develop several Korn scripts to integrate workflows on Control-M.
- Performance Tuning on the Targets, Sources, Mappings and Sessions used Techniques like reusable Transformations, reusable Caches and partitioning etc.
- Developed test plans, test scenarios, test cases as part of Unit testing.
- Responsible for setting up Production Jobs and created Production Support Document.
Confidential, Richmond VA
ETL Designer and developer
Responsibilities:
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping.
- Created reusable transformations, and implemented various slowly changing dimensional mappings.
- Designed error checking mechanisms to reduce failures and monitor error logs.
- Develop several Korn scripts to integrate workflows on Control-M.
- Performance Tuning on the Targets, Sources, Mappings and Sessions used Techniques like reusable Transformations, reusable Caches and partitioning etc.
Confidential, Richmond VA
ETL Designer and developer
Responsibilities:
- Development of high performing, easily maintainable Abinitio graphs and shell scripts.
- Involves staging data file from source (vendor), cleaning data, publishing on Confidential data servers and later performing various operations on it before finally loading into data warehouse.
- Participated in all stages of project grouped into various stories, project being agile.
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping.
- Created reusable transformations, and implement various slowly changing dimensional mappings.
- Designed error checking mechanisms to reduce failures and monitor error logs.
- Develop several Korn scripts to integrate workflows on Control-M.
- Performance Tuning on the Targets, Sources, Mappings and Sessions used Techniques like reusable Transformations, reusable Caches and partitioning etc.
- Developed test plans, test scenarios, test cases as part of Unit testing.
- Responsible for setting up Production Jobs and created Production Support Document.
Confidential, Richmond VA
ETL designed and developed using Teradata SQL scripts
Responsibilities:
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping, setting up production jobs and dependencies using control-m.
- Performance tuning by the use of collect statistics, creation of volatile and global temporary tables, defining new primary indexes with keeping join keys in mind, use of driver tables to restrict the data and make the aggregations run faster.
Confidential, Richmond VA
ETL design and development
Responsibilities:
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping, setting up production jobs and dependencies using control-m.
- Staging, cleansing, transforming and loading are main objectives.
- Various error handling and exception handling codes are implemented in Abinitio graphs and UNIX shell scripting.
- Psets and Generic graphs are used to handle processing of five different files through one Abinitio graph.
- Design got highly appreciated from client and set standards for every further projects.
Confidential
Responsibilities:
- Use of driver table to restrict the data in joins. Performance tuning is done by making sure that join btw two tables happens always on keys, this is made sure by creation of either volatile and/or global temporary tables. Collect statistics is also done to ensure faster fetch from queries on tables.
- Various aggregation functions are used to present the composite data into PL table which further made tasks of data analysts’ easier.
- Use of constructs like GROUP BY, CASE WHEN, CAST, STRING functions, SUM, MAX, MIN, AVG, COUNT, LIKE etc are common.
- Mocking up data and creating data scenarios were important parts.
Confidential
Responsibilities:
- Responsible for Requirements gathering, Technical System Design, Development, Testing, Code Review, Code migration, Supporting System testing, Job scheduling, Offshore Coordination, prepared source to target mapping, setting up production jobs and dependencies using control-m.
Confidential
Responsibilities:
- Enhancement or break-fix (defects) from various data warehouse applications across Confidential were assigned to us. Break-fixes could vary from functional defect, control-m dependency change, small enhancement, performance improvement etc.Requests always involved lot of analysis as was always in existing functionality. Impact analysis was a major task.
