We provide IT Staff Augmentation Services!

Data Engineer/sr Etl Consultant Resume

4.00/5 (Submit Your Rating)

San Ramon, CA

SUMMARY

  • An experienced Big Data Engineer, ETL architect and DWH Consultant with 12+ years of experience in various Data Warehousing engagements across Technologies and various domains
  • Robust work expertise in architecting end - to-end solutions with flair in designing DWBI engagements
  • Utilized the Big Data technology, Apache HIVE & SQOOP for mining valuable information out from Datawarehouse/Datamarts
  • Good exposure to NoSQL databases, HBase and Cassandra
  • Using MySQL Database with MYSQL Workbench for the last 5+ years
  • Experienced in architecting highly scalable, distributed systems using different tools as well as designing, optimizing processes in the Hadoop and Informatica ecosystems
  • Excellent understanding of database concepts and writing PL SQL, SQL query tuning and performance tuning of queries
  • Experienced with most of ETL tools like Informatica Power center, DPI & BODS
  • Sound knowledge in Teradata Transactional loading utilities BTEQ & TPUMP and bulk utilities like FastLoad & MultiLoad
  • Exposure to AWS - EC2, S3
  • Most of data exercise has been pulling data from XML and JSON events to Oracle and SQL Server
  • Experience in Parsing XML and JSON events, APIs to flat DB tables in Oracle using Python libraries/modules
  • Explored most of UNIX commands and python libraries/modules which helps to automate DB tables archival
  • Proficiency in SDLC and Agile processes and act as scrum lead to deliver user stories
  • Created Dashboards on Financial portfolios using Tableau 8.x & 9.x
  • Good understanding of Data Modeling and Erwin tool
  • Knowledge on FDRs (Full Dressed Rehearsal) for go-live DW projects
  • Proficient in Test Conditions, Test Cases, Test Procedures & Test Plan preparation

TECHNICAL SKILLS

  • Informatica Power Center 8.6
  • DPA (Decision Point Administrator)
  • Oracle 10g
  • PL/SQL
  • Oracle ERP
  • OM
  • PO
  • SCM
  • Item Master
  • OBIEE 10g
  • Hyperion Brio
  • Toad
  • Unix Shell Script
  • Win XP

PROFESSIONAL EXPERIENCE

Confidential, San Ramon, CA

Data Engineer/Sr ETL Consultant

Responsibilities:

  • Collaborating with business team to understand the requirements
  • Creating technical design documents to narrate the requirements with ETL and Data Engineering solutions
  • Designed and developed remarkable banking business systems like Decision Tree, FDIC, Credit Cards Takedowns using ETL tool Informatica, IDQ, SQL & PL/SQL
  • Cleansed raw he data using Informatica Data Quality (IDQ) to remove duplicates, streamline into single record out of multiple matches
  • Developing end to end complex data transformations from flat files to Data Marts using ETL tool Informatica with SCD1 & SCD2 table designs
  • Creating data pipelines for unstructured data using Luigi, Python package
  • Loading huge files into Teradata DB by using FastLoad utilities for the first time and for further loadings, MultiLoad utility is being used
  • Done POC on web crawling and pulled data and stored in HBase (NoSQL) database for high availability and fast access for ingesting into HDFS
  • Extracting valuable banking requirements of Customers like, loans, mortgages, FD information out from Datawarehouse and DataMarts table by importing them to HDFS using SQOOP and did analysis using HIVE and Python
  • Creating dashboards using Tableau 9.x and Python Tinkter packages over the fact tables for QTD, YTD and ITD metrics for deposits, loans and mortgages
  • Creating detailed Unit Test documents with all scenarios before moving the code to Test environment
  • Creating detailed implementation documents to share with production support team with the schedules, dependencies, recoverable mechanisms etc.

Environment: Apache HIVE, Python 3, Tableau 9.x, Informatica 9.6, IDQ 9.6, MySQL 5.7, MySQL Workbench 6.3, 11.5, SVN, Unix Shell Script, Tidal, Windows 7

Confidential, San Francisco, CA

Data Engineer/Sr ETL Consultant

Responsibilities:

  • Designed ETL framework to move the data from Flat File to Datasets
  • Created data pipelines for unstructured data using Luigi, Python package for Financial data sources
  • Created financial reports using Python Tinkter packages
  • The framework includes design and development of ETL mappings from Flat File to ODS, ODS to Dimensions, Dimensions to Fact table / .csv target flat file
  • Created dashboards using Tableau 8.x for opportunity portfolio that gives more insight to the management to get the prospectus customers
  • An Unix script written to post the final target file to dataset which is formatted through JSON files
  • The design done by creating Data Model for new tables joining with existing dimensional and fact tables
  • Did a thorough analysis on data and its granularity levels before pitching into development
  • Written shell scripts using AWK & SED commands to feed the Flat files into readable semicolon delimited formats and remove unwanted commas in between the fields
  • Performed QA’s within our peers
  • Deployed Shell Scripts, DMLs, DDLs and XMLs for Go-Live on SIT environment
  • Integrated the Informatica jobs with Cisco Tidal Enterprise Scheduler tool, version 6.0.3.197

Environment: Apache HIVE, Python 2, Salesforce Wave reportings, Informatica 9.6, MySQL 5.7, MySQL Workbench 6.3, Perforce, Unix Shell Script, Tidal, Windows 7

Confidential, South San Francisco, CA

ETL Informatica Tech Lead

Responsibilities:

  • Involved in all sprint meetings where the project related specs and docs are shared and discussed with the team.
  • Interacted closely with the BSAs to understand the step by step business requirements to create the tech specs and STMs
  • Imported large datasets from HDFS to Informatica using Sqoop & Power Exchange
  • Done analytics on huge datasets using Hive QL
  • Supported Map Reduce jobs which was used to analyze unstructured data like Aprimo data which mostly contains HCPs feedback and profiles etc.
  • Created complex mappings where the flow involves Landing, Delta, History, S1, S2 & DM schemas.
  • Designed complex ETL logics by dividing the flow into simple steps with feasible ETL approaches that brought down the job timings well before the threshold timings.
  • Extensively used PL/SQL Stored Procedure, Normalizer, Joiners and other regular transformations like, Expression, Lookups, Aggregator, Filters, Routers, etc.,
  • Used dynamic session parameters widely
  • Assisted SAP BO team in modeling the reports datamart in efficient ways
  • Worked on Sales Force UI & APEX since SFDC is one of the source systems for AGGS Spend. Hands-on with sales force relationships and join queries
  • Conducting regular technical meetings within the team where I would share the technical suggestions and clarify their functional queries.

Environment: Hadoop, Hive, Pig, Sqoop, Informatica Power Center 9.1, MySQL 5.7, SAP Business Objects XI3.1, Sales Force, Aprimo, Redwood (Scheduler), MySQL Workbench, SQL*Loader, Unix Shell Script, Win 7

Confidential, Cupertino, CA

ETL Informatica Lead/ Data Analyst

Responsibilities:

  • After studying the complete application, identified the bottlenecks in the system and suggested predominant performance improvements on PL/SQL part like using BULK Collects, bind variables, NOCOPY parameter, SIMPLE INTEGER in place of NUMBER data type etc.
  • Processed the monthly and yearly jobs in parallel processing using Greenplum
  • Requested AWR reports on daily basis and compared the report whenever any new code or major changes went live on system
  • Used TKPROF utility to understand the AWR report in a better way and suggested to use bind variables to make less I/O, implemented using appropriate index based on the cardinality of the data
  • Independently owned the validation part in of iCas whenever there are changes in the upstream systems like POS, Zingers. Recently owned POS R12.3 project
  • Handled Brickhouse 1.5 project upgrade
  • SPOC for Sawdust projects - which is the validation part in iCas when the Confidential store got inaugurated in Sweden recently
  • Owner for Database refresh in test environments. Whenever the data is refreshed from prod to test, carried out the scrubbing part of prod customer information and encoding and decoding all fee columns.
  • Written an automated shell script calling an stored procedure for post refresh activities like deletion of some prod only data in test systems and updation of email notifications addresses.
  • Liaised closely with business people while month end closings and provided the solutions immediately to solve the issue on the fly
  • Closely monitored the systems during the high volume season like iPhone 5 release. Mitigated the out of memory issue during that time by providing a feasible method.
  • Conducted sessions with our offshore team on daily basis on particular topics and shared the knowledge.

Environment: Oracle 11g, PL/SQL, Toad, SQL*Loader, Unix Shell Script, Mac OS 10.7.4.CLIENT: Cisco, San Jose, CA Mar 2012 - Jun 2012

Confidential

ETL Informatica/ BIDW Lead

Responsibilities:

  • Redesigned the source system and target loading architecture by creating materialized views in place of ordinary tables that saved 70% of job timings
  • Modified the SQL queries used in Informatica to pick up the smallest table as driving table to convert the joins as HASH join which results in fastest execution of the queries
  • Involved in the requirement phase with IT Team
  • Created the BRD & FSDs on their requirement. Once it got reviewed by IT team, started to prepare the ETL design of the requirement
  • Shared the BRDs & FSDs with offshore team and explained the requirements and assigned tasks accordingly on ETL & reporting
  • Conducted regular check point meetings with IT team
  • Developed most complex ETLs for GFV (Global Fund Volume) Bookings
  • Prepared unit test cases, regression test cases on Unit and SIT before moving to UAT
  • During UAT fixed the defects raised by UAT team then and there
  • Once UAT is signed off, prepared KT document for ITDS team and given a KT to them and got KT sign-off as well.
  • Co-ordinated with ITDS (Support) team during migration code from UAT environment to production
  • Made two successful Go-lives (Backlog & GFV Bookings)

Environment: Informatica Power Center 9.1, Oracle 10g, PL/SQL, SAP Business Objects XI3.1, Toad, SQL*Loader, UNIX Shell Script, $Universe ($U) Win XP

Confidential, Milpitas, CA

BIDW / ETL Informatica Lead

Responsibilities:

  • Understood the complete architecture of EDW
  • Closely worked with ERP source systems team in optimizing the creation of records for DW
  • Involved in DB performance tuning during month end and quarter end closes when high volume of data flown into system by reducing the Physical reads and increased logical reads by tuning the top 10 queries listed in AWR.
  • Supported end users by resolving the service tickets with high commitments and responsibilities
  • Handled a 6 member team at offshore, India by assigning tickets after gathering enough inputs from users.
  • Resolved the P1 & P2 tickets 24/7 then and there without any time on show stopper issues. Never breached P1 & P2s
  • Worked on some vital Change Requests from end to end to change for the existing logic of several mappings and reports on Informatica and OBIEE respectively as requested by corporate users.
  • Provided solutions to make the provisioning of access request hassle free, which is now under consideration by JDSU.

Environment: Informatica Power Center 8.6, DPA (Decision Point Administrator), Oracle 10g, PL/SQL, Oracle ERP (OM, PO, SCM, Item Master), OBIEE 10g, Hyperion Brio, Toad, Unix Shell Script, Win XP

Confidential

Senior Informatica Developer

Responsibilities:

  • Analyzed the existing architecture of CDW
  • Used Materialized views extensively to improve the data movement between source and target systems.
  • With the help of Teradata utilities like TPump, Fast Load and with BTEQ queries migrated the Siebel source system to TD data base.
  • Cleansed the legacy data using Informatica transformations such as sorters, expression transformations etc., usually cleansed customer data like addresses, telephone, zip code etc
  • Assured the quality of data using Informatica transformations e.g. the commas between numeric fields, dashes between phone numbers, and product names such as MOL into Max OnLine.
  • Developed several new mappings to pull data from legacy to new architecture
  • Owner for bridge tables where the bridge tables serve as the reference for the legacy keys to current Siebel keys
  • Involved in FDR3 cycle executions and fixed the runtime issues then and there
  • Done performance tuning for long running jobs during FDR3 cycle and executed the jobs separately through an adhoc workflows.

Environment: Informatica Power Center 8.6, Siebel, Oracle 10g, PL/SQL, Toad, UNIX Shell Script, Win XP

Confidential, Redwood Shores, CA

ETL Informatica Developer

Responsibilities:

  • As a Senior Systems Analyst for ETLAnalyzed the existing architecture
  • Documented the analysis with complete detailed architecture and provided separate architectures for the proposed MDM
  • Created a new database called R12 DB using data modeling tool Erwin for both logical and physical models
  • Analyzed all ERP source systems existing in Oracle 11i and provided feasible solutions to move into Oracle R12.
  • Designed high level ETL design in all feasible without affecting the performance.

Environment: Informatica Power Center 8.6, Oracle 9i, PL/SQL, Toad, Unix Shell Script, Win XP, HFM, Digital River, Demantra

Confidential 

Informatica Developer/ Database Analyst

Responsibilities:

  • Developed complex logic by using Oracle Procedures and Packages
  • Cleansed the source data and assured the quality of the data at the staging level
  • Developed complex logic by using transformations in Informatica
  • Providing Technical Solutions to the Team
  • Development work for new enhancements - Oracle PL/SQL, Informatica
  • Testing - unit testing & integration testing
  • Application Support
  • Analyzing data-issues
  • Coordinating with source systems to resolve data issues
  • Performance tuning and providing technical solutions to critical issues
  • Involved in production support activities for SLA project both in ODS and ETL
  • Debugging and fixing the issues in UNIX shell scripts

Environment: Informatica Power Center 7.1/8.1, Oracle 9i, PL/SQL, Toad, UNIX Shell Script, Win XP

We'd love your feedback!