We provide IT Staff Augmentation Services!

Etl Talend Architect Resume

3.00/5 (Submit Your Rating)

Richmond, VA

SUMMARY

  • Over 18 years of IT experience in development and Implementation of various projects using Oracle, DB2, SQL Server, Teradata, NoSQL DBs, AWS, SQL, PL/SQL, Talend, Informatica Power Centre, Shell Scripting, Sun Solaris and Windows.
  • Strong working experience in Data Modeling, dimensional modeling, cube/dimension design, database procedures, design, development and OLAP analytical reports.
  • Expertise in Server Installations, Metadata management, Informatica and Data warehouse environments.
  • Extensive experience working with AWS deployment pipelines / EC2 Instances / S3
  • Designed and Defined data retention policies on cloud data platforms
  • Delivered optimum benefits and has full life cycle implementation exposure: Business Requirements Definition, Modeling, Logical Database Design, Mapping, Gap Analysis, Data Conversion, Validation, Testing, Documentation, Production Migration, Post Implementation Support, Assessments, Enhancement and Refinements.
  • Thorough knowledge of Database architectures, designs, modeling, setups, tuning, migrations, integrations, data merger and consolidations.
  • Has deep understanding and knowledge of Database Administration activities like Installation and Creation of database, database objects, backup & recovery.
  • Experience in Optimizing the Performance of SQL scripts and Oracle database/application tuning.
  • Hands on experience in all aspects of Software Development Life Cycle (S.D.L.C).
  • Excellent interpersonal skills, motivated, energetic team player with demonstrated proficiency for learning new tools and business environment with excellent communication and analytical skills.
  • Able to manage changing responsibilities and deliver time critical projects on schedule.
  • Proven ability to accept challenges and produce results in a multi - tasked environment.

PROFESSIONAL EXPERIENCE

ETL Talend Architect

Confidential, Richmond, VA

Environment: Oracle 19c, Talend, Teradata, Talend on AWS, MongoDB, Erwin, MS-Visio, TOAD, Data pumps, MS-Project, UNIX Shell Script, AIX-Unix and Windows

Responsibilities:

  • Extensively worked on Talend TAC, Nexus, SVN repository setups taking into consideration the load triggers, user management, Port Handling
  • Worked withConfidential to set up clustered environment in lieu of limited Hardware resources, such that all the non-prod/test environments are handled on one VM but still have the delineation of multiple environments.
  • Setup Local and Remote Talend projects such that the development team can develop projects locally on their respective machines and then move the code to Remote projects as they get over the initial skeleton development
  • Command level execution of Talend jobs such that the clients, testing partners does not need to have Talend studio installations or need to be knowledge of Talend
  • Extensively worked with NOSQL components of Talend to insert data to Mongo Arrays overcoming the tool limitations
  • Designed and developed Dynamic data type jobs to be able to fetch 120+ delimited file data to go into multiple tables, all within one job
  • Designed and developed Oracle procedure to enable sync between two different schemas on incremental basis. Not needing to explicitly mention Oracle table and column names was able to sync on incremental basis and not compromising on the performance
  • Designed and have set up Remote Projects such that every developer works on a shared repository like Informatica such that the Client machine failures wouldn’t impact the ETL jobs, and the code is shared across the teams
  • Administered Talend Installations and Patches
  • Talend migrations between Projects, repositories and from on-premises to Cloud (AWS)
  • Extensively worked with XML sources, considering all the looping and validation constraints of XML and be able to load the underlying Oracle tables
  • Designed and developed ELT components to be able to sync massive volumes of data between schemas in Oracle
  • Designed scheduling mechanism using Talend Triggers and EFT components for secured file transfers and job triggering to work hand-in-hand
  • Designed Publish mechanism to TAC such that every developer once unit tested will publish his code to TAC - Snapshots and Repositories
  • Extensively worked with Nexus and complications around to enable HA of Nexus for all Artifact repository needs

ETL Informatica Architect

Confidential, Richmond, VA

Environment: Oracle 12c, Informatica, AWS(S3), MongoDB, Erwin, MS-Visio, TOAD, Data pumps, MS-Project, UNIX Shell Script, AIX-Unix and Windows

Responsibilities:

  • Designed and developed the ETL architecture to denormalize the Source Feed data to validate every column of the row of data to perform the predefined validity checks and create Exception logs for the error records substituting the value with predefined Default values
  • Worked with Admin teams. to set up clustered environment in lieu of limited Hardware resources, such that all the non-prod/test environments are handled on one VM but still have the delineation of multiple environments.
  • Pushed couple of complex and process intensive integration maps to cloud leveraging the Informatica Cloud platform / Informatica Intelligent Cloud Services (IICS)
  • Designed and defined strategies to choose the execution engine for our Hadoop targets using Data Engineering Integration (DEI)
  • Developed dynamic mappings to switch between varying data formats Text & Avro using Data Engineering Integration (DEI)
  • Configured DEI to choose Spark as the execution engine for all the underlying Big Data jobs
  • Built couple of dynamic jobs using DEI to process switching data sources between files and Databases (files are used for data layovers and ad-hoc reconciliation loads)
  • Leveraged DEI capabilities with codeless advanced integration transformations to extract data using templates from mapping designer
  • Developed a data driven approach to switch between on-premises and cloud-based source to extract data and to load to cloud targets using DEI advanced capabilities
  • Configured the DB connections to MongoDB and the connector for AWS S3 using access key ID and secret keys, updated the security.jar file and updated the client side and Server-Side components enabling the reading and writing of data to/from AWS S3
  • Created Target Mongo DB definitions considering all the collections of the Mongo DB irrespective of the presence of data values to those data elements
  • Design and Develop of ETL architecture with failure recovery methodology and dynamic switching of DB connections pointing to the Contingent Databases
  • Design and Develop of Archival methodology for faster Data Access, Data Loads, and less DB maintenance
  • Oracle Database design for consolidating the existing redundant databases
  • Developed scripts to automate Oracle data dumps using Data pumps
  • Designed the future ETL job scheduling for the future loads which enables the administrators to add job to the existing schedule not intervening with dependency between the loads
  • Designed and proposed the Incremental strategy for the data loads to the DataMart sourcing from the Stage Output which is basically an accumulation of all the Intra Days Data
  • Designed and proposed an automated Data Quality check as opposed to the existing manual data quality check.
  • Interview Business users and Subject Matter Experts (SME) to collect the business requirements and to gather the ‘Better to have’ abilities from the current system.

Sr. Data warehouse Developer

Confidential, Boston, MA

Environment: ORACLE 10g, SQL server 2000, Informatica- Power Center 9i, SQL, PL/SQL, Erwin, DB2, MS-Visio, TOAD, SQL*Loader, UNIX Shell Script, AIX-Unix and Windows

Responsibilities:

  • Active participation in the Design and development of the Schema for the ODS and star schema for the Data warehouse.
  • Set up of centralized Informatica repository to handle ETL operations across the company
  • Capturing of changed data (CDC) from source systems with no update dates on them
  • Synchronization of two ODS spread at NJ and NE on nightly basis
  • Study and analysis of the business requirements, understanding client's data warehousing requirements and system analysis of business process and requirements of Data marts.
  • ETL architecture to cater to the extraction requirements from the Source systems to Staging, from Staging to ODS and ODS to the Centralized Data Warehouse.
  • Audit and Error handling and fuller utilization of Versioning ability of Informatica
  • Design and develop the Control Tables to achieve the incremental data loads
  • Configuration of Database connections from Informatica repository to diversified Databases using native connections and ODBC connections
  • Building of Informatica mappings to load the Staging Area sourcing the Subscriber tables of SQL server 2000.
  • Used the Materialized View log ability to capture the changed data (CDC) from Oracle sources
  • Design and Develop Informatica Workflow to respect the dependencies between the sessions and for proper scheduling of Data Loads.
  • Managed the Informatica Repository as an Informatica Admin.

Sr. Data warehouse Developer

Confidential, Charlotte, NC

Environment: ORACLE 10g, TrueComp 3.7, Informatica- Power Center 9i, Epiphany, SQL, PL/SQL, Erwin, DB2, MS-Visio, TOAD, SQL*Loader, UNIX Shell Script, HP-Unix and Windows

Responsibilities:

  • Design and develop the Schema for the Data warehouse (Star schema) using the native keys and surrogate keys using ERWIN.
  • Study and analysis of the business requirements, understanding client's data warehousing requirements and system analysis of business process and requirements of Data marts.
  • Preparing of High-Level Design Documents (HLD) and Low-Level Documents (LLD)
  • Building of Informatica mappings to load the Reference Data and for building the hierarchy for the reference data for Account Type Hierarchy, Product Hierarchy and SA Agreements.
  • Building of Informatica mappings to load the Transactional data into TrueComp Schema from the legacy systems.
  • Building of Informatica mappings to load the TrueComp Data Mart from the TrueComp Repository.
  • Creating of Universes to source Data from both Data Marts and Data Warehouse for the reporting requirements.
  • Design and Develop Informatica Workflow to respect the dependencies between the sessions and for proper scheduling of Data Loads.
  • Also worked on developing technical specifications for ETL process, mappings and to load into Data warehouse.
  • Managed the Informatica Repository as an Informatica Admin.
  • Develop mappings in Informatica using transformations like Look ups, aggregator, ranking, Expressions, functions, Filters, SQL overrides in lookups and data flow management into multiple targets using routers and reusable components like Maplets and Re-usable transformations.

Sr. Data Warehousing / Informatica Consultant

Confidential, CA

Environment: ORACLE 9i, Informatica Power Center, Business Objects 6, Dollar Universe, Business Objects Designer, SQL, PL/SQL, ERwin, DB2, MS-Visio, TOAD, SQL*Loader, Autosys 3.4.2.35, UNIX Shell Script, HP-Unix and Windows 2000/NT

Responsibilities:

  • Created Informatica mappings with complex transformation logic to load the Cisco data warehouse (CDW) tables on an incremental basis.
  • Developed and implemented UPROCS using Dollar Universe for active Scheduling of the Load Process executing the Informatica Sessions and PL/SQL packages on a daily basis, where failures and restartability were majorly emphasized.
  • Scheduling of dimensional data loads through Autosys.
  • Extracted, scrubbed and transformed data from Fixed-width Flat Files, Delimited Flat Files, COBOL Files (Using Power Connect for mainframes) and loaded staging tables using Informatica.
  • Optimizing of Informatica mappings to load huge volumes of data.
  • Developed and implemented the multi-threading concept of Informatica sessions for faster completion of data loads.
  • Developed logic for Auto Analyzing of all the major fact tables on periodic intervals.
  • Creating and optimizing PLSQL Stored Package/Procedure.
  • Developed logic for dynamic aggregation of data from views using CASE logic.
  • Business Objects Universe, reporting, BCA, Supervisor.
  • Attending to issues arising from data inconsistency at source systems.
  • Closely worked with down stream systems for enhancements to current logic.
  • Supporting daily loads and debugging errors, back outs and partial load fixes.
  • Developed and implemented logic to synchronize the Data Warehouse data with the source systems to capture all the deletes happening at the source systems after data is loaded to the data warehouse.

Team Lead

Confidential, Arlington, VA

Environment: ORACLE 9i, Informatica Power Center, Business Objects 6, Web Intelligence Server, Business Objects Designer, SQL, PL/SQL, Erwin, MS-Visio, TOAD, SQL*Loader, UNIX Shell Script, HP-Unix and Windows

Responsibilities:

  • Developed System Design document, technical documents, manual metrics algorithm document and production support document.
  • Developed Informatica mappings and mapplets to load data into the Data Warehouse.
  • Developed expression, router, filter and lookup transformations.
  • Developing and configuring Informatica Sessions/batches. Created complex packages and stored procedures in Oracle to apply business rules to produce desired results from the FACT/Lookup tables in Data Mart Area.
  • Load data using Informatica mappings to Data Warehouse (From Transient to Staging to Data Warehouse to Data Mart), reporting and resolving any discrepancies with the data provider.
  • Debugging Informatica mappings.
  • Generate ASCII files for the data load team using ORACLE UTL FILE package.
  • Generated test scripts for testing the output and making available data for testing.
  • Telnetting, FTPing and taking backup of the generated results on a regular basis to/from Unix Server.
  • Developed Packages/Stored Procedures and Informatica Mappings as per new CCRs (Change Control Requests) and Maintenance.
  • Autosys for Scheduling the Load Process.
  • Created expression, lookup and router transformations to check customer information and add customers those are not present already.

We'd love your feedback!