We provide IT Staff Augmentation Services!

Etl Talend Developer Resume

Plymouth Meeting, PA

SUMMARY:

  • ETL Talend Developer 8+ years of experience as a Technical analyst, safety Developer and administrator.
  • Extensive experience in IBM Info Sphere Data Stage - ETL integration with SAP bank analyzer functional module performing data extract, loading and transformation process for financial general ledger accounts data.
  • Experience in designing and implementing Data Warehouse applications mainly using ETL tool Talend Data Fabric for Big data integration and data ingestion.
  • Worked with different components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)
  • Worked with different tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Well versed with XML technology and DOM and SAX parsers
  • Experience in UNIX file/dataset management to keep the load ready data for all financial transactional data.
  • Experience on waterfall and Agile methodology project implementation.
  • Involved in complete Software Delivery Life Cycle (SDLC) in large data warehouse environment for financial data system.
  • Thorough knowledge in data warehousing, dimensional modeling, data integration, data virtualization, data synchronization, star schema, waterfall schema, ETL development & performance tuning, BI data analysis, SAP integrating, DFS & HDFS cluster segregation.
  • Worked with senior Bio-statisticians and clinical data managers to provide ETL programs in analyzing the data, generating safety and efficacy loads, summary tables.
  • Excellent oral and written communication skills. A quick learner with an eye for detail and excellent analytical and quantitative skills. Ability to work independently and adept at managing multiple competing tasks.

PROFESSIONAL EXPERIENCE:

ETL Talend Developer

Confidential - Plymouth Meeting, PA

Responsibilities:

  • Architect, design, construct, test, tune, and deploy ETL infrastructure based on the Hadoop ecosystem based technologies.
  • Work closely with administrators, architects, and application teams to insure applications are performing well and within agreed upon SLAs.
  • Work closely with Management and Data Scientist teams to achieve company business objectives.
  • Collaborate with other technology teams and architects to define and develop solutions.
  • Deploy ETL code that aligns with the ETL target state architecture standards and development standards.
  • Research and experiment with emerging ETL technologies and tools related to Big Data.
  • Contribute to the Big Data open source ecosystem.
  • Work with the team to establish and reinforce disciplined software development, processes, standards, and error recovery procedures are deployed; ensuring a high degree of data quality.
  • Maintain, tune, and support the ETL platform on a day-to-day basis to insure high availability.

ETL Talend Developer

Confidential - Pittsburgh, PA

Responsibilities:

  • Developed complex ETL jobs from various sources such as SQL server, Postgressql and other files and loaded into target databases using Talend ODS ETL tool.
  • Created Talend jobs using the dynamic schema feature.
  • Created the Talend jobs for Store Support Center, Coupons jobs and Promotions etc.
  • Interact with business community and gathered requirements based on changing needs. Incorporated identified factors into Talend jobs to build the Data Mart.
  • Performance tuning - Using the tmap cache properties, Multi-threading and Parallelize components for better performance in case of huge source data. Tuning the SQL source queries to restrict unwanted data in ETL process.
  • Involved in Preparing Detailed design and technical documents from the functional specifications.
  • Prepared low level design documentation for implementing new data elements to EDW.
  • Provided Knowledge Transfer to the end users and created extensive documentation on the design, development, implementation, daily loads and process flow of the mappings.
  • Have used AWS components (Amazon Web Services) - Downloading and uploading data files (with ETL) to AWS system using S3 talend components.
  • Used more components in Talend and Few to be mentioned: tMap, tFilterRow, tjava, toracle, txmlMap, tdelimited files, tlogrow, tlogback, components etc. in many of my Jobs Design
  • Worked on Joblets (reusable code) & Java routines in Talend.
  • Design, Develop and Test ETL processes in order to meet project requirements
  • Created Projects in TAC and Assign appropriate roles to Developers and integrated SVN (Subversion)
  • Used to be On call Support if the Project is deployed to further Phases
  • Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis (Cron Trigger)

Environment: Talend Open studio V (5.6), UNIX, AWS-S3, Microsoft SQL Server management Studio, Postgres SQL,Netezza Database, Oracle Database, XML Processing.

ETL Talend Developer/Admin

Confidential

Responsibilities:

  • Participated in all phases of development life-cycle with extensive involvement in the definition and design meetings, functional and technical walkthroughs.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
  • Created and managed Source to Target mapping documents for all Facts and Dimension tables.
  • Design, develop, validate and deploy the Talend ETL processes for the DWH team using HADOOP (PIG, HIVE) on Hadoop.
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
  • Developed multiple tracks (Modules) for the data migration between multiple different systems like MRA, TIPP to NRSC, LIM to JAS, EPM to MR, ONW to TIPP.
  • Used ETL methodologies and best practices to create Talend ETL jobs. Followed and enhanced programming and naming standards.
  • Involved in Data Extraction from Flat files and XML files using Talend by using Java as Backend Language.
  • Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
  • Develop ETL mappings for various Sources (.TXT, .CSV, XML) and also load the data from these sources into relational tables with Talend Enterprise Edition.
  • Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
  • Handled importing of data from various data sources, performed transformations using Hive, MapReduce, Spark and loaded data into HDFS.
  • Experience in using DOM4J parsing and XML Bean to process, validate, parse and extract data from XML files.
  • Implemented installation and configuration of multi-node cluster on cloud using AWS on EC2.
  • Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)
  • Copy data to AWS S3 for storage and use COPY command to transfer data to Redshift. Used Talend connectors integrated to Redshift.
  • Developed jobs to expose HDFS files to Hive tables and Views depending up on the schema versions.
  • Imported data from RDBMS (MySQL, Oracle) to HDFS and vice versa using Sqoop. (Big Data ETL tool) for Business Intelligence, visualization and report generation.
  • Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
  • Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
  • Automated SFTP process by exchanging SSH keys between UNIX servers. Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.

Environment: Talend Data Integration 6.4.0, Talend Enterprise Big Data Edition, TalendAdministrator Console, XML, Oracle 11g, Hive, HDFS, Sqoop, SQL Navigator, Toad, Control M, Putty, Winscp.

ETL Talend Developer

Confidential

Responsibilities:

  • Interacted with business team to understand business needs and to gather requirements.
  • Designed target tables as per the requirement from the reporting team and also designed Extraction, Transformation and Loading (ETL) using Talend.Worked on Data Integration from different source systems.
  • Created Technical Design Document from Source to stage and Stage to target mapping. Worked with Talend Studio (Development area) & Admin Console (Admin area)
  • Created Java Routines, Reusable transformations, Joblets using Talend as an ETL Tool.
  • Created Complex Jobs and used transformations like tMap, tOracle (Components), tLogCatcher, tStatCatcher, tFlowmeterCatcher, File Delimited components and Error handling components (tWarn, tDie)
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Identified performance issues in existing sources, targets and Jobs by analyzing the data flow, evaluating transformations and tuned accordingly for better performance.
  • Manage all technical aspects of the ETL Jobs process with other team members.
  • Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading.
  • Created contexts to use the values throughout the process to pass from parent child to child jobs and child to parent jobs.Worked on Joblets (reusable code) & Java routines in Talend.
  • Expertise in Service Oriented Architecture (SOA) and Involved in publishing Web Services that include components like WSDL, SOAP, UDDI.
  • Performed Unit testing and created Unix Shell Scripts and provided on call support. Schedule Talend Jobs using Job Conductor (Scheduling Tool in Talend) - available in TAC.
  • Used XML for ORM mapping relations with the java classes and the database.
  • Retrieved data from Oracle and loaded into SQL Server data Warehouse.
  • Created many complex ETL jobs for data exchange and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structure.
  • Monitoring the Data Quality, generating weekly/monthly/yearly statistics reports on production processes - success / failure rates for causal analysis as maintenance part and Enhancing existing production ETL process.

Environment: Talend Platform 6.2, Big Data, UNIX, Oracle, XML, Oracle, TAC.

ETL Talend Developer

Confidential

Responsibilities:

  • Worked closely with Business Analysts to review the business specifications of the project and also to gather the ETL requirements.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
  • Created and managed Source to Target mapping documents for all Facts and Dimension tables
  • Analyzing the source data to know the quality of data by using Talend Data Quality.
  • Involved in writing SQL Queries and used Joins to access data from Oracle, and MySQL.
  • Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
  • Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Experienced in using debug mode of talend to debug a job to fix errors.
  • Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
  • Conducted JAD sessions with business users and SME's for better understanding of the reporting requirements.
  • Developed Talend jobs to populate the claims data to data warehouse - star schema.
  • Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.
  • Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
  • Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.

Environment: Talend Data Integration 5.5.1, Talend Enterprise Big Data Edition 5.1, XML, Talend Administrator Console, MS SQL Server 2012/2008, Oracle 11g, Hive, HDFS, Sqoop, TOAD, UNIX.

Software Engineer

Confidential

Responsibilities:

  • Interacted with Data Modellers and Business Analysts to understand the requirements and the impact of the ETL on the business.
  • Designed ETL specification documents for all the projects. Created Tables, Keys (Unique and Primary) and Indexes in the DB2 server.
  • Extracted data from Flat files, DB2, SQL and Oracle to build an Operation Data Source. Applied business logic to load the data into Global Data Warehouse.
  • Extensively worked on Facts and Slowly Changing Dimension (SCD) tables.
  • Extensively used the Add Currently Processed Flat File Name port to load the flat file name and to load contract number coming from flat file name into Target.
  • Worked on complex Source Qualifier queries, Pre-and Post SQL queries in the Target.
  • Extensively used workflow variables, mapping parameters and mapping variables.
  • Created sessions, batches for incremental load into staging tables and scheduled them to run daily.
  • Prepared migration document to move the mappings from development to testing and then to production repositories.

Environment: Informatica Power Center 8.6.1/9.1.0, Oracle 11g, SQLServer2008, IBM (DB2), MS Access, Windows XP, Toad, Tidal, SQL developer.

Hire Now