Talend Developer Resume
Santa Clara, CA
PROFESSIONAL SUMMARY:
- Proficient ETL Developer over 5 years of experience in Data Warehouse, ETL Maintenance, Master Data Management (MDM) strategy, Data Quality and Big Data Eco Systems.
- Expertise with Talend Data Integration 6.2.1 frequently used components (tOracleInput, tMysqlnput, tMap, tSoap, tESBConsumer, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, t SetGlobalVar, tHashInput & tHashOutput, tDie and more.
- Experienced in working with Horton works distribution of Hadoop, HDFS, MapReduce, Hive, Sqoop, Flume, Pig, HBase, and MongoDB.
- Experience in dealing with structured and semi - structured data in HDFS.
- Experience in integration of various data sources like Teradata, SQL Server, Oracle, DB2, Netezza and Flat Files.
- Extracted data from multiple operational sources for loading staging area, Data warehouse, Data Marts using SCDs (Type 1/Type 2/ Type 3) loads.
- Expert in using the Talend Troubleshooting and DataStage to understand the errors in Jobs and used the tMap/expression editor to evaluate complex expressions and look at the transformed data to solve mapping issues.
- Created complex mappings in Talend using components like: tMap, tJoin, tReplicate, tParallelize, tAggregateRow, tDie, tUnique, tFlowToIterate, tSort, tFilterRow, tWarn, tbuffer, tcontextload.
- Strong Oracle/SQL Server database programming - development of stored procedures, triggers and views.
- Extensive experience in using Talend features such as context variables, triggers, connectors for Database and flat files.
- Hands on Experience on many components which are there in the palette to design Jobs & used Context Variables to Parameterize Talend Jobs.
- Experience in Debugging, Error Handling and Performance Tuning of sources, targets, Jobs etc.
- Extensive experience in Relational and Dimensional Data modelling for creating Logical and Physical Design of Database and ER Diagrams using data modelling tools like ERWIN and ER Studio.
- Worked extensively schema for Data Warehouse, ODS architecture by using tools like Erwin data modeler, Power Designer, Embarcadero E-R Studio and Microsoft Visio.
- Having good knowledge in Normalization and De-Normalization techniques for optimum on XML data and XSD schema designing.
- Involved in preparing test plans and cases for unit testing based on requirements.
- Debugging, Logging, and Testing, demonstrates the different methods for finding problems within Talend code.
- Experience in scheduling tools Autosys, Control M & Job Conductor ( Talend Admin Console).
- Experience in working with parallel extender for splitting bulk data into subsets to distribute the data to all available processors to achieve best job performance.
- Self-motivated, excellent written/verbal communication and team work skills yet flexible to independent responsibilities.
- Fast learner with excellent analytical skills and good communication skills.
WORK EXPERIENCE:
Talend Developer
Confidential, Santa Clara, CA
Responsibilities:
- Implemented File Transfer Protocol operations using Talend Studio to transfer files in between network folders.
- Developed complex ETL mappings for Stage, Dimensions, Facts and Data marts load Worked on Data Migration using export/import.
- Created Talend jobs using the dynamic schema feature.
- Load and transform data into HDFS from large set of structured data /Oracle/Sql server using Talend Big data studio.
- Used Big Data components (Hive components) for extracting data from hive sources.
- Wrote HiveQL queries using joins and implemented in tHiveInput component.
- Utilized Big Data components like tHiveInput, tHiveOutput, tHDFSOutput, tHiveRow, tHiveLoad, tHiveConnection, tOracleInput, tOracleOutput, tPreJob, tPostJob, tLogRow.
- Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
- Performance tuning - Using the tmap cache properties, Multi-threading and tParallelize components for better performance in case of huge source data. Tuning the SQL source queries to restrict unwanted data in ETL process.
- Used more components in Talend and Few to be mentioned: tjava, toracle, txmlMap, tdelimited files, tlogrow, tlogback components etc. in many of my Jobs Design.
- Worked on Joblets (reusable code) & Java routines in Talend.
- Implemented Talend POC to extract data from Salesforce API as an XML Object & .csv files and load data into SQL Server Database.
- Experienced in writing SQL Queries and used Joins to access data from Oracle, and MySQL.
- Implemented Error Logging, Error Recovery, and Performance Enhancement's & created Audit Process (generic) for various Application teams.
- Experience in using Repository Manager for Migration of Source code from Lower to higher environments.
- Created Projects in TAC and Assign appropriate roles to Developers and integrated SVN (Subversion).
- Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis (Cron Trigger).
Environment: Talend Platform for Big Data 5.6.2, Enterprise Platform for Data integration and MDM (V6.1.1,5.5.1, 5.6.1), UNIX, Oracle 11g, SQL Server 2012, Microsoft SQL Server management Studio, WINDOWS XP.
Talend Developer
Confidential, Rochester, MN
Responsibilities:
- Worked with Data Mapping Team to understand the source to target mapping rules.
- Analyzed the requirements and framed the business logic for the ETL process using Talend.
- Involved in the ETL design and its documentation.
- Developed Jobs in Talend Enterprise edition from stage to source, intermediate, conversion and Target.
- Worked on Talend ETL to load data from various sources to Oracle DB. Used tmap, treplicate, tfilterrow, tsort and various other features in Talend.
- Worked on Talend ETL and used features such as Context variables, Database components like tMSSQLInput, tOracleOutput, file components, ELT components etc.
- Followed the organization defined Naming conventions for naming the Flat file structure, Talend Jobs and daily batches for executing the Talend Jobs.
- Worked on Context variables and defined contexts for database connections, file paths for easily migrating to different environments in a project.
- Implemented Error handling in Talend to validate the data Integrity and data completeness for the data from the Flat File.
- Used Talend components such as tmap, tFileExist, tFileCompare, tELTAggregate, tOracleInput, tOracleOutput etc.
- Designed and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2 to capture the changes.
- Followed the organization defined Naming conventions for naming the Flat file structure, Talend Jobs and daily batches for executing the Talend Jobs.
- Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
- Involved in automation of FTP process in Talend and FTPing the Files in UNIX.
- Created Talend Development Standards. This document describes the general guidelines for Talend developers, the naming conventions to be used in the Transformations and development and production environment structures.
- Extracted data from Oracle as one of the source databases.
- Optimized the performance of the mappings by various tests on sources, targets and transformations.
Environment: Talend Data integration 5.6.1, Oracle 11g, MS SQL Server 2012/2008, PL/SQL, Agile Methodology, T-SQL, SSIS, TOAD, AIX, Shell Scripts, Autosys.
ETL Developer
Confidential, Park, NJ
Responsibilities:
- Developed Data-architecture, ETL Batch processes in Talend, SQL-server and PostgreSQL for Call-center systems including architecture for parsing documents related to Call-center notes.
- Designed and implemented new environments and ETL frameworks for conformed data delivery and Call-Center-Analytics in Tableau and custom-Analytics platform.
- Worked on Dimensional Models, SCDs, Error-event schema, Audit dimensions to satisfy business requirement and worked with development teams to implement the Data-Models, Mapping-docs and ETL to satisfying technical requirements.
- Developed jobs, components and Job lets in Talend.
- Created complex mappings in Talend using tHash, tDenormalize, tMap, tUnique Row. tPivot To Columns Delimited as well as custom component such as tUnpivot Row.
- Used tStats Catcher, tDie, tLog Row to create a generic job let to store processing stats into a Database table to record job history.
- Created Talend Mappings to populate the data into dimensions and fact tables.
- Developed complex Talend ETL jobs to migrate the data from flat files to database.
- Implemented custom error handling in Talend jobs and also worked on different methods of logging.
- Created Talend jobs to load data into various Oracle tables. Utilized Oracle stored procedures and wrote few Java code to capture global map variables and use them in the job.
- Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
Environment: Talend Open Studio 5.0.1, Informatica Power center, UNIX, Oracle, SQL Server, TOAD, AutoSys.
ETL Informatica Developer
Confidential
Responsibilities:
- Interacted with the business users on regular basis to consolidate and analyze the requirements.
- Identified the Entities and the relationships between the Entities to develop a logical model and later translated into physical model.
- Used Normalization up to 3NF and De-normalization for effective performance.
- Involved in implementation of the Test cases and Test Scripts.
- Tested the data and data integrity among various sources and targets.
- Tested to verify that all data were synchronized after the data is troubleshoot, and used SQL to verify/validate test cases.
- Written Test Cases for ETL to compare Source and Target database systems and check all the transformation rules.
- Defects identified in testing environment where communicated to the developers using defect tracking tool HP Quality Center.
- Performed Verification, Validation, and Transformations on the Input data.
- Tested the messages published by INFORMATICA and data loaded into various databases.
- Written Test Cases for ETL to compare Source and Target database systems and check all the transformation rules.
- Extracted data from databases like Oracle, SQL server and DB2 using Informatica to load it into a single repository for data analysis.
- Worked on multiple data marts in Enterprise Data Warehouse (EDW).
- Worked on Informatica Power Center Designer tools like Source Analyzer, Target Designer, Transformation Developer, Mapping Designer and Mapplet Designer.
- Worked on Informatica Power Center Workflow Manager tools like Task Designer, Workflow Designer, and Worklet Designer.
- Designed and developed Informatica power center medium to complex mappings using transformations such as the Source Qualifier, Aggregator, Expression, Lookup, Filter, Router, Rank, Sequence Generator, Stored Procedure and Update Strategy.
- Worked as a key project resource taking day-to-day work direction and accepting accountability for technical aspects of development.
- Developed the business rules for cleansing/validating/standardization of data using Informatica Data Quality.
- Designed and developed multiple reusable cleanse components.
Environment: Erwin r7.3, SQL/MS SQL Server, MS Analysis Services, Windows NT, MS Visio, XML, Informatica.