We provide IT Staff Augmentation Services!

Etl/ Talend Developer Resume

5.00/5 (Submit Your Rating)

Oakland, CA

SUMMARY:

  • Over 9+ years of IT industry experience in all aspects of Analysis, Design, Testing, Development, Implementation and Support of Relational Database (OLTP), Data Warehousing Systems (OLAP) and Data Marts in various domains.
  • Around 3+ years of experience with Talend Open Studio &Talend Enterprise platform for Data Management.
  • Experience in working with Data Warehousing Concepts like OLAP, OLTP, Star Schema, Snow Flake Schema, Logical Data Modeling, Physical Modeling and Dimension Data Modeling.
  • Utilized t - Stats Catcher, t-Die, t-Log Row to create a generic job to store processing stats.
  • Experienced in creating Generic schemas and creating Context Groups and Variables to run jobs against different environments like Dev, Test and Prod.
  • Thorough knowledge of addressing Performance Issues and Involved in query tuning, Index tuning, Data profiling and other database related activities.
  • Extensively created mappings in Talend using t-Map, t-Join, t-Replicate, t-Parallelize, t-Java, t-Java row, t-Die, t-Aggregate Row, t-Warn, t-Log Catcher, t-Filter, t-Global map etc.
  • Wrote Hive and Pig queries for data analysis to meet the business requirements.
  • Experienced in scheduling Talend jobs using Talend Administration Console (TAC)
  • Experience with Talend DI Installation, Administration and development for Confidential and application integration.
  • Expertise in Data modelling techniques like Data Modeling- Dimensional/ Star Schema, and Snowflake modelling, Slowly Changing Dimensions (SCD Type 1, Type 2, and Type 3)
  • Tracking Daily data load, Monthly data extracts and send to client for their verification.
  • Strong experience in designing and developing Business Intelligence solutions in Data Warehousing using ETL Tools.
  • Excellent understanding and best practice of Data Warehousing Concepts, involved in Full Development life cycle of Data Warehousing.
  • Involved in extracting user's data from various data sources into Hadoop Distributed File Systems (HDFS)
  • Automated all the jobs, for pulling data from FTP server to load data into Hive tables, using Oozie workflows.
  • Experienced in analyzing, designing and developing ETL strategies and processes, Writing ETL specifications.
  • Experience with MapReduce, Pig, Programming Model, Installation and Configuration of Hadoop, HBase, Hive, Pig, Sqoop and Flume using Linux commands.
  • Experienced in working with different data sources like Flat files, Spreadsheet files, log files and Databases.
  • Worked extensively with slowly changing dimensions.
  • Hands-on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, systems integration and user acceptance testing.
  • Excellent interpersonal and communication skills, and is experienced in working with senior level managers, business people and developers across multiple disciplines

PROFESSIONAL EXPERIENCE:

ETL/ Talend Developer

Confidential, Oakland, CA

Responsibilities:

  • The project was aimed to support existing clients by changing from manual Process to Automation through ETL development using Talend&Cognos Data Manager
  • Closely worked with Data Architects in designing of tables and even involved in modifying technical Specifications.
  • Designed and Implemented the ETL process using Talend Enterprise Big Data Edition to load the data from Source to Target Database.
  • Involved in Data Extraction from Oracle, Flat files and XML files using Talend by using Java as Backend Language.
  • Used tWaitForFile component for file watch event jobs.
  • Used over 20+ Components in Talend Like (tMap, tfilelist, tjava, tlogrow, toracleInput, toracleOutput, tsendEmailetc)
  • Used debugger and breakpoints to view transformations output and debug mappings.
  • Develop ETL mappings for various Sources (.TXT, .CSV, .XML) and also load the data from these sources into relational tables with Talend Enterprise Edition.
  • Worked on Global Context variables, Context variables, and extensively used over 100+components in Talend to create jobs
  • Created child jobs to use them in parent job in using tRunJob.
  • Extracting transformed data from Hadoop to destination systems, as a one-off job, batch process, or Hadoop streaming process.
  • Worked on Error handling techniques and tuning the ETL flow for better performance.
  • Worked Extensively TAC (Admin Console), where we Schedule Jobs in Job Conductor
  • Extensively Used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tOracleInput, tOracleOutput, tfileList, tDelimited etc.
  • Worked with Oracle SQL Developer while Implementing Unit Testing of ETL Talend Jobs.
  • Scheduling the ETL mappings on daily, weekly, monthly and yearly basis.
  • Working on POC Big Data like loading the data into HDFS and creating Map Reduce Jobs
  • Worked on the project documentation and also prepared the Source Target mapping specs with the business logic and also involved in data modeling
  • Worked on migrating data warehouses from existing SQL Server to Oracle database.
  • Implemented Performance tuning in Mappings and Sessions by identifying the bottlenecks and Implemented effective transformation Logic.
  • Created Workflows using various tasks like sessions, control, decision, e-mail, command, worklets, and assignment and worked on scheduling of the workflows.
  • Verify the logs to confirm all the relevant jobs are completed successfully and timely and involved in production support to resolve the production issues.
  • Migrated the code and release documents from DEV to QA (UAT) and to Production.

Environment: Talend Platform for Big Data 6.2, Talend Open Studio 5.0.1, Cognos Data Manager, Cognos 10.2.2, UNIX, Oracle 12c, TAC (Admin Center), SQL Server, TOAD, Autosys.

Talend Developer

Confidential, Fremont, CA

Responsibilities:

  • Responsible for building scalable distributed data solutions using Hadoop. Continuous monitoring and managing the Hadoop cluster through Cloudera Manager
  • Upgraded the Hadoop Cluster from CDH3 to CDH4, setting up High Availability Cluster and integrating HIVE with existing applications.
  • Designed Oozie jobs for the auto processing of similar data. Worked on project to retrieve log messages procured by leveraging Spark Streaming.
  • Collect the data using Spark Streaming.
  • Analyzed the data by performing Hive queries and running Pig scripts to know user behavior.
  • Extensively used for all and bulk collect to fetch large volumes of data from table.
  • Installed Oozie workflow engine to run multiple Hive and Pig jobs.
  • Developed Pig scripts in the areas where extensive coding needs to be reduced.
  • Performed transformations, cleaning and filtering on imported data using Hive, Map Reduce, and loaded final data into HDFS.
  • Handled importing of data from various data sources using Sqoop, performed transformations using Hive, MapReduce and loaded data into HDFS.
  • Created HBase tables to store various data formats of PII data coming from different portfolios.
  • Configured Sqoop and developed scripts to extract data from MySQL into HDFS.
  • Worked on analyzing Hadoop cluster and different big data analytic tools including Pig HBase database and Sqoop.
  • Created HBase tables to store various data formats of PII data coming from different portfolios. Data processing using SPARK.
  • Parsed high-level design specification to simple ETL coding and mapping standards.
  • Cluster co-ordination services through Zookeeper.
  • Developed complex Talend jobs mappings to load the data from various sources using different components.
  • Design, develop and implement solutions using Talend Integration Suite.
  • Built big Data solutions using HBase handling millions of records for the different trends of data and exporting it to Hive.
  • Tested the data coming from the source before processing.
  • Debugged the technical issues and errors was resolved.

Environment: Hadoop (Cloudera), Talend ETL Tool, HDFS, MapReduce, Pig, Hive, Sqoop, HBase, Oozie, Flume, Zookeeper, java, SQL, Scripting, Spark.

ETL/Talend Developer

Confidential, Mount Laurel, NJ

Responsibilities:

  • Worked closely with Business Analysts to review the business specifications of the project and also to gather the ETL requirements.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
  • Created and managed Source to Target mapping documents for all Facts and Dimension tables
  • Analyzing the source data to know the quality of data by using Talend Data Quality.
  • Involved in writing SQL Queries and used Joins to access data from Oracle, and MySQL.
  • Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
  • Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput&tHashOutput and many more)
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Experienced in using debug mode of talend to debug a job to fix errors.
  • Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
  • Conducted JAD sessions with business users and SME's for better understanding of the reporting requirements.
  • Developed Talend jobs to populate the claims data to Confidential - star schema.
  • Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.
  • Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
  • Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.

Environment: Talend Data Integration 5.5.1, Talend Enterprise Big Data Edition 5.1, Talend Administrator Console, MS SQL Server 2012/2008, Oracle 11g, Hive, HDFS, Sqoop, TOAD, UNIX.

Informatica Developer

Confidential, Morristown, NJ

Responsibilities:

  • Interacted with Data Modellers and Business Analysts to understand the requirements and the impact of the ETL on the business.
  • Designed ETL specification documents for all the projects.
  • Created Tables, Keys (Unique and Primary) and Indexes in the DB2 server.
  • Extracted data from Flat files, DB2, SQL and Oracle to build an Operation Data Source. Applied business logic to load the data into Global Confidential .
  • Extensively worked on Facts and Slowly Changing Dimension (SCD) tables.
  • Extensively used the Add Currently Processed Flat File Name port to load the flat file name and to load contract number coming from flat file name into Target.
  • Worked on complex Source Qualifier queries, Pre-and Post SQL queries in the Target.
  • Extensively used workflow variables, mapping parameters and mapping variables.
  • Created sessions, batches for incremental load into staging tables and scheduled them to run daily.
  • Prepared migration document to move the mappings from development to testing and then to production repositories.

Environment: Informatica Power Center 8.6.1/9.1.0, Oracle 11g, SQLServer2008, IBM (DB2), MS Access, Windows XP, Toad, Tidal, SQL developer

Database Developer

Confidential

Responsibilities:

  • Responsible for requirement analysis of the application .ETL tool Informatica was used to load strategic source data to build the data marts.
  • An operational data store was created. Metadata Build up was designed for performing data mapping.
  • Also involved in Mass data loads, refreshing the data in various applications, Performance evaluations, modifying the existing code to accommodate new features.
  • Used various Transformations like Aggregator, Router, Expression, Source Qualifier, Filter, Lookup, Joiner, Sorter, XML Source qualifier, Stored Procedure and Update Strategy.
  • Worked extensively on Flat Files, as the data from various Legacy Systems are flat files.
  • Have setup Test and Production Environment for all mappings and sessions.
  • Created and configured Sessions in Workflow Manager and Server Manager.
  • Debugged the sessions using Debugger and monitored Workflows, Worklets and Tasks by Workflow Monitor.
  • Created and used mapping parameters, mapping variables using Informatica mapping designer to simplify the mappings.
  • Developed Oracle stored procedures, packages and triggers for data validations

Environment: Informatica Power Center 8.6.1, Oracle 11g, SQLServer2008, MS Access, Windows XP, Toad, SQL developer

We'd love your feedback!