We provide IT Staff Augmentation Services!

Talend Consultant Resume

0/5 (Submit Your Rating)

Austin, TX

SUMMARY

  • Having 7+years of total experience in Information Technology including Data Warehouse/Data Mart development using Talend DI/BDE and Informatica Power Center.
  • 4 years of experience in using Talend Data Integration tool as applied to BI data analytics, reporting and dashboard projects.
  • Expertise in designing/building Enterprise Data Warehouses (EDW), Operational Data Store (ODS), Data Marts, and Decision Support Systems (DSS) using Multidimensional and Ralph Kimball Dimensional modeling (Star and Snowflake schema) Concepts
  • Highly proficient in the integration of various data sources involving multiple relational databases like Oracle, MS SQL Server, Teradata, and non - relational sources like COBOL Files and Flat File.
  • Strong understanding of the principles of DW using fact tables, dimension tables and star/snowflake schema modeling.
  • Developed slowly changing dimension (SCD) mappings using type-I, type-II, and type-III methods.
  • Expertise on Exception Handling Mappings for Data Quality, Data Cleansing and Data Validation.
  • Excellent knowledge in identifying performance bottlenecks and tuning the ETL Loads for better performance and efficiency.
  • Expertized in creating specifications documents from source to Target systems.
  • Experience in UNIX shell scripting, CRON, FTP and file management in various UNIX environments.
  • Experience in using Talend Big Data components to create connections to various third-party tools used for transferring, storing or analyzing big data, such as HDFS Components, Hive, Sqoop, MongoDB and BigQuery to quickly load, extract, transform and process large and diverse data sets.
  • Experience in Trouble shooting and implementing Performance tuning at various levels such as Source, Target, Mapping, Session and System in ETL Process. Have heavily worked on performance tuning long running Informatica mappings using Pushdown optimization and session partitioning.
  • Excellent knowledge in identifying performance bottlenecks and tuning the ETL Loads for better performance and efficiency.
  • Good exposure in overall SDLC including requirement gathering, development, testing, debugging, deployment, documentation and production support.
  • Able to work under tight deadlines and rapidly changing priorities with proactive, creative & focused approach to business needs with high analytical skill and team playing skills.

TECHNICAL SKILLS

ETL Tools: Talend DI/BDE 6.3/6.0.1/5. X, InformaticaPowercenter V8.X

DW Tools: Erwin, ER/Studio, MS Visio, Teradata studio, Teradata SQL Assistant

RDBMS: Oracle 10G/9i/8.x, Teradata v14, MS SQL Server 2012/2015

Languages: SQL, PL/SQL, C, C++, VB, Shell Scripting, Java and XML.

Operating Systems: Microsoft Windows, MS-DOS, UNIX and Linux

Development Tools: TOAD, SQL Plus, SQL Developer, MS Visual Studio, Autosys

PROFESSIONAL EXPERIENCE

Confidential, Austin, TX

Talend Consultant

Responsibilities:

  • Interacting with the clients on a regular basis to discuss day-to-day issues and matters.
  • On-Call/Production Support provided during day-time and off-hours.
  • Played as administrator for setting up development, QA, UAT and PROD environments for Talend, Postgres for development and documenting install plan at client locations.
  • Setup ETL Framework, best practices around Talend for data integration implementation.
  • Responsible for installing Talend on multiple environments, creating projects, setting up user roles, setting up job servers, configure TAC options, adding Talend jobs, job failures, on-call support and scheduling etc.
  • Excellent experience working on tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Developed jobs to expose HDFS files to Hive tables and Views depending up on the schema versions.
  • Created Hive tables, partitions and implemented incremental imports to perform ad-hoc queries on Structured data. Developed jobs to move inbound files to HDFS file location based on monthly, weekly, daily and hourly partitioning.
  • Worked extensively on design, development and deployment of talend jobs to extract data, filter the data and load them into datalake.
  • Manage and Review Hadoop log files and hands on with executing Linux and HDFS Commands
  • Responsible for writing Talend Routines in Java.
  • Developed ODS/OLAP data model in Erwin and also created source to target mapping documents.
  • Experience working with web services using tSOAP components for sending XML requests and receiving response XML files. Expertized in reading XMLs files on a loop and sending to webservice end point for generating output XML files. And also used Advanced XML mappers for parsing multiple loop elements.
  • Responsible for digging into PL/SQL code for investigating data issues.
  • Involved in the development of Talend Jobs and preparation of design documents, technical specification documents.
  • Implemented job parallelism in Talend BDE 6.0.1.
  • Experience working with Big data components for extracting and loading data into HDFS file system.
  • Production Support activities like application checkout, batch cycle monitoring and resolving User Queries.
  • Responsible for deploying code to different environments using GIT.

Environment: Talend Big Data 6.0.1/Talend DI 6.0.1/6.3, Jaspersoft Professional 6, Postgres 9.4, Oracle 11g, Web services (SOAP), GIT, Jira, Jenkins, AWS

Confidential, Chicago, IL

Talend Developer/Admin

Responsibilities:

  • Interacted with the Business users to identify the process metrics and various key dimensions and measures. Involved in the complete life cycle of the project.
  • Lead couple of integration projects from gathering requirements phase to implementation phase which enabled Customers to see the data daily basis or weekly basis with very minimal manual intervention rather than spending quite amount of manual hours.
  • Developed talend jobs to push data into consolidated stage (for all feeds), which is sauce for MDM process.
  • Worked on MDM modeling through MDM perspective through Talend 5.5.1 suite and developed jobs to push data to MDM.
  • Implemented different matching rules, data validation rule to get the golden record.
  • Used Talend most used components (tMap, tDie,tConvertType, tSOAP, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie and many more).
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems Including RDBMS, XML, CSV, and Flat file structures.
  • Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
  • Developed PL/SQL Packages.
  • Conducted JAD sessions with business users and SME's for better understanding of the reporting requirements.
  • Developed Talend jobs to populate the claims data to data warehouse - star schema.
  • Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.
  • Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
  • Experienced with Java transformations for calling Hive views to extract data from Hadoop systems.
  • Expertized in developing shell scripts in UNIX.
  • Created Talend mappings using the Transformations.
  • Involved in analyzing and extracting mongo application collections into ODS by using hive views.
  • Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
  • Involved in Unit testing, User Acceptance Testing to check whether the data loads into target are accurate, which was extracted from different source systems according to the user requirements.
  • Responsible for prioritizing the issues and assign them to the production support team and planning the deployment of fixes for the same.

Environment: Talend Data Integration 6.1/5.5.1, Talend MDM 5.5,Talend Enterprise Big Data Edition 5.5.1, MS SQL Server 2012/2008, Oracle 11g, Cloudera, Hive, HDFS, Sqoop, TOAD, UNIX.

Confidential

ETL Developer

Responsibilities:

  • Worked with Business Analyst to design Business Requirement Documents.
  • Created a data dictionary to map the business requirements to attributes to design logical data model implementing star schema.
  • Responsible in producing report analysis by column level for all oracle tables, primary key analysis, foreign key analysis, cross domain analysis.
  • Provided tips for the DW team to improve the performance of the ETL jobs.
  • Developed templates for Dimensional tables, Slowly Changing dimension tables, Fact tables.
  • Designed and Developed ETL logic for implementing CDC by tracking the changes in critical fields required by the user.
  • Developed standard and re-usable mappings and mapplets using various transformations like expression, aggregator, joiner, source qualifier, router, lookup Connected/Unconnected, and filter.
  • Extensive use of Persistent cache to reduce session processing time.Identified performance issues in existing sources, targets and mappings by analyzing the data flow, evaluating transformations and tuned accordingly for better performance.
  • Maintained warehouse metadata, naming standards and warehouse standards for future application development.
  • Used Workflow Manager for creating, validating, testing and running the sequential and concurrent sessions and scheduling them to run at specified time and as well to read data from different sources and write it to target databases.
  • Implemented type II slowly changing dimension techniques on changing attributes.
  • Implemented Retry logic for reprocessing records in next run if the foreign key data missed in first run due to timing issues.
  • Automated coding process to create ETL jobs from Enterprise to Stage for faster delivery.
  • Developed UNIX shell scripts for automating process.

Environment: Informatica 8.5, Oracle 9i/10g, Teradata, SQL Plus, Teradata SQL Assistant, Control M, PVCS.

We'd love your feedback!