We provide IT Staff Augmentation Services!

Talend-big Data Developer Resume

5.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY:

  • Around 5 Years of IT experience in all aspects of Analysis, Design, Testing, Development, Implementation and Support of ETL, Data Warehousing Systems and Data Marts in various domains.
  • Expertise in Data Warehouse/Data mart, ODS, OLTP and OLAP implementations teamed with project scope, Analysis, requirements gathering, data modeling, Effort Estimation, ETL Design, development, System testing, Implementation and production support.
  • Around 4 years of experience with Talend Open Studio & Talend Enterprise platform.
  • Experienced in working with Talend for Big Data, Talend for Data Integration.
  • Experience in using cloud components and connectors to make API calls for accessing data from cloud storage Amazon S3 in Talend Studio.
  • Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType,, tflowtoIterate, tAggregate, tSortRow, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tWarn, tLogCatcher, tMysqlScd, tFilter, tGlobalmap, tDie etc.
  • Expertise in Data modeling techniques like Data Modeling - Dimensional/ Star Schema and Snowflake modeling, Slowly Changing Dimensions (SCD Type 2).
  • Experienced in Waterfall, Agile methodologies.
  • Experienced with Talend Big Data, Hadoop, Hive and used Talend Big data components like tHDFSInput, tHDFSOutput,, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput. Experience in using spark SQL
  • Experience in development and design of ETL (Extract, Transform and Loading data) methodology for supporting data transformations and processing, in a corporate wide ETL Solution using Informatica Power Center.
  • Created mappings using Lookup, Aggregator, Joiner, Expression, Filter, Router, Update strategy and Normalizer Transformations. Developed reusable Transformation and Mapplets.
  • Strong Experience with shell scripting, understanding of approaches for business intelligence, data warehouse.
  • Experience in data migration with data from different application into a single application.
  • Hands-on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, system integration and user acceptance testing.

TECHNICAL SKILLS:

Big Data: HDFS, Hive

Databases: Teradata, Oracle12c/11x/10g,Microsoft SQL Server

T: SQL, PL/SQL

Scripting: UNIX shell scripting

Languages: SQL, Python.

SQL Navigator, Putty, MS: Office.JIRA,winscp

Environment: Windows, UNIX & Linux

PROFESSIONAL EXPERIENCE:

Confidential, Chicago IL

Talend-Big Data Developer

Responsibilities:

  • Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirements.
  • Closely worked with Data Architects in designing of tables and even involved in modifying technical Specifications.
  • Involved in Extraction, Transformation and Loading of data from multiple source systems to Aws S3.
  • Involved in the Development of copying Data from AWS S3 to Redshift using the Talend Process.
  • Involved in writing custom copy command query’s and used the context extensively and implemented in tRedshift Row Component.
  • Extensively used t system component to push large sets of data to s3.
  • Developed a Talend Code for S3 Tagging in the Process of Moving data from source to S3
  • Utilized Talend components like tS3Put, tS3Get, tS3File List, tRedshiftRow, tRedshiftUnload, tRedshiftBulkExec, tFlowToIterate, tSetGlobalVar,tsnowflakeinput,tsnowflakeoutout,tsnowflakerow.
  • Experienced in creating external staging tables in snowflake to move data between snowflake and Aws S3.
  • Experienced in writing complex snow SQL queries in tsnowflakerow.
  • Experienced in loading parquet format files to S3 using Big Data Batch Jobs.
  • Extensively used AWS Athena to Query Parquet format files in S3.
  • Extensively used copy command in snowflake to copy data to Aws S3 using customized file formats like CSV and PARQUET.
  • Involved in Integrating IAM Roles in Talend Components.
  • Experienced in creating standard jobs. Involved in development of big data batch Jobs.
  • Extensively used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tOracleInput, tOracleOutput, tFileList etc, tS3put, tS3get, tReplicate, tSortrow, tDenormalize, tNormalize, tRedshiftRow
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tHiveOutput, tHiveRow, tHive Connection .
  • Experienced in executing the jobs in parallel using tparallelize component.
  • Used debugger and breakpoints to view transformations output and debug mappings.
  • Load and transform data into HDFS from large set of structured data /Oracle/SQL server using Talend Big data studio.
  • Worked on Global variables, Context variables, and extensively used tcontext load in most of the jobs.
  • Experienced in creating reusable jobs for error handling.
  • Experienced in tuning the ETL Jobs for better performance.
  • Extensively worked with TAC (Talend Administrator Console) for scheduling jobs using the execution plan.

Environment: Talend Data Integration 6.1/5.5.1, Talend Enterprise Big Data Edition 5.5.1, Talend Administrator Console, Oracle 11g, Hive, HDFS, SQL Navigator, Toad, Putty, Winscp

Confidential, Nashville TN

Talend-Big Data Developer

Responsibilities:

  • Participated in all phases of development life-cycle with extensive involvement in the definition and design meetings, functional and technical walkthroughs.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components
  • Created and managed Source to Target mapping documents for all Facts and Dimension tables
  • Used ETL methodologies and best practices to create Talend ETL jobs. Followed and enhanced programming and naming standards.
  • Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
  • Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport
  • Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
  • Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Created Implicit, local and global Context variables in the job. Worked on Talend Administration Center (TAC) for scheduling jobs and adding users.
  • Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
  • Developed stored procedure to automate the testing process to ease QA efforts and also reduced the test timelines for data comparison on tables.
  • Automated SFTP process by exchanging SSH keys between UNIX servers. Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
  • Involved in production n deployment activities, creation of the deployment guide for migration of the code to production, also prepared production run books.

Environment: Talend Data Integration 6.4, Talend Enterprise Big Data Edition 6.4, Talend Administrator Center, Oracle 11g, Hive, HDFS, Netezza, SQL Navigator, Toad, Control M, Putty, Winscp.

Confidential

Talend Developer

Responsibilities:

  • Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirements.
  • Developed jobs using various components in Talend.
  • Designed ETL Jobs/Packages using Talend Integration Suite (TIS).
  • Created complex mappings in Talend using Talend components like tHash, tDenormalize, tMap, tUniqueRow.
  • Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats into a Database table to record job history.
  • Created Talend Mappings to populate the data into dimensions and fact tables.
  • Frequently used Talend Administrative Console (TAC)
  • Developed complex Talend ETL jobs to migrate the data from flat files to database.
  • Implemented custom error handling in Talend jobs and worked on different methods of logging.
  • Created ETL/Talend jobs both design and code to process data to target databases.
  • Created Talend jobs to load data into various Oracle tables. Utilized Oracle stored procedures and wrote few Java code to capture global map variables and use them in the job.
  • Successfully Loaded Data into different targets from various source systems like Oracle Database, DB2, Flat files, XML files etc. into the Staging table and then to the target database.
  • Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
  • Performed Unit testing and System testing to validate data loads in the target.

Environment: Talend 5.5.2, UNIX, Shell script, SQL Server, Oracle, Business ObjectsNotus Technologies, Hyderabad

Confidential

Jr. SQL/BI Developer

Responsibilities:

  • Responsible for designing and developing of mappings, Mapplets, sessions and work flows for load the data from source to target database using Informatica Power Center and tuned mappings for improving performance.
  • Created database objects like views, indexes, user defined functions, triggers and stored procedures.
  • Involved in ETL process from development to testing and production environments.
  • Extracted date from various sources like Flat files, Oracle and loaded it into Target systems using Informatica 7.x.
  • Developed mappings using various transformations like update strategy, lookup, stored procedure, router, joiner, sequence generator and expression transformation.
  • Developed PL/SQL triggers and master tables for automatic creation of primary keys.
  • Used Informatica Power Center Workflow Manager to create sessions, batches to run with the logic embedded in the mappings.
  • Tuned mappings and SQL queries for better performance and efficiency.
  • Automated existing ETL operations using Autosys.
  • Created & Ran shell scripts in UNIX environment.
  • Created and ran the Workflows using Workflow manager in Informatica Maintained stored definitions, transformation rules and targets definitions using Informatica repository manager.
  • Created tables and partitions in database Oracle.

Environment: Informatica Power Center 9.x, Oracle, SQL developer, MS Access, PL/SQL, UNIX Shell Scripting, SQL Server, Windows XP

We'd love your feedback!