We provide IT Staff Augmentation Services!

Talend-big Data Developer Resume

4.00/5 (Submit Your Rating)

Charlotte, NC

SUMMARY:

  • Over 8 years of IT industry experience in all aspects of Analysis, Design, Testing, Development, Implementation and Support of Relational Database, Data Warehousing Systems and Data Marts in various domains.
  • Around 5 years of experience with Talend Open Studio & Talend Enterprise platform for Data Management.
  • Expertise in extracting data from various sources into Hadoop Distributed File Systems (HDFS).
  • Experience in using cloud components and connectors to make API calls for accessing data from cloud storage Amazon S3 in Talend open Studio.
  • Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType,, tflowtoIterate, tAggregate, tSortRow, tFlowMeter, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tWarn, tLogCatcher, tMysqlScd, tFilter, tGlobalmap, tDie etc
  • Created Talend ETL jobs to receive attachment files from pop e - mail using tPop, tFileList and tFileInputMail and then loaded data from attachments into database and achieved the files.
  • Strong understanding of NoSQL databases like HBase, MongoDB.
  • Expertise in Data modeling techniques like Data Modeling- Dimensional/ Star Schema and Snowflake modeling, Slowly Changing Dimensions (SCD Type 2).
  • Excellent working experience in Waterfall, Agile methodologies.
  • Proficient in performance analysis, monitoring and SQL query tuning using Collect Statistics, Hints and SQL Trace both in Teradata as well as Oracle.
  • Experience in Hadoop Big Data Integration with Data stage ETL on performing data extract, loading and transformation process for automobile ERP data.
  • Experience of Hadoop Architecture and various components such as HDFS, Name Node, Data Node, Job Tracker, Task Tracker, YARN and Map Reduce
  • Well versed with Talend Big Data, Hadoop, Hive and used Talend Big data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Experience in development and design of ETL (Extract, Transform and Loading data) methodology for supporting data transformations and processing,
  • ETL Solution using Informatica PowerCenter and IDQ tool.
  • Created mappings using Lookup, Aggregator, Joiner, Expression, Filter, Router, Update strategy and Normalizer Transformations. Developed reusable Transformation and Mapplets..
  • Strong Experience with shell scripting, understanding of approaches for business intelligence, data warehouse.
  • Self-Starter and Team Player with excellent communication, organizational and interpersonal skills with the ability to grasp things quickly

TECHNICAL SKILLS:

Big Data: HDFS, HBase, Hadoop MapReduce, Zookeeper, Hive, Impala, Sqoop, Flume, Oozie.

Databases: Netezza, Teradata & utilities (BTEQ, FASTLOAD, FASTEXPORT, MULTILOAD, TRUMP), Oracle12c/11x/10g, DB2, Microsoft SQL Server, Hive, Impala, Sybase.

Programming: T-SQL, PL/SQL, HTML, XML.

Scripting: Korn shell script & Windows batch scripting, JavaScript

Languages: SQL, Java, Python.

Other Tools: SQL Navigator, Putty, MS-Office, VMWare Workstation

Environment: Windows, UNIX (SunSolaris10, HP, AIX) & Linux

PROFESSIONAL EXPERIENCE:

Confidential, Charlotte, NC

Talend-Big Data Developer

Responsibilities:

  • Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirements.
  • Closely worked with Data Architects in designing of tables and even involved in modifying technical Specifications.
  • Involved in Extraction, Transformation and Loading of data.
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tHiveOutput, tHiveRow, tHiveConnection .
  • Designed and Implemented the ETL process using Talend Enterprise Big Data Edition to load the data from Source to Target Database.
  • Involved in Data Extraction from Flat files and XML files using Talend by using Java as Backend Language.
  • Using Talend to load the data into our warehouse systems
  • Used over 20+ Components in Talend Like ( tMap, Tfilelist, Tjava, Tlogrow, ToracleInput, ToracleOutput, tsendEmail etc)
  • Used debugger and breakpoints to view transformations output and debug mappings.
  • Load and transform data into HDFS from large set of structured data /Oracle/Sql server using Talend Big data studio.
  • Develop ETL mappings for various Sources ( .TXT, .CSV, XML ) and load the data from these sources into relational tables with Talend Enterprise Edition.
  • Worked on Global Context variables, Context variables, and extensively used over 50+components in Talend to create jobs.
  • Extracting transformed data from Hadoop to destination systems, as a one-off job, batch process, or Hadoop streaming process.
  • Worked on Error handling techniques and tuning the ETL flow for better performance.
  • Worked Extensively TAC (Admin Console), where we Schedule Jobs in Job Conductor.
  • Extensively Used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tOracleInput, tOracleOutput, tfileList etc .
  • Migrated the code and release documents from DEV to QA (UAT) and to Production.
  • Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type-2.
  • Created complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.

Environment: Talend Data Integration 6.1/5.5.1, Talend Enterprise Big Data Edition 5.5.1, Talend Administrator Console, Oracle 11g, Hive, HDFS, Sqoop, Netezza, SQL Navigator, Toad, Control M, Putty, Winscp

Confidential, Detroit,MI

Talend-Big Data Developer

Responsibilities:

  • Participated in all phases of development life-cycle with extensive involvement in the definition and design meetings, functional and technical walkthroughs.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components
  • Created and managed Source to Target mapping documents for all Facts and Dimension tables
  • Used ETL methodologies and best practices to create Talend ETL jobs.
  • Followed and enhanced programming and naming standards.
  • Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
  • Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
  • Used Talend most used components ( tMap, tDie, tConvertType, tLogCatcher, tRowGenerator, tHashInput & tHashOutput and many more).
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Created Implicit, local and global Context variables in the job.
  • Worked on Talend Administration Console (TAC ) for scheduling jobs and adding users.
  • Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
  • Developed stored procedure to automate the testing process to ease QA efforts and reduced the test timelines for data comparison on tables.
  • Automated SFTP process by exchanging SSH keys between UNIX servers.
  • Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
  • Involved in production n deployment activities, creation of the deployment guide for migration of the code to production, also prepared production run books.

Environment: Talend Data Integration 6.1/5.5.1, Talend Enterprise Big Data Edition 5.5.1, Talend Administrator Console, Oracle 11g, Hive, HDFS, Sqoop, Netezza, SQL Navigator, Toad, Control M, Putty, Winscp

Confidential, Tampa

Talend / ETL Developer

Responsibilities:

  • Closely worked with Data Architects in designing of tables and even involved in modifying technical Specifications.
  • Involved in the Huge development of a ETL Process Which Dynamically fetches the Source Data and Inserts the files in to S3.
  • Involved in the Development of copying Data from S3 to RedShift using the Talend Process.
  • Involved in writing custom copy command query’s and used the context extensively and implemented in tRedshift Row Component.
  • Extensively used t system component to push large sets of data to s3.
  • Developed a Talend Code for S3 Tagging in the Process of Moving data from source to S3
  • Involved in Extraction, Transformation and Loading of data.
  • Utilized Talend components like tS3Put, tS3Get, tS3File List, tRedshift Row, tRedshiftUnload, tRedshift BulkExec, tflowtoiterate, tSetGlobalVar.
  • Integrated Redshift SSO cluster with Talend
  • Involved in Integrating IAM Roles in Talend Components.
  • Used Components in Talend Like (tMap, Tfilelist, Tjava, Tlogrow, ToracleInput, ToracleOutput, tsendEmail etc, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tRedshiftrow, tsystem) .
  • Involved in Creating the Talend process which captures the Job Exception and Table Level Stats.
  • Involved in Performance Tuning of Data Ingestion Process to S3 in all the Job Servers
  • Responsible for the Scheduling and Deployment in TAC.
  • Involved in implementing the custom build schema converter to generate create table DDL for redshift.
  • Used T parallelize component in running the Jobs in Parallel.
  • Created Talend Jobs for Batch Processing of Data. Worked on Multi-Threading in Talend.

Environment: Talend 5.5/5.0, Oracle 11g, Teradata SQL Assistant, HDFS, MS SQL Server 2012/2008, PL/SQL, Agile Methodology, TOAD, ERwin, AIX, Shell Scripts, AutoSys, SVN.

Confidential, Bergen,IL

Talend Developer

Responsibilities:

  • Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirements.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components. Created and managed Source to Target mapping documents for all Facts and Dimension tables
  • Analyzing the source data to know the quality of data by using Talend Data Quality.
  • Involved in writing SQL Queries and used Joins to access Data from Oracle, and MySQL. Assisted in migrating the existing data center into the AWS environment.
  • Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer of project from development to testing environment and then to production environment.
  • Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
  • Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more)
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Experienced in using debug mode of Talend to debug a job to fix errors.
  • Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
  • Conducted JAD sessions with business users and SME's for better understanding of the reporting requirements.
  • Developed Talend jobs to populate the claims data to data warehouse - star schema.
  • Used Talend Admin Console Job conductor to schedule ETL Jobs on daily, weekly, monthly and yearly basis.
  • Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
  • Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.

Environment: Talend Enterprise Big Data Edition 5.1, Talend Administrator Console, MS SQL Server 2012/2008, Oracle 11g, Hive, HDFS, Sqoop, TOAD, UNIX Enterprise Platform for Data integration.

Confidential

Jr. Java Developer

Responsibilities:

  • Involved in Analysis, design and coding on Java and J2EE Environment.
  • Implemented strutsMVC framework.
  • Maintained the business standards and the requirements inEJB and deployed them on to Web Logic Application Server.
  • Designed, developed and implemented the business logic required for Security presentation controller.
  • Set up the deployment environment on Web Logic Developed system preferences UI screens using JSP and HTML.
  • Developed UI screens using Swing components like JLabel, JTable, JScrollPane, JButtons, JTextFields, etc.
  • Used JDBC to connect to Oracle database and get the results that are required.
  • Designed asynchronous messaging using Java Message Service (JMS).
  • Configured JMS Queues in Web Sphere Application Server.
  • Developed web Components using JSP, Servlets and Server side components using EJBunderJ2EE Environment.
  • Integrated a crystal report to the application, which will show all the details of the vouchers on the paper and used for future reference for the workers.0
  • Developed a web service, which will retrieve the details from the Database.
  • Prepared complete documentation of the Web Service developed.
  • Prepared a complete documentation of the project and test case documentation with sample inputs
  • Version control is done by using CVS.

Environment: Java (JDK1.4), J2EE, EJB, JSP, Servlets, JMS, XML, HTML, CSS, PL/SQL, Web Services, JNDI, Struts, UML, J2EE Design Patterns, JavaScript, WebLogic server,JDBC, Crystal reports XI, WindowsXP,SQLServer2000, Ant Scripts, LDAP, Java Beans, Hibernate, CVS.

We'd love your feedback!