We provide IT Staff Augmentation Services!

Talend-big Data Integration Developer Resume

Chicago, IL

SUMMARY:

  • Around 8+ years IT experience in Design, Development, Maintenance, Enhancements and Production Support of which includes Data warehousing and Legacy applications using ETL tools like Talend /Informatica/SAS.
  • Experience in all aspects of Analysis, Design, Testing, Development, Implementation and Support of Relational Database (OLTP), Data Warehousing Systems (OLAP) and Data Marts in various domains.
  • Experience in working with Data Warehousing Concepts like OLAP, OLTP, Star Schema, Snow Flake Schema, Logical Data Modelling, Physical Modelling and Dimension Data Modelling.
  • Utilized tStat Catcher, tDie, tLog Row to create a generic job to store processing stats.
  • Involved in extracting user's data from various data sources into Hadoop Distributed File Systems (HDFS)
  • Experience in adhering software methodologies like waterfall, Agile.
  • Excellent understanding and knowledge of NOSQL databases.
  • Experienced in creating Generic schemas and creating Context Groups and Variables to run jobs against different environments like Dev, Test and Prod.
  • Thorough knowledge of addressing Performance Issues and Involved in query tuning, Index tuning, Data profiling and other database related activities.
  • Extensively created mappings in Talend using tMap, tJoin, tReplicate, tParallelize, tJava, tJava row, tDie, tAggregate Row, tWarn, tLog Catcher, tFilter, tGlobal map etc.
  • Wrote Hive and Pig queries for data analysis to meet the business requirements.
  • Experienced in scheduling Talend jobs using Talend Administration Console (TAC)
  • Experience with Talend DI Installation, Administration and development for data warehouse and application integration.
  • Tracking Daily data load, Monthly data extracts and send to client for their verification.
  • Strong experience in designing and developing Business Intelligence solutions in Data Warehousing using ETL Tools.
  • Excellent understanding and best practice of Data Warehousing Concepts, involved in Full Development life cycle of Data Warehousing.
  • Experienced in working with different data sources like Flat files, Spreadsheet files, log files and Databases.
  • Hands - on experience across all stages of Software Development Life Cycle (SDLC) including business requirement analysis, data mapping, build, unit testing, systems integration and user acceptance testing.
  • Excellent interpersonal and communication skills, and is experienced in working with senior level managers, business people and developers across multiple discipline

SKILL:

Databases: Teradata,Oracle12c/11x/10g, DB2, Microsoft SQL Server, Hive, Sybase,Redshift DB

Programming: T-SQL, PL/SQL, HTML, XML.

Environment: Windows, UNIX & Linux

Scripting: Korn shell script & Windows batch scripting, JavaScript

Languages: SQL, Java, .Net, BASE SAS.

Other Tools: SQL Navigator, Putty, MS-Office.

PROFESSIONAL EXPERIENCE:

Talend-Big Data Integration Developer

Confidential, Chicago, IL

Responsibilities:

  • Responsible to build the jobs by looking to the ETL Specification documents.Worked closely with Business Analysts to review the business specifications of the project and to gather the ETL requirements.
  • Closely worked with Data Architects in designing of tables and even involved in modifying technical Specifications.
  • Involved in Extraction, Transformation and Loading of data.
  • Experienced in Creating the Context load using the parameter variables.
  • Worked on Batch Jobs and scheduling,
  • Used Components in Talend Like (tMap, Tfilelist, Tjava, Tlogrow, ToracleInput, ToracleOutput, tsendEmail etc)
  • Extensively Used components like tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator tfileList etc.
  • Utilized Big Data components like tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport, tSqoopExport, tHiveOutput, tHiveRow, tHiveConnection.
  • Data ingestion with different data sources and load into redshift.
  • Developed jobs to send and read data from AWS S3 buckets using components like tS3Connection, tS3BucketExist, tS3Get, tS3Put.
  • Experienced in implementing the project end to end Solutions.
  • Involved in Data Extraction from Flat files and XML files using Talend and Java as Backend Language.
  • Used Talend to load the data into our warehouse systems
  • Used debugger and breakpoints to view transformations output and debug mappings.
  • Develop ETL mappings for various Sources (.TXT, .CSV, XML) and load the data from these sources into relational tables with Talend Enterprise Edition.
  • Worked on Global Context variables, Context variables, and extensively used over 70+components in Talend to create jobs.
  • Worked on Error handling techniques and tuning the ETL flow for better performance.
  • Worked Extensively TAC (Admin Console), where we Schedule Jobs in Job Conductor.
  • Migrated the code and release documents from DEV to QA (UAT) and to Production.
  • Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type-2.

Environment: Talend Enterprise Big Data Edition 6.5, Talend Administrator Console, MS SQL Server, Oracle 11g, HDFS, Hive, Amazon Redshift.

Talend Developer

Confidential, DesMoines, IA

Responsibilities:

  • Participated in all phases of development life-cycle with extensive involvement in the definition and design meetings, functional and technical walkthroughs.
  • Created Talend jobs to copy the files from one server to another and utilized Talend FTP components
  • Created and managed Source to Target mapping documents for all Facts and Dimension tables
  • Used ETL methodologies and best practices to create Talend ETL jobs.
  • Followed and enhanced programming and naming standards.
  • Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
  • Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
  • Used Talend most used components (tMap, tDie, tConvertType, tLogCatcher, tRowGenerator, tHashInput & tHashOutput and many more).
  • Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Created Implicit, local and global Context variables in the job.
  • Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
  • Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
  • Involved in production n deployment activities, creation of the deployment guide for migration of the code to production, also prepared production run books.

Environment: Talend Data Integration 6.1/5.5.1,, Talend Administrator Console, Oracle 11g, Hive, HDFS, SQL Navigator, Toad, Putty, Winscp.

SQL/ETL Inforamtica Developer

Confidential

Responsibilities:

  • Coordinated with Business Users for requirement gathering, business analysis to understand the business requirement and to prepare Technical Specification documents (TSD) to code ETL Mappings for new requirement changes.
  • Analyze and create low level design document (LLD) and mapping document.
  • Performed analysis, design, development, Testing and deployment for Ingestion, Integration, provisioning using Agile Methodology.
  • Created jobs to perform record count validation and schema validation.
  • Developed jobs that are used for ETL purposes.
  • Administer all requests and analyze issues and provide efficient resolution for same.
  • Design all program specifications and perform required tests in same.
  • Design all programs and systems and associate documentation for same.
  • Prepare all program and system implementation for all informatics programs.
  • Monitor all production issues and inquiries and provide efficient resolution for same.
  • Performed unit testing and integration testing after the development and got the code reviewed.
  • Designed and Developed Oracle PL/SQL and UNIX Shell Scripts, Data Import/Export.
  • Identified and fixed the Bottle Necks and tuned the Mappings and Sessions for improving perform.
  • Tuned both ETL process as well as Databases.

Environment: Talend Studio 5.2.2 /5.6, Talend Studio Big Data Platform, HBase, XML files, Flat files, HL7 files, HDFS, Hive, Oracle 11g, Business Objects, UNIX, WinSCP, Clear Case, Clear Quest, Erwin, PL/SQL, Toad, Windows 7 Pro, TFS, JIRA.

SAS Developer/SQL/ETL Informatica

Confidential

Responsibilities:

  • Plan Design and implement application database code objects, such as stored procedure and views.
  • Worked on different ETL tools like Informatica, BASE SAS fot the data movement.
  • Build and Maintain SQL scripts, indexes and complex queries for data analysis and extraction.
  • Created complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
  • Created contexts to use the values throughout the process to pass from parent to child jobs and child to parent jobs.
  • Developed joblets that are reused in different processes in the flow.
  • Created, modified/Updates SAS code to in corporate changes in new releases of the product.
  • Produced ad hoc reports in SAS using .txt files upon request for in-depth analysis.
  • Data manipulation by merging, appending and sorting datasets.
  • Extensive use of PROC SQL to create subsets of data for further analysis.
  • Extensive use of DATA STEP, PROC MEANS, PROC SQL, MACROS, PROC FREQ, PROC TABULATE, DATA NULL, PROC MEANS, PROC SUMMARY, RETAIN.
  • Generated output files in various formats like text, MS Excel.
  • Worked on big project like large group Migration, where the data has to be moved from Staging data base to the Target Production data.
  • Validated the data and worked on large number of defect fixes to satisfy the business for the required exact formats for the data.
  • Figured out the problem to fix the defects, if the data is not reflected in the target as expected like Code changes, timing issues (when running the Cycles) etc.
  • Worked on CDR (Change Data Requests), Installation Documents, and attended the meetings to promote the code to the production (live).

Environment: XML files, Flat files, HL7 files, HDFS, Hive, Oracle 11g, Business Objects, UNIX, WinSCP, Clear Case, Clear Quest, Erwin, PL/SQL, Toad, Windows 7 Pro, TFS, JIRA.

Hire Now