We provide IT Staff Augmentation Services!

Talend Etl Lead Resume

5.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY

  • Experienced ETL Lead and Data Engineer well - versed in several Databases, Big Data technologies, NoSQL platforms and proficient in various extraction tools.
  • Extensive experience in handling large and complex Business Intelligence/Data Warehouse and Big Data Projects. Adept in design, development, deployment and delivering scalable, efficient and elegant enterprise solutions.
  • 9 plus years of development and design of ETL methodology for supporting data transformations and processing, in a corporate wide ETL Solution using Informatica PowerCenter
  • 4 plus years of experience wif Talend 6.4/6.3/6.2/6.1
  • 9 plus years of SQL wif Oracle 7/8i/9i/10g/11g (SQL, PL/SQL, Stored Procedures, Functions), MS-Access 2002/2003, MS SQL Server 2000/7/12, Teradata, Netezza, DB2, SQL*Plus, SQL*Loader, TOAD. MS SQL, Sybase,
  • 4 plus years of experience wif BigData Hadoop using Hive, Sqoop, PIG, Flume and Spark on Hortonworks and Cloudera framework
  • 4 plus years of SQL experience wif MongoDB, HBase, Cassandra
  • 2 plus years of experience wif cloud-based ecosystem like AWS Amazon Redshift/S3
  • Expert understanding of ETL techniques and best practices to handle extremely large volume of data.
  • Experience in both relational and dimensional data modeling. Good noledge on Data Warehousing Architecture & Designing Star Schema, Snow flake Schema.
  • Experience in Big Data technologies and Hadoop ecosystem components like Spark, HDFS, MapReduce, Pig, Hive, YARN, Sqoop,, Flume, Kafka and NoSQL systems like HBase, Cassandra
  • Experienced in creating complex data ingestion pipelines, data transformations, data management and data governance at enterprise level.
  • Experience in development of logical and physical data models.
  • Experience in agile development framework using scrum and Kanban.
  • Experience of data analysis and modeling at the conceptual, logical and physical levels.
  • Experience in programming language like C++, Java and Python.
  • Experience wif data management standards.
  • Extensive noledge of Bill Inmon and Ralph Kimball methodologies, business process re-engineering, and database design methodologies and tools
  • Proficient wif database architecture tools and programs
  • Expert in data analysis, migration, integration, and data warehousing concepts
  • Extensive experience in Data Modeling, Relations Database Architecture and implementation
  • Performed Data Profiling / Data Cleansing of several disparate data sources fixed width, delimited, single/ multi source flat files and data based on MS SQL Server, DB2 UDB, Sybase and developed Data Migration strategies
  • Experience in integration of various data sources like Oracle, Sybase, DB2, SQL Server, Flat Files, XML files.
  • Expertise in data integration development, testing and support using Informatica 9.x PowerCenter and PowerExchange, DX, IDQ data profiling/data cleansing, B2B, and a much rare skill in the market DT, MDM exposure
  • Expertise in support and quality validation thru test cases for all stages Unit and Integration testing
  • Expertise in system integration using the Informatica for extraction, loading and transformation of data from systems like Oracle Financials, Siebel, Hyperion and other external third parties for loading the data into data warehouse
  • Professional development experience in full Software Development Life Cycle (SDLC) for Data Warehousing, Data Integration & DSS using (ETL) Informatica Power Center
  • Strong academic background in software engineering and familiarity wif various SDLC models like RUP & Agile methodologies, Iterative, spiral and Modified-V
  • Experience in Production Support to tier 2 specialist
  • An excellent team member wif:
  • An ability to perform individually as well as ability to work in-group
  • Excellent problem solving wif good analytical and programming skills
  • Good time management skills and strong communication skills
  • Quick leaner and initiative to learn new technology and tools quickly

TECHNICAL SKILLS

Big Data Ecosystem: Hadoop, HDFS, MapReduce, YARN, Hive, Hue, Pig, Sqoop, Spark, Tez, HBase, Cassandra, Hortonworks Data Platform 2.3/Cloudera Data Platform

ETL Tools: Informatica Power Center 9.6/9.5/9.1/8.5/8.1.1/8.0/7.1/7.0/6.2/6.1/5.1.2/5.1.1/4.7 Talend 6.4/6.3/6.2/6.1 /.60/5.6/5.5/5.3Informatica BDE, Informatica B2B, IDQ, SSIS 2008

Cloud Infrastructure: AWS Redshift,S3,EC2

Programming Languages: C, C++, Java, Unix Shell Scripting, PL/SQL, Pig Latin, Python, HiveQL, ANSI SQLHTML, CSS, Java Script, C++, Java, HTML, PL/SQL, SQL, PHP, AJAX, Visual Basic

Scripting Languages: Shell Scripting, Java script.

Database: MongoDB, Cassandra, HBase, Redshift, Oracle 11g/10g/9i, MySQL, Microsoft SQL Server 2012, Teradata. Oracle 7/8i/9i/10g/11g (SQL, PL/SQL, Stored Procedures, Functions), MS-Access 2002/2003, MS SQL Server 2000/7/12, Teradata, Netezza, SQL*Plus, SQL*Loader, TOAD. MS SQL, Sybase

Data Modelling Tool: E-R Modelling, ERWin, PowerDesigner

Visualization: Tableau, QlikView

PROFESSIONAL EXPERIENCE

Confidential, Chicago, IL

Talend ETL lead

Responsibilities:

  • Interacted wif various business people in External Vendors side and gathered the business requirements and translated them into technical specifications
  • Work wif Subject Matter Experts to clarify open and scenario questions
  • Worked wif team, lead developers, Interfaced wif business analysts, coordinated wif management and understand the end user experience
  • Worked wif data analysts to implement ETL specification documents, shell scripts, and stored procedures to meet business requirements based on the Hadeoop ecosystem based technologies
  • Involved in deciding data loading strategies in data lake
  • Involved in HBase table design and Talend development standards
  • Import files received from GE on HDFS
  • Loaded unstructured data into HDFS
  • Implemented Partitioning, Dynamic Partition, and Buckets in Hive for efficient data access
  • Worked on Talend BigData HDFS components such tHDFSConnection, tHDFSPut, tHDFSInput, tHDFSOutput tocreate tables in Hive for Transportation Finance portfolio
  • Worked on Talend BigData sqoop components such as tSqoopExport, tSqoopImport, tSqoopIm port AllTables to load tables in Hive from Netezza
  • Integrated Hadoop jobs wif Kerberos security for Hive
  • Worked on TAC to publish jobs in Nexus, scheduling them, and actively monitoing in Activity Monitoring Console and checking logs
  • Maintain, tune, and support Talend ETLs on a day to day basis and ensure high availability
  • Integrated Talend jobs to AWS Redshift for future cloud integration
  • Provided support for production and development environments and developed product support documentation

Confidential, Libertyville, IL

Talend ETL developer

Responsibilities:

  • Interacted wif various business people in External Vendors side and gathered the business requirements and translated them into technical specifications
  • Worked wif data analysts to implement ETL specification documents, shell scripts, and stored procedures to meet business requirements based on the Hadeoop ecosystem based technologies
  • Designed and developed ETL jobs to ingest data into Hadoop Ecosystem
  • Worked on Talend components like tReplace, tMap, tSort, and tFilterColumn, tFilterRow, etc.
  • Used datatabase components such tMSSQLInput, tMSSQLRow, etc. to connect FIS servers and source data
  • Worked on Talend BigData HDFS components such tHDFSConnection, tHDFSPut, tHDFSInput, tHDFSOutput to place FIS flat files on HDFS
  • Worked on Talend BigData Hive copmponents such tHiveConnection, tHiveCreateTable, tHiveLoad, tHiveInput, tHiveClose to load some of the flat files into Hive as external Hive tables
  • Worked on Talend BigData sqoop components such as tSqoopExport, tSqoopImport, tSqoopImportAllTables to load Oracle data mart tables in to Hive
  • Performance tune Talend ETL jobs using parallelization
  • Integrated Talend jobs to AWS S3 for cloud integration
  • Created triggers to kick of Talend ETLs
  • Provided support for production and development environments and developed product support documentation

Confidential, Little Rock, AR

Informatica ETL Lead

Responsibilities:

  • Interacted wif various business people in External Vendors side and gathered the business requirements and translated them into technical specifications.
  • Worked alongside data architects to create full netezza datawarehouse consisting of dimensions, facts, views, and snapshots required for OBIEE reporting
  • Interacted wif Netezza utility tools such as Aginity workbench and worked on NZSQL and NZload to perform data loads into ODS from Oracle, DB2 and FlatFiles
  • Experienced working wif team, lead developers, Interfaced wif business analysts, coordinated wif management and understand the end user experience.
  • Worked wif data analysts to implement Informatica mappings and workflows, shell scripts and stored procedures to meet business requirements.
  • Documented business requirements, discussed issues to be resolved and translated user input into ETL design documents.
  • Responsible for creating interfaces using different types of mappings using various transformations like Expression, filter, router, Aggregator, Look up, Joiner, Stored Procedure, Update Strategy, Etc.
  • Extensively worked on Mapping Variables, Mapping Parameters, Workflow Variables and Session Parameters for the delta process to extract only the additional data added during dat period.
  • Used debugger in identifying bugs in existing mappings by analyzing data flow, evaluating transformations.
  • Designed and Developed the Informatica workflows/sessions to extract, transform and load the data into Target.
  • Worked on mappings to dynamically generate parameter files used by other mappings.
  • Extensive performance tuning by determining bottlenecks at various points like targets, sources, mappings and sessions.
  • Implemented the batch process to track historical and incremental loads for each processes defined
  • Involved in performance tuning of the ETL process by addressing various performance issues at the extraction and transformation stages.
  • Documented the mappings used in ETL processes including the Unit testing and Technical document of the mappings for future reference.
  • Involved in ETL process from development to testing and production environments.
  • Extensively worked wif Unix Shell scripting to validate and verify the data in the flat files generated by the ETL process.
  • Wrote UNIX shell scripts as part of the ETL process to compare control totals, automate the process of loading, pulling and pushing data from and to different servers.
  • Created the UNIX shell scripts to send out E-mails on the success of the process indicating the destination folder where the files are available.
  • Developed post-session and pre-session shell scripts for tasks like merging flat files after creating and deleting temporary files, changing the file name to reflect the file generated date.
  • Involved in migration of Informatica mapping from Development to Production environment.
  • Coordinated wif the informatica administration team during deployments.
  • Provide production support and noledge transfer to the operations team

Confidential, Waukegan, IL

Senior Informatica Lead/Developer

Responsibilities:

  • Interact directly wif business users and performed business discovery and define functional specifications
  • Work alongside data modeler to prepare Sales and Marketing Data Warehouse and convert the data mart from Logical design to Physical design, defined data types, Constraints, Indexes, and generated schema in the Teradata Database
  • Preparation of technical specification for the development of Informatica Extraction, Transformation and Loading (ETL) mappings to load data into various tables in Teradata Data Marts and defining ETL standards
  • Implemented the batch process to track historical and incremental loads for each processes defined
  • Performance tuning by removing lookups and performing a join wif source query, performing session partitions over index, persistent cache memory, and index cache
  • Created various UNIX Shell Scripts for pre and post-sessions and embedded automated load success/failures wif notification through email
  • Project migration across different environments (Dev, QA, Test, and Production)
  • Prepared load plan for all the SMDW jobs and automated them prior to production through Autosys
  • Provide production support and noledge transfer to the operations team
  • Involved in different phases of the project life cycle for implementation from DB2 to Enterprise Teradata Data warehouse
  • Involved in a follow on phase of the Retail Expansion project non as Incentive Compensation and Sales Crediting currently on going

Confidential, Elk Grove Village, IL

Informatica Developer

Responsibilities:

  • Designed a high level ETL architecture design to integrate data amongst three different sources(DB2, FlatFiles, Oracle) to data mart
  • Preparation of technical specification for the development of Informatica Extraction, Transformation and Loading (ETL) mappings to load data into various tables in Data Marts and defining ETL standards
  • Extracted Source data from various database including SQL Server, Oracle and DB2
  • Designed and implemented ETL processes to extract historical and latest Contract, Transaction, Benefit, Producer, Demographics, Loans/Rider information from ODS tables
  • Designed and developed complex join, lookup transformations (business rules) to generate consolidated data
  • Prepared the schedules for the historical load run and incremental load runs
  • Developed complex mappings using connected/unconnected lookups, union, aggregator, filter, sorter, router, joiner, sequence generator, update strategy and reusable transformations to extract data from ODS tables
  • Prepared Workflow process to run sessions based on load dependency
  • Migrated mappings, sessions and workflows from Development to Test environment
  • Tuned performance of Informatica sessions for large data files by increasing block size, data cache size and, sequence buffer length
  • Automated load run on Informatica sessions through UNIX kron, PL/SQL scripts and implemented pre and post-session scripts, also automated load failures wif successful notification through email
  • Created TEMPeffective Test Cases and did Unit and Integration Testing to ensure the successful execution of data loading process
  • Provided support for production and development environments and developed product support documentation
  • Furnished support to dedicated tier 2 production specialist

We'd love your feedback!