We provide IT Staff Augmentation Services!

Senior Etl/talend Developer Resume

5.00/5 (Submit Your Rating)

Birmingham, AL

SUMMARY:

  • High skilled ETL/BI developer with 6+ years of experience in tools like Talend, Informatica Power Center (9.5.1, 9.x, 8.x, 7.x), Informatica IDQ/Informatica Data Explorer (IDE), Informatica MDM, and in developing &Administering ETL mappings.
  • Good experience in all phases of software development life cycle (SDLC) including system design, development, integration, testing, deployment and delivery of applications.
  • Experience in Big Data technologies like Hadoop/Map Reduce, Pig, Hive, and sqoop.
  • Experienced in integration of various data sources like Oracle 11g/10g/9i, IBM DB2, MS SQL Server, My SQL, Snowflake, Teradata, Netezza, XML files, Mainframe sources into staging area and different target databases.
  • Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType, tflowtoIterate, tAggregate, tSortRow, tFlowMeter, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tLogCatcher, tFilter, tGlobalmapetc.
  • Expertise in Informatica MDM Hub Match and Merge Rules, Batch Jobs and Batch Groups.
  • Experience using Informatica IDQ for qualifying the data content and MDM to filter duplicate data as well as to deploy the project as well as Meta Data Management.
  • Created Snowflake Schemas by normalizing the dimension tables as appropriate, and creating a Sub Dimension named Demographic as a subset to the Customer Dimension.
  • Experience in all stages of ETL - requirement gathering, designing and developing various mappings, unit testing, integration testing and regression testing.
  • Extensive Experience in designing and developing complex mappings applying various transformations such as Expression, Aggregator, Lookup, Source qualifier, Update strategy, Filter, Router, Sequence generator, Rank, Stored procedure, Joiner and Sorter.
  • Experienced in Data Analysis/Validation and Profiling based on the business and functional requirements of the Project.
  • Good experience in installation of Talend and Informatica Power Exchange.
  • Hands on experience in Pentaho Business Intelligence Server Studio.
  • Hands on experience in developing and monitoring SSIS/SSRS Packages and outstanding knowledge of high availability SQL Server solutions, including replication.
  • Proficient in the implementation of Data Cleanup procedures, Stored Procedures, Scripts and execution of Test plans for loading the data successfully into the various Target types (Relational and flat file).
  • Experienced in writing stored procedures/SQL scripts with various Teradata utilities like MLOAD, FASTLOAD, TPUMP, and FASTEXPORT.
  • Good experience in developing jobs for OLTP & OLAP databases.
  • Extensive experience in SQL scripting, shell scripting in Linux and windows based environments.
  • Experience in working on Enterprise Job scheduling tools like Autosys.
  • Good communication and interpersonal skills, ability to learn quickly, with good analytical reasoning and adaptive to new and challenging technological environment.
  • Strong Team working spirit, relationship management and presentation skills.

TECHNICAL SKILLS:

Operating Systems: Windows 2008/2007/2005/ NT/XP, UNIX, MS-DOS

ETL Tools: Talend6.x/5.x, Informatica Power Center 9.x/8.x, I MDM, SSIS, Ab-Initio.

Databases;: Oracle 12c/11g/10g, Snowflake, MS SQL Server 2012 /2008/2005, DB2 v8.1, Netezza, Teradata, Hbase.

Methodologies: Data Modeling - Logical Physical, Dimensional Modeling - Star / Snowflake

Programming Skills: C++, Eclipse, Shell Scripting (K-Shell, C-Shell), PL/SQL, Hadoop, Pig, Hive, Hcatalog, JAVA (Eclipse IDE and Net Beans IDE, JAVA Script, CSS.

Web services: Microsoft Azure.

Testing Tools: QTP, WinRunner, LoadRunner, Quality Center, Test Director, Clear test, Clear case.

PROFESSIONAL EXPERIENCE:

Confidential, Birmingham, AL

Senior ETL/Talend Developer

Responsibilities:

  • Participated in Client Meetings and gathered business requirements and analyzed them.
  • Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.
  • Design and Implement ETL processes to import data from and into Microsoft Azure.
  • Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
  • Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
  • Develop stored procedures/views in Snowflake and use in Talend for loading Dimensions and Facts.
  • Develop merge scripts to UPSERT data into Snowflake from an ETL source.
  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
  • Created complex mappings in Talend using tMap, tJoin, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
  • Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
  • Developed jobs to move inbound files to vendor server location based on monthly, weekly and daily frequency.
  • Implemented Change Data Capture technology in Talend in order to load deltas to a Data Warehouse.
  • Perform ETL using different sources like databases, flat files, and xml files.
  • Migrated Snowflake database to Windows Azure and updating the Connection Strings based on requirement.
  • Managed and reviewed Hadoop log files.
  • Wrote ETL jobs to read from web APIs using REST and HTTP calls and loaded into HDFS using java and Talend.
  • Shared responsibility for administration of Hadoop, Hive and Pig and Talend.
  • Tested raw data and executed performance scripts.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.

Environment: Talend6.x, HDFS, HBase, MapReduce, Java scripts, Snowflake, Eclipse, XML, JUNIT, Microsoft Azure, Hadoop, Apache Pig, Hive, Elastic Search, Web Services, Microsoft Office.

Confidential,

Raleigh, NC

Senior Talend Developer

Responsibilities:

  • Interact with Solution Architects and Business Analysts to gather requirements and update Solution Architect Document.

  • Engaged in various data warehouse and migration projects in understanding the business requirements and analyzing the design documents created by Data Architects and then generate a technical Low level Design document with the approach for the project.
  • Worked hands on Datastage 8.7 ETL migration to Talend Studio ETL process.
  • Performed analysis, design, development, Testing and deployment for Ingestion, Integration, provisioning using Agile Methodology.
  • Attended Daily Scrum meetings to provide update on the progress of the user stories Rally and to the Scrum Master and also to notify blocker and dependency if any.
  • Experienced in creating Generic schemas and creating Context Groups and Variables to run jobs against different environments like Dev, Test and Prod.
  • Created Talend Mappings to populate the data into dimensions and fact tables.
  • Broad design, development and testing experience with Talend Integration Suite and knowledge in Performance Tuning of mappings.
  • Experienced in Talend Data Integration, Talend Platform Setup on Windows and UNIX systems.
  • Created complex mappings in Talend 6.0.1/5.5 using tMap, tJoin, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
  • Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
  • Experience in using Repository Manager for Migration of Source code from Lower to higher environments.
  • Developed jobs to move inbound files to vendor server location based on monthly, weekly and daily frequency.
  • Implemented Change Data Capture technology in Talend in order to load deltas to a Data Warehouse.
  • Created jobs to perform record count validation and schema validation.
  • Created contexts to use the values throughout the process to pass from parent to child jobs and child to parent jobs.
  • Developed joblets that are reused in different processes in the flow.
  • Developed error logging module to capture both system errors and logical errors that contains Email notification and also moving files to error directories.
  • Provided the Production Support by running the jobs and fixing the bugs.
  • Experienced in using Talend database components, File components and processing components based up on requirements.
  • Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
  • Performed unit testing and also integration testing after the development and got the code reviewed.
  • Responsible for code migrations from Dev. to QA and production and providing operational instructions for deployments.

Environment: Talend Studio 6.0.1/5.5, Oracle 11i, XML files, Flat files, HL7 files, JSON, TWS, Hadoop 2.4.1, HDFS, Hive 0.13, HBase 0.94.21, Talend Administrator Console, IMS, Agile Methodology, HPSM.

Confidential. Richardson, TX

ETL Developer

Responsibilities:

  • Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.

  • Prepare high level design documents, detail design documents, business requirement documents, technical specifications, table level specs and test plan documents.
  • Extract data from legacy systems to staging area and then cleanse, homogenize, process and load into the data warehouse.
  • Use JIRA to create, implement and deploy ETL related stories.
  • Participate in daily scrum and bi weekly iteration planning as part of agile environment.
  • Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
  • Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
  • Develop stored procedures/views in SQL Server and use in Talend for loading Dimensions and Facts.
  • Develop MERGE scripts to UPSERT data into SQL Server from an ETL source.
  • Work on writing pig/hive/hcatalog scripts to process huge data files and export the data into FTP locations.
  • Perform ETL using different sources like databases, flat files, xml, and avro files.
  • Wrote ETL jobs to read from web APIs using REST and HTTP calls and loaded into HDFS using java and Talend.
  • Write custom java code in Talend to achieve complex functionality that is not available in ETL tool.
  • Used over 20+ Components in Talend Like (tMap, Tfilelist, Tjava, Tlogrow, ToracleInput, ToracleOutput, tsendEmail etc)
  • Used debugger and breakpoints to view transformations output and debug mappings.
  • Implemented Performance tuning in Mappings and Sessions by identifying the bottlenecks and Implemented effective transformation Logic.
  • Write shell scripts to encrypt and decrypt files containing Personal Identifiable Information (PII) and call the scripts from ETL tool.
  • Manage ETL jobs using Talend Administrator Console (TAC) in development and production environment and administer the Talend ETL tool in development and production environments.
  • Working on POC wrt BigData like loading the data into HDFS and creating Map Reduce Jobs
  • Develop ETL mappings for various Sources (.TXT, .CSV,.XML) and also load the data from these sources into relational tables with Talend Enterprise Edition.
  • Created specs and rpms and deployed shell/pig/hive scripts using continuous integrating with Jenkins.
  • Use GitHub and svn as version control for the code and implemented branching for different environments.

Environment: Talend 5.x, JIRA, Java, GIT, Sql Server, MySQL, Putty, XML, JUNIT, Hadoop, Jenkins, Apache Pig, Hive, Elastic Search, Web Services, Microsoft Office, SOAPUI/REST.

Confidential

Informatica Developer

Responsibilities:

  • Extensively used Informatica Power Center for extracting, transforming, and loading databases from sources including Oracle, DB2, and Flat files.

  • Participated in User meetings, gathering requirements.
  • Worked with business analysts, application developers, production teams and across.
  • Translating user inputs into ETL design docs.
  • Participate in Design Reviews of Data model and Informatica mapping design.
  • Functional units to identify business needs and discuss solution options.
  • Performed match/merge and ran match rules to check the effectiveness on data.
  • Created mappings using various Transformations like Aggregator, Expression, Filter, Router, Joiner, Lookup, Update strategy, Source Qualifier, Sequence generator, Stored Procedure and Normalizer.
  • Extensively used debugger for troubleshooting issues and checking session stats/logs.
  • Performed validation and testing of Informatica mapping against the pre-defined ETL design standards.
  • Implemented Performance tuning at database level and at Informatica level. Reduce load time by using partitions and concurrent sessions running at a time.
  • Involved in writing SQL, PL/SQL, Stored Procedures, Triggers and Packages in Data.
  • Warehouse environments that employ Oracle.
  • Developed Interfaces using UNIX Shell Scripts to automate the bulk load and Update.
  • Using shell scripts to automate the export data into flat files for backup and delete data.
  • From staging tables for the given time period.
  • Extensive experience in performance tuning, identifying bottlenecks and resolving to improve Performance at database level, Informatica mappings and session level.
  • Involved in the optimization of SQL queries which resulted in substantial performance.
  • Improvement for the conversion processes.
  • Involved in migration of objects in all phases (DEV, INT and PRD) of project and trained.
  • Used debugger in identifying bugs in existing mappings by analyzing data flow, evaluating Transformations.

Environment: Informatica Power Center 8.1, MS SQL Server, Oracle, Flat files, Erwin, UNIX Shell Scripting, Toad, Autosys, Windows 2003 Server.

We'd love your feedback!