We provide IT Staff Augmentation Services!

Sr. Etl/talend Developer Resume

5.00/5 (Submit Your Rating)

Chicago, IL

SUMMARY:

  • High skilled ETL/BI developer with 8+ years of experience in tools like Talend, Informatica PowerCenter (10.x, 9.x, 8.x, 7.x), Informatica IDQ/Informatica Data Explorer (IDE), Informatica MDM, and in developing & Administering ETL mappings.
  • Good experience in all phases of software development life cycle (SDLC) including system design, development, integration, testing, deployment and delivery of applications.
  • Extracted the data using Pentaho Data Integration Designer Kettle from the flat files and other RDBMS databases like SQL Server 2005/2008 and Oracle 10g into staging area and populated onto Data warehouse.
  • Used Kettle designer to design and create Pentaho Transformations and jobs.
  • Created source and target connections in Talend Integration cloud to pull files from mainframe and place it into ESB Server.
  • Experience in Big Data technologies like Hadoop/Map Reduce, Pig, Hive, sqoop.
  • Experienced in integration of various data sources like Oracle 11g/10g/9i, IBM DB2, MS SQL Server, My SQL, Snowflake, Teradata, Netezza, XML files, Mainframe sources into staging area and different target databases.
  • Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType,, tflowtoIterate, tAggregate, tSortRow, tFlowMeter, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tLogCatcher, tFilter, tGlobalmapetc.
  • Expertise in Informatica MDM Hub Match and Merge Rules, Batch Jobs and Batch Groups.
  • Experience using Informatica IDQ for qualifying the data content and MDM to filter duplicate data as well as to deploy the project as well as Meta Data Management.
  • Created Snowflake Schemas by normalizing the dimension tables as appropriate, and creating a Sub Dimension named Demographic as a subset to the Customer Dimension.
  • Experience in all stages of ETL - requirement gathering, designing and developing various mappings, unit testing, integration testing and regression testing.
  • Extensive Experience in designing and developing complex mappings applying various transformations such as Expression, Aggregator, Lookup, Source qualifier, Update strategy, Filter, Router, Sequence generator, Rank, Stored procedure, Joiner and Sorter.
  • Experienced in Data Analysis/Validation and Profiling based on the business and functional requirements of the Project.
  • Good experience in installation of Talend and Informatica Power Exchange.
  • Hands on experience in Pentaho Business Intelligence ServerStudio.
  • Hands on experience in developing and monitoring SSIS/SSRS Packages and outstanding knowledge of high availability SQL Server solutions, including replication.
  • Proficient in the implementation of Data Cleanup procedures, Stored Procedures, Scripts and execution of Test plans for loading the data successfully into the various Target types (Relational and flat file)
  • Experienced in writing stored procedures/SQL scripts with various Teradata utilities like MLOAD, FASTLOAD, TPUMP, FASTEXPORT.
  • Good experience in developing jobs for OLTP & OLAP databases.
  • Extensive experience in SQL scripting, shell scripting in Linux and windows based environments.
  • Experience on Data Modeling Tools Erwin, Visio, Sybase Power designer.
  • Experience in working on Enterprise Job scheduling tools like Autosys.

PROFESSIONAL EXPERIENCE:

Sr. ETL/Talend Developer

Confidential, Chicago, IL

Responsibilities:

  • Participated in Requirement gathering, Business Analysis, User meetings and translating user inputs into ETL mapping documents.
  • Worked closely with Business Analysts to review the business specifications of the project and also to gathered ETL requirements.
  • Implemented new users, projects, tasks within multiple different environments of TAC.
  • Created complex mappings in Talend using tMap, tJoin, tXML, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
  • Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
  • Developed complex Talend ETL jobs to migrate the data from flat files to database.
  • Pulled files from mainframe into Talend execution server using multiple ftp components.
  • Developed complex Talend ETL jobs to migrate the data from flat files to database.
  • Developed Talend ESB services and deployed them on ESB servers on different instances.
  • Worked on Talend Data Integration cloud components like tfileinputpositional, tfilelist, tActioninput, tActionoutput, tActionReject, tActionLog, tActionFailure, jClou dLog, tsplitrow, tcontextLoad, tFileArchive, tFileDelete, tFixedFlowInput.
  • Worked with Data mapping team to understand the source to target mapping rules.
  • Created source and target connections in Talend Integration cloud to pull files from mainframe and place it into ESB Server.
  • Published 200 plus jobs into Talend integration cloud.
  • Worked on dynamic schema in Integration Actions enable web users to define new columns on the run when they create Flows.
  • Created test cases for Integration Actions.
  • Integration Actions are published to Talend Integration Cloud for Cloud users to build Flows.
  • Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer ofproject from development to testing environment and then to production environment.
  • Performed Unit testing and System testing to validate data loads in the target.

ENVIRONMENT: Talend 6.3.1, Talend integration cloud, Java scripts, SQL Server, GIT, Eclipse, XML, JUNIT, xmlfiles, flat files, mainframe files, ssis, microsoft office, Web Services.

Sr. ETL/Talend Developer

Confidential, Denver, CO

Responsibilities:

  • Participated in Client Meetings and gathered business requirements and analyzed them.
  • Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.
  • Design and Implement ETL processes to import data from and into Microsoft Azure.
  • Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
  • Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
  • Develop stored procedures/views in Snowflake and use in Talend for loading Dimensions and Facts.
  • Develop merge scripts to UPSERT data into Snowflake from an ETL source.
  • Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW tables and historical metrics.
  • Created complex mappings in Talend using tMap, tJoin, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
  • Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
  • Developed jobs to move inbound files to vendor server location based on monthly, weekly and daily frequency.
  • Implemented Change Data Capture technology in Talend in order to load deltas to a Data Warehouse.
  • Perform ETL using different sources like databases, flat files, xml files.
  • Migrated Snowflake database to Windows Azure and updating the Connection Strings based on requirement.
  • Managed and reviewed Hadoop log files.
  • Wrote ETL jobs to read from web apis using REST and HTTP calls and loaded into HDFS using java and Talend.
  • Shared responsibility for administration of Hadoop, Hive and Pig and Talend.
  • Tested raw data and executed performance scripts.
  • Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.

ENVIRONMENT: Talend6.3.1, Pentaho, HDFS, HBase, MapReduce, Java scripts, Snowflake, Eclipse, XML, JUNIT, Microsoft Azure, Hadoop, Apache Pig, Hive, Elastic Search, Web Services, Pentaho kettleMicrosoft Office.

Sr. ETL/Talend Developer

Confidential, Birmingham, AL

Responsibilities:

  • Involved in building the ETL architecture and Source to Target mapping to load data into Data warehouse.
  • Performed data manipulations using various Talend components like tMap, tJavaRow, tjava, tOracleRow, tOracleInput, tOracleOutput, tMSSQLInput and many more.
  • Designed and customized data models for Data warehouse supporting data from multiple sources on real time.
  • Designed ETL process using Talend Tool to load from Sources to Targets through data Transformations.
  • Extensive experience on Pentaho designer, Pentaho kettle, Penatho BI server, BIRT report designer
  • Developed advanced Oracle stored procedures and handled SQL performance tuning.
  • Involved in creating the mapping documents with the transformation logic for implementing few enhancements to the existing system.
  • Monitored and supported the Talend jobs scheduled through Talend Admin Center (TAC)
  • Developed the Talend mappings using various transformations, Sessions and Workflows. Teradata was the target database, Source database is a combination of Flat files, Oracle tables, Excel files and Teradata database.
  • Loaded data in to Teradata Target tables using Teradata utilities (FastLoad, MultiLoad, and FastExport) Queried the Target database using Teradata SQL and BTEQ for validation.
  • Used Talend to Extract, Transform and Load data into Netezza Data Warehouse from various sources like Oracle and flat files.
  • Created connection to databases like SQL Server, oracle, Netezza and application connections.
  • Created mapping documents to outline data flow from sources to targets.
  • Prepare the Talend job level LLD documents and working with the modeling team to understand the Big Data Hive table structure and physical design.
  • Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.
  • Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend.
  • Maintained stored definitions, transformation rules and targets definitions using Informatica repository Manager.
  • Used various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings.
  • Developed mapping parameters and variables to support SQL override.
  • Developed Talend ESB services and deployed them on ESB servers on different instances.
  • Created mapplets & reusable transformations to use them in different mappings.
  • Developed mappings to load into staging tables and then to Dimensions and Facts.
  • Developed the Talend jobs and make sure to load the data into HIVE tables & HDFS files and develop the Talend jobs to integrate with Teradata system from HIVE tables.
  • Worked on different tasks in Workflows like sessions, events raise, event wait, decision, e-mail, command, worklets, Assignment, Timer and scheduling of the workflow.
  • Unit testing, code reviewing, moving in UAT and PROD.
  • Designed the Talend ETL flow to load the data into hive tables and create the Talend jobs to load the data into Oracle and Hive tables.
  • Migrated the code into Testing and supported QA team and UAT (User)
  • Created detailed Unit Test Document with all possible Test cases/Scripts.
  • Working with high volume of data and tracking the performance analysis on Talend job runs and session.
  • Conducted code reviews developed by my team mates before moving the code into QA.
  • Experience in Batch scripting on windows, Windows 32 bit commands, Quoting, Escaping.
  • Used Talend reusable components like routines, context variable and globalMap variables.
  • Provided support to develop the entire warehouse architecture and plan the ETL process.
  • Knowledge on Teradata Utility scripts like FastLoad, MultiLoad to load data from various source systems to Teradata.
  • Modified existing mappings for enhancements of new business requirements.
  • Prepared migration document to move the mappings from development to testing and then to production repositories.
  • Configured the hive tables to load the profitability system in Talend ETL Repository and create the Hadoop connection for HDFS cluster in Talend ETL repository.
  • Works as a fully contributing team member, under broad guidance with independent planning & execution responsibilities.

Environment: Talend, Hive, Pig, Hadoop, Sqoop, PL/SQL, Oracle 12c/11g/, Erwin, Autosys, SQL Server 2012, Teradata, Netezza, Sybase, SSIS, UNIX, Profiles, Role hierarchy, Workflow & Approval processes, Data Loader, Reports, Custom Objects, Custom Tabs, Data Management, Lead processes, Record types.

ETL Developer

Confidential, Richardson, TX

Responsibilities:

  • Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.
  • Prepare high level design documents, detail design documents, business requirement documents, technical specifications, table level specs and test plan documents.
  • Extract data from legacy systems to staging area and then cleanse, homogenize, process and load into the data warehouse.
  • Use JIRA to create, implement and deploy ETL related stories.
  • Participate in daily scrum and bi weekly iteration planning as part of agile environment.
  • Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
  • Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
  • Develop stored procedures/views in SQL Server and use in Talend for loading Dimensions and Facts.
  • Develop MERGE scripts to UPSERT data into SQL Server from an ETL source.
  • Work on writing pig/hive/hcatalog scripts to process huge data files and export the data into FTP locations.
  • Perform ETL using different sources like databases, flat files, xml, avro files.
  • Wrote ETL jobs to read from web apis using REST and HTTP calls and loaded into HDFS using java and Talend.
  • Write custom java code in Talend to achieve complex functionality that is not available in ETL tool.
  • Write shell scripts to encrypt and decrypt files containing Personal Indentifiable Information (PII) and call the scripts from ETL tool.
  • Manage ETL jobs using Talend Administrator Console (TAC) in development and production environment and administer the Talend ETL tool in development and production environments.
  • Created specs and rpms and deployed shell/pig/hive scripts using continuous integrating with Jenkins.
  • Use GitHub and svn as version control for the code and implemented branching for different environments.
  • ENVIRONMENT: Talend 5.x, JIRA, Java, GIT, Sql Server, MySQL, Putty, XML, JUNIT, Hadoop, Jenkins, Apache Pig, Hive, Elastic Search, Web Services, Microsoft Office, SOAPUI/REST.
  • ETL Developer
  • Western Union Englewood, California - Englewood, CA
  • Responsibilities:
  • Responsible for writing requirements for automation of online contract management tool.
  • Involved in designing MDM data model and created base objects, mappings, defined trust setting for sources, customized user exists, customized IDD applications.
  • Performed match/merge and ran match rules to check the effectiveness of MDM process on data
  • Involved in designing, Implementing the Inbound and Outbound interfaces to send data to and from the MDM System to the sources.
  • Involved in the UAT testing of the Items Data Mart price files tool automation project.
  • Responsible for handling Supply Chain Management Data Warehouse and related data base and online solutions.
  • Involved in maintenance of reporting tools and processes utilizing Data Warehouse data.
  • Responsible for generating pricing reports from Finance & Procurement Data Warehouse and Data Mart using Business Objects.
  • Responsible for managing the online contract management repository and items data mart pricing information.
  • Work on formulating the ETL rules using Informatica and AbInitio.
  • Used Informatica Power Center, tool for migrating data from various OLTP databases to the data mart.
  • Worked with different sources like Oracle, flat files, XML files, DB2, MS SQL Server.
  • Designed and developed jobs and transformations to load data into dimensions and facts.
  • Worked on Data Analytics/Visualization Project using informatica (ETL and Data Visualization Suite)
  • Extracted data from Sales department to flat files and load the data to the target database.
  • Developed complex mappings using Informatica to load Dimension & Fact tables as per STAR schema techniques.
  • Contribute to the entire implementation process including driving the definition of improvements based on business need and architectural improvements.
  • Extracted data from sources like fixed width and Delimited Flat files transformed the data according the business requirement and then loaded into Target Data mart.
  • Created various Informatica mappings and mapplets to load the data mart. The mappings involved extensive use of transformations like Aggregator, Filter, Router, Expression, Joiner, Sequence generator.
  • Configured the mappings to handle the updates to preserve the existing records using Update Strategy Transformation (SCD Type-2)
  • Used IDQ for generation of various reports and graphs based on data used for reporting.
  • Used Debugger to identify the errors in the mappings and fix them.
  • Performed unit testing to validate mappings and populate the database.
  • Used TOAD and SQL Plus to write queries and interact with Oracle database.
  • Implemented several different tasks (Session, Assignment, Command, Decision, Timer, Email, Event-Raise, Event-Wait, Control) in the workflow.
  • Involved in writing UNIX scripts and used them to automate the scheduling process.
  • Error Handling handled in Informatica Mappings & sessions and also through validation script.
  • Co-ordinated with DBA Team, Functional Team and offshore teams.
  • Automation of job processing, establish automatic email notification to the concerned persons.
  • Written documentation to describe program development, logic, coding, testing, changes and corrections.
  • Used SQL Trace, Analyze, Explain Plan Utilities, and Hints for the performance tuning.
  • Managed the Metadata associated with the ETL processes used to populate the data warehouse.
  • Tested the data and data integrity among various sources and targets. Associated with Production support team in various performances related issues.
  • Worked with business users and gather data cleansing, data reconciliation and data convergence requirements.

Environment: Informatica Power Center 9.x, IDQ, MDM, Data explorer, Oracle 10g/11g, SQL Server 2012, PL/SQL, Flat files, Teradata, PLSQL Developer.

ETL Developer

Confidential

Responsibilities:

  • Using Informatica PowerCenter Designer analyzed the source data to Extract & Transform from various source systems (oracle 10g, DB2, SQL server and flat files) by incorporating business rules using different objects and functions that the tool supports.
  • Using Informatica PowerCenter created mappings and mapplets to transform the data according to the business rules.
  • Used various transformations like Source Qualifier, Joiner, Lookup, sql, and router, Filter, Expression and Update Strategy.
  • Implemented slowly changing dimensions (SCD) for some of the Tables as per user requirement.
  • Write Teradata SQL, BTEQ, MLoad, OLELoad, FastLoad, and FastExport for Ad-hoc queries, and build UNIX shell script to perform ETL interfaces BTEQ, FastLoad or FastExport, via Hummingbird and Control-M software.
  • Developed Stored Procedures and used them in Stored Procedure transformation for data processing and have used data migration tools.
  • Created and ran the Workflows using Workflow manager in Informatica Maintained stored definitions, transformation rules and targets definitions using Informatica repository manager.
  • Documented Informatica mappings in Excel spread sheet.
  • Tuned the Informatica mappings for optimal load performance.
  • Have used BTEQ, FEXP, FLOAD, MLOAD Teradata utilities to export and load data to/from Flat files.
  • Analyzed, identified, fixed bad data and imported data from SalesForce - CRM to Oracle. Upstream data integration and migration processes in predefined schemas.
  • Involved in creating the Unix Scripts and jobs to handle Informatica workflows and Teradata utilities like Bteq, Mload, Fast Export and TPT scripts.
  • Extensively involved in Data Extraction, Transformation and Loading (ETL process) from Source to target systems using Informatica.
  • Created and Configured Workflows and Sessions to transport the data to target warehouse Oracle tables using Informatica Workflow Manager. Informatica Web-services and web-portal applications.
  • Handful experience on Windows 32 bit commands, Quoting, Escaping.
  • Managed the migration of SQL Server 2008 databases to SQL Server 2012.
  • Have generated reports using OBIEE 10.1.3 for the future business utilities.
  • Created SQL scripts to load the custom data into Development, Test and production Instances using Import/Export. Created scripts to create custom Tables and Views.
  • This role carries primary responsibility for problem determination and resolution for each SAP application system database server and application server.
  • Worked along with UNIX team for writing UNIX shell scripts to customize the server scheduling jobs.
  • Constantly interacted with business users to discuss requirements.

Environment: Informatica Power Center Designer, Informatics Repository Manager, OBIEE Oracle10g/9i, and DB2 6.1, Erwin, TOAD, and SAP Version: 3.1.H, Unix- SunOS, PL/SQL, and SQL Developer, Java/J2EE, Struts, JDBC, PL/SQL, JUNIT, ANT, HTML, DHTML, JSP, JavaScript, XML, Oracle, Apache Tomcat, MS Excel.

ETL developer

Confidential

Responsibilities:

  • Created Technical Design Specifications, Unit test document based on functional design specifications provided by Business Analyst.
  • Designed and developed ETL Processes based on business rules, job control mechanism using Informatica Power Center.
  • Worked extensively on complex mappings using source qualifier, joiner, expressions, aggregators, filters, Lookup, update strategy, stored procedure transformations, etc.
  • Used workflow monitor to monitor the jobs, reviewed session/workflow logs that were generated for each session to resolve issues, used Informatica debugger to identify issues in mapping execution.
  • Re-engineered lots of existing mappings to support new/changing business requirements.
  • Monitored production jobs on a daily basis and worked on issues relating to the job failure and restarted failed jobs after correcting the errors.
  • Developed reusable transformations, mapplets, sessions, worklets to make Informatica code very modular and reuse as required
  • Performed unit testing, system integration testing, and supported user acceptance testing.
  • Performance tuned SQL statements, Informatica mappings, used Informatica parallelism options to speed up data loading to meet defined SLA
  • Using Informatica Powercenter8.1.1 to make the changes to the existing ETL mappings in each of the environments.
  • Collaborated with Project Manager, Tech Lead, Developers, QA teams and Business SMEs to ensure delivered solutions optimally support the of business outcomes.
  • SCRUM and AGILE approach was used to solve defects and monitor daily progress.
  • Supported Informatica, non-Informatica code migration between environments (DEV/QA/PRD)
  • Developed PL/SQL procedures for processing business logic in the database and use them as a Stored Procedure Transformation.
  • Developed Oracle PL/SQL Packages, Procedures, Functions and Database Triggers.
  • Data Warehouse Data modeling based on the client requirement using Erwin (Conceptual, Logical and Physical Data Modeling)
  • Dimensional Modeling for creation of Star Schema and Snow Flake Schema.

Environment: Informatica Power Center 8.1.1, Repository Manager, Designer, Work Flow Manager, Oracle 9i/10g, SQL Server 2008/2005, Teradata, XML Files, Flat Files, CSV files, PL/SQL (Stored Procedure, Trigger, Packages), Erwin,, MS Visio, TOAD, Windows.

We'd love your feedback!