Sr. Etl/talend Developer Resume
Chicago, IL
SUMMARY:
- High skilled ETL/BI developer with 8+ years of experience in tools like Talend, Informatica PowerCenter (10.x, 9.x, 8.x, 7.x), Informatica IDQ/Informatica Data Explorer (IDE), Informatica MDM, and in developing & Administering ETL mappings.
- Good experience in all phases of software development life cycle (SDLC) including system design, development, integration, testing, deployment and delivery of applications.
- Extracted the data using Pentaho Data Integration Designer Kettle from the flat files and other RDBMS databases like SQL Server 2005/2008 and Oracle 10g into staging area and populated onto Data warehouse.
- Used Kettle designer to design and create Pentaho Transformations and jobs.
- Created source and target connections in Talend Integration cloud to pull files from mainframe and place it into ESB Server.
- Experience in Big Data technologies like Hadoop/Map Reduce, Pig, Hive, sqoop.
- Experienced in integration of various data sources like Oracle 11g/10g/9i, IBM DB2, MS SQL Server, My SQL, Snowflake, Teradata, Netezza, XML files, Mainframe sources into staging area and different target databases.
- Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType,, tflowtoIterate, tAggregate, tSortRow, tFlowMeter, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tLogCatcher, tFilter, tGlobalmapetc.
- Expertise in Informatica MDM Hub Match and Merge Rules, Batch Jobs and Batch Groups.
- Experience using Informatica IDQ for qualifying the data content and MDM to filter duplicate data as well as to deploy the project as well as Meta Data Management.
- Created Snowflake Schemas by normalizing the dimension tables as appropriate, and creating a Sub Dimension named Demographic as a subset to the Customer Dimension.
- Experience in all stages of ETL - requirement gathering, designing and developing various mappings, unit testing, integration testing and regression testing.
- Extensive Experience in designing and developing complex mappings applying various transformations such as Expression, Aggregator, Lookup, Source qualifier, Update strategy, Filter, Router, Sequence generator, Rank, Stored procedure, Joiner and Sorter.
- Experienced in Data Analysis/Validation and Profiling based on the business and functional requirements of the Project.
- Good experience in installation of Talend and Informatica Power Exchange.
- Hands on experience in Pentaho Business Intelligence ServerStudio.
- Hands on experience in developing and monitoring SSIS/SSRS Packages and outstanding knowledge of high availability SQL Server solutions, including replication.
- Proficient in the implementation of Data Cleanup procedures, Stored Procedures, Scripts and execution of Test plans for loading the data successfully into the various Target types (Relational and flat file)
- Experienced in writing stored procedures/SQL scripts with various Teradata utilities like MLOAD, FASTLOAD, TPUMP, FASTEXPORT.
- Good experience in developing jobs for OLTP & OLAP databases.
- Extensive experience in SQL scripting, shell scripting in Linux and windows based environments.
- Experience on Data Modeling Tools Erwin, Visio, Sybase Power designer.
- Experience in working on Enterprise Job scheduling tools like Autosys.
PROFESSIONAL EXPERIENCE:
Sr. ETL/Talend Developer
Confidential, Chicago, IL
Responsibilities:
- Participated in Requirement gathering, Business Analysis, User meetings and translating user inputs into ETL mapping documents.
- Worked closely with Business Analysts to review the business specifications of the project and also to gathered ETL requirements.
- Implemented new users, projects, tasks within multiple different environments of TAC.
- Created complex mappings in Talend using tMap, tJoin, tXML, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
- Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
- Developed complex Talend ETL jobs to migrate the data from flat files to database.
- Pulled files from mainframe into Talend execution server using multiple ftp components.
- Developed complex Talend ETL jobs to migrate the data from flat files to database.
- Developed Talend ESB services and deployed them on ESB servers on different instances.
- Worked on Talend Data Integration cloud components like tfileinputpositional, tfilelist, tActioninput, tActionoutput, tActionReject, tActionLog, tActionFailure, jClou dLog, tsplitrow, tcontextLoad, tFileArchive, tFileDelete, tFixedFlowInput.
- Worked with Data mapping team to understand the source to target mapping rules.
- Created source and target connections in Talend Integration cloud to pull files from mainframe and place it into ESB Server.
- Published 200 plus jobs into Talend integration cloud.
- Worked on dynamic schema in Integration Actions enable web users to define new columns on the run when they create Flows.
- Created test cases for Integration Actions.
- Integration Actions are published to Talend Integration Cloud for Cloud users to build Flows.
- Prepared ETL mapping Documents for every mapping and Data Migration document for smooth transfer ofproject from development to testing environment and then to production environment.
- Performed Unit testing and System testing to validate data loads in the target.
ENVIRONMENT: Talend 6.3.1, Talend integration cloud, Java scripts, SQL Server, GIT, Eclipse, XML, JUNIT, xmlfiles, flat files, mainframe files, ssis, microsoft office, Web Services.
Sr. ETL/Talend Developer
Confidential, Denver, CO
Responsibilities:
- Participated in Client Meetings and gathered business requirements and analyzed them.
- Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.
- Design and Implement ETL processes to import data from and into Microsoft Azure.
- Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
- Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
- Develop stored procedures/views in Snowflake and use in Talend for loading Dimensions and Facts.
- Develop merge scripts to UPSERT data into Snowflake from an ETL source.
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW tables and historical metrics.
- Created complex mappings in Talend using tMap, tJoin, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
- Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
- Developed jobs to move inbound files to vendor server location based on monthly, weekly and daily frequency.
- Implemented Change Data Capture technology in Talend in order to load deltas to a Data Warehouse.
- Perform ETL using different sources like databases, flat files, xml files.
- Migrated Snowflake database to Windows Azure and updating the Connection Strings based on requirement.
- Managed and reviewed Hadoop log files.
- Wrote ETL jobs to read from web apis using REST and HTTP calls and loaded into HDFS using java and Talend.
- Shared responsibility for administration of Hadoop, Hive and Pig and Talend.
- Tested raw data and executed performance scripts.
- Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
ENVIRONMENT: Talend6.3.1, Pentaho, HDFS, HBase, MapReduce, Java scripts, Snowflake, Eclipse, XML, JUNIT, Microsoft Azure, Hadoop, Apache Pig, Hive, Elastic Search, Web Services, Pentaho kettleMicrosoft Office.
Sr. ETL/Talend Developer
Confidential, Birmingham, AL
Responsibilities:
- Involved in building the ETL architecture and Source to Target mapping to load data into Data warehouse.
- Performed data manipulations using various Talend components like tMap, tJavaRow, tjava, tOracleRow, tOracleInput, tOracleOutput, tMSSQLInput and many more.
- Designed and customized data models for Data warehouse supporting data from multiple sources on real time.
- Designed ETL process using Talend Tool to load from Sources to Targets through data Transformations.
- Extensive experience on Pentaho designer, Pentaho kettle, Penatho BI server, BIRT report designer
- Developed advanced Oracle stored procedures and handled SQL performance tuning.
- Involved in creating the mapping documents with the transformation logic for implementing few enhancements to the existing system.
- Monitored and supported the Talend jobs scheduled through Talend Admin Center (TAC)
- Developed the Talend mappings using various transformations, Sessions and Workflows. Teradata was the target database, Source database is a combination of Flat files, Oracle tables, Excel files and Teradata database.
- Loaded data in to Teradata Target tables using Teradata utilities (FastLoad, MultiLoad, and FastExport) Queried the Target database using Teradata SQL and BTEQ for validation.
- Used Talend to Extract, Transform and Load data into Netezza Data Warehouse from various sources like Oracle and flat files.
- Created connection to databases like SQL Server, oracle, Netezza and application connections.
- Created mapping documents to outline data flow from sources to targets.
- Prepare the Talend job level LLD documents and working with the modeling team to understand the Big Data Hive table structure and physical design.
- Involved in Dimensional modeling (Star Schema) of the Data warehouse and used Erwin to design the business process, dimensions and measured facts.
- Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend.
- Maintained stored definitions, transformation rules and targets definitions using Informatica repository Manager.
- Used various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings.
- Developed mapping parameters and variables to support SQL override.
- Developed Talend ESB services and deployed them on ESB servers on different instances.
- Created mapplets & reusable transformations to use them in different mappings.
- Developed mappings to load into staging tables and then to Dimensions and Facts.
- Developed the Talend jobs and make sure to load the data into HIVE tables & HDFS files and develop the Talend jobs to integrate with Teradata system from HIVE tables.
- Worked on different tasks in Workflows like sessions, events raise, event wait, decision, e-mail, command, worklets, Assignment, Timer and scheduling of the workflow.
- Unit testing, code reviewing, moving in UAT and PROD.
- Designed the Talend ETL flow to load the data into hive tables and create the Talend jobs to load the data into Oracle and Hive tables.
- Migrated the code into Testing and supported QA team and UAT (User)
- Created detailed Unit Test Document with all possible Test cases/Scripts.
- Working with high volume of data and tracking the performance analysis on Talend job runs and session.
- Conducted code reviews developed by my team mates before moving the code into QA.
- Experience in Batch scripting on windows, Windows 32 bit commands, Quoting, Escaping.
- Used Talend reusable components like routines, context variable and globalMap variables.
- Provided support to develop the entire warehouse architecture and plan the ETL process.
- Knowledge on Teradata Utility scripts like FastLoad, MultiLoad to load data from various source systems to Teradata.
- Modified existing mappings for enhancements of new business requirements.
- Prepared migration document to move the mappings from development to testing and then to production repositories.
- Configured the hive tables to load the profitability system in Talend ETL Repository and create the Hadoop connection for HDFS cluster in Talend ETL repository.
- Works as a fully contributing team member, under broad guidance with independent planning & execution responsibilities.
Environment: Talend, Hive, Pig, Hadoop, Sqoop, PL/SQL, Oracle 12c/11g/, Erwin, Autosys, SQL Server 2012, Teradata, Netezza, Sybase, SSIS, UNIX, Profiles, Role hierarchy, Workflow & Approval processes, Data Loader, Reports, Custom Objects, Custom Tabs, Data Management, Lead processes, Record types.
ETL Developer
Confidential, Richardson, TX
Responsibilities:
- Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.
- Prepare high level design documents, detail design documents, business requirement documents, technical specifications, table level specs and test plan documents.
- Extract data from legacy systems to staging area and then cleanse, homogenize, process and load into the data warehouse.
- Use JIRA to create, implement and deploy ETL related stories.
- Participate in daily scrum and bi weekly iteration planning as part of agile environment.
- Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
- Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
- Develop stored procedures/views in SQL Server and use in Talend for loading Dimensions and Facts.
- Develop MERGE scripts to UPSERT data into SQL Server from an ETL source.
- Work on writing pig/hive/hcatalog scripts to process huge data files and export the data into FTP locations.
- Perform ETL using different sources like databases, flat files, xml, avro files.
- Wrote ETL jobs to read from web apis using REST and HTTP calls and loaded into HDFS using java and Talend.
- Write custom java code in Talend to achieve complex functionality that is not available in ETL tool.
- Write shell scripts to encrypt and decrypt files containing Personal Indentifiable Information (PII) and call the scripts from ETL tool.
- Manage ETL jobs using Talend Administrator Console (TAC) in development and production environment and administer the Talend ETL tool in development and production environments.
- Created specs and rpms and deployed shell/pig/hive scripts using continuous integrating with Jenkins.
- Use GitHub and svn as version control for the code and implemented branching for different environments.
- ENVIRONMENT: Talend 5.x, JIRA, Java, GIT, Sql Server, MySQL, Putty, XML, JUNIT, Hadoop, Jenkins, Apache Pig, Hive, Elastic Search, Web Services, Microsoft Office, SOAPUI/REST.
- ETL Developer
- Western Union Englewood, California - Englewood, CA
- Responsibilities:
- Responsible for writing requirements for automation of online contract management tool.
- Involved in designing MDM data model and created base objects, mappings, defined trust setting for sources, customized user exists, customized IDD applications.
- Performed match/merge and ran match rules to check the effectiveness of MDM process on data
- Involved in designing, Implementing the Inbound and Outbound interfaces to send data to and from the MDM System to the sources.
- Involved in the UAT testing of the Items Data Mart price files tool automation project.
- Responsible for handling Supply Chain Management Data Warehouse and related data base and online solutions.
- Involved in maintenance of reporting tools and processes utilizing Data Warehouse data.
- Responsible for generating pricing reports from Finance & Procurement Data Warehouse and Data Mart using Business Objects.
- Responsible for managing the online contract management repository and items data mart pricing information.
- Work on formulating the ETL rules using Informatica and AbInitio.
- Used Informatica Power Center, tool for migrating data from various OLTP databases to the data mart.
- Worked with different sources like Oracle, flat files, XML files, DB2, MS SQL Server.
- Designed and developed jobs and transformations to load data into dimensions and facts.
- Worked on Data Analytics/Visualization Project using informatica (ETL and Data Visualization Suite)
- Extracted data from Sales department to flat files and load the data to the target database.
- Developed complex mappings using Informatica to load Dimension & Fact tables as per STAR schema techniques.
- Contribute to the entire implementation process including driving the definition of improvements based on business need and architectural improvements.
- Extracted data from sources like fixed width and Delimited Flat files transformed the data according the business requirement and then loaded into Target Data mart.
- Created various Informatica mappings and mapplets to load the data mart. The mappings involved extensive use of transformations like Aggregator, Filter, Router, Expression, Joiner, Sequence generator.
- Configured the mappings to handle the updates to preserve the existing records using Update Strategy Transformation (SCD Type-2)
- Used IDQ for generation of various reports and graphs based on data used for reporting.
- Used Debugger to identify the errors in the mappings and fix them.
- Performed unit testing to validate mappings and populate the database.
- Used TOAD and SQL Plus to write queries and interact with Oracle database.
- Implemented several different tasks (Session, Assignment, Command, Decision, Timer, Email, Event-Raise, Event-Wait, Control) in the workflow.
- Involved in writing UNIX scripts and used them to automate the scheduling process.
- Error Handling handled in Informatica Mappings & sessions and also through validation script.
- Co-ordinated with DBA Team, Functional Team and offshore teams.
- Automation of job processing, establish automatic email notification to the concerned persons.
- Written documentation to describe program development, logic, coding, testing, changes and corrections.
- Used SQL Trace, Analyze, Explain Plan Utilities, and Hints for the performance tuning.
- Managed the Metadata associated with the ETL processes used to populate the data warehouse.
- Tested the data and data integrity among various sources and targets. Associated with Production support team in various performances related issues.
- Worked with business users and gather data cleansing, data reconciliation and data convergence requirements.
Environment: Informatica Power Center 9.x, IDQ, MDM, Data explorer, Oracle 10g/11g, SQL Server 2012, PL/SQL, Flat files, Teradata, PLSQL Developer.
ETL Developer
Confidential
Responsibilities:
- Using Informatica PowerCenter Designer analyzed the source data to Extract & Transform from various source systems (oracle 10g, DB2, SQL server and flat files) by incorporating business rules using different objects and functions that the tool supports.
- Using Informatica PowerCenter created mappings and mapplets to transform the data according to the business rules.
- Used various transformations like Source Qualifier, Joiner, Lookup, sql, and router, Filter, Expression and Update Strategy.
- Implemented slowly changing dimensions (SCD) for some of the Tables as per user requirement.
- Write Teradata SQL, BTEQ, MLoad, OLELoad, FastLoad, and FastExport for Ad-hoc queries, and build UNIX shell script to perform ETL interfaces BTEQ, FastLoad or FastExport, via Hummingbird and Control-M software.
- Developed Stored Procedures and used them in Stored Procedure transformation for data processing and have used data migration tools.
- Created and ran the Workflows using Workflow manager in Informatica Maintained stored definitions, transformation rules and targets definitions using Informatica repository manager.
- Documented Informatica mappings in Excel spread sheet.
- Tuned the Informatica mappings for optimal load performance.
- Have used BTEQ, FEXP, FLOAD, MLOAD Teradata utilities to export and load data to/from Flat files.
- Analyzed, identified, fixed bad data and imported data from SalesForce - CRM to Oracle. Upstream data integration and migration processes in predefined schemas.
- Involved in creating the Unix Scripts and jobs to handle Informatica workflows and Teradata utilities like Bteq, Mload, Fast Export and TPT scripts.
- Extensively involved in Data Extraction, Transformation and Loading (ETL process) from Source to target systems using Informatica.
- Created and Configured Workflows and Sessions to transport the data to target warehouse Oracle tables using Informatica Workflow Manager. Informatica Web-services and web-portal applications.
- Handful experience on Windows 32 bit commands, Quoting, Escaping.
- Managed the migration of SQL Server 2008 databases to SQL Server 2012.
- Have generated reports using OBIEE 10.1.3 for the future business utilities.
- Created SQL scripts to load the custom data into Development, Test and production Instances using Import/Export. Created scripts to create custom Tables and Views.
- This role carries primary responsibility for problem determination and resolution for each SAP application system database server and application server.
- Worked along with UNIX team for writing UNIX shell scripts to customize the server scheduling jobs.
- Constantly interacted with business users to discuss requirements.
Environment: Informatica Power Center Designer, Informatics Repository Manager, OBIEE Oracle10g/9i, and DB2 6.1, Erwin, TOAD, and SAP Version: 3.1.H, Unix- SunOS, PL/SQL, and SQL Developer, Java/J2EE, Struts, JDBC, PL/SQL, JUNIT, ANT, HTML, DHTML, JSP, JavaScript, XML, Oracle, Apache Tomcat, MS Excel.
ETL developer
Confidential
Responsibilities:
- Created Technical Design Specifications, Unit test document based on functional design specifications provided by Business Analyst.
- Designed and developed ETL Processes based on business rules, job control mechanism using Informatica Power Center.
- Worked extensively on complex mappings using source qualifier, joiner, expressions, aggregators, filters, Lookup, update strategy, stored procedure transformations, etc.
- Used workflow monitor to monitor the jobs, reviewed session/workflow logs that were generated for each session to resolve issues, used Informatica debugger to identify issues in mapping execution.
- Re-engineered lots of existing mappings to support new/changing business requirements.
- Monitored production jobs on a daily basis and worked on issues relating to the job failure and restarted failed jobs after correcting the errors.
- Developed reusable transformations, mapplets, sessions, worklets to make Informatica code very modular and reuse as required
- Performed unit testing, system integration testing, and supported user acceptance testing.
- Performance tuned SQL statements, Informatica mappings, used Informatica parallelism options to speed up data loading to meet defined SLA
- Using Informatica Powercenter8.1.1 to make the changes to the existing ETL mappings in each of the environments.
- Collaborated with Project Manager, Tech Lead, Developers, QA teams and Business SMEs to ensure delivered solutions optimally support the of business outcomes.
- SCRUM and AGILE approach was used to solve defects and monitor daily progress.
- Supported Informatica, non-Informatica code migration between environments (DEV/QA/PRD)
- Developed PL/SQL procedures for processing business logic in the database and use them as a Stored Procedure Transformation.
- Developed Oracle PL/SQL Packages, Procedures, Functions and Database Triggers.
- Data Warehouse Data modeling based on the client requirement using Erwin (Conceptual, Logical and Physical Data Modeling)
- Dimensional Modeling for creation of Star Schema and Snow Flake Schema.
Environment: Informatica Power Center 8.1.1, Repository Manager, Designer, Work Flow Manager, Oracle 9i/10g, SQL Server 2008/2005, Teradata, XML Files, Flat Files, CSV files, PL/SQL (Stored Procedure, Trigger, Packages), Erwin,, MS Visio, TOAD, Windows.