Etl/talend Developer Resume
Portland, OR
SUMMARY
- Over 5+ years of experience in ETL architecture, design and development of complex IT applications in large enterprises by using Informatica PowerCenter, SQL, UNIX etc.
- Experience in designing and customized data models for Data warehouse supporting data from multiple sources on real time. Involved in building teh ETL architecture and Source to Target mapping to load data into Data warehouse. Created mapping documents to outline data flow from sources to targets.
- Created Talend ETL jobs to receive attachment files from pop e - mail using tPop, tFileList and tFileInputMail and then loaded data from attachments into database and achieved teh files.
- Strong working experience in Data Warehousing concepts, Star Schema and Snowflake Schema methodologies. Experience in Data Modeling, Dimensional modeling of large databases. Working Experience wif SQL, PL/SQL, Informatica, Business Objects.
- Hands-on working experience in Dimensional Data Modeling, Data Cleansing, Standardization and Migration, and Data Staging of operational sources using ETL processes for data warehouses.
- Well versed wif Talend Big Data, Hadoop, Hive and used Talend Big data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
- Extensive experience in writing UNIX shell scripts and automation of teh ETL processes using UNIX shell scripting.
- Experience in all phases of Data warehouse development from requirements gathering for teh data warehouse to develop teh code, Unit Testing and Documenting.
- Experience wif Informatica complex mappings using different transformations like Expressions, Filters, Joiners, Routers, Union, Lookup, Stored Procedure, Aggregator, Update Strategy, Normalizer, Sorter, HTTP, XML, and SQL.
- Expertise in working wif relational databases such as Oracle, SQL Server 2016/14/12, DB2 8.0/7.0, UDB, MS Access and Teradata.
- Experienced in teh Fast-paced environments like Agile and Scrum Methodologies.
- Strong communication, collaboration and team building skills wif proficiency in grasping new technical concepts quickly.
- Self-starter always inclined to learn new technologies and Team Player wif very good communication, organizational and interpersonal skills.
TECHNICAL SKILLS
ETL Tools: Informatica 10.x/ 9.6/9.1/8.6/8.5.1 (PowerCenter/Power Mart), IDQ, Talend Data Integration/ Big Data Integration/ Data Quality/ESB 6.1/5.5/5.0, Talend Administrator Console
Data Modeling: Erwin 4.0/3.5, Star Schema Modelling
Databases: Oracle 11g/10g/9i/8i, My SQL Server 16/14/12/08, DB2
Languages: SQL, PL/SQL, Unix Shell Script, Visual Basic
Tools: Toad, SQL* Loader, Cognos 7.0/6.0
Operating Systems: Windows, UNIX, MS-DOS, Mac
Scheduling Tools: Autosys r11.3/10.X, Control-M 9/8/6.1.X
PROFESSIONAL EXPERIENCE
Confidential, Portland, OR
ETL/Talend Developer
Responsibilities:
- Participated in all phases of development life-cycle wif extensive involvement in teh definition and design meetings, functional and technical walkthroughs.
- Created Talend jobs to copy teh files from one server to another and utilized Talend FTP components
- Created and managed Source to Target mapping documents for all Facts and Dimension tables
- Used ETL methodologies and best practices to create Talend ETL jobs. Followed and enhanced programming and naming standards.
- Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
- Design and Implemented ETL for data load from heterogeneous Sources to SQL Server and Oracle as target databases and for Fact and Slowly Changing Dimensions SCD-Type1 and SCD-Type2.
- Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
- Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
- Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more).
- Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
- Created Implicit, local and global Context variables in teh job. Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
- Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
- Developed stored procedure to automate teh testing process to ease QA efforts and also reduced teh test timelines for data comparison on tables.
- Automated SFTP process by exchanging SSH keys between UNIX servers. Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
- Involved in production n deployment activities, creation of teh deployment guide for migration of teh code to production, also prepared production run books.
Environment: Talend Data Integration 6.1/5.5.1, Talend Enterprise Big Data Edition 5.5.1, Talend Administrator Console, Oracle 11g, Hive, HDFS, Sqoop, Netezza, SQL Navigator, Toad, Control M, Putty, Winscp.
Confidential, Lafayette, LA
ETL/Talend Developer
Responsibilities:
- Participated in JAD sessions wif business users and SME's for better understanding of teh reporting requirements.
- Design and developed end-to-end ETL process from various source systems to Staging area, from staging to Data Marts.
- Analyzing teh source data to no teh quality of data by using Talend Data Quality.
- Broad design, development and testing experience wif Talend Integration Suite and noledge in Performance Tuning of mappings.
- Developed jobs in Talend Enterprise edition from stage to source, intermediate, conversion and target.
- Involved in writing SQL Queries and used Joins to access data from Oracle, and MySQL.
- Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats.
- Solid experience in implementing complex business rules by creating re-usable transformations and robust mappings using Talend transformations like tConvertType, tSortRow, tReplace, tAggregateRow, tUnite etc.
- Developed Talend jobs to populate teh claims data to data warehouse - star schema.
- Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested teh mappings.
- Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats into a Database table to record job history.
- Integrated java code inside Talend studio by using components like tJavaRow, tJava, tJavaFlex and Routines.
- Experienced in using debug mode of talend to debug a job to fix errors. Created complex mappings using tHashOutput, tHashInput, tNormalize, tDenormalize, tMap, tUniqueRow. tPivotToColumnsDelimited, etc.
- Used tRunJob component to run child job from a parent job and to pass parameters from parent to child job.
- Created Context Variables and Groups to run Talend jobs against different environments.
- Used tParalleize component and multi thread execution option to run subjobs in parallel which increases teh performance of a job.
- Implemented FTP operations using Talend Studio to transfer files in between network folders as well as to FTP server using components like tFileCopy, TFileAcrchive, tFileDelete, tCreateTemporaryFile, tFTPDelete, tFTPCopy, tFTPRename, tFTPut, tFTPGet etc.
- Experienced in Building a Talend job outside of a Talend studio as well as on TAC server.
- Experienced in writing expressions wif in tmap as per teh business need. Handled insert and update Strategy using tmap. Used ETL methodologies and best practices to create Talend ETL jobs.
- Extracted data from flat files/ databases applied business logic to load them in teh staging database as well as flat files.
Environment: Talend 5.5/5.0, Oracle 11g, Teradata SQL Assistant, HDFS, MS SQL Server 2012/2008, PL/SQL, Agile Methodology, Informatica, TOAD, ERwin, AIX, Shell Scripts, AutoSys, SVN.
Confidential, Plano TX
ETL Developer
Responsibilities:
- Participate in design and analysis sessions wif business analysts, source-system technical teams, and end users.
- Worked on Informatica Power Center tool - Source Analyzer, Data warehousing designer, Mapping &Mapplet Designer and Transformation Designer.
- Designed and developed ETL processes based on business rules, job control mechanism using Informatica Power Center.
- Implemented ETL Informatica designs and processes for loading data from teh sources to target warehouse.
- Involved in Designing teh Data warehouse using Informatica ETL tool by using Source Analyzer, Warehouse Designer, Mapping Designer & Mapplet and Transformations.
- Design teh overall ETL/ELT strategy in Informatica defining teh logical, physical schemas and design decisions to support efficient real time data integration.
- Developed mappings to load into staging tables and then to Dimensions and Facts.
- Developed Informatica Mappings to populate teh data into dimension and Fact tables for data classifications to end developers.
- Developed ETL procedures to transform teh data in teh intermediate tables according to teh business rules and functionality requirements.
- Documentation of Technical specification, business requirements, functional specifications for teh development of Informatica Extraction, Transformation and Loading (ETL) mappings to load data into various tables.
- Release management of teh ETL jobs (Informatica mappings) across environment repositories using Informatica Repository Manager.
- Optimized teh ETL mappings performance to load data efficiently and utilizing teh ELT approach and implementing Push down optimization/pass through to load data efficiently.
- Used various transformations like Filter, Expression, Sequence Generator, Update Strategy, Joiner, Stored Procedure, and Union to develop robust mappings in teh Informatica Designer. Used Type 1 SCD and Type 2 SCD mappings to update slowly Changing Dimension Tables.
- Created Stored Procedures to transform teh Data and worked extensively in PL/SQL for various needs of teh transformations while loading teh data.
- Proficient in Agile development and Scrum methodology.
- Tested data and data integrity among various sources and targets.
Environment: Informatica Poer Center, Business Objects, Data Ware House, SSIS, UNIX, Teradata, SQL and Windows.
Confidential, McLean VA
ETL Developer
Responsibilities:
- Analyzed business requirements and worked closely wif various application teams and business teams to develop ETL procedures that are consistent across all applications and system.
- Experience in writing Informatica ETL design documents, establish ETL coding standards and perform Informatica mapping reviews.
- Extensively worked on Power Center Client Tools like Repository Admin Console, Repository Manager, Designer, Workflow Manager, and Workflow Monitor.
- Analyzed teh source data coming from different sources (Oracle, DB2, XML, Flat files) and worked on developing ETL mappings.
- Developed complex Informatica Mappings, reusable Mapplets and Transformations for different types of tests in research studies on daily and monthly basis.
- Implemented mapping level optimization wif best route possible wifout compromising wif business requirements.
- Created Sessions, reusable worklets and workflows in Workflow Manager and Scheduled workflows and sessions at specified frequency.
- Worked on fixing invalid Mappings, testing of Stored Procedures and Functions, and Integration Testing of Informatica Sessions.
- Responsible for teh Performance tuning at teh Source Level, Target Level, Mapping Level and Session Level.
- Worked extensively on SQL, PL/SQL, and UNIX shell scripting.
- Performed Data profiling for data quality purposes.
- Proven Accountability including professional documentation, and weekly status report.
- Documented flowcharts for teh ETL (Extract Transform and Load) flow of data using Microsoft Visio and created metadata documents for teh Reports and teh mappings developed and Unit test scenario documentation for teh mentioned.
Environment: Informatica PowerCenter, Oracle 11g, Workflow Manager, Workflow Monitor, Informatica Power Connect / Power Exchange, UNIX Scripting, DB2, Toad.