Senior Etl/talend Developer Resume
Birmingham, AL
SUMMARY:
- High skilled ETL/BI developer with 6+ years of experience in tools like Talend, Informatica Power Center (9.5.1, 9.x, 8.x, 7.x), Informatica IDQ/Informatica Data Explorer (IDE), Informatica MDM, and in developing &Administering ETL mappings.
- Good experience in all phases of software development life cycle (SDLC) including system design, development, integration, testing, deployment and delivery of applications.
- Experience in Big Data technologies like Hadoop/Map Reduce, Pig, Hive, and sqoop.
- Experienced in integration of various data sources like Oracle 11g/10g/9i, IBM DB2, MS SQL Server, My SQL, Snowflake, Teradata, Netezza, XML files, Mainframe sources into staging area and different target databases.
- Expertise in creating mappings in TALEND using tMap, tJoin, tReplicate, tParallelize, tConvertType, tflowtoIterate, tAggregate, tSortRow, tFlowMeter, tLogCatcher, tRowGenerator, tNormalize, tDenormalize, tSetGlobalVar, tHashInput, tHashOutput, tJava, tJavarow, tAggregateRow, tLogCatcher, tFilter, tGlobalmapetc.
- Expertise in Informatica MDM Hub Match and Merge Rules, Batch Jobs and Batch Groups.
- Experience using Informatica IDQ for qualifying the data content and MDM to filter duplicate data as well as to deploy the project as well as Meta Data Management.
- Created Snowflake Schemas by normalizing the dimension tables as appropriate, and creating a Sub Dimension named Demographic as a subset to the Customer Dimension.
- Experience in all stages of ETL - requirement gathering, designing and developing various mappings, unit testing, integration testing and regression testing.
- Extensive Experience in designing and developing complex mappings applying various transformations such as Expression, Aggregator, Lookup, Source qualifier, Update strategy, Filter, Router, Sequence generator, Rank, Stored procedure, Joiner and Sorter.
- Experienced in Data Analysis/Validation and Profiling based on the business and functional requirements of the Project.
- Good experience in installation of Talend and Informatica Power Exchange.
- Hands on experience in Pentaho Business Intelligence Server Studio.
- Hands on experience in developing and monitoring SSIS/SSRS Packages and outstanding knowledge of high availability SQL Server solutions, including replication.
- Proficient in the implementation of Data Cleanup procedures, Stored Procedures, Scripts and execution of Test plans for loading the data successfully into the various Target types (Relational and flat file).
- Experienced in writing stored procedures/SQL scripts with various Teradata utilities like MLOAD, FASTLOAD, TPUMP, and FASTEXPORT.
- Good experience in developing jobs for OLTP & OLAP databases.
- Extensive experience in SQL scripting, shell scripting in Linux and windows based environments.
- Experience in working on Enterprise Job scheduling tools like Autosys.
- Good communication and interpersonal skills, ability to learn quickly, with good analytical reasoning and adaptive to new and challenging technological environment.
- Strong Team working spirit, relationship management and presentation skills.
TECHNICAL SKILLS:
Operating Systems: Windows 2008/2007/2005/ NT/XP, UNIX, MS-DOS
ETL Tools: Talend6.x/5.x, Informatica Power Center 9.x/8.x, I MDM, SSIS, Ab-Initio.
Databases;: Oracle 12c/11g/10g, Snowflake, MS SQL Server 2012 /2008/2005, DB2 v8.1, Netezza, Teradata, Hbase.
Methodologies: Data Modeling - Logical Physical, Dimensional Modeling - Star / Snowflake
Programming Skills: C++, Eclipse, Shell Scripting (K-Shell, C-Shell), PL/SQL, Hadoop, Pig, Hive, Hcatalog, JAVA (Eclipse IDE and Net Beans IDE, JAVA Script, CSS.
Web services: Microsoft Azure.
Testing Tools: QTP, WinRunner, LoadRunner, Quality Center, Test Director, Clear test, Clear case.
PROFESSIONAL EXPERIENCE:
Confidential, Birmingham, AL
Senior ETL/Talend Developer
Responsibilities:
- Participated in Client Meetings and gathered business requirements and analyzed them.
- Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.
- Design and Implement ETL processes to import data from and into Microsoft Azure.
- Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
- Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
- Develop stored procedures/views in Snowflake and use in Talend for loading Dimensions and Facts.
- Develop merge scripts to UPSERT data into Snowflake from an ETL source.
- Created Hive queries that helped market analysts spot emerging trends by comparing fresh data with EDW reference tables and historical metrics.
- Created complex mappings in Talend using tMap, tJoin, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
- Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
- Developed jobs to move inbound files to vendor server location based on monthly, weekly and daily frequency.
- Implemented Change Data Capture technology in Talend in order to load deltas to a Data Warehouse.
- Perform ETL using different sources like databases, flat files, and xml files.
- Migrated Snowflake database to Windows Azure and updating the Connection Strings based on requirement.
- Managed and reviewed Hadoop log files.
- Wrote ETL jobs to read from web APIs using REST and HTTP calls and loaded into HDFS using java and Talend.
- Shared responsibility for administration of Hadoop, Hive and Pig and Talend.
- Tested raw data and executed performance scripts.
- Provided design recommendations and thought leadership to sponsors/stakeholders that improved review processes and resolved technical problems.
Environment: Talend6.x, HDFS, HBase, MapReduce, Java scripts, Snowflake, Eclipse, XML, JUNIT, Microsoft Azure, Hadoop, Apache Pig, Hive, Elastic Search, Web Services, Microsoft Office.
Confidential,Raleigh, NC
Senior Talend Developer
Responsibilities:
-
Interact with Solution Architects and Business Analysts to gather requirements and update Solution Architect Document.
- Engaged in various data warehouse and migration projects in understanding the business requirements and analyzing the design documents created by Data Architects and then generate a technical Low level Design document with the approach for the project.
- Worked hands on Datastage 8.7 ETL migration to Talend Studio ETL process.
- Performed analysis, design, development, Testing and deployment for Ingestion, Integration, provisioning using Agile Methodology.
- Attended Daily Scrum meetings to provide update on the progress of the user stories Rally and to the Scrum Master and also to notify blocker and dependency if any.
- Experienced in creating Generic schemas and creating Context Groups and Variables to run jobs against different environments like Dev, Test and Prod.
- Created Talend Mappings to populate the data into dimensions and fact tables.
- Broad design, development and testing experience with Talend Integration Suite and knowledge in Performance Tuning of mappings.
- Experienced in Talend Data Integration, Talend Platform Setup on Windows and UNIX systems.
- Created complex mappings in Talend 6.0.1/5.5 using tMap, tJoin, tReplicate, tParallelize, tJava, tjavarow, tJavaFlex, tAggregateRow, tDie, tWarn, tLogCatcher, etc.
- Created joblets in Talend for the processes which can be used in most of the jobs in a project like to Start job and Commit job.
- Experience in using Repository Manager for Migration of Source code from Lower to higher environments.
- Developed jobs to move inbound files to vendor server location based on monthly, weekly and daily frequency.
- Implemented Change Data Capture technology in Talend in order to load deltas to a Data Warehouse.
- Created jobs to perform record count validation and schema validation.
- Created contexts to use the values throughout the process to pass from parent to child jobs and child to parent jobs.
- Developed joblets that are reused in different processes in the flow.
- Developed error logging module to capture both system errors and logical errors that contains Email notification and also moving files to error directories.
- Provided the Production Support by running the jobs and fixing the bugs.
- Experienced in using Talend database components, File components and processing components based up on requirements.
- Responsible for developing, support and maintenance for the ETL (Extract, Transform and Load) processes using Talend Integration Suite.
- Performed unit testing and also integration testing after the development and got the code reviewed.
- Responsible for code migrations from Dev. to QA and production and providing operational instructions for deployments.
Environment: Talend Studio 6.0.1/5.5, Oracle 11i, XML files, Flat files, HL7 files, JSON, TWS, Hadoop 2.4.1, HDFS, Hive 0.13, HBase 0.94.21, Talend Administrator Console, IMS, Agile Methodology, HPSM.
Confidential. Richardson, TX
ETL Developer
Responsibilities:
-
Design, develop, test, implement and support of Data Warehousing ETL using Talend and Hadoop Technologies.
- Prepare high level design documents, detail design documents, business requirement documents, technical specifications, table level specs and test plan documents.
- Extract data from legacy systems to staging area and then cleanse, homogenize, process and load into the data warehouse.
- Use JIRA to create, implement and deploy ETL related stories.
- Participate in daily scrum and bi weekly iteration planning as part of agile environment.
- Research, analyze and prepare logical and physical data models for new applications and optimize the data structures to enhance data load times and end-user data access response times.
- Create pig and hive scripts to process various types of data sets and load them into data warehouse built on Hive.
- Develop stored procedures/views in SQL Server and use in Talend for loading Dimensions and Facts.
- Develop MERGE scripts to UPSERT data into SQL Server from an ETL source.
- Work on writing pig/hive/hcatalog scripts to process huge data files and export the data into FTP locations.
- Perform ETL using different sources like databases, flat files, xml, and avro files.
- Wrote ETL jobs to read from web APIs using REST and HTTP calls and loaded into HDFS using java and Talend.
- Write custom java code in Talend to achieve complex functionality that is not available in ETL tool.
- Used over 20+ Components in Talend Like (tMap, Tfilelist, Tjava, Tlogrow, ToracleInput, ToracleOutput, tsendEmail etc)
- Used debugger and breakpoints to view transformations output and debug mappings.
- Implemented Performance tuning in Mappings and Sessions by identifying the bottlenecks and Implemented effective transformation Logic.
- Write shell scripts to encrypt and decrypt files containing Personal Identifiable Information (PII) and call the scripts from ETL tool.
- Manage ETL jobs using Talend Administrator Console (TAC) in development and production environment and administer the Talend ETL tool in development and production environments.
- Working on POC wrt BigData like loading the data into HDFS and creating Map Reduce Jobs
- Develop ETL mappings for various Sources (.TXT, .CSV,.XML) and also load the data from these sources into relational tables with Talend Enterprise Edition.
- Created specs and rpms and deployed shell/pig/hive scripts using continuous integrating with Jenkins.
- Use GitHub and svn as version control for the code and implemented branching for different environments.
Environment: Talend 5.x, JIRA, Java, GIT, Sql Server, MySQL, Putty, XML, JUNIT, Hadoop, Jenkins, Apache Pig, Hive, Elastic Search, Web Services, Microsoft Office, SOAPUI/REST.
Confidential
Informatica Developer
Responsibilities:
-
Extensively used Informatica Power Center for extracting, transforming, and loading databases from sources including Oracle, DB2, and Flat files.
- Participated in User meetings, gathering requirements.
- Worked with business analysts, application developers, production teams and across.
- Translating user inputs into ETL design docs.
- Participate in Design Reviews of Data model and Informatica mapping design.
- Functional units to identify business needs and discuss solution options.
- Performed match/merge and ran match rules to check the effectiveness on data.
- Created mappings using various Transformations like Aggregator, Expression, Filter, Router, Joiner, Lookup, Update strategy, Source Qualifier, Sequence generator, Stored Procedure and Normalizer.
- Extensively used debugger for troubleshooting issues and checking session stats/logs.
- Performed validation and testing of Informatica mapping against the pre-defined ETL design standards.
- Implemented Performance tuning at database level and at Informatica level. Reduce load time by using partitions and concurrent sessions running at a time.
- Involved in writing SQL, PL/SQL, Stored Procedures, Triggers and Packages in Data.
- Warehouse environments that employ Oracle.
- Developed Interfaces using UNIX Shell Scripts to automate the bulk load and Update.
- Using shell scripts to automate the export data into flat files for backup and delete data.
- From staging tables for the given time period.
- Extensive experience in performance tuning, identifying bottlenecks and resolving to improve Performance at database level, Informatica mappings and session level.
- Involved in the optimization of SQL queries which resulted in substantial performance.
- Improvement for the conversion processes.
- Involved in migration of objects in all phases (DEV, INT and PRD) of project and trained.
- Used debugger in identifying bugs in existing mappings by analyzing data flow, evaluating Transformations.
Environment: Informatica Power Center 8.1, MS SQL Server, Oracle, Flat files, Erwin, UNIX Shell Scripting, Toad, Autosys, Windows 2003 Server.