Sr Talend Developer Resume
Philadelphia, PA
SUMMARY
- Having 8 years of total experience in Information Technology including Data Warehouse/Data Mart development using Talend DI/BDE and Informatica Power Center.
- 6 years of experience in using Talend Open Studio/DI/BDE tool as applied to Data Migration, BI data analytics, and dashboard projects.
- Highly proficient in the integration of various data sources involving multiple relational databases like Oracle, MS SQL Server, Teradata, and non - relational sources like XML Files and Flat File.
- Strong understanding of the principles of DW using fact tables, dimension tables and star/snowflake schema modeling.
- Expertise on Exception Handling Mappings for Data Quality, Data Cleansing and Data Validation.
- Excellent knowledge in identifying performance bottlenecks and tuning the ETL Loads for better performance and efficiency.
- Hands on experience on major components in Hadoop Ecosystem like Hadoop MapReduce, HDFS, YARN, Cassandra, Hive, Pig, HBase, Sqoop, Oozie, Kafka.
- Experience in Spark to process large stream of data.
- Experience in UNIX shell scripting,Python scripting, CRON, FTP and file management in various UNIX environments.
- Experience in using Talend Big Data components to create connections to various third-party tools used for transferring, storing or analyzing big data, such as HDFS Components, Hive, Sqoop, MongoDB and Big Query to quickly load, extract, transform and process large and diverse data sets.
- Experience in Trouble shooting and implementing Performance tuning at various levels such as Source, Target, Mapping, Session and System in ETL Process.
- Good exposure in overall SDLC including requirement gathering, development, testing, debugging, deployment, documentation and production support.
- Worked on building a Data Warehouse and implementing an ETL solution for loading Data warehouse.
- Well-versed at building the Star/Snowflake schema, Fact/Dimensional tables used in relational,
- Dimensional and multidimensional data modelling.
- Developed slowly changing dimension (SCD) mappings using type-I, type-II, and type-III methods.
- Extracted, transformed and loaded (ETL) between different systems and formats such as XML Files, flat files, Oracle, SQL Server, PostGre SQL etc...
- Extensive understanding of managing and automating Control flow, Data flow, Events and
- Logging for Talend Jobs.
- Developed complex SQLs for data extraction, data validation and experience with tuning SQLs.
- Strong experience in error handling, product support, debugging, testing and error logging for ETL jobs.
- Very strong in creating and maintaining database objects like Indexes, Views, Stored Procedure,
- User Defined Functions, Triggers in Oracle & SQL Server.
- Efficient in building recursive queries using CTEs (Common Table Expressions).
TECHNICAL SKILLS
Operating Systems: Microsoft Windows, UNIX, MS-DOS
ETL Tools: Talend, Talend Open Studio, Talend DI/BDE, Informatica Power Center 9.x/8.x
Databases: Oracle 12c/11g/10g/9i, MS SQL Server /2005, DB2 v8.1, Teradata.
Methodologies: Data Modeling - Logical Physical Dimensional Modeling - Star / Snowflake
Languages: SQL, PL/SQL, UNIX, Shell scripts, Python, C++, Web Services, HTML.
Scheduling Tools: TOAD, SQL Plus, SQL Developer, Autosys, Control-M
PROFESSIONAL EXPERIENCE
Confidential, PHILADELPHIA, PA
Sr Talend Developer
Responsibilities:
- Worked closely with development lead to finalize on estimated/actual project completion dates for process design, development, testing and implementation
- Worked with support team to handover ETL to perform UAT and production role
- Developed process jobs as per design, prepare unit test cases, prepare test data, and Perform unit testing
- Worked with BIS project manager to finalize on estimated/actual project completion dates
- Performed final code review before moving to SIT environment and fix issue in warranty Phase
- Extensively worked on Talend Management Console and Talend Cloud.
- Deployed jobs to Talend Cloud and migrated them to various environments
- Performed technical analysis and prepare recommended data acquisition design approach
- Participated in a team environment for the design, development, and implementation of data warehousing projects
- Designed the system components / ETL automation for the extract/transform or conversion of data from source systems to the target application (including historization, surrogation, data quality, error handling components)
- Designed and developed complex ETL mappings, sessions, workflows and identify areas of optimizations
- Estimation of level of effort associated with new ETL jobs
- Worked extensively on Oracle, UNIX server and SQL server
- Sound knowledge of data warehouse architecture theories
- Demonstrated working experience in large and complex data and Cloud environments (AWS)
- Worked on ETL architecture and data schemas
- Developed ETL architecture in accordance with business requirements.
- Provided technical support in all phases on enterprise architect life cycle.
- Provided an architecture solution that is reusable, maintainable, and scalable.
- Experienced in ETL / Data Warehouse Management
- Developed data migration solution in highly demanding environment and provide hands on guidance to other team members
- Quickly learnt business processes and how the various business units interact with data
- Demonstrated critical thinking, analytical skills, and employ judgment to offer thoughtful, concise solutions to business client data problems.
- Excellent working experience in Waterfall, Agile methodologies. Proficient in performance analysis, monitoring and SQL query tuning using EXPLAINPLAN, Collect Statistics, Hints and SQL Trace both in Teradata as well as Oracle.
- Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
- Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput & tHashOutput and many more).
- Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
- Worked on various Talend components such as tMap, tFilterRow, tAggregateRow, tFileExist, tFileCopy, tFileList, tDie etc.
- Involved in production n deployment activities, creation of the deployment guide for migration of the code to production, also prepared production run books.
- Collaborated with functional experts and business users to develop architectural requirements to ensure client satisfaction with solution
- Provided production support related to ETL schedules, tasks and work with other BI/ IT team members to resolve data refresh issues
Confidential, NJ
Talend Developer/Admin
Responsibilities:
- Interacting with the clients on a regular basis to discuss day-to-day issues and matters.
- On-Call/Production Support provided during day-time and off-hours.
- Played as administrator for setting up development, QA, UAT and PROD environments for Talend, Postgres for development and documenting install plan at client locations.
- Setup ETL Framework, best practices around Talend for data integration implementation.
- Responsible for installing Talend on multiple environments, creating projects, setting up user roles, setting up job servers, configure TAC options, adding Talend jobs, job failures, on-call support and scheduling etc.
- Troubleshooting, debugging Talend issues, while maintaining the health and performance of the ETL environment.
- Excellent experience working on tHDFSInput, tHDFSOutput, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
- Developed jobs to expose HDFS files to Hive tables and Views depending up on the schema versions.
- Developed jobs to send and read data from AWS S3 buckets using components like tS3BucketExist, tS3Connection, tS3Get, tS3Put.
- Created Hive tables, partitions and implemented incremental imports to perform ad-hoc queries on Structured data. Developed jobs to move inbound files to HDFS file location based on monthly, weekly, daily and hourly partitioning.
- Manage and Review Hadoop log files and hands on with executing Linux and HDFS Commands.
- Experience building reusable framework for consuming and providing REST API’s using Java/ Node.js / PHP / Swagger
- Responsible for writing Talend Routines in Java.
- Demonstrated experience working with REST API design and Implementation.
- Experience using Talend against both on premise databases such as Oracle, Microsoft SQL Server, DB2, MySQL as well as cloud platforms preferably AWS.
- Implemented ODS/OLAP data mappings in Talend and created source to target mapping documents.
- Implemented job parallelism in Talend BDE 6.0.1.
- Experience working with Big data components for extracting and loading data into HDFS file system.
- Provided assistance on advancement systems with banner products.
- Automated and modified payroll data processes with SQL and Banner tools.
- Experience working with web services using tSOAP components for sending XML requests and receiving response XML files. Expertized in reading XMLs files on a loop and sending to webservice end for generating output XML files. Also used Advanced XML mappers for parsing multiple loop elements.
- Experience in reading JSON/XML requests and generating the JSON/XML responses. also responsible for validating JSON schema and handling exceptions in Talend.
- Involved in the development of Talend Jobs and preparation of design documents, technical specification documents.
- Production Support activities like application checkout, batch cycle monitoring and resolving User Queries.
- Responsible for deploying code to different environments using GIT.
Environment: Talend Big Data 6.0.1/Talend DI 6.0.1/6.3, Hive, HBase, HDFS, Spark, java, Sqoop, PIG, UNIX, Postgres 9.4, Oracle 11g, Web services (SOAP), GIT, Jira, Jenkins, AWS.
Confidential
Data warehouse Developer
Responsibilities:
- Source systems data are exported as Power Exchange data maps from IBM Mainframes which maintains legacy data. Data is staged in Load Ready / Staging tables, finally load the given data model using PDO logic.
- Involved in discussion of various business users Data Analysts from ISD team.
- Worked in an Agile project life cycle environment.
- Conduct source system and source table analysis to develop the necessary data
- Assisted the Data Architect in creating the logical/physical data models.
- Extensively worked on Teradata tables, created Teradata BTEQ procedures, and used TPT Reader/Writer utilities for Bulk Loading.
- Created several Staging, Historical and Daily Incremental and ETL maps.
- Worked extensively on Aggregated/Summarized data.
- Created DataStage Mappings for Error Handling/Audit Balance control flows.
- Prototyping ETL mappings and workflows for Slowly Changing Dimensions (SDCs),
- Push Down Optimization (PDO) and corresponding SQL code (Aggregations, Transformations, Rollups, Inserts, Updates, Deletes)
- Performance tuning and troubleshooting various DataStage mappings.
Environment: Windows 7, Linux, DataStage, Erwin, Teradata 13, Oracle 10G, Flat Files, XML Files, ESP scheduling tool, Teradata SQL Assistant.