Sr. Talend Bigdata/aws Consultant Resume
Dallas, TX
SUMMARY
- Over 6 years of IT experience in the Analysis, Design, Development, Testing and Implementation of business application systems for Financial,, Aviation and and sports entertainment industries using Talend Real Time Big Data Platform 7.0.1 and 7.2.1.
- Proficient in Talend admin (Usergroups, project authorization project, locks, Licenses, job conductor, big data streaming, execution plan, logging, Activity Monitoring console, audit, drools and migration check)
- Expertise with Talend Data Integration and ESB frequently used components like tMap, cTalendJob, tJMSOutput, tContextLoad, tFixedFlowInput, tXMLMap, tJavaRow, tJavaFlex, tRestClient, tSOAP, tESBProviderRequest, tESBConsumer, tRunJob,tHDFSConnection, tCacheIn, tSparkSQL and more.
- Good Knowledge in creating route using a data service, WSDL and calling a SOAP web service
- Well versed in developing various database objects like packages, stored procedures, functions, triggers, tables, indexes, constraints, views in MS SQL Server and Confidential .
- Experienced in ingesting log data into HDFS using Kafka and tuning Spark for Apache Spark Big Data jobs and running Spark over YARN.
- Wrote ETL jobs to read from REST client API in JSON and loaded that JSON data into MySQL table.
- Created Unix Scripts and run them using tSystem for reading the Data from flat files and archiving the Flat files at the specified server.
- Worked in Agile environment and using Atlassian tool set including JIRA, Confluence, Bitbucket and Jenkins for Continuous Integration.
- Extensive development, support and maintenance experience working in all phases of the Software Development Life Cycle (SDLC) especially in Data Warehousing and Business Intelligence.
- Automation, Scheduling and monitoring of jobs with scheduler AutoSys.
- Deployed Talend jobs on various environments including dev, test and production environments
TECHNICAL SKILLS
ETL Tools: Talend Real Time Big Data Platform 7.2.1,SQL Server Integration Services (SSIS)
Databases: SQL Server 2016, Confidential, Confidential, MySQL, Redshift, MongoDB
Big Data Ecosystems: Spark SQL, HDFS, YARN, Amazon Elastic MapReduce (EMR)
Programming: SQL, PL/SQL,Java,Python
Scheduling Tools: Autosys, Control - M
BI and Analytics: Birst, Dremio,Denodo
Version Control Tools: GIT
Operating Systems: Win 10, Linux, macOS
Other Tools: Putty, JIRA,Confluence
Industry Verticals: , Aviation, Financial, Sports Entertainment
PROFESSIONAL EXPERIENCE
Confidential, Dallas, TX
Sr. Talend BigData/AWS Consultant
Responsibilities:
- Worked with Business Analysts and Architect on Commissions, Credit Monitoring and Anti Money Laundering projects to understand the data and coordinate with upstream teams to create ETL Framework/s which retrieves data from hive to CSV/JSON and induct data to EDW system for business reporting using Talend Real-Time Big Data Platform 7.2.1
- Created Big Data batch job/s to integrate messages from Kafka to Data-ware house using Spark and read data from S3 and store data into HDFS in parquet format.
- Experienced in using different loading utilities like API and TPT, MULTILOAD, SQLLDR and BCP to load data into Confidential, Confidential and SQL Server database systems.
- Created ETL jobs to read from REST API source and stored it in Confidential database system for consumption.
- Orchestrated multiple ETL jobs using AWS Step Functions and AWS Lambda.
- Involved in production and deployment activities, creation of the deployment guide for migration of the code to production and made sure production run book is up to date.
Environment: Talend Real Time Big Data Platform for Big Data 7.31, Apache Spark, Hive, HDFS, REST API using Postman, AWS CLI, AWS Redshift, Confidential, Confidential, MongoDB, S3, Athena, EMR, JIRA, GIT, UDeploy
Confidential, Dallas, TX
Sr. ETL/AWS Consultant
Responsibilities:
- Provided ETL solutions for various projects at Confidential, including international and US venue source systems, vendor data and Confidential data into an analytical presentation layer for consumption. Delivered outstanding solutions utilizing multiple leading technologies in marketplace.
- Developed Talend Big Data ETL jobs for a rapidly expanding company. Key member of the SQL Server to AWS migration team. Migrated all of our current tables, data into the AWS data lake ecosystem, using S3, JSON, Talend, Athena and Redshift using the ingest, model, enhance, transform, delivery method.
Environment: Talend Real Time Big Data Platform for Big Data 6.4.1, Spark, MS SQL Server 2016, REST API, AWS CLI, AWS Redshift, S3, Athena, EMR, Cloudwatch, SNS, JIRA, Git, Datagrip, Talend Administration Console.
Confidential, Seattle, WA
Sr. ETL Developer/
Responsibilities:
- Key player in developing Talend jobs to ingest Saudi Arabian Airlines (SVA) and Aircraft Maintenance and Engineering System (AMOS) data into the ADW Inbound layer to support operational metrics for BGS Fleet Care program using Talend Real-time Big Data Platform.
- Currently involved in the development of Flight Following application to support flight scheduling and flight execution data from the Airline Customer to the Gold Care AMOS application by creating services and routes using cAMPQ, cTalendJob, cProcessor, tESBProviderRequest, tJMSOutput, tXMLMap and tESBProviderResponse.
- Implemented FTP operations using Talend Studio to transfer files in between network folders as well as to FTP
- Server using components like tFileCopy, TFileArchive, tFile Delete, tCreateTemporaryFile, tFTPDelete, tFTPCopy.
- Experienced in using debug mode of Talend to debug a job to fix errors. Created complex mappings using tHashOutput, tHashInput, tNormalize, tDenormalize, tMap, tUniqueRow, pivot to Columns Delimited etc.
- Used tRunJob component to run child job from a parent job and to pass parameters from parent to child job. Created Context Variables and Groups to run Talend jobs against different environments.
- Written complex PL/SQL procedures for specific requirements
- Used tParallelize component and multi thread execution option to run sub jobs in parallel to increase job Performance.
- Performed data manipulations using various components like tMap, tJavaRow, tFixedFlowInput, tOracleRow, tJoin, tReplicate, tFilter and many more.
- Used reusable components like routines, context variable and global Map variables.
- Published Talend Jobs as SNAPSHOT to TAC, created CRON Triggers and did a release candidate to support monthly releases as a Talend Administrator.
- Maintaining the project in GitLab VCS as owner and merging the code back to master from individual branches after the end of each release.
- Monitored the daily runs, weekly runs and adhoc runs to load data into the target systems.
Environment: Talend Real Time Big Data Platform for Big Data 6.4.1, MS SQL Server 2016, MySQL, Confidential, Confidential, ActiveMQ, Java, JIRA, Git, Talend Administration Console.
Confidential, SynapsisI, NC
ETL/DatabaseDeveloper
Responsibilities:
- Participate in project Sprint Planning, Refinement and Retro.
- Coordinated with Product owner on a regular basis to discuss the business requirements.
- Involved in the Talend Development Standards document which describes the general guidelines for Talend developers, the naming conventions to be used in the Transformations and development and production environment structures.
- Created complex mappings in Talend 6.2.1 Big Data Edition using tMap, tParallelize, tJavatAggregateRow, tFixedFlowInput, tFlowToIterate, tMSSqlInput, tMSSqlRow etc.
- Created Talend jobs using the dynamic schema feature.
- Worked on Joblets & Java routines in Talend
- Published events to Kafka from Talend and consumed events from Kafka.
- Implemented Error handling in Talend to validate the data Integrity and data completeness for the data from the Flat File.
- Optimized the performance of the mappings by various tests on sources, targets and transformations.
- Wrote complex SQL queries to take data from various sources and integrated it with Talend.
- Developed Type-1 and Type-2 SCD for mappings to current and historical data.
- Incorporated business logic for Incremental data loads on a daily basis.
- Written complex PL/SQL procedures for specific requirements.
- Used Parameter Variables and Mapping variables for incremental data feeds.
- Schedule Talend Jobs using Job Conductor (Scheduling Tool in Talend) - available in TAC.
- Worked with Microsoft SQL Server management Studio while Implementing Unit Testing.
- Provided support to QA team during testing of solutions.
Environment: Talend Data Integration 6.4.1, MS SQL Server 2012, MySQL, Apache Kafka, Java, JIRA, Git, Talend Administration Console.
Confidential, Boston, MA
ETL/Database Developer
Responsibilities:
- Assisted in creating fact and dimension tables as per requirement.
- Used Informatica client tools such as Source Analyzer, Warehouse designer, Mapping designer, Mapplet Designer, Transformation Developer.
- Developed various transformations like Source Qualifier, Update Strategy, Lookup & Expression transformation, Expressions and Sequence Generator for loading the data into target table.
- Developed numerous Complex Informatica Mappings and reusable Transformations.
- Migrated existing Informatica Jobs to Talend Using ETL Talend Studio for Data Integration 6.0.1
- Designed and created complex source to target mappings using various transformations inclusive of but not limited to Aggregator, Joiner, Filter, Source Qualifier, Expression and Router Transformations.
- Configured mappings to handle updates to preserve existing records using Slowly Changing Dimensions SCD Type-2.
- Created stored procedures and functions, as per requirements in PL/SQL.
- Performed Performance tuning of targets, sources, mappings and session.
- Used UNIX scripting while writing cron jobs for batch processing.
- Used Test Director to log and keep a track of defects. Worked with Test team in resolving defects.
- Documented Technical specifications, business requirements and functional specifications for development of ETL mappings to load data into various tables.
Environment: Talend Open Studio for Data Integration 6.0.1, Informatica Powercenter, Confidential 9i, DB2 UDB, Confidential, UNIX, SQL Server, PL/SQL, SQL*Loader.