Etl Tech Lead Resume
Detroit, MI
SUMMARY
- 8 years of experience in full life cycle of software project development in various areas like design, Applications development of Enterprise Data Warehouse on large scale development efforts leveraging industry standard using Talend
- 6+ years of experience using Talend Data Integration/Big Data Integration (7.x/6.x/5.x) / Talend Data Quality.
- Extensive knowledge of business process and functioning of Event Management/Marketing, Health Care, Manufacturing, Mortgage, Financial, Retail and Insurance sectors.
- Extensive experience in ETL methodology for performing Data Profiling, Data Migration, Extraction, Transformation and Loading using Talend and designed data conversions from wide variety of source systems including Netezza, Oracle, DB2, SQL server, Teradata, Hive, Hana and non - relational sources like flat files, XML and Mainframe files.
- Expertise in processing data from HDFS, HBase using Hive, Sqoop and Pig components.
- Used Spark and MapReduce frame works to configure the big data batch jobs.
- Strong understanding of NoSQL databases like HBase, MongoDB.
- Expertise in Data modeling techniques like Data Modeling- Dimensional/ Star Schema and Snowflake modeling, Slowly Changing Dimensions (SCD Type 1, Type 2, and Type 3).
- Excellent working experience in Waterfall, Agile methodologies.
- Created mappings using Lookup, Aggregator, Joiner, Expression, Filter, Router, Update strategy and Normalizer Transformations. Developed reusable Transformation and Mapplets.
- Expertise in deploying from DEV to QA, UAT and PROD with both Deployment group and Import/Exports method.
- Experience in gathering and writing detailed business requirement and translating them into technical specifications and design.
- Expertise in understanding and supporting the client with requirements definition, analysis, design, testing, system documentation and user training.
- Extensive experience in gathering requirements and documenting the same as per the industry best practices.
- Designed the data conversion strategy, development of data mappings, source data profiling and the design of Extraction, Transformation and Load (ETL) routines for migrating data from non-relational or source relational to target relational.
- Strong analytical, logical, Communication and problem-solving skills and ability to quickly adapt to new technologies by self-learning.
TECHNICAL SKILLS
ETL/BI Tools: Talend Data Integration/ Big Data Integration/ Data Quality / 7.1/6.3/6.1/5.5/5.0 , Talend Administrator Console, Talend Repository manger, Tableau.
Databases: Redshift, Salesforce, Oracle 11g/10g, DB2 9.x/8.x, Teradata, Netezza, SQL Server, Hive, HBase, MongoDB, RDBMS, MDM.
Programming Languages: XML, PL/SQL, SQL,SOQL,Apache Spark SQL,Hive SQL, T-SQL, Unix/LINUX shell programming and scripting
Development Tools: Teradata SQL Assistant, SQL developer, SQL Server Management Studio, Toad, iSeries navigator, QMF for windows, SQL*Loader, Aginity Workbench
Operating Platforms: Windows 10/8/ 7/XP/2000, Unix, Linux, AIX, Mac OS
Other Tools: AutoSys, CA7 Job Scheduler, Maximo, Jira, Tivoli Ticketing System, Altona XML Spy, SQL *Plus, Putty
PROFESSIONAL EXPERIENCE
Confidential, Detroit, MI
ETL Tech Lead
Responsibilities:
- As an ETL Tech lead, lead the ETL development for enhancements in Data warehouse.
- Working closely with the business users to understand the requirements and converting them into project level technical capabilities.
- Worked with business analysts to identify the appropriate data elements for required capabilities.
- Update the status and plan the releases through the scrum meetings.
- Coordinating with team and providing the inputs.
- Worked with source teams to find out the source team changes.
- The project involved developing mappings for moving data from Flat files to Staging Area (STG) and then to Data Warehouse (DWH) and then to Data Mart.
- Developing the ETL detail design documents for each target tables (Fact and dimension tables).
- Creating primary objects (tables, views, indexes) required for the application
- Designing ETL jobs. Used Talend as ETL tool.
- Designed and developed complex mapping for varied transformation logic like Expression, Filter, Aggregator, Joiner Update Strategy, Unconnected and Connected lookups
- Used Debugger to troubleshoot logical errors and runtime errors.
- Designed and developed common modules for error checking (e.g. to check if the reject recordsoutput file is empty and to check if there are duplicate natural keys in a given table.)
- Performed the tuning at source, Target and mappings using Indexes, Hints and Partitioning in DB2, SQL Server.
- Prepared Test Plans for Unit test and System Integrated testing and Workflows.
- Co-ordinate with the QA team in various testing phases by resolving the defects and ensuring smooth execution of the test plans.
- Creating the deployment documents and migrating the code to the production environment.
- Investigating and fixing the bugs occurred in the production environment and providing the on-call support.
Environment: Talend Data Integration 7.3.1/7.1.1 , TalendBig Data Platform 7.3.1/7.1.1 , Amazon Web Services, Aginity Workbench, Talend Administrator Console, Talend Repository Manger,SQl, Hive, HDFS, XML files, Flat files, Putty.
Confidential, Reston, VA
Sr. ETL Developer
Responsibilities:
- Participated in JAD sessions with business users and SMEs for better understanding of the requirements.
- Design and developed end-to-end ETL process from various source systems to Staging area, from staging to Data Marts.
- Analyzing the source data to know the quality of data by using Talend Data Quality.
- Broad design, development and testing experience with Talend Integration Suite and knowledge in Performance Tuning of mappings.
- Developed jobs in Talend Enterprise edition from stage to source, intermediate, conversion and target.
- Involved in writing SQL Queries and used Joins to access data from Redshift DB.
- Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats.
- Solid experience in implementing complex business rules by creating re-usable transformations and robust mappings using Talend transformations like tConvertType, tSortRow, tReplace, tAggregateRow, tUnite etc.
- Developed Talend jobs to populate the claims data to data warehouse - star schema.
- Creating Cluster metdata manually and automatically
- Schedule Big Data Job execution from the Talend Administration Center (TAC) and enriching logs.
- Developed mappings to load Fact and Dimension tables, SCD Type 1 and SCD Type 2 dimensions and Incremental loading and unit tested the mappings.
- Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats into a Database table to record job history.
- Integrated java code inside Talend studio by using components like tJavaRow, tJava, tJavaFlex and Routines.
- Experienced in using debug mode of talend to debug a job to fix errors.
- Created complex mappings using tHashOutput, tHashInput, tNormalize, tDenormalize, tMap, tUniqueRow. tPivotToColumnsDelimited, etc.
- Used tRunJob component to run child job from a parent job and to pass parameters from parent to child job.
- Used Link and HDFS commands on EMR and Spark Shell script.
- Created Context Variables and Groups to run Talend jobs against different environments.
- Used tParalleize component and multi thread execution option to run subjobs in parallel which increases the performance of a job.
- Implemented FTP operations using Talend Studio to transfer files in between network folders as well as to FTP server using components like tFileCopy, TFileAcrchive, tFileDelete, tCreateTemporaryFile, tFTPDelete, tFTPCopy, tFTPRename, tFTPut, tFTPGet etc.
- Experienced in Building a Talend job outside of a Talend studio as well as on TAC server.
- Experienced in writing expressions with in tmap as per the business need.
- Handled insert and update Strategy using tmap.
- Used ETL methodologies and best practices to create Talend ETL jobs.
- Extracted data from flat files/ databases applied business logic to load them in the staging database as well as flat files.
Environment: Talend Data Integration 7.1.1/6.3.1 , TalendBig Data Platform 7.1.1/6.3.1 , Amazon Web Services, Aginity Workbench, Talend Administrator Console, Talend Repository Manger,SQl, Hive, HDFS, XML files, Flat files, Putty.
Confidential, Lombard, IL
ETL Developer
Responsibilities:
- Developed high level technical design specification and low-level specifications based on the business requirements.
- Developed jobs, components and Joblets in Talend. Designed ETL Jobs/Packages using Talend Integration Suite (TIS).
- Design and developed end-to-end ETL process from various source systems to Staging area, from staging to Data Marts.
- Involved in writing SQL Queries and used Joins to access data from Oracle,MySQL, Redshift.
- Expedited in writing SOQL Queries on Salesforce objects.
- Worked on various salesforce.com Standard Objects like Accounts, Contacts, Cases, Leads, Campaigns, Reports and Dashboards.
- Created buttons (URL, VFP and Link), custom objects, page layouts, validation rules and workflows in the salesforce UI.
- Used salesforce Data Loader for insert, update, and bulk import or export of data from Salesforce.com Objects.
- Used it to read, extract, and load data from comma-separated values (CSV) files.
- Defined lookup and master-detail relationships on the objects and created junction objects to establish connectivity among objects.
- Used tStatsCatcher, tDie, tLogRow to create a generic joblet to store processing stats.
- Solid experience in implementing complex business rules by creating re-usable transformations and robust mappings using Talend transformations like tConvertType, tSortRow, tReplace, tAggregateRow, tUnite etc.
- Created Talend Mappings to populate the data into dimensions and fact tables.
- Developed complex Talend ETL jobs to migrate the data from flat files to database.
- Integrated java code inside Talend studio by using components like tJavaRow, tJava, tJavaFlex and Routines.
- Experienced in using debug mode of talend to debug a job to fix errors.
- Created complex mappings using tHashOutput, tHashInput, tNormalize, tDenormalize, tMap, tUniqueRow. tPivot To Columns Delimited, etc.
- Have used AWS components (Amazon Web Services) - Downloading and uploading data files (with ETL) to AWS system using S3 components.
- Used more components in Talend and Few to be mentioned: tjava, toracle, txmlMap, tdelimited files, tlogrow, tlogbackcomponents etc. in many of my Jobs Design
- Collected data using Spark Streaming from AWS S3 bucket in near-real-time and performs necessary
- Implemented Talend jobs to Extract data from Redshift as a flat files and load data into Salesforce objects using tSalesforceBulkExec, tSalesforceConnection,tSalesforceInputetc components
- Used tHDFSConfiguration and tS3Configuration components to access the data from s3 and HDFS when we run the job on Amazon EMR.
- Storing and reading the parquet files and csv files using tFileOutParquet, tFileInputParquet,tfileInputDelimited and tfileOutputDelimited.
- Developed a process to identify the personal health information (PHI) and mask them with the predefined reference keys.
- Created and Subscribed SNS Topics for the email notifications from the jobs on failure.
- Expertise in deploying from DEV to QA, UAT/CERT and PROD with both Deployment group and Import/Exports method.
- Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
- Worked on Talend Repository Manger (TRM) for moving the jobs from one environment to other.
- Strong experience in working with Elastic MapReduce(EMR)and setting up environments on Amazon AWS EC2 instances
Environment: Talend Data Integration 7.1.1/6.3.1 , TalendBig Data Platform 7.1.1/6.3.1 , Amazon Web Services, Aginity Workbench, Talend Administrator Console, Talend Repository Manger, Oracle 11g, SQl, SOQL, APEX, Visual Force, Data Loader, Hive, HDFS, XML files, Flat files, Putty.
Confidential, Irving, TX
ETL Developer
Responsibilities:
- Participated in all phases of development life-cycle with extensive involvement in the definition and design meetings, functional and technical walkthroughs.
- Storing and processing tables from HDFS using Hive Components.
- Connecting to Hadoop cluster to Process Data on HDFS using Big Data Batch Jobs.
- Converting normal jobs on talend to bigdata batch jobs.
- Ingested the data from various heterogeneous sources into HDFS.
- Generating and monitoring raw logs and enriched logs using the batch jobs and reports.
- Creating Hive and Pig metadata manually.
- Created tables, partitions and views on Hive and loaded data from various sources.
- Utilized Big Data components like tHDFSInput, tHDFSOutput, tPigLoad, tPigFilterRow, tPigFilterColumn, tPigStoreResult, tHiveLoad, tHiveInput, tHbaseInput, tHbaseOutput, tSqoopImport and tSqoopExport.
- Used ETL methodologies and best practices to create Talend ETL jobs.
- Created and deployed physical objects including custom tables, custom views, stored procedures, and Indexes to SQL Server for Staging and Data-Mart environment.
- Created Talend jobs to copy the files from one server to another and utilized Talend FTP components.
- Extensively used tMap component which does lookup & Joiner Functions, tjava, tOracle, txml, tdelimtedfiles, tlogrow, tlogback components etc. in many of my Jobs Created and worked on over 100+components to use in my jobs.
- Used Talend most used components (tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tSetGlobalVar, tHashInput&tHashOutput and many more).
- Created many complex ETL jobs for data exchange from and to Database Server and various other systems including RDBMS, XML, CSV, and Flat file structures.
- Created Implicit, local and global Context variables in the job.
- Worked on Talend Administration Console (TAC) for scheduling jobs and adding users.
- Developed stored procedure to automate the testing process to ease QA efforts and reduced the test timelines for data comparison on tables.
- Worked Extensively on Talend Admin Console and Schedule Jobs in Job Conductor.
- Involved in production n deployment activities, creation of the deployment guide for migration of the code to production, also prepared production run books.
Environment: Talend Data Integration 6.1/5.5.1, TalendBig Data Platform 6.0.1/5.5, Talend Administrator Console, Oracle 11g, Hive, HDFS, Sqoop, SQL Navigator, Mongo DB 3.0.12, XML files, Flat files, HL7 files, JSON 2.4.1, Toad, Control M, Putty, Winscp.
Confidential
SQL Developer
Responsibilities:
- Analyzed reports and fixed bugs in stored procedures using SSRS.
- Used complex expressions to group data, filter and parameterize reports.
- Created linked reports and managed snapshots using SSRS.
- Performed various calculations using complex expressions in the reports and created report models.
- Generated Complex SSRS reports like reports using Cascading parameters, Snapshot reports Drill-down Reports, Drill-Through Reports, Parameterized Reports and Report Models and ad hoc reports using SSRS based on Business Requirement Document.
- Conducted performance tuning of complex SQL queries and stored procedures by using SQL Profiler and index tuning wizard.
- Provided Production support to analyze and fix the problems and errors on daily basis by modifying SSIS Packages and Stored Procedure if necessary.
- Designed and developed Tables, Stored procedures, Triggers and SQL scripts using TSQL, Perl and Shell scripting for enhancements and maintenance of various database modules
Environment: MS SQL Server 2005/2008, SSRS, SSIS, SSAS, T-SQL, Erwin, SQL Explorer.