Specialist Data Engineering Resume
Tampa, FL
SUMMARY
- Specialize in design and development and implementation of complex, large warehouse in SQL Server, AWS.
- Experience in designing streaming data pipeline in AWS using services like AWS Kinesis, AWS Lambda, S3, Secrets Manager, SQS, SNS, Redshift.
- Experience in designing stored procedures in MySQL for complex business scenarios.
- Experience in writing Python scripts to solve complex business problems.
- Responsible for architecting ETL jobs in Talend and integrating them with the AWS Services.
- Strong experience in using ETL methodologies for supporting Data Extraction, Data Migration, Data Transformation, Data Integration, and loading using Talend 6.2/6.5/7.1/7.3.
- Involved in various projects related to Data Modeling, System/Data Analysis, Design and Development for both OLTP and Data warehousing environments.
- Experience in using Talend Data Studio,TBDandTalend Data Fabrictools (Talend DI, Talend MDM, Talend DQ, Talend Data Preparation, TAC)
- Experienced in using AllTalend DI,MDM, DQ, DPcomponents.
- Familiar with terraforming the services in AWS to build the data pipeline.
- Experienced inmatch/mergeinMDMto runmatch rulesto check the effectiveness ofMDMprocess on data.
- Experienced inData Ingestionprojects to inject data intoData Lakeusing multiple sources systems usingTalend Bigdata
- Experienced in scheduling Talend jobs usingTalend Administration Console (TAC).
- Experienced in scheduling Talend jobs usingAutosys Portal.
- Experienced in ETLTalend Data Fabriccomponents and used features ofContext Variables, MySQL, Oracle, Hive Database components.
- Familiar with dataarchitectureincludingdata ingestion pipeline design, Hadoop information architecture,data modellingand data mining, machine learning and advanced data processing and Experience optimizing ETL workflows.
- Good experience in Big Data and Hadoop Ecosystem components like, Pig, Hive, Sqoop, Flume, MapReduce.
- Experience inDebugging, Error HandlingandPerformance Tuningof sources, targets, Jobs etc.
- Expert in all areas of SQL Server development including tables, user defined functions, views, indexes, stored procedures, triggers, clustering, T - SQL, relational database models, data dictionaries, and data integrity.
- Experience in Design and implementation of PL/SQL Stored Procedures, Views, Functions, Packages, Ref Cursors, Constraints, Database Triggers for Oracle Relational databases
- Experience in Interacting with Business users to analyze the business process and transforming requirements into screens, performing ETL, documenting and rolling out the deliverables.
- Experience in managing Alerts, Operators and Scheduled Jobs through SQL Server Agent.
- Good team player, strong interpersonal and communication skills combined with self-motivation, initiative, and the ability to think outside the box.
TECHNICAL SKILLS
ETL Tools: SQL Server Integration Services (SSIS), Talend 6.2, Talend 6.5, Talend 7.1, Talend 7.3
Programming Languages: Python, T-SQL, C#, SQL, PL/SQL
GUI: DTS Import/Export, SQL server Integration Services (SSIS), SQL Server Reporting Services (SSRS), SQL Server Analysis Services (SSAS) DBForge, DBeaver
Application Software: Microsoft Visual Studio, Talend.
Databases: MS SQL Server 2016, 2012,2008/2005, Oracle 19c/12g/11g, AWS Redshift, SQL Workbench, Aginity Workbench, Toad, MYSQL
Modelling Tool: Database Designer
PROFESSIONAL EXPERIENCE
Confidential, Tampa, FL
Specialist Data Engineering
Responsibilities:
- Gathering development requirement from client’s functional team and from functional design specification documents.
- Application Monitoring in NAM/EMEA for UAT/PROD environments.
- Helped troubleshoot Talend batch jobs/other application components in case of issues highlighted. Analyzed, troubleshoot and resolve defects coming in daily/health check reports.
- Created logical and physical objects- table spaces, tables, indexes, referential integrity constraints and triggers to development needs.
- Optimized the oracle PL/SQL queries by performance tunning
- Data load through TALEND workflows and mapping
- Involved in performance tuning using Indexes
- Develop scripts to automate ETL jobs using Shell Scripting.
- Executed Jobs through Autosys Portal
- Attending daily status meeting and other meetings with end client to take proper clarifications
Environment: Talend 7.3, Tectia.
Confidential, GA
Talend Developer
Responsibilities:
- Built an audit framework on Oracle/PLSQL that logs each time a Talend batch & DI/Big Data Job is run in production.
- All the context values, db. connection is stored in external files and using joblet’s pulling all the context information into the main job.
- Executed the jobs through TAC by publishing the job to Nexus Artifact Repository and creating an Artifact task.
- Used over 20+ Components in Talend Like (tMap, tjava, tlogrow, toracleInput, toracleoutput, tsendEmail etc)
- Used debugger and breakpoints to view transformations output and debug mappings.
- Worked on Error handling techniques and tuning the ETL flow for better performance.
- Worked Extensively TAC (Admin Console), where we Schedule Jobs in Job Conductor.
- Extensively Used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, thashoutput, thashinput, tRowGenerator, tOracleInput, tOracleOutput, tfileList etc.
- Migrated the code and release documents from DEV to QA (UAT) and to Production.
Environment: Talend 7.1, Oracle Server.
Confidential, Atlanta, GA
Data Engineer
Responsibilities:
- Worked with stakeholders to gather the requirements for streaming application.
- Created a Kinesis data stream with 4 shards to receive streaming data from sensors of vendors.
- Kinesis stream triggers the Lambda in one environment. Then the same data must be fanned out to multiple environments. Hence created the fan-out Lambda.
- Also, kinesis firehose dumps the data to S3 in few scenarios. Created S3 event trigger to trigger the Lambda function.
- Raw data ingested is dumped with minor transformations to staging table in MySQL.
- Created multiple Stored Procedures in MySQL to process the raw data. And scheduled the stored procedures to run one after another for complete data processing via AWS CloudWatch & Lambda.
- Created failure detection in every process to make the process robust.
- Helped in terraforming the entire process.
- Automated the processes to run the pipeline end to end.
Environment: AWS Kinesis, Lambda, S3, Glue, Redshift, MySQL, Python, CloudWatch.
Confidential, Atlanta, GA
Talend Developer
Responsibilities:
- Built an audit framework on SQL Server that logs each time a Talend batch & DI/Big Data Job is run in production.
- All the context values are stored in SQL Server database, database credentials are stored in a file on Talend Job Server which was read in DI job using tfileinputdelimited and loaded into contexts using tcontextload. Database connections were created using tmssqlconnection.
- Built 2 Talend DI Framework jobs - The main framework job logs batch run details into batch master & batch run log tables using tmssqlrow and trigger the child framework job which logs job run details into job run log & job master tables and trigger multiple Spark Batch Jobs parallelly using tparallelize.
- Designed 3 different kinds of Spark Batch jobs for different business tables. Job 1 - Reads data from SQL SERVER and writes it to s3 (tJDBCConfiguration, tS3Configuration, tHDFSConfiguration, tJDBCInput, tSQLRow, tFileOutputDelimited). Job 2 - Reads data from the s3 file, lookups data from s3 files, applies ods transformation logics and writes it to s3 again (tMap, tPartition, tFileInputDelimited). Job 3 - Reads data from Redshift Spectrum® Tables and applies dimension & fact logics and writes the data to s3, then copies the data from s3 to Redshift DW (tRedshiftConfiguration, tRedshiftInput)
- Designed a DI job to copy data from s3 to Redshift tables.
- Executed the jobs through TAC by publishing the job to Nexus Artifact Repository and creating an Artifact task.
- Involved in Data Extraction from Flat files and XML files using Talend by using Java as Backend Language.
- Using Talend to load the data into our warehouse systems
- Used over 20+ Components in Talend Like (tMap, Tfilelist, Tjava, Tlogrow, ToracleInput, ToracleOutput, tsendEmail etc)
- Used debugger and breakpoints to view transformations output and debug mappings.
- Load and transform data into HDFS from large set of structured data /Oracle/Sql server using Talend Big data studio.
- Develop ETL mappings for various Sources ( .TXT, .CSV, XML ) and load the data from these sources into relational tables with Talend Enterprise Edition.
- Worked on Global Context variables, Context variables, and extensively used over 50+components in Talend to create jobs.
- Extracting transformed data from Hadoop to destination systems, as a one-off job, batch process, or Hadoop streaming process.
- Worked on Error handling techniques and tuning the ETL flow for better performance.
- Worked Extensively TAC (Admin Console), where we Schedule Jobs in Job Conductor.
- Extensively Used Talend components tMap, tDie, tConvertType, tFlowMeter, tLogCatcher, tRowGenerator, tOracleInput, tOracleOutput, tfileList etc.
- Migrated the code and release documents from DEV to QA (UAT) and to Production.
Environment: Talend Data Integration 6.2,6.5,7.1.
Confidential, Atlanta, GA
Senior SQL BI Developer
Responsibilities:
- Involved in gathering the requirements, conducted meetings with client and documented the requirements and ITIL assessment & implementation of project.
- Responsible for the design, development and implementation of mappings/transformations based on source-to-target specifications, defining ETL (extract, transform, load) and ETL development standards and conventions
- Created views to reduce apparent database complexity for end users and preventing sensitive columns from being selected.
- Created the indexes on columns having high selectivity and low density to speed up the execution of data retrieval queries and stored procedures.
- Modified current database by adding tables, altering referential integrity and relationships to support online system.
- Created SSIS Packages to load data from different sources like Flat Files, Excel files and ERP’s like Oracle, DB2 and Sql Server.
- Created Jobs to load data using packages, Scheduled, and maintained SSIS packages.
- Optimized the performance of SSIS packages.
- Identified, tested, and resolved database performance issues (monitoring and tuning) to ensure database optimization.
- Created Ad-hoc reports, Tabular reports, Drill down and Drill through reports.
- Generated reports using SSRS, Excel (Pivot) that could be used to send information to diverse group of users like VP, AVP, CD, Coach, Lead, SA, SR, Office Clerk.
- Developed MDX Scripts to create datasets to perform reporting and included interactive drill down reports, report models and dashboard reports.
- Scheduled reports for daily, weekly, monthly reports for executives, Business analyst and customer representatives for various categories and regions based on business needs using SQL Server Reporting services (SSRS).
- Developed new mappings using SSIS Look Up/Aggregation transformations to match the needs of the business case.
- Converted data using Derived Column, Data conversion transformation and Script task.
- Used ETL (Extraction, Transformation, Loading) from heterogeneous source systems. Assisted DBA in using the repository manager to give permissions to users to create new users and repositories.
- Converted DTS packages to SSIS 2012
- Created DTSX Packages to import and export data from Excel Spreadsheets, People Soft, Text files and CSV files.
- Executed the DTSX Packages by using DTExec Command.
- Designed slowly changing dimension transformations (SCD) and customized the existing mappings.
- Developed Data Mart using Star Schema. Reviewed the Current Data Mart and OLAP Cubes and Proposed Corrections and enhancements. Environment: SQL Server 2012/2008 Enterprise Edition, Toad, Microsoft Visual Basic, VBScript, XML, Microsoft Visio, Windows 2003, Microsoft Reporting Service.
Environment: SQL Server 2012/2008 Enterprise Edition, Toad, Microsoft Visual Basic, VBScript, XML, Microsoft Visio, Windows 2003, Microsoft Reporting Service.
Confidential
SQL Server Developer
Responsibilities:
- Develop, optimize, and review queries for data retrieval and update.
- Worked closely with team to get various requirements. Involved in design, implementation of Technical Specifications, Design Documents and Approach Documents.
- Actively involved in designing databases for the system.
- Involved in ER diagrams (Physical and Logical using Erwin) and mapping the data into database objects.
- Managing historical data from various heterogeneous data sources (i.e. Excel, DB2 and Access).
- Extracted data using SSIS from OLTP to OLAP.
- Created complex stored procedures and functions for integrity checks of transferred data.
- Performed unit testing and SQL performance tuning.
- Tuned SQL statements Using Indexes and stored procedures.
- Developed data access queries/programs using SQL Query to run in production on a regular basis and assists end users with development of complex Ad Hoc queries.
- Worked on implementation and maintenance of disaster recovery solutions such as replication and log shipping.
- Responsible for query optimization using MS SQL tools such as Query Analyzer, SQL Profiler, index tuning wizard.
- Created alerts, notifications, and emails for system errors, insufficient resources, fatal database errors, hardware errors, and security breach.
Environment: SQL Server 2005, TFS, T-SQL, SSIS, Windows 2003, Shell Scripting.