We provide IT Staff Augmentation Services!

Sr. Etl Developer Resume

3.00/5 (Submit Your Rating)

Farmington Hills, MI

PROFESSIONAL SUMMARY:

  • 8 years of ETL Developer experience in analysis, design and development experience in Data Warehousing and Big Data projects.
  • Strong background in Data processing with proficiency in multiple ETL tools such as Talend, and IBM DataStage.
  • Experience in building Hadoop based ETL workflows to ingest, transform and aggregate data using Spark and Hive.
  • Deep understanding of  Hadoop Architecture and various components such as  HDFS, Job Tracker, Task Tracker, Name Node, Data Node and MapReduce concepts along with HIVE, Pig, Sqoop, Zookeeper. 
  • Strong understanding of Data Warehousing and Big Data Hadoop ecosystems.
  • Expert knowledge of SQL (Oracle, SQL Server) and NoSQL (HBase) databases. Ability to write queries to research, analyze, troubleshoot data issues.
  • Expertise in working with  Cognos  Enterprise Planning and  Cognos  TM1 tools to create budget and forecasting applications. 
  • Worked extensively on loading data from various sources such as text, CSV, RDBMS systems to Cubes by using  ETL, (Turbo Integrate Process)
  • Familiar with data architecture including data ingestion pipeline design,  Hadoop information architecture, data modelling and data mining, machine learning and advanced data processing. Experience optimizing  ETL, workflows.
  • Familiarity and working knowledge of reporting and analytics tools such as Cognos and Tableau.
  • Good knowledge of data warehouse concepts, data integration and data management practices. Experience in Dimensional data modeling, Star and snowflake schemas.
  • Excellent working knowledge of UNIX Shell Scripting and automation of  ETL, processes using Autosys on platforms such as UNI
  • Participated in complete Software Development Life Cycle (SDLC) of Data Integration and Data warehousing projects in an Agile work environment. 
  • Mentored junior staff, perform code reviews, led onsite and offshore team members.
  • Supported applications in production, performed root cause analysis for repeated issues.

TECHNICAL SKILLS :

ETL: Talend 6.3, SSIS, IBM, DataStage 9.1 Analytics Tableau, SSAS, Cognos 10.2,

Database: Oracle (PL/ SQL), SQL Server 2016 (T SQL) Big Data Hadoop, Hive, HBase, Sqoop, HiveQL, SparkSQL

Operating Systems: Windows 7/XP/NT/2000, UNIX (AIX, Linux), MS - DOS 

Scripting/ Others: Python, Git, XML, Agile

PROFESSIONAL EXPEREINCE:

Confidential, Farmington Hills, MI

Sr. ETL Developer

Responsibilities:

  • Extensively used Talend Big Data platform to develop Big Data batch jobs using components such as tMap, tHDFSGet, tHDFSInput, tHDFSConnection, tHDFSPut, tHDFSList, tHIVEInput, tHIVEOutput etc.
  • Created Design and mapping documents and explained to Development team to create the  DataStage Jobs.
  • Developed Python / SparkSQL scripts to load data into HDFS as HIVE tables.
  • Created Hive tables and extensively worked with HiveQL for analysis, transformation and verification of data. Developed common ETL and written Python code to format XML documents which can help to source data from different platforms.
  • Analyzed the SQL  scripts and designed the solution to implement using PySpark 
  • Worked on reading multiple data formats on  HDFS using PySpark.
  • Developed  ETL, jobs using various stages like ODBC Connector, Lookup, Join, Aggregator, Transformer, Sort, Remove Duplicate, Dataset etc. 
  • Performed Data profiling through Information Analyzer client 
  • Performance tuning of the  ETL, jobs and also problem analysis and Issue resolution.  
  • Modularized ETL jobs to increase reusability and ease of maintenance.
  • Scheduled ETL batch jobs using Talend Administration Center.
  • Responsible for knowledge transfer to team members on DataStage, Teradata and other products that are in use for this project. 
  • Extensively used the Hue browser for interacting with  Hadoop components. 
  • Documented the systems processes and procedures for future references. 
  • Reviewed existing programs to identify areas of improvement to reduce complexity and increase processing speed. 

Environment: ETL, Hortonworks (HDP) Hadoop, Talend, HBase, Teradata, Spark, Python(PySpark), Data stage, XML, SQL Server, Agile/Scrum, Unix.

Confidential, Texas 

ETL Developer 

Responsibilities:

  • Created Hortonworks cluster and HDFS connection metadata in Talend.
  • Worked with tMap, tHDFSConnection, tHDFSPut, tFileList, tSort, tHive components.
  • Created Hive tables, HiveQL queries to analyze data on HDFS.
  • Designing and Developing Datastage jobs based on the design document and mapping sheet. 
  • Created Hive tables and extensively worked with HiveQL for analysis, transformation and verification of data.
  • Developed common ETL and written Python code to format XML documents which can help to source data from different platforms. 
  • Modularized ETL jobs to increase reusability and ease of maintenance. 
  • Scheduled ETL batch jobs using Talend Administration Center. 
  • Worked with NoSQL column family database Hbase for storing data in Hadoop. 
  • Modifying the Unix shell scripts which are used to trigger  DataStage jobs 
  • Worked with NoSQL column family database Hbase for storing data in Hadoop.
  • Scheduled ETL batch jobs using Talend Administration Center.
  • Used Pig as  ETL, tool to do transformations, event joins, filter and some pre-aggregations. 
  • Supported applications in production, performed root cause analysis for repeated issues. 

Environment: ETL, Hortonworks (HDP) Hadoop, HDFS, Talend, HBase, Sqoop, SQL Server, Agile, Unix.

Confidential, Manhattan, New York

ETL Developer

Responsibilities:

  • Gathered business and technical requirements through (Joint Application Development) sessions and executive interview to gather business and technical requirements.
  • Running Publish to move the application data from  Cognos Planning to  Cognos BI publish containers for reporting purpose on daily basis. 
  • Developed & Designed functional and technical specification documents, mappings and exception
  • handler for ETL processes and outbound interfaces, in short provided Best Practices guidelines for all ETL developers and Worked on MDM data models.
  • Created, updated and maintained  ETL, technical documentation. 
  • Created Technical design documents TDDs based on BRD/ FRDs. 
  • Designed transformation rules and processes to derive the correct data from the extracted data and transform into the required format and structure to support the business requirements.
  • Used DataStage as an  ETL, tool to extract data from sources systems, loaded the data into the IBM DB2 database and Installation and configuration of database server. 
  • Deployment of packages, folders, reports and cube.
  • Extracts the data from Different source systems and load them in Oracle Database 
  • Created  ETL, Framework tables which are static which will have all the details about source system 
  • Created an UNIX Script which will check for the details in the Framework tables and calls the  DataStage job with appropriate invocation ID.  
  • Worked with Sorter, Sequential file, Datasets, Lookup, Merge, Join, Funnel, Transformer, ODBC, Oracle enterprise, Remove Duplicates Stage etc.
  • Worked on Analysis, Profiling of structured and unstructured data and used data quality techniques such as standardization and matching.
  • Created, updated and maintained  ETL, technical documentation. 
  • Monitored workflows Using Workflow monitor. 
  • Performed Unit testing of mappings and sessions. 
  • Developed and customized  Cognos Impromptu Reports and queried from different database tables as per requirement. Also, built Multi-dimensional cubes using  Cognos Transformer.

Environment: ETL, IBM DataStage 9.1, Oracle 11g, MDM, Cognos 10.2, Agile, UNIX (Linux).

Confidential, Indianapolis, Indiana

ETL Developer 

Responsibilities:

  • Involved in all phases of SDLC (Software Development Life Cycle) including Requirement collection, Design and analysis of Customer specification, Development and Customization of the application. 
  • Lead global teams, provided hands-on participation, technical guidance and leadership to separate data.
  • Re-architected DataStage jobs to maintain logical separation as de-merger progressed.
  • Resolved data issues in Cognos reports post data separation.
  • Extensive experience with Data Profiling. 
  • Defined Target Load Order Plan for loading Target when control table logic is used. 
  • Configured the sessions using Workflow manager to have Multiple Partitions on Source data and to improve performance. 
  • Worked on DataStage upgrade from version 8.1 to version 9.1.
  • Used ETL  ( SSIS ) to develop jobs for extracting, cleaning, transforming and loading data into data warehouse. 
  • Used  DataStage stages namely Datasets, Sort, Lookup, Peek, Standardization, Row Generator stages, Remove Duplicates, Filter, External Filter, Aggregator, Funnel, Modify, and Column Export in accomplishing the ETL  coding. 
  • Managed migration of Data Center and DataStage Applications from SunGard to KDC domain.
  • Migrated existing DataStage jobs, installed ODBC drivers and tested jobs in the new environment.
  • Optimized and automated ETL process using new DataStage 9.1 features and Unix Shell scripts.
  • Lead offshore and onshore developers and coordinated the ETL work.
  • Create joins and sub-queries for complex queries involving multiple tables. 
  • Resolved the defects from UAT, and delivered an ETL system that is automated, stable and efficient.

Environment: ETL, IBM DataStage 9.1, DataStage 8.1, UNIX (Linux), Oracle 9i, Oracle 11g, Cognos 10.2, Agile, UNIX.

Confidential, Charlotte, NC

DW & BI Consultant

Responsibilities:

  • Extensively worked on Data Acquisition and Data Integration of the source data from all regions.
  • Developed & Designed functional and technical specification documents, mappings and exception handler for ETL processes and outbound interfaces, in short provided Best Practices guidelines for all ETL developers.
  • Developed ETL jobs using IBM DataStage and Oracle for integrating BAAN and Pcard data and implemented business rules.
  • Data Extraction, Transformation and Loading from source systems. 
  • System Testing of the Data Mart and Data Warehouse  
  • Developed DataStage Parallel Extender jobs using stages like Aggregator, Join, Merge, Lookup, Source dataset, Row generator, Column Generator, Change Capture, Copy, Funnel, Peek stages etc.
  • Provided technical support post production.
  • Design the  ETL, processes using  DataStage tool to load data to target DB2 database. 
  • Developed and customized  Cognos  Impromptu Reports and queried from different database tables as per requirement. Also, built Multi-dimensional cubes using  Cognos  Transformer.
  • Conducted unit test and involved on preparing the test data. 
  • Daily production support for all jobs.

Environment: IBM DataStage 8.1, ETL, UNIX (AIX), SQL, Oracle 9i, Cognos, Agile, Windows.

We'd love your feedback!