We provide IT Staff Augmentation Services!

Etl Developer Resume

5.00/5 (Submit Your Rating)

PERSONAL SUMMARY:

  • An experienced ETL developer with good knowledge in Data warehousing & ETL
  • Hands on experience in tools like BO Data Services, Informatica, Teradata, Hadoop, Oracle, Redshift, AWS EMR, AWS S3, Autosys, Hive etc.
  • Expertise in application design and architecture of complex ETL jobs & workflows which supports billons of records.
  • Expertise in script creation and enhancing existing scripts. (Unix shell scripts, Python scripts & java programs, metadata management.)

SKILL:

BI ETL: BO Data Services, Informatica, Teradata utilities, Crystal reports, Apache Storm, Datanet, Horizon

Databases: Oracle, Teradata, PL/SQL coding, Hadoop, Redshift, AWS S3, EMR

Utilities: Autosys, Gitlab, Netty, HIVE

Scripting: Unix shell scripting

Languages: Python, Java, Html, Visual Basic, Visual C++

Platforms: Linux (Redhat), AIX, Windows, Macintosh

PROFESSIONAL EXPERIENCE:

ETL Developer

Confidential

Responsibilities:

  • During the month end, the team compares the amounts between COW and the USLBI to ensure there is no material variance.
  • In case of any issues, the team has to drill down to shipment id level and find the root cause with the help of excel.
  • Due to the huge volume, it is a cumbersome process to do with excel.
  • This project allowed the accounting team to get a high level overview as well as in case of discrepancy drill down to the data at shipment id level.
  • Used Datanet and Horizon as the ETL tools to pull the COW data to EDX and transform using hive queries.
  • Used Datanet jobs to pull the USLBI data to EDX and reconcile the data between COW and USLBI.
  • Then load the final data in redshift on top of which the OBIEE reports will run
  • Horizon uses S3 as an intermediate data storage and used EMR for processing power

Confidential

ETL Developer

Responsibilities:

  • The BI team wanted to identify the pattern of queries used in the Datanet jobs that access a set of tables in the finance space.
  • Once the list is available, the team will generate a template that can be used by the application team so that the performance of those Datanet jobs will be better and the load on the database will reduce.
  • Created a stored procedure to extract the table names from the queries and compared with the tables.
  • Retrieved the metrics of the job name that uses the above jobs with additional information like user id, query.

Confidential

ETL Developer

Responsibilities:

  • Since there were no finite set of objects, had to scan through each objects to check the presence of company code column and notify the BI team about columns that has potential risk of 2 character limitation.
  • Different teams used different naming convention for the company codes. There were no final list for company codes. Hence there were high risk of missing out some company code fields.
  • Created a stored procedure to identify the potential company code columns by identifying the string columns with character length less than four. Matched the sample 1000 values against the company code values.
  • Once the final list of tables were derived, queried the Informatica metadata tables and identified the Informatica mappings and workflows.
  • Built query to Identify the ports used as company codes from the list of Informatica mapping list derived.
  • Provided the final report of the Oracle tables and Informatica Objects that has potential issue of limiting the company code size to two.
  • Provided an additional metrics of the objects that are currently in use and the objects that has not been access even once for a year. This helped the BI team to deprecate those objects and free up space.

ETL Developer

Confidential

Responsibilities:

  • Data from source Oracle is imported using ETL tool Apache storm. Data from flat file and SAP was pulled using Business Objects Data Services.
  • The data from the source system would be pulled to a file and the file will be loaded to the Hadoop core tables.
  • The business logic is applied on the semantic layer on top of core tables and would be loaded to a semantic table.
  • Created hql scripts to apply the business logic on hadoop.
  • The final data from the semantic tables are pulled to a delimited file and send to the external vendor.
  • Autosys jobs are used to schedule the ETL jobs and the core loading jobs.

ETL Developer

Confidential

Responsibilities:

  • The reports will be created on top of these core tables with minimal business logic applied.
  • The script will determine which utility to be used to load the data according to the volume.
  • Creating ETL jobs that will be used as a standard by the development team.
  • Created shell scripts and python scripts to for the manipulating the source data file.
  • Built PL/SQL procedures for applying business logic.
  • Communicate with the business users to understand the requirement.
  • Producing Documentation of the each activity including guidelines regarding the usage.
  • Created scripts to encrypt the output files created by BODS with the GPG encryption format.
  • The script deployment from one environment to another was done with the help of versioning tool like SVN and gitlab.

ETL Developer

Confidential

Responsibilities:

  • Extend service excellence program to iOS devices
  • Provision flexibility in defining and managing service excellence metrics
  • Enable rule based eligibility of repairs that supports exclusion, inclusion and overrides
  • Provide visibility to Service excellence metrics, scores
  • Enable alerting that includes key service excellence score to drive actions
  • BODS 4.0 was introduced to Confidential as POC to reduce the turn around time to pull data from SAP to Teradata for Confidential care.
  • Previously the data from the SAP R/3 systems was pulled to BW so that Informatica 8.5 can pull data and push to a file. This file will be loaded to Teradata core tables.
  • But with the introduction of BODS 4.0, we were able to pull data from SAP R/3 with the help of extractors that removed the dependency of loading data to BW and the use of Informatica.
  • We completed one project end to end and set this project as a template to the project teams as a .
  • Worked as a BODS administrator for 4 years starting with the version BODS 4.0.
  • Setup the BODS environment from scratch identifying and setting up the prerequisites.
  • Installed and upgraded BODS tool in Linux based environment with the front end on Windows.
  • We started the project by installing BODS 4.0 till BODS 4.2 thru BODS 4.1 with different Service Packs and Patches installed during the process.
  • We identified different bugs and reported to SAP and got the bugs fixed by installing the patches or service packs provided by SAP.
  • Worked on creating users and different groups under BODS with full access, execute access, CMC access….
  • Created repositories for different development teams.
  • Worked on migration of objects from one environment to other with the help of central repository with different migration options.
  • Created the framework scripts that will work with the Confidential ’s ETL metadata.
  • Developed projects with BODS scripts, workflows, dataflow & exception handling.
  • Created custom functions in BODS so that it can be shared among the application teams.
  • These projects were shared with the development team as a template and helped them in debugging the issues in case if they are stuck up with any issues.
  • Created shell scripts for supporting different application teams requirement like checksum validation to make sure that the output file that was created by BODS and after sftp is the same and intact.
  • Researched on different options for disaster recovery and setup a disaster recovery system if the existing system is down.
  • Maintaining maximum availability of supported services for users.
  • Worked on setting up jobs that will pull data from HANA, flat files, SAP ERP, & Oracle and loaded data to HANA, Teradata, delimited files and Oracle
  • Had implemented auditing in the dataflow to keep track of the data.
  • Worked in in - cooperating Teradata utilities to work with BODS for fast loading.
  • Teradata stored procedures was created as part of the project development for Confidential care.
  • The data that are pulled from various sources are dumbed into a file using any of the ETL tools like Informatica or BODS and the output file will be loaded to Teradata using the Teradata utilities. Once the data is loaded, the business logic will be applied on the Teradata core tables with the help of procedures, views.
  • The reports will be created on top of these core tables with minimal business logic applied.
  • Created Teradata utilities like tpump and fastload for loading the data to the target Teradata tables.
  • Used bteq scripts for customized data load.
  • Setup SSL for the web server (tomcat) for BODS.
  • Enabled load balancing in the job server and CMC level.
  • Scheduled the BODS jobs with the help of an external scheduler named Autosys.
  • Created a wrapper script to invoke the corresponding BODS job while an Autosys job will get triggered.
  • Setup of automatic cleanup script that is triggered to execute everyday which will cleanup files older that 7 days with the help of crontab, which was recently migrated to Autosys.
  • Enabled access restriction so that the server SOX compliant.
  • Created scripts which will allow the support team to restart the BODS application remotely.
  • Enabled monitoring for system resources and sent alert via mails or tickets to intimate the resource utilization if went past the threshold.
  • We generated reports from the BODS internal metadata tables to get the statistics of job execution to depict a graph in-order to forecast the growth.
  • Documented each and every aspect of BODS including the setup, upgrade, scripts… and got great appreciation from the client.
  • Pulled data from non SAP sources via Informatica 8.5 and loaded the data to file which in-turn was loaded to the Teradata core tables.
  • Worked on the Teradata semantic procedure creation to load data to the semantic layer applying the business logic.
  • Created Informatica workflows that pulls data from Teradata and loads it to Oracle tables so that the external team could access those Oracle tables.
  • The scripts migration from one environment to another was done with the help of SVN

Teradata Developer

Confidential

Responsibilities:

  • Worked extensively in Informatica by creating mappings and workflows.
  • The data was pulled from different source like SAP BW, XML, Oracle and flat files with the help of Informatica.
  • Created complex Teradata procedures that were properly tuned for performance.
  • Worked on existing Unix shell scripts and enhanced it to accommodate the project requirements.
  • Scheduled the Informatica and Teradata jobs using the scheduler Autosys.
  • Worked on different Teradata utilities mainly like bteq, tpump and fastload for loading data to the Teradata database.
  • Involved in the support, maintenance and development of critical systems.
  • Carrying out design, construction, review and defect analysis activities.
  • Responsible for database design and development and Application Tuning
  • Providing suitably tested fixes to defects within the sanctioned delivery process.
  • Producing and updating relevant documentation.
  • Developing in accordance with policies regarding secure coding & standards.
  • Meeting clients face to face and on site to ascertain requirements.
  • Develop software in accordance with agreed design with consideration for usability, software quality and future maintenance.
  • Worked in Teradata query tuning to improve the performance of the existing projects.
  • The scripts migration from one environment to another was done with the help of SVN

SOFTWARE DEVELOPER

Confidential

Responsibilities:

  • Understood the various modules for both the products that was developed using Visual Basic 6.
  • Worked on the salary module in the product that will allow the employee to select the pay frequency Monthly or weekly.
  • Added an additional feature of choice of the currency
  • Had worked with the team on calculating the overtime calculation with a complex logic for workers in different time zones.
  • Created reports for each module with Crystal reports.
  • Worked on modules like Basic Salary and Pay components.
  • Both the tools were developed with VB 6 as front end and backend Oracle and MS SQL server with reporting tool as crystal reports 7.I have worked on all the areas.

SOFTWARE DEVELOPER

Confidential

Responsibilities:

  • Developed a web-based project (Task scheduler) to distribute the tasks among the employees and keep track of the progress of each tasks. This project helped the management to get an insight about the project deliverables.
  • Designed and developed another web based project (Placement Management System) for carrying out the placement activity.
  • Both were developed with J2EE and java script with MSSQL server as backend.

We'd love your feedback!