We provide IT Staff Augmentation Services!

Lead Etl Developer Resume

4.00/5 (Submit Your Rating)

Atlanta, GA

SUMMARY

  • Over 15 plus years of extensive IT experience with special emphasis on Software Architecture, Analysis, Design, Development, Integration testing, end to end testing, Production support and worked in Agile/Scrum iterative development model, and Software development life cycle (SDLC) waterfall methodologies.
  • Certified in Informatica Designer, IBM Web Sphere Data Stage Designer, Oracle 9i/11G PL/SQL Programming and Teradata V2R5 SQL Specialist tools.
  • Extensively worked in various functional areas and domains like Banking, Financial, Telecom, Retail, Hotels &Hospitality, Automobile, etc.
  • Experienced in Designing Process Flow and scheduling of the system between various Interfaces i.e. Project High - Level Design and Low-Level design. Involved in Dimensional Data modeling, Star Schema/snowflakes schema, Bill Inmon, Ralph Kimball DW approaches, dimension and Fact tables.
  • Expertise in Extract Transform and loading of data (ETL Processes), Decisions Support systems, Data Warehousing & Online Analytical Processing technology using DataStage, Informatica, Teradata, Talend, snowflake, data lakes, data pipelines, Amazon web services (AWS) and Big Data Hadoop tools along with other BI Tools.
  • Programming experience in Teradata, Oracle 12c/ Exadata, Oracle PL/SQL, MS SQL Server, Unix scripting, Java, Python etc. Involved in Database scripting, performance tuning of the queries and analyzing the SQL and PL/SQL scripts and fixing the data and logical issues, data quality and end to end validations.
  • Bigdata Hadoop framework data integration, migration of historical data, ETL, and importing and exporting of data using Sqoop to load data from oracle to Hadoop HDFS file system and HIVE SQL interface to query data from HDFS file system environment. Knowledge in Java, Python, Hadoop YARN, MapReduce programs and scripts for data processing and knowledge in Spark and Scala. Apache Hadoop Hue (Hadoop user experience) tool for checking the Bigdata Hadoop HDFS folders in various data lakes and Oozie for workflow and scheduling of various jobs.
  • Extensive knowledge in Amazon web services RDS oracle instances and Amazon Redshift data warehouse databases, AWS Cloud Watch to monitor various services and instances. Worked on the development environment AWS EC2 instance, S3 storage service, Oracle RDS instance, redshift, Elastic map reduce (EMR) cluster. Gained knowledge in Microsoft Azure databases, Logic Apps and migration services.
  • Effective in cross-functional and global/international environments with Onsite/Offshore development models and to manage multiple tasks & assignments concurrently both as a leader and team player with effective communication skills including mentoring developers.

TECHNICAL SKILLS

DW and ETL Tools: Worked in Data Stage Designer, Manager, Director, Informatica Power Center Mapping Designer, Workflow Manager and Monitor, Abinitio, Teradata Utilities (Fast Load, Multi Load, Fast Export, Tpump, TPT scripts, BTEQ scripting), Oracle Exadata with SQL Developer, PL/ SQL scripting, Toad, Teradata SQL Assistant, Unix Shell Scripting, Talend.

Database Systems: Worked in Teradata 15, Oracle 11g/ 12c / Exadata, PL/SQL, Microsoft SQL Server, Sybase DB2 UDB, Oracle Enterprise Manager (OEM), Teradata Viewpoint.

Scheduling Tools: Autosys, CA Workload Automation (CAWA), Tiwoli workload scheduler (TWS), Cisco Dollar Universe tool, Confidential ESP Scheduling, Amex EngineG.

Versioning Tools: GIT Hub, Serena Change Man DS Client, CVS code versioning tool, Power systems and Visual Source Safe tools, TortoiseSVN versioning tool, Google Bit Bucket versioning.

Business Intelligence Tools: Knowledge in Business Objects XI R2, Micro strategy, OBIEE, Tableau Data virtualization.

Programming Languages: Knowledge in Core Java, Java Scripts, Python, C# .Net, UNIX and Shell Scripting, Spark, Scala.

Amazon Web Services (AWS) and Big data services: EC2, S3, RDS, Redshift, DynamoDB, Elastic Map Reduce (EMR) and Hadoop framework HDFS, Map Reduce, Hive, Scoop, Oozie, Pig, HBase, Spark, Cassandra, etc.

Google Cloud Platform (GCP) and Azure Services: Compute Engine, Google Storage Buckets, Bit Buckets, Cloud SQL, Big Table, Big Query, Cloud Data flow, Data Proc, Cloud Publisher Subscriber services, Azure databases, Logic Apps and migration services.

Supporting and Process Tools: Knowledge in Version One Agile Software, Rally Agile tool, Confluence and Jira Agile, Microsoft Project Plan, BMC Remedy Incident and Change Management requests, CA Erwin Data modeling, Gromit tool for Data Architect & Metadata data management and Microsoft Project Plan (MPP) tools, MSBI (SSIS, SSRS), Amazon Web services (AWS).

PROFESSIONAL EXPERIENCE

Confidential, Atlanta, GA.

Data Architect / Lead ETL Developer

Responsibilities:

  • Requirements gathering and designing data conversion process, created High level interface design document along with low level Detail Design documents. Understanding the existing system and domain knowledge along with process flows and technical architecture of the applications. Followed the data architect standards, procedures, guidelines and worked with various teams for reviews and approval process.
  • Worked on agile projects, back-log stories and performed grooming, prioritizing prior to the Sprint Planning/Iterations and worked with product owner, scrum master, solution architects, business teams, various testing teams in Unit testing, UAT, Integration testing, E2E testing.
  • Development, testing of informatica mappings, mapplets, Source qualifier override queries, reusable transformations, lookups and complex business logic development, debugging, validations of data from source system to target systems, created several sessions with partitioning, push down optimization and kick off workflows development with various tasks flows.
  • Developed informatica mapping with change data capture logic for incremental data loads, implemented slowly changing dimensions for history data loading and parameter file generations for Dev, QA and Production environments. Creation of concurrent worklets and execution of concurrent workflows for various business to business financial end clients.
  • Worked on new Teradata tables, base views, business views, financial views with partitioning, UPI, USI, compression values, data modelling in CA Erwin, reviewing with Teradata DBA and Architect forum and materializing in Dev, QA, QI, QS and production environments.
  • Development of Teradata MultiLoad, TPump, BTEQ scripts, Teradata FastLoad, Teradata parallel transporter (TPT), stored procedures, macros with PL/SQL scripts, query tuning, performance improvement, and reviewing the scripts.
  • Worked on development and creation of various data conversion Oracle Packages header/body procedures, Stored Procedures development with complex transformation logics, PL/SQL with business logic transformations.
  • Developed Oracle PLSQL scripts with data merge, insert select logic scripts, oracle analytical functions for complex logic and stored procedures, dynamic SQL’s, calling autonomous stored procedures in packages. Performance tuning of the stored procedures, query tuning, adding indexes, gather stats, fixes and maintenance of the PLSQL objects.
  • Developed Hive scripts to perform transformations on the data ingested from upstream systems, and load data into google cloud storage and then into the hive internal and external tables.
  • Knowledge on Google Cloud Platform (GCP) Storage Buckets, Bit Buckets for code migrations, Cloud Data flow, Big Table, Big Query services to load data into cloud data lakes.
  • Actively worked on performance tuning of Hive queries by breaking up the complex queries into smaller queries where only required columns are passed to subsequent queries, applying filters early in the process etc.
  • Worked with downstream application teams with Business Objects intelligence developers populate data in Dashboards, Static Reports and Ad-hoc reports for generation of summary and detailed reports.
  • Unit Testing, source and target database tables data validations, verification and validation of the code developed and compare the results with existing system data and target system.
  • Involved in discussions and meetings with client interfacing teams, business teams, offshore, near shore team members and system integration testing teams to identify and resolve issues.

Environment: Informatica 10.1, Teradata 14, Oracle 12c, PL/SQL, Hadoop, Syncsort DMX BigData Integration ETL tool, Google Cloud Platform (GCP), Hive, Sqoop, Business Objects XI, Confluence and Rally Agile Tool, GIT Hub versioning, Bit Buckets versioning, CA Workload Automation (CAWA), UNIX Shell Scripting, SharePoint portal, Wiki, Microsoft Teams.

Confidential, Alpharetta, GA.

Data Architect / Senior ETL Developer

Responsibilities:

  • Requirement gathering and designing data conversion process, created High level interface design document along with low level Detail Design documents. Understanding the existing system and domain knowledge along with process flows and technical architecture of the applications. Followed the data architect standards, procedures, guidelines and worked with various teams for reviews and approval process.
  • Worked on agile projects, back-log stories and performed grooming, prioritizing prior to the Sprint Planning/Iterations and worked with product owner, scrum master, solution architects, various testing teams in Unit testing, UAT, Integration testing, E2E testing.
  • Development, testing of informatica mappings, mapplets, Source qualifier override queries, reusable transformations, lookups and complex business logic development, debugging, validations of data from source system to target systems, created several sessions with partitioning, push down optimization and kick off workflows development with various tasks flows.
  • Developed informatica mapping with change data capture logic for incremental data loads, implemented slowly changing dimensions for history data loading and parameter file generations for Dev, QA and Production environments. Creation of concurrent worklets and execution of concurrent workflows for various business to business financial end clients.
  • Worked on development and creation of various data conversion Oracle Packages header/body procedures, Stored Procedures development with complex transformation logics, PL/SQL with business logic transformations, object synonyms, grants between various schemas for 3X sources prod 3xconv, conv transporter, prod 4xconv and WI application target systems, data comparison, data fix scripts, data validation to verify differences between source and target systems data population.
  • Developed Oracle PLSQL scripts with data merge, insert select logic scripts, oracle analytical functions for complex logic and stored procedures, dynamic SQL’s, calling autonomous stored procedures in packages. Performance tuning of the stored procedures, query tuning, adding indexes, gather stats, fixes and maintenance of the PLSQL objects.
  • Involved in the development of Sqoop scripts to import and export data to/from relational databases. Developed and tested incremental loads using Sqoop from upstream systems into Hadoop and export results from Hadoop to Oracle.
  • Developed Hive scripts to perform transformations on the data ingested from upstream systems, perform change data capture and slowly changing dimension data capture to take care of occasional corrections / updates to past data.
  • Actively worked on performance tuning of Hive queries by breaking up the complex queries into smaller queries where only required columns are passed to subsequent queries, applying filters early in the process etc.
  • Involved in analyzing existing scripts and converting them to Hive scripts. Developed Oozie workflows by creating workflow XMLs, job property files and tested the workflows successfully. Created Autosys jobs to schedule and run Oozie workflows.
  • Worked with downstream application teams with Business Objects intelligence developers populate data in Dashboards, Static Reports and Ad-hoc reports for generation of summary and detailed reports.
  • Used Apache Hadoop Hue (Hadoop user experience) tool for checking the Bigdata Hadoop HDFS, Oozie workflows in data lakes. HIVE SQL interface to query data from target Hive tables, HDFS file system environment.
  • Unit Testing, source and target database tables data validations, verification and validation of the code developed and compare the results with existing system data and target system.
  • Involved in discussions and meetings with client interfacing teams, business teams, offshore team members and system integration testing teams to identify and resolve issues.

Environment: CDH 5.7, HDFS, Oracle 12c, PL/SQL, Informatica 10.1, Teradata, Hive, Sqoop, Oozie, Amazon Web Services (AWS), Business Objects XI, Confluence and Jira Agile Tool, GIT Hub versioning, Java, Autosys, OBIEE Reporting, UNIX Shell Scripting.

Confidential, Alpharetta, GA

Data Architect / Senior ETL Developer

Responsibilities:

  • Involved in analyzing the business requirement of the project and performed the impact analysis of the components, analyzed the system requirement documents and created the High level and low-level design documents. Gained domain knowledge along with process flows and technical architecture of the existing system from various source systems to target data ware house system and to data marts.
  • Based on the understanding of the new architecture of the project, created the Application Interface Design documents and then created the Application Detail Design Documents along with the Run Book details, production support handover documents, SLA documents. Followed the data architect standards, procedures, guidelines and maintained the check list of the DA process and worked with various teams for the reviews and approval process.
  • Created the logical and physical data models for the new tables, views and modified the existing data models like EDWDB, EDWFINDB, EDWAIMDB etc according to the data model standards in CA Erwin Data model tool and followed the processes set by the Data Architect teams.
  • Data Architect specifications are created for each new, impacted tables and columns, views, updated the transformation logic along with the metadata for each of the column details in Gromit tool and also make sure that the Metadata is being maintained in the enterprise metadata system.
  • Data element mapping from source to target with business rules, transformation logic where included in technical design documents and has conducted several approval reviews meetings with business teams, upstream and downstream teams.
  • Conducted several technical review sessions with Technical Forum team, principal technical architects, system analysts and developers for the end to end development flow with source system tables, file handling, extraction of the data with load ready files, processing of data in staging layer, transformation logic with business rules, loading of data in to base tables and then to target tables in the target warehouse and creation of the base views, custom views and business user views. Also, to create the extract and export scripts for processing of the data in the reporting cubes.
  • Created the new Teradata tables, base views, business views, financial views with partitioning, UPI, USI, compression values, data modelling in CA Erwin, reviewing with Teradata DBA and Architect forum and materializing in Dev, QA and production environments. Development of Teradata MultiLoad, TPump, BTEQ scripts, Teradata FastLoad, Teradata parallel transporter (TPT), stored procedures, macros with PL/SQL scripts, query tuning, performance improvement, and reviewing the scripts.
  • Development of informatica mappings, sessions, workflows, reusable transformations, debugging, validations of data from source system to target systems. Worked on debugging of mappings and validations of sessions, fixing any of the data issues, improving the performance of the mappings with query tuning, session partitioning, push down optimization and re-designing of mapping logic, etc.
  • Involved in Bigdata Hadoop data integration, migration of historical data, ETL, and importing and exporting of data using Sqoop to load data from oracle to Hadoop HDFS file system and HIVE SQL interface to query data from HDFS file system environment. Worked on Java and MapReduce programs to perform the business logic and load the results into HDFS files.
  • Worked on Apache Hadoop Hue (Hadoop user experience) tool for checking the Bigdata Hadoop HDFS folders in various data lakes. Used Oozie tool for workflow and scheduling of various jobs.
  • Worked with various Business intelligence reporting and application teams on OBIEE and Micro strategy reporting tools populate data in Universe data models, Dashboards, Static Reports and Ad-hoc reports.
  • Discussions and meetings with offshore and near shore development teams and system integration testing teams, source and target systems and along with end users and various business teams to identify and resolve issues.

Environment: Informatica, Teradata 14, Rally Agile Tool, CA Erwin Data modeling, Gromit Data Analyst and metadata tool, UNIX Shell Scripting, Oracle Database, OBIEE Reporting, Tiwoli workload scheduler (TWS) Scheduling, Micro strategy, Data lakes and Bigdata Hadoop framework, Java, Scala, Amazon Web Services (AWS)4)

Confidential, Phoenix, AZ.

Senior Integration Engineer / Data Engineer

Responsibilities:

  • Extensively worked on the requirement analysis, design, data element mapping from source to target and technical design documents in Agile/Scrum development model. Understood the existing End to End Architecture of the ETL Technical Process flow of various projects and existing systems which are being maintained and supported.
  • Worked on the logical and physical data models for the tables, views according to the data model standards in Erwin Data model tool and maintained the various project data models. Prepared High level, Detail level design documents, Data Architect specifications are created for new and existing tables, views, updated the metadata for each of the column and businesses rules.
  • Extracted the LAWSON financial system data and SSMS corporate wide and organizational data and implemented the ETL implementation logic using Data stage jobs and sequences and loaded the data into staging area, operational data source (ODS) system and then target dimension and fact tables.
  • Extensively used DataStage Designer, Manager, Director, Administrator for creating and implementing jobs and sequences. Implemented DataStage parallel jobs stages like DB2 Connectors, lookups, transformations, joiners, modulus source extract partitioning, merge, Oracle connectors, pivot, remove dups, filter, etc. Created several sequence jobs with various activity stages and dependencies and developed the control flows.
  • Developed DataStage parallel jobs and Extensively used pre and post SQL merge statements, gather stats, disable indexes, partitioning of the job’s flows, full and incremental loads etc. to load various ODS target tables along Enterprise Data Warehouse Dimension and Fact tables using surrogate keys.
  • Implemented many stored procedures, views, materialized views and many of the scripts to process and load data into target data base system. Actively worked on the performance tuning, analyzing the statistics, tuning the SQL and PL SQL queries, partitioning of the tables, using the parallel hints and effective joins and archival of the historical table data. Also worked on loading of data CLOB related data from source to target system.
  • Created the data stage parallel jobs and focused on performance tuning and created the sequences for loading of the Master and Lookup table’s data, exception handling of the data then loading the transaction table’s data and later Dimension and Fact tables Full and Incremental Data Loads which got incremental extracts from source system data.
  • Extensively worked with Enterprise Application Studio (EAS) Application development teams and BI Micro strategy and OBIEE developers populate data in Dashboards, Static Reports and Ad-hoc reports with prompts and filters applied Confidential end user level and generate Charts and Graphs for analysis.
  • Worked on fixing the data related issues and debugging the scripts and modifying the logic and make sure the data is matching with the source system to the target system.
  • Worked on the back-log stories and performed grooming, prioritizing prior to the Sprint Planning/Iterations and worked with testing teams in Unit testing, UAT, Integration, E2E testing. Followed Daily Standup’s (DSU), grooming, planning, and attended retrospectives ceremonies with the team to ensure successful project forecasting and realistic commitments.
  • Performed Incremental delivery of the project with Scrum demos and constant reviews with the customers and end uses to get proper feedback and any further improvements to the deliverables. Collaboratively worked with development teams, product owner, scrum master, functional teams, and managers, cross functional teams to gain the business needs, requirements and then develop and deliver the successful products.
  • Estimated the stories and tasks from the backlog items associated the current sprints, worked and updated the status of tasks to be in-progress, completed, ready to validate and Accepted and closed work items and constantly update the efforts and to do hours remaining and reflect breakdown chart and velocity of the scrum project.
  • Responsible for the performing the production support on daily and weekly basis and to close the stories as and when they are being worked and validated. Coordinate and conduct meetings with offshore development teams and worked with various cross functional teams located across different locations. Proactively documented the issues/ solutions in various applications, ETL and database scripts for efficient resolutions of recurring and to minimize future issues.

Environment: IBM Info Sphere Data Stage, Oracle Database 11g / 12c, Oracle Exadata, Teradata 14, Sybase DB2/UDB MS SQL Server 2005, Version One Agile Tool, CA Erwin Data modeling, MicroStrategy, UNIX Shell Scripting, ESP Scheduling, OBIEE Reporting.

Confidential, Bentonville, AR.

Senior Developer

Responsibilities:

  • Designed the Efficient Architecture of the ETL Technical Process flow (i.e) High level Design documents for implementation across multiple countries.
  • Estimated the overall activities and time needed to complete the project.
  • Technical Integration specifications were developed and documented for the low-level detailed design.
  • Created several project issue tracking tickets and change controls, raised emergency change controls during holiday focus dates for the project and also worked on the HP quality center for raising and fixing defects.
  • Performed data transformation, error handling and loaded the data into the staging database.
  • Focused on reusability and global implementation of the project for various countries, implemented Parameter sets to parameterize the jobs and sequencers and reused Perl scripts and UNIX scripts.
  • Extensively worked on DataStage parallel jobs and implemented Lookup stages, Join Stage, ODBC, DB2 API, Change Capture, Remove Duplicate Stage, Sort stage, Transformer, Aggregator stage and Flat File Stage. Created several sequence jobs with various activity stages and dependencies and developed the control flows. Worked on DataStage Designer, Manager, Director, Administrator for creating and implementing jobs and sequences.
  • Developed DataStage parallel jobs and Extensively used pre and post SQL merge statements, gather stats, disable indexes, partitioning of the job’s flows, full and incremental loads etc. to load various ODS target tables along Enterprise Data Warehouse Dimension and Fact tables using surrogate keys.
  • Performed history data loads, performance data loads and used various Database stages like DB2 API, ODBC connectors, DB2 Enterprise and transformation stages to implement various business logics.
  • Documented Unit test Cases as part of the development phase and performed End to end testing and full system integration testing.
  • Coordination with Onsite-Offshore development team members, assigning the tasks to make sure the deliverables and milestones are met.
  • Data Validation rules, Auditing of source system and target system files and table data. Code migration to Pilot and production environment and also maintain and support the project.

Environment: IBM Web sphere data stage (Designer, Administrator, Director), Teradata, DB2/ UDB, SQL Server 2005, UNIX Shell Scripting, Erwin Data modeling, Perl scripting Mainframe JCL scheduling.

Confidential

Senior Developer

Responsibilities:

  • Designed the Architecture of ETL process of the Data warehouse and discussed with the technical teams and business teams. Preparation of Project Initiation Document and Technical design documents.
  • Development of jobs, creating of tables, stored procedures for the various source systems. Estimate the efforts and timelines for the change requests and performing the impact analysis and implementing the change requests and improving the performance of the workflows. Reviewing of the Source System Tables Mapping Specification excel sheets and making modifications to the transformation logics as per the business rules.
  • Worked on development of informatica mappings, mapplets, Source qualifier override queries, reusable transformations, lookups and complex business logic development, debugging, validations of data from source system to target systems, created several sessions with partitioning and kick off workflows development with various tasks flows.
  • Developed informatica mapping with change data capture logic for incremental data loads, implemented slowly changing dimensions for history data loading and parameter file generations for Dev, QA and Production environments. Creation of concurrent worklets and execution of concurrent workflows for various business to business financial end clients.
  • Developed and loaded data using Teradata Fast load, Multi Load, Fast export, BTEQ scripts for loading data from Microsoft SQL server to Teradata and used Teradata Parallel Transporter (TPT) wrapper scripts to extract the data then to load into target.
  • Document and version control of the modifications made in the code to keep track of the changes made to code. Prepared the data validation for the interfaces and comparison of results.
  • Monitored the UAT runs on daily basis and make sure the production and UAT environment are in synchronization. Created the Autosys jobs on development, test & production environments.
  • Deployment of code on development environment and preparing the deployment plan for migration of code on test and production environment

Environment: Informatica Power center, Oracle 9i, Teradata, Microsoft SQL Server 2000, Unix Shell scripting, Autosys scheduling tool.

Confidential

Software Developer

Responsibilities:

  • Developing mappings and workflows as per the EDW2B ETL Design Specification for different releases. Performed Source system analysis (SSA) and prepared Low-level design (LDD) documents.
  • Worked with informatica mapping transformations including router transformation, update strategy, expression transformation, lookup transformation, sequence generator, aggregator transformation and sorter transformations etc.
  • Performed the performance and tuning Confidential source, Target levels using Indexes, Hints and Partitioning in databases and Informatica mappings. Worked Process Control Table that would maintain the status of all the change data capture (CDC) jobs.
  • Writing BTEQ scripts for different scenarios in the project and performing data cleansing, Bug Fixing and sanity checks.
  • Reviewing and verifying the mappings and workflows as per the Physical Data Model (PDM) and Standard check list. Performing Unit testing and integration testing for the mappings that has been developed. Validating of data and reconciliation of data in production environment.
  • Prepared Code Migration Document (CMD) for different releases and successfully deployed code in to test and production environments. Managed Alliance cases relative to priority 3(P3) and priority 4(P4) resources. Root cause analysis of Alliance Cases as per the user requirements and monitoring EDW Case activities.
  • Provided production support for different type of issues that has been observed by end users. Handled change requests as per new customer requirements. Prepared Unit Test Case documents and dependency document for the existing mappings and workflows.

Environment: Informatica Power center, Teradata SQL Assistant, BTEQ scripts, TOAD for Oracle 9i, Kintana migration tool to deploy scripts in various environments, Dollar Universe scheduling, UNIX scripting.

We'd love your feedback!