Sr. Etl Developer/ Aws Data Engineer Resume
Jersey City, NJ
SUMMARY
- 15+ years of IT experience in Analysis, design, development, implementation, maintenance and support with experience in Datastage, SAP DataServices(SAP BODS), AWS, Python, SQL, UNIX Shell Scripting and BizTalk Server 2013.
- Extensive experience in Extraction, Transformation and Loading(ETL) data from various data sources in to Data Warehouse and Data Marts using using Datastage and SAP DataServices(SAP BODS).
- Experience in using various stages in Datastage such as Joiner, Transformer, Aggregator, Lookup, Sequence Generator, Filter, Sorter, remove duplicates, Funnel and Copy etc.
- Worked extensively with Business Objects Data Services(SAP BODS) to create complex Data flows using Query, Key Generation, SQL Transform, Table Comparison, Case, Validation, Merge, etc.,
- Experience in Error Handling & Debugging and implemented various Performance Tuning techniques on Sources, Targets, Mappings, and Workflows/Sequences in Datastage/SAP BODS ETL mappings to increase the through put.
- Expertise in Implementing CDC - Change Data Capture.
- Expertise in implementing Slow Changing Dimension (SCD) Type 1 & Type 2 using Datastage and SAP BODS to maintain history in Dimension tables.
- Solid expertise in using different ETL Scheduling tools Autosys and Conrol-M along with Datastage Scheduler and SAP BODS Scheduler.
- Strong experience in using Database systems like Oracle, DB2, SQL Server and integrating them with Datastage and SAP BODS.
- Experience in integrating with UNIX Shell scripts with SAP BODS jobs.
- Performed end to end Data Migration Implementation Projects using SAP BODS.
- Experience in file formats like text files, Excel files and Central Repository usage in SAP BODS.
- Experience in working with Datastage migrations.
- Experience in following ETL Code standards & Naming Procedure as per organization standards.
- Strong experience in writing UNIX Shell Scripts for different purposes in DWH.
- Experience in UNIX Shell scripts on validating files and generating the reports for reconciliations.
- Worked with ETL Admin team for code migrations and ETL code deployments.
- Used Version control tools for maintaining the ETL code versions for migrations.
- Involved in Production code deployments and post production code & data validations.
- Hands on experience in change management and mentoring team members.
- I have excellent analytical, problem solving; oral communication and team lead skills, and is willing to work in independently or in a team environment.
- Experience in Amazon Web Services(AWS) like EC2,SQS, ECS, ELB, S3, Cloud Watch, AWSAuto Scaling, Lambda, SNS, Dynamo DB, CodeCommit, CodeBuild, CodePipeline.
- Experience in using Terraform for automation of CI/CD Process.
- Using other services like - Cloud Watch, SNS etc. for monitoring and alerting in AWS.
- Worked on SNS and SQS by pushing the events and calling the lambda functions.
- Experience in installing in IBM MQ clients on containers.
- Experience in using Python to Build ECS Services and developing lambda function.
- Worked to connect MQ Queues using Python for inbound and Outbound.
- Experience in design and migrating the projects to AWS enviroment.
- Worked on several python packages like Nameko, Boto3, Numpy, Zeep, Request, Pandas, Flask, Cachetools, PyYAML etc.
- Experience in using GIT Commands.
TECHNICAL SKILLS
ETL and Data Warehousing Tools: Datastage, SAP DataServices(SAP BODS), Informatica,BizTalk Server 2013
Cloud Technologies: AWSSQS, ECS, EC2, ELB, S3, Cloud Watch, AWSAuto Scaling, Lambda, SNS, Dynamo DB, CodeCommit, CodeBuild, CodePipeline, GIT, AWSCLI
Programming languages: Python,Unix&Linux shellscripts and SQL
Databases: Oracle, DB2, SQL Server
Automation and CI/CD Tools: Terraform, Git
Operating Systems: Windows, UNIX, RedHat
Scheduling Tools: Autosys, Control - M and CA Work Station
Version Control Tools: GIT(Github), CVS, PVCS
Education: Master’s in computer applications from Osmania University, 2005, India.
PROFESSIONAL EXPERIENCE
Confidential, Jersey City, NJ
Sr. ETL Developer/ AWS Data Engineer
Responsibilities:
- Extensively involved in gathering requirements from business, analysis, design, development, testing, Effort Estimation, Application code review, Work allocation, Bug tracking, Bug fixing, Onsite offshore coordination.
- Designed the jobs, workflows, data flows and used various transformations to load the data from Clarity Source into staging area, dimensions and fact tables using SAP Data Services.
- Worked in implementation of custom functions, initialization scripts for global variables that are used during course of Job Execution.
- Having daily standup and attending planning and pre planning meetings as per Agile.
- Tuned jobs for performance optimize the resources; push-down the load on to the source database / applications.
- Performed unit testing / integration testing on the data which is loaded into the tables.
- Analyze data in source systems to assess data cleanliness and make the recommendations for addressing within the source systems or as part of the ETL process.
- Designed the ETL code using SAP Data Services to implement Type II Slowly Changing Dimensions with Surrogate keys.
- Created Data Services Job to load the data into Enterprise data warehouse which involved extensive use of simple and complex transformations like Key Generation, SQL Transform, Table Comparison, Case, Validation, Merge, etc., in the Data flows.
- Developed test cases to validate the jobs and validated the jobs in the development environment, then moved and validated the same in QA environment before deploying them into the Production environment.
- Troubleshooting the issues that are raised during the execution of the SAP BODS jobs in production.
- Monitoring and fixing application issues to ensure day to day healthcare data is available for reporting purposes.
- Worked on SAP ECC Configurations and Database systems.
- Integrated UNIX Shell scripts with SAP BODS Jobs.
- Scheduled SAP BODS jobs using Autosys Scheduler.
- Implemented Error handling and reconciliations using SAP BODS.
- Extensively involved in gathering requirements from business, analysis, design, development, testing, Effort Estimation, Application code review, Work allocation, Bug tracking, Bug fixing, Onsite offshore coordination.
- Developed complex jobs using different transformations/ Stages like Joiner, Transformer, Aggregator, Lookup, Sequence Generator, Filter, Sorter, remove duplicates, Funnel, Copy etc.
- Designed and developed the jobs for transforming, integrating, and loading data using Data Stage Designer.
- Developed Job Sequences for automating and scheduling the DataStage Jobs.
- Responsible for preparing the design document and mapping documents.
- Involved in the design review, code review with the clients for the developed design/codes.
- Developing DataStage jobs for the complex requirements and providing critical decision in development phase.
- Handled multiple datastage migration projects.
- Worked on Autosys which involved in creation of Autosys jobs, execution.
- Designed ETL process as per the requirements and documented ETL process using MS VISIO. Preparation of design guidelines for ETL process development using DataStage Parallel Extender and Shell scripts.
- Provided tool mentoring, actively pursuing, and promoting knowledge transfer to share/ evolve best practices.
- Developed shell scripts for enhancing the jobs. Developed Job Sequences for automating and scheduling the DataStage Jobs.
- Prepared Unit test cases, Integration test cases and executed them during Code migration phase.
- Writing technical documents like technical specifications, Gap analysis documents and deployment documents for every code release.
- Involved in Different reviews like Internal and external code review, weekly status calls, issue resolution meetings and onsite code acceptance meetings
- Experience generating and interpreting mapping documentation, and translation into detailed design specifications and ETL code.
- Involved in Production support activities.
- Worked on migrating parallel jobs as service in IBM Quality stage and able to fix issues while connecting external systems.
- Worked on migrating all UNIX scripts to RED HAT Linux AWS server and tested with various scenarios.
- Experience writing UNIX Shell Scripts to call Datastage jobs.
- Experience in using SAP BAPI and ABAP Stages.
- Developing and maintaining applications written for Amazon Simple Storage Service, Amazon DynamoDB, Amazon Simple Queue Service, Amazon Simple Notification Service.
- Worked on python code for building the AWS repositories.
- Created SQS standard queues for reading the messages and processing it through python code.
- Created Success/Error Queues for posting the messages respectively and updating the status in Dynamo DB table.
- Worked on installing IBM MQ client on Containers.
- Worked on connecting IBM MQ for pushing and reading XML messages using Python in AWS.
- Developing with RESTful API interfaces
- Created AWS Lambda function for extracting the data from SAP database and post the data to AWS S3 bucket on scheduled basis using AWS cloud watch event.
- Worked on several python packages like Nameko, Boto3, Numpy, Zeep, Request, Pandas, Flask, Cachetools, PyYAML etc.
- Worked on error handling in Python code and creating common format for handling various errors.
- Worked on creating Terraform files for building AWS SQS Queues, ECS Clusters, AutoScaling groups, SNS, LAMBDA, Dynamo DB and Cloud Watch Event.
- Worked on deployment of code to AWS Code Commit using GIT commands (pull, fetch, push and commit. etc) from AWS CLI.
- Involved in monitoring cloud watch events and fixing the performance issues raised during Python HTTP service scale up.
- Worked on fixing performance issue raised while reading the data from s3 bucket and processed it in Python code by using python cache.
- Using Python unit test library, created various unit test cases for testing python code and generating report with details
- Worked on SQS Extended Client for handling huge messages which are more than 256KB.
Environment: AWS(SQS, ECS, EC2, ELB, S3, Cloud Watch, AWSAuto Scaling, Lambda, SNS, Dynamo DB, CodeCommit, CodeBuild, CodePipeline), GIT, AWSCLI, Unix/Linux, Python 3.6, Shell scripting, Terraform, YAML, JSON, XML, SAP WSDL, DataStage11.7, SAP Data Services (SAP BODS),BizTalk Server2013, SQL Server, Oracle, Autosys.
Confidential, Parsippany, NJ
Senior ETL Developer
Responsibilities:
- Involved in Business analysis and requirements gathering.
- Designed and developed the jobs for transforming, integrating, conforming, and loading data using Data Stage Designer
- Participated in design, development and implementation of complex jobs and sequencers using DataStage 9.1.
- Ensured that the mappings are correct and conducted data validation testing
- Validating the load process of ETL to make sure the target tables are populated according to the data mapping provided that satisfies the transformation rules.
- Validating the Archive process to purge the data that meet the defined business rules.
- Prepared test plans/test schedules with inputs from Project manager and development team
- Worked on Autosys which involved in creation of Autosys jobs, execution
- Worked on issues with migration from development to testing
- Designed ETL process as per the requirements and documented ETL process using MS VISIO. Preparation of design guidelines for ETL process development using DataStage Parallel Extender and Shell scripts.
Environment: DataStage 9.1, SAP, UNIX, Autosys.
Confidential, Phoenix, AZ
ETL Lead
Responsibilities:
- Lead Designer role comprises the list of responsibilities such as Requirement Analysis, Application design, Effort Estimation, Application code review, Work allocation, Bug tracking, Bug fixing, Onsite offshore coordination
- Lead Designer role comprises the list of responsibilities such as Requirement Analysis, Application design, Effort Estimation, Application code review, Work allocation, Bug tracking, Bug fixing, Onsite offshore coordination
- Responsible for preparing the design document, mapping document for the Listbill extract file, self-administration extract file, adjustments extract file (Billfile header section,Listbill detail record section, contact record section, billing file trailer section)
- Involved in the design review, code review with the clients for the developed design/codes.
- Developing DataStage jobs for the complex requirements and providing critical decision in development phase.
- Developed complex jobs using different transformations/ Stages like Joiner, Transformer, Aggregator, Lookup, Sequence Generator, Filter, Sorter, remove duplicates, Funnel, Copy etc.
- Creating a common shared container for collecting the statistics of the table loaded in the database.
- Creating common job control sequencers for automatically archiving all the required files after each run.
- Creating unique values in a sequencer before inserting into any database tables.
- Developing UNIX Scripts which can be coupled with the DataStage for more commonly used scenarios.
- Optimizing the sequencers/jobs to increase the performance.
- Developing shared container and common jobs for the whole project.
Environment: DataStage9.1, Unix, CA Workstation, ETL Job Control Application.
Confidential, Parsippany, NJ
ETL Developer
Responsibilities:
- Involved in Business analysis and requirements gathering.
- Designed and developed the jobs for transforming, integrating, conforming, and loading data using Data Stage Designer
- Participated in design, development and implementation of complex jobs and sequencers using DataStage 8.1.
- Ensured that the mappings are correct and conducted data validation testing
- Validating the load process of ETL to make sure the target tables are populated according to the data mapping provided that satisfies the transformation rules.
- Validating the Archive process to purge the data that meet the defined business rules.
- Prepared test plans/test schedules with inputs from Project manager and development team
- Worked on Autosys which involved in creation of Autosys jobs, execution
- Worked on issues with migration from development to testing
- Designed ETL process as per the requirements and documented ETL process using MS VISIO. Preparation of design guidelines for ETL process development using DataStage Parallel Extender and Shell scripts.
- Provided tool mentoring, actively pursuing, and promoting knowledge transfer to share/ evolve best practices.
- Developed shell scripts for enhancing the jobs. Developed Job Sequences for automating and scheduling the DataStage Jobs.
Environment: IBM Infosphere suite 8.1, SAP, UNIX, Autosys.
Confidential
ETL Developer
Responsibilities:
- Extracting, cleansing, transforming, integrating, and loading data into data warehouse using Designer
- Analyzed the business requirements and framing the business logic for the ETL process
- Extensively used Data Stage Designer to design and develop Server and PX jobs to migrate data from Data base into the Data Warehouse.
- Extensively used Processing Stages- Join, Funnel, Filter, Aggregator, Sort, Remove Duplicates, Copy, Transformer and Lookup. Develop/Debug- Row Generator & Peek stage. File Set Stages- Dataset, Sequential File, Database Stages-Teradata
- Involved in the testing of the various jobs developed and maintaining the test log
- Extensively worked on understanding the business requirements.
- Developed several Parallel Jobs using Data stage 7.5.
- Involved in writing DB2 scripts for loading data to target data warehouse for BTEQ, Fast Load and Multi Load
- Creating appropriate indexes dependent on table situation and requirements.
- Implementing Performance Tuning at various levels.
Environment: IBM Infosphere suite 8.1, DB2, UNIX, Control -M.
Confidential
ETL Developer
Responsibilities:
- Requirements Gathering, design, development, Unit testing, UAT, Production roll-out, enhancements and Production support.
- Involved in all the phases of the SDLC requirement as to understand the requirements and determine the best approach for timely delivery of in turn.
- Ensuring timely deliveries of work items to the Client.
- Involved in Implementing ETL standards and best practices within our portfolio.
- Used the DataStage Designer to develop processes for extracting, cleansing, transforming, integrating, and loading data into data warehouse database.
- Reusing the logic from DataStage jobs in real time.
- Developed jobs using different types of stages -- Sequential File, Transformer, Aggregator, Merge, Link Partitioner and Link Collector and Hashed File.
- Extensively worked on Error handling, cleansing of data, Creating Hash files and performing lookups for faster access of data.
- Used DataStage Manager for importing metadata from repository, new job categories and creating new data elements.
- Extensively used DataStage Director for Job Scheduling, emailing production support for troubleshooting from LOG files.
Environment: IBM Infosphere suite 7.5, DB2, UNIX, Autosys.
Confidential
ETL Developer
Responsibilities:
- Requirements Gathering, design, development, Unit testing, UAT, Production roll-out, enhancements and Production support.
- Involved in all the phases of the SDLC requirement as to understand the requirements and determine the best approach for timely delivery of in turn.
- Ensuring timely deliveries of work items to the Client.
- Involved in Implementing ETL standards and best practices within our portfolio.
- Responsible for all the phases of building the data marts like analyzing the business requirements, technical specifications, ETL process design, performance enhancement and maintenance.
- Actively involved in transforming and loading the data into different dimensions.
- Act on various stages like Aggregator, Sort, Transformer and Sequential file.
- Worked on Slowly Changing Dimensions techniques to maintain the historical details of the data. SCD Type 1 and SCD Type 2 were used as a response to the change in the source data.
- Worked on improving the performance of the designed jobs by using various performance tuning strategies.
Environment: Informatica, Oracle 9i, UNIX.
