Sr. Big Data/etl Tester Resume
Philadelphia, PA
SUMMARY
- Over 7 years of IT experience, working as Hadoop and ETL QA on various Data Warehouse Environments in multiple domains.
- Experience in full software Development Life Cycle (SDLC) and Quality Assurance Life Cycle (QALC) analysis, design, development, testing, implementation and maintenance in data warehousing environment.
- In depth technical knowledge and understanding of Data Warehousing, Data Validations, SQL, and Hadoop
- Designed and implemented distributed data processing pipelines using Spark, Hive, Sqoop, Python, and other tools and languages prevalent in the Hadoop ecosystem.
- Expertise in understanding Business Requirements, Functional Requirements, decomposing into HLD's and ETL Specifications into Test Cases for positive and negative test conditions.
- Extensive experience in ETL process consisting of data transformation, sourcing, mapping, conversion and loading.
- Delivered Test Data Management Strategy, TDM processes, Gap Analysis, Data Sub Setting Strategy
- Experience working with Siperian Hub MDM (Master Data management) for customer integration model and tested the same.
- Experience in Data Analysis, Data Validation, Data modeling, Data Cleansing, Data Verification and identifying data mismatch.
- Exposure in Databricks in processing and transforming GMI feeds which are massive in nature
- Experience in API Testing using Postman tool
- Experience in testing real time kafka applications
- Experience in gathering functional requirements from users and designing data model for ETL development.
- Experience in writing and development of test cases, test scripts and test procedures.
- Extensive experience in performing different types of testing including White Box, Black Box, System, Functional, Integration, Regression, Smoke, GUI, Database and User Acceptance Testing (UAT) and Unified Functional Testing (UFT).
- Experience in Data Profiling, Data Extraction, Data Integration, Data Testing and Data Cleansing.
- Excellent communication skills, team player and hard working.
- Strong technical knowledge of UNIX shell scripts as part of the Hadoop and ETL process, automate the process of loading, pulling the data
- Application Data warehousing experience in Banking, Healthcare and Insurance.
- Passionate worker to contribute individually and as part of a team.
TECHNICAL SKILLS
Scripting Languages: UNIX Shell Scripting
Testing Tools: Quick Test Pro (QTP), WinRunner, LoadRunner, Rational Clear Quest, Postman, Confluent Manager
Databases: Teradata, Oracle & SQL Server, NOSQL, DB2
Hadoop/Big Data: HDFS, HIVE, Spark, Kafka
Languages: SQL, Unix Shell scripting, PL/SQL, C,Python
Operating Systems: Win XP, 7, 8 and 10, UNIX, Linux
Defect Tracking Tool: Quality Center (QC), Test Director, Rational Clear Quest, ReQTest.
Methodologies: Agile, Scrum
PROFESSIONAL EXPERIENCE
Confidential, Philadelphia, PA
Sr. Big Data/ETL Tester
Responsibilities:
- Executed Test Scripts and Test Cases using manual testing procedures.
- Involved in planning and implementation solutions for Data Migration project.
- Experience working with Siperian Hub MDM (Master Data management) for customer integration model and tested the same
- Experience in RESTFULL web service Testing using Rest Assured framework java. Validated JSON formatted data, different http status code like 200, 201, 400, 415, 500 etc.
- Experience in testing streaming Kafka applications
- Dashboards and data extract Data - bricks and automate the jobs in the data-bricks.
- Involved in all over the quality controls (Hadoop & DB2), data validations, Key business checks, RTAS implementation of Hadoop ETL flow and DB2 landing zone/target tables.
- Experience in Testing Informatica ETL pipelines
- Performed data validations between summary files and extract files after ETL flow.
- Prepared the Linux Shell scripts with profile files by maintaining all global environment/application level variables.
- Demonstrated ability working and adapting to Big Data tools such as Databricks, bolb storage, Data bricks, HDFS, Pig, MapReduce Hive.
- Created and used automated script to ease complex workflow.
- Attended the daily Bug review meetings, weekly status meetings and walkthroughs and interacted with Business Analysts and Developers for resolving Defects.
- Used Microsoft excel and Word to document the test status reports and the Test Data
- Writing Test Hive queries to test development algorithms in Hadoop cluster.
- Coordinated with all the teams, representatives, and downstream users involved in testing for sign off approvals.
Environment: /Tools: Hadoop, Sqoop, AQT, HiveQL, JAVA, Data Bricks, UNIX, DB2, Tableau, Flat files, Shell Scripting, Putty, WinSCP, TOAD, Agile, Beyond Compare, Postman, Confluent Manager, Kafka, Informatica
Confidential, Kenilworth, New Jersey
Big Data/ETL Tester
Responsibilities:
- Designed and executed various well described Test scripts for System testing based on the Functional Design Document.
- Writing Test Hive queries to test development algorithms in Hadoop cluster
- Execution of cluster batch jobs &Hadoop workflows to generate test data in Hive & Hbase tables
- Imported and exported data into HDFS and Hive using Sqoop.
- Used Data Lake concepts to store data in HDFS
- Using TDCH to load the data in to TeraData from HDFS and Validating the Data in Teradata.
- Creates, prepares, reviews and executes quality assurance activities, including the Test Plan, Test cases, Test scripts, Test Summary and regression testing using complex SQL
- Preparation of Test Cases based on ETL Specification Document, Use Cases, Low Level Design document.
- Performed Testing in Slowly changing Dimension (Type1 and Type2) and Change Data Capture (CDC)
- Automate the Testcases in Spark Environment for getting fast Results.
- Good work Experience in stream Sets for Loading the files to HDFS and perform testing on StremSets
- Good work Experience in ICEDQ Automation Tool.
- Developed complex HQL scripts to transform and process the data per business requirements using multiple staging tables, joins and complex hive analytic and windowing functions.
Environment: SQL, HDFS, Hive, Spark, Python, Beeline, Sqoop, Putty, Shell Scripting, Teradata, Cntrl M, Stremsets, Bamboo, Jira, Confluence, HDFS
Confidential, Atlanta, GA
Big Data Tester
Responsibilities:
- Test cases, Test Data preparation, reviews & execution of Test scripts. Creating Test case & Test scripts by Agile Methodology.
- Provided subset of Functional test cases for UAT testing.
- Identified the test cases that needs to be automated using QTP.
- Involved in writing and executing script for regression and functional testing
- As an ETL Tester responsible for the understanding the business requirements, ETL and subsequently design ETL testing strategy using SQL scripts.
- Logged defects in Jira and created the sub task for the work.
- Involved in data integrity, backend, data driven, regression and functional testing.
- Performed manual testing executing all the test cases in ALM before switching to automation testing.
- Involved in testing data mapping and conversion in a server-based data warehouse.
- Used Microsoft excel and Word to document the test status reports and the Test Data
- Writing Test Hive queries to test development algorithms in Hadoop cluster
- Execution of cluster batch jobs &Hadoop workflows to generate test data in Hive & Hbase tables
- Imported and exported data into HDFS and Hive using Sqoop.
- Used Data Lake concepts to store data in HDFS
Environment: Hive, SQL, Beeline, Sqoop, Shell Scripting, oracle, Crontab, Bamboo, Jira, Confluence, HDFS
Confidential
Business Analyst/ETL Tester
Responsibilities:
- Interact with customers and/or business owners and quickly identifying the underlying business needs and requirements.
- Analyze the business and technical requirements and identify any gaps between customer requirements and the Acclaim platform.
- Create user stories and use case documents during requirements gathering
- Involved in unit testing, systems testing, integrated testing and user acceptance testing.
- Involved in Data Profiling, Data Extraction, Data Integration, Data Testing and Data Cleansing.
- Produce metrics, which provide management insight into the requirements process, to include, status, size, volatility, quality, effort using Visio.
- Perform data validation and root cause analysis using Excel and SQL
- Identify options for potential solutions and estimate efforts in conjunction with the other teams.
- Analyzed the “As is” and “To be” system documents to show the current and proposed functionalities by following the required HIPAA standards
- Plan and define system requirements to wireframe with Use Case, Use Case Scenario and Use Case Narrative using the UML (Unified Modelling Language) methodologies
- Create change controls and obtain approvals from customers. Work with the product owners to in corporate new features into product enhancement roadmap.
- Meet with stakeholders and business owners to understand requirements and obtain signoffs.
- Function as a point of contact for product related issues, new user training and future enhancements
- Hands on experience over MS Project tool for managing timeline and reporting.
- Conduct pharma processes, workflow, process diagram and GAP analyses to derive requirements for system enhancements
Confidential
Business Analyst
Responsibilities:
- Conducted the requirements and design walkthroughs with designers, developers and management staff
- Analyzed and prioritized user and business requirements as system requirements that must be included while developing the software.
- Conducted workflow, process diagram and GAP analyses to derive requirements for existing systems enhancements
- Functioned as the primary liaison between the business line, operations, and the technical areas throughout the project cycle.
- Tracking and managing requirement changes.
- Involved in walkthroughs and meetings with development team to discuss related issues.
- Assisting project management and working across multiple geographies and time zones
- Collaborating with business partners and global stakeholders to drive cross-workstream escalation and issue resolution
- Creating management reports for leadership teams
- Developing workshop, meeting material, presentation and reports with input from teams
- Establishing common processes throughout the enterprise
- Facilitating communications between project teams, management and stakeholders
- Overseeing work related to risks / issues reporting, meeting preparation support, workforce Supporting various project activities (UAT, SIT, Capacity Planning, formation and review of FRD and BRD documents, etc.)