Etl/hadoop Tester Resume
Nyc, NY
SUMMARY
- Over 8 years of Professional Experience in Software Quality Assurance and Testing in different environments and platforms.
- Expertise in Automated Software Quality Testing & Manual Testing of Web based and Client - Server applications.
- Experienced in Full Life cycle implementations of CRM & Big Data solutions (HADOOP) gathering Requirements from Clients and perform customizations/ setting up configurations and testing for Performance.
- Strong understanding of HADOOP Eco-System HDFS/MapReduce, JAVA Related Projects and other Hadoop related.
- Well versed with various types Software Development methodologies- Waterfall, Agile, RUP, Iterative and Extreme Programming.
- Strong Experience in testing and configuration of Apache & cloudera distribution Hadoop Namenode, secondary Namenode, jobtracker, and datanodes.
- Strong Experience in Hadoop ecosystem like Hive, flume, Pig, sqoop. Etc.
- Extensive experience in reviewing Business Requirement Documents, Software Requirement Documents and preparing Test Cases, Test scripts and Execution.
- Experience with Agile Methodology.
- Good Knowledge on Business intelligence, OLAP, Dimensional modeling, Star and Snowflake schema, extraction, transformation and loading (ETL) process
- Strong technical Knowledge of UNIX Utilities, Shell Script to automate process.
- Ability to develop complicated SQL script for Data validation testing by running SQL script, procedures.
- Experience in data retrieval methods using Universes, Personal Data files, Stored Procedures, and free hand SQL.
- Automated and scheduled the Informatica jobs using UNIX Shell Scripting.
- Experienced with ETL tools like Informatica, SSIS, Ab Initio and Data stage.
- Experience in testing Business Report developed in Cognos
- Experience in Black box testing with a complete QA cycle - from testing, defect logging and verification of fixed bugs
- Extensive experience in Functional testing, Integration/System testing, Regression testing and User Acceptance testing.
- Well versed in GUI application testing, Database testing and Front-end testing.
TECHNICAL SKILLS
Data Warehousing: Informatica 8.6.1/8.1/7.1/6/5.1.1/1.75 , Data Stage 8.x
Reporting Tools: OBIEE 10.3.1.4, Business Objects 6.5/XIR3
Data Modeling: Star-Schema Modeling, Snowflake-Schema Modeling, FACT and dimension tables, Pivot Tables, Erwin
Testing Tools: Win Runner, Load Runner, Test Director, Mercury Quality Center, Rational Tools
RDBMS: Oracle 10g/9i/8i/7.x, MS SQL Server 2005/2008, UDB DB2, Sybase, Teradata V2R6, MS Access 2008
Programming: UNIX Shell Scripting, Korn Shell, C Shell, Bourne Shell, Bash, SQL, SQL*Plus, PL/SQL, TOAD, C++
Web Technologies: JavaScript, HTML 4.0, and DHTML, .NET, Java, J2EE, XML, XSD, XSLT
Environment: UNIX, MVS, HP-UX, IBM AIX 4.2/4.3, Hyperion, Novell NetWare, Win 3.x/95/98, NT 4.0, Sun-Ultra, Sun-Spark, Sun Classic, and SCO
MPP Databases: Netezza NPS 8050
PROFESSIONAL EXPERIENCE
Confidential, NYC, NY
ETL/Hadoop Tester
Responsibilities:
- Analyzed business requirements, system requirements, data mapping requirement specifications, and responsible for documenting functional requirements and supplementary requirements in Quality Center 9.2
- Tested ETL jobs as per business rules using ETL design document
- Assisted in creating fact and dimension table implementation in Star Schema model based on requirements.
- Expert in writing Complex SQL/PLSQL Scripts in querying Teradata and Oracle.
- Defined data requirements and elements used in XML transactions.
- Tested the database schema with help of data architects using ERWIN.
- Involved in Testing the Hadoop cluster infrastructure, capacity planning and build plan for Hadoop cluster installations.
- Tested & Certified various Shared Script (Shell) to minimize the manual Administration activity and to maximize throughput via HPSA/Opsware.
- Involved in the testing of Data Mart using Power Center.
- Identified and Documented additional data cleansing needs and consistent error patterns that could diverted by modifying ETL code.
- Setting up the Hadoop Clusters & HDFS/MapReduce Jobs.
- Good Knowledge on Hadoop Cluster architecture and monitoring the clusters and Huge Data sets integration.
- Data is exported from MS SQL Server and other application databases into Hadoop.
- Automated workflows using shell scripts pull data from various databases into Hadoop.
- Used QTP for Checkpoints, Regular Expression and Accessed Data Tables for multiple applications.
- Involved in writing test scripts and functions in Test Script Language using QTP for automated testing.
- Experience in using the Form editor and the Outline Editor in SoapUI Pro to improve the productivity and make Web Service Testing easier
- Worked in an Agile technology with Scrum.
- Responsible for different Data mapping activities from Source systems to Teradata.
- Queried Teradata Database and validated the data using SQL Assistant.
- Effectively distributed responsibilities, arranged meetings and communicated with team members in all phases of the project.
- Used import and export facilities of the application to download/upload XMLs of failed test cases so as to re-verify.
- Scheduled the jobs using Auto sys and automated the jobs to be ran at specific time and automated the reports.
- Writing UNIX scripts to perform certain tasks and assisting developers with problems and SQL optimization.
- Converted SQL queries results into PERL variable and parsed the same for multiple reasons.
- Extensively used Autosys for automation of scheduling jobs on daily, bi-weekly, weekly monthly basis with proper dependencies.
- Wrote complex SQL queries using joins, sub queries and correlated sub queries
- Performed Unit testing and System Integration testing by developing and documenting test cases in Quality Center.
- Did Unit testing for all reports and packages.
- Designed and developed UNIX shell scripts as part of the ETL process, automate the process of loading, pulling the data.
- Involved in extensive DATA validation using SQL queries and back-end testing.
- Used Query Studio to test ad hoc reports
- Responsible for migrating the code changes from development environment to SIT, UAT and Production environments.
- Validated cube and query data from the reporting system back to the source system.
- Tested analytical reports using Analysis Studio.
Environment: Apache/Cloudera HADOOP, HIVE, PIG, SQOOP, FLUME, Flat files, Perl, Erwin 4.0, DTS, MS SQL Server 2008, Oracle 10g, SQL, PL/SQL, IBM DB2 8.0, AGILE, Teradata V2R6, Teradata SQL Assistant, COBOL, Mercury Quality Center 11, QTP 10.0,Autosys, XML, XSLT, XML Spy 2008, TSO, ISPF, OS/z, JCL, Mainframes, Toad, Unix Shell Scripting, Windows XP/2000
Confidential, Warren, NJ
ETL QA Tester
Responsibilities:
- Created test data for testing specific ETL flow.
- Designed the Load Runner scenarios with various user-mix incrementing in constant ratios, till the peak load is achieved.
- Extracted certain column of data from a number to files using PERL.
- Used PERL for automating all the types of modules at a time.
- CLOUDERA MANAGER is used to deploy/administrate and monitor the health and performance of the Hadoop components.
- Good Knowledge on Hadoop Cluster architecture and monitoring the clusters and Huge Data sets integration.
- Data is exported from MS SQL Server and other application databases into Hadoop
- Responsible for Data mapping testing by writing complex SQL Queries using WINSQL.
- Experience in creating UNIX scripts for file transfer and file manipulation.
- Validating the data passed to downstream systems
- Worked with data validation, constraints, source to target row counts.
- Developed automated Test Scripts using QTP.
- Used data conversion to protect the data and prevent data loss.
- Strong in writing SQL queries and makes table queries to profile and analyze the data in MS Access.
- Formulate methods to perform Positive and Negative testing against requirements.
- Performed Manual Testing of the application Front-End and Functionality. Identified the critical test scripts to be automated.
- Performed Verification, Validation, and Transformations on the Input data (Text files, XML files) before loading into target database.
- Perform Functional, Data Validation, Integration, regression and User Acceptance testing.
- Used TOAD, DB Artisan tools to connect to Oracle Database to validate data that was populated by ETL applications
- Involved in testing Unix Korn Shell wrappers to run various ETL Scripts to load the data into Target Database (Oracle).
- Worked on Clear Quest to log defects and track resolution till the closing of defect after retesting.
- Created test cases and executed test scripts using Quality Center.
- Used Microsoft excel and Word 2003 to document the test status reports and the Test Data.
- Used Quality Center to track and report system defects.
- Involved in testing the XML files and checked whether data is parsed and loaded to staging tables
Environment: Apache Hadoop 2.2.0, Informix, DB2, Java, XML, XSLT, XML Spy 2008, SQL, SQL Server 2000/2005, PERL, Teradata V2R6 (MLOAD, FLOAD, FAST EXPORT, BTEQ), Win Runner, Load Runner Teradata SQL Assistant 7.0, XML, XSLT, XML Spy 2008, IBM AIX 5.3, UNIX, Shell Scripting, QTP 9.2,WinSQL, Ultra editor, Rumba UNIX Display, Mercury Quality Center 8.2
Confidential, Farmingdale, NY
QA Tester
Responsibilities:
- Assisted in creating fact and dimension table implementation in Star Schema model based on requirements.
- Develop test plans based on test strategy. Created and executed test cases based on test strategy and test plans based on ETL Mapping document.
- Wrote complex SQL queries for querying data against different data bases for data verification process.
- Designed the data flow diagrams using MS VISIO.
- Prepared the Test Plan and Testing Strategies for Data Warehousing Application.
- Preparation of technical specifications and Source to Target mappings.
- Extensively used SQL programming in backend and front-end functions, procedures, packages to implement business rules and security
- Followed the Mapping documents and HIPAA implementation guides for HIPAA Transaction Mapping.
- Involved in testing HIPAA EDI Transactions and mainly focused on 837 and 835.
- Involved in Data mapping to/from legacy to Facets that will be used to populate the oracle database
- Involved in testing HIPAA EDI Transactions and mainly focused on 837 and 835
- Used Foresight Desktop Validator for HIPAA IG validation.
- Written test cases to test the application manually in Quality Center and automated using Quick Test Pro.
- Worked with SSIS system variable, passing variables between packages.
- Created cascading prompts at the universe level. These cascading prompts were used within full client and thin client reports to narrow down the selection parameters.
- Tested different types of reports, like Master/Detail, Cross Tab and Charts (for trend analysis).
- Developed automated test scripts from manual test cases for Regression testing based on the requirement documents using Quick Test Professional.
- Written Test Plans and Test Cases on Mercury’s Test Director Tool.
- Defects identified in testing environment where communicated to the developers using defect tracking tool Mercury Test Director.
- Optimized QTP scripts for Regression testing of the application with various data sources and data types.
- Executed regression tests at each new build in QTP.
- Developed scripts, utilities, simulators, data sets and other programmatic test tools as required executing test plans.
- Tested a number of complex ETL mappings, mapplets and reusable transformations for daily data loads.
- Designed and supported SQL 2000 Reporting services, Integration services and Analysis services.
- Creating test cases for ETL mappings and design documents for production support.
- Setting up, monitoring and using Job Control System in Development/QA/Prod.
- Extensively worked with flat files and excel sheet data sources. Wrote scripts to convert excel to flat files.
- Scheduling and automating jobs to be run in a batch process.
- Effectively communicate testing activities and findings in oral and written formats.
- Reported bugs and tracked defects using Test Director 6.5
- Extensively used Informatica power center for Extraction, Transformation and Loading process (ETL).
- Worked with ETL group for understating mappings for dimensions and facts.
- Extracted data from various sources like Oracle, flat files and SQL Server.
- Worked on issues with migration from development to testing.
- Extensively tested several Cognos reports for data quality, fonts, headers & cosmetic.
- Compared the actual result with expected results. Validated the data by reverse engineering methodology i.e. backward navigation from target to source.
Environment: SAS/Base 8.1, SAS/Macros, SAS/ETL, PERL, UNIX Shell Scripting, Informatica Power Center 7.1 (Power Center Designer, workflow manager, workflow monitor), Mercury Test Director 6.5, QTP 7.2, SQL *Loader, Cognos 7.0, Oracle8i, SQL Server 2000, Erwin 3.5, Windows 2000, TOAD 7, Business Objects 6.1, Teradata V2R4
Confidential - Princeton, NJ
ETL/BI Test Analyst
Responsibilities:
- Wrote SQL Statements to extract Data from Tables and to verify the output Data of the reports
- Performed back-end testing on the Oracle database by writing SQL queries.
- Collected, analyzed and reported testing metrics, executed SQL queries
- Built and maintained effective QA test lab environments
- Developed SQL queries in Toad and DB Artisan to achieve the data transformations
- Participated in various meetings and discussed Enhancement Request issues
- Conducted Regression testing after the bugs have been fixed by the development team
- Generated reports from Test Director.
- Responsible for source control, versioning, and configuration management of test scripts, test results, defects
- Developed Test Strategy, Test plan, Test cases, and Test scripts
- Created test plan and executed test cases using Rational Test Manger
- Conducted Functionality, Security, and End to End testing
- Reported bugs using Test Director and generated the defect reports for review
- Defect reporting and tracking was also done using Test Director.
- Maintained the test traceability matrix
- Performed Regression Testing on weekly builds.
- Monitored the workflow transformations in Informatica work flow monitor
- Verified backend testing for compliance with FDA's 21 CFR Part 11 Electronic Signatures rules
Environment: Windows 2000 Server, SQL Server, Oracle 9i, Documentum 5i, Rational Test Manager, Rational Requisite Pro, Rational Clear Case, Rational Robot, ETL Informatica, Data Stage, Test Director, Siebel E-Clinical 7.5, Toad, Java, J2EE, JSP