- January 14, 2013
- Posted by: InApp
- Category: Independent Software Testing
- Metrics can be defined as “STANDARDS OF MEASUREMENT”.
- Metric is a unit used for describing or measuring an attribute.
- Test metrics are the means by which the software quality can be measured.
- Test provides the visibility into the readiness of the product , and gives clear measurement of the quality and completeness of the product.
Why we Need Metrics?
“You cannot improve what you cannot measure.”
“You cannot control what you cannot measure”
AND TEST METRICS HELPS IN
Take decision for next phase of activities
Evidence of the claim or prediction
Understand the type of improvement required
Take decision on process or technology changemet
Type of Test Metrics
Base Metrics (Direct Measure)
Base metrics constitute the raw data gathered by a Test Analyst throughout the testing effort. These metrics are used to provide project status reports to the Test Lead and Project Manager; they also feed into the formulas used to derive Calculated Metrics.
Ex: # of Test Cases, # of Test Cases Executed
Calculated Metrics (Indirect Measure)
Calculated Metrics convert the Base Metrics data into more useful information. These types of metrics are generally the responsibility of the Test Lead and can be tracked at many different levels (by module, tester, or project).
Ex: % Complete, % Test Coverage
Metrics life Cycle
Test Plan Coverage on Functionality:
Total number of requirement v/s number of requirements covered through test scripts.
- (No of requirements covered / total number of requirements) * 100
Define requirements at the time of Effort estimation
Example: Total number of requirements estimated are 46, total number of requirements tested 39, blocked 7…define what is the coverage?
Note: Define requirement clearly at project level
Test case defect Density:
Total number of errors found in test scripts v/s developed and executed.
- (Defective Test Scripts /Total Test Scripts) * 100
Example: Total test script developed 1360, total test script executed 1280, total test script passed 1065, total test script failed 215
So, test case defect density is
215 X 100
—————————- = 16.8%
This 16.8% value can also be called as test case efficiency %, which is depends upon total number of test cases which uncovered defects
Defect Slippage Ratio:
Number of defects slipped (reported from production) v/s number of defects reported during execution.
- Number of Defects Slipped / (Number of Defects Raised – Number of Defects Withdrawn)
Example: Customer filed defects are 21, total defect found while testing are 267, total number of invalid defects are 17
So, Slippage Ratio is
[21/(267-17) ] X 100 = 8.4%
Number of requirements agreed v/s number of requirements changed.
- (Number of Requirements Added + Deleted + Modified) *100 / Number of Original Requirements
- Ensure that the requirements are normalized or defined properly while estimating
Example: VSS 1.3 release had total 67 requirements initially, later they added another 7 new requirements and removed 3 from initial requirements and modified 11 requirements.
So, requirement Volatility is
(7 + 3 + 11) * 100/67 = 31.34%
The Review Efficiency is a metric that offers insight on the review quality and testing.
Some organization also use this term as “Static Testing” efficiency and they are aiming to get min of 30% defects in static testing.
Review efficiency=100*Total number of defects found by reviews/Total number of project defects.
Example: A project found total 269 defects in different reviews, which were fixed and test team got 476 defects which were reported and valid.
So, Review efficiency is [269/(269+476)] X 100 = 36.1%
Efficiency & Effectiveness of Processes:
- Effectiveness: Doing the right thing. It deals with meeting the desirable attributes that are expected by the customer.
- Efficiency: Doing the thing right. It concerns the resources used for the service to be rendered