Prioritizing Test Cases for Regression Testing - PowerPoint PPT Presentation

About This Presentation
Title:

Prioritizing Test Cases for Regression Testing

Description:

... for Regression Testing. Sebastian Elbaum University of Nebraska, Lincoln ... Alexey Malishevsky Oregon State University. Gregg Rothermel Oregon State University ... – PowerPoint PPT presentation

Number of Views:289
Avg rating:3.0/5.0
Slides: 26
Provided by: ELB90
Category:

less

Transcript and Presenter's Notes

Title: Prioritizing Test Cases for Regression Testing


1
Prioritizing Test Cases for Regression Testing
ISSTA 2000
  • Sebastian Elbaum University of Nebraska, Lincoln
  • Alexey Malishevsky Oregon State University
  • Gregg Rothermel Oregon State University

2
Defining Prioritization
  • Test scheduling
  • During regression testing stage
  • Goal maximize a criterion/criteria
  • Increase rate of fault detection
  • Increase rate of coverage
  • Increase rate of fault likelihood exposure

3
Prioritization Requirements
  • Definition of goal
  • Increase rate of fault detection
  • Measurement criterion
  • Of faults detected over life of test suite
  • Prioritization technique
  • Randomly
  • Total statements coverage
  • Probability of exposing faults

4
Previous Work
  • Goal
  • Increase rate of fault detection
  • Measurement
  • APFD
  • weighted average of the
  • percentage of
  • faults detected over life of test suite
  • Scale 0 - 100 (higher means faster detection)

5
Previous Work (2)
Measuring Rate of Fault Detection
A-B-C-D-E
C-E-B-A-D
E-D-C-B-A
6
Previous Work (3)
Prioritization Techniques
Label Prioritize on
1 random randomized ordering
2 optimal optimize rate of fault detection
3 sttotal coverage of statements
4 staddtl coverage of statements not yet covered
5 stfep probability of exposing faults
6 stfepaddtl probability of faults, adjusted to consider previous test cases
7
Summary Previous Work
  • Performed empirical evaluation of general
    prioritization techniques
  • Even simple techniques generated gains
  • Used statement level techniques
  • Still room to improve

8
Research Questions
  1. Can version specific TCP improve the rate of
    fault detection?
  2. How does fine technique granularity compare with
    coarse level granularity?
  3. Can the use of fault proneness improve the rate
    of fault detection?

9
Addressing RQ
  • New family of prioritization techniques
  • New series of experiments
  • Version specific prioritization
  • Statement
  • Function
  • Granularity
  • Contribution of fault proneness
  • Practical implications

10
Additional Techniques
Label Prioritize on
7 fntotal coverage of functions
8 fnaddtl coverage of functions not yet covered
9 fnfeptotal probability of exposing faults
10 fnfepaddtl probability of exposing faults, adjusted to consider previous tests
11 fnfitotal probability of fault likelihood
12 fnfiaddtl probability of fault likelihood, adjusted to consider previous tests
13 fnfifeptotal combined probabilities of fault existence and fault exposure
14 fnfifepaddtl combined probabilities of fault existence/exposure, adjusted on previous coverage
11
Family of Experiments
  • 8 programs
  • 29 versions
  • 50 test suites per program
  • Branch coverage adequate
  • 14 techniques
  • 2 control techniques optimal random
  • 4 statement level
  • 8 function level

12
Generic Factorial Design
Techniques
Programs
29 Versions
50 Test Suites
Independence of changes
Independenceof suite composition
Independence of code
13
Experiment 1a Version Specific
  • RQ1 Prioritization works on version specific at
    stat. level.
  • ANOVA Different average APFD among stat. level
    techniques
  • Bonferroni St-fep-addtl significantly better

Group Technique Value
A St-fep-addtl 78.88
B St-fep-total 76.99
B St-total 76.30
C St-addtl 74.44
Random 59.73
14
Experiment 1b Version Specific
  • RQ1 Prioritization works on version specific at
    function level.
  • ANOVA Different average APFD among function
    level techniques
  • Bonferroni Fn-fep not significantly different
    than Fn-total

Group Technique Value
A Fn-fep-addtl 75.59
A Fn-fep-total 75.48
A Fn-total 75.09
B Fn-addtl 71.66
15
Experiment 2 Granularity
  • RQ2 Fine granularity has greater prioritization
    potential
  • Techniques at the stat. level are significantly
    better than functional level
  • However, best functional level are better than
    worse statement level

16
Experiment 3 Fault Proneness
  • RQ3 Incorporating fault likelihood did not
    significantly increased APFD.
  • ANOVA Significant differences in average APFD
    values among all functional level techniques
  • Bonferroni Surprise. Techniques using fault
    likelihood did not rank significantly better

Group Technique Value
A Fn-fi-fep-addtl 76.34
A B Fn-fi-fep-total 75.92
A B Fn-fi-total 75.63
A B Fn-fep-addtl 75.59
A B Fn-fep-total 75.48
B Fn-total 75.09
C Fn-fi-addtl 72.62
C Fn-addtl 71.66
  • Reasons
  • For small changes fault likelihood does not seem
    to be worth it.
  • We believe it will be worthwhile for larger
    changes. Further exploration required.

17
Practical Implications
APFD Optimal 99 Fn-fi-fep-addtl
98 Fn-total 93 Random 84
Time Optimal 1.3 Fn-fi-fep-addtl 2.0
(.7) Fn-total 11.9 (10.6) Random
16.5 (15.2)
18
Conclusions
  • Version specific techniques can significantly
    improve rate of fault detection during regression
    testing
  • Technique granularity is noticeable
  • In general, statement level is more powerful but,
  • Advanced functional level techniques are better
    than simple statement level techniques
  • Fault likelihood may not be helpful

19
Working on
  • Controlling the threats
  • More subjects
  • Extending model
  • Discovery of additional factors
  • Development of guidelines to choose best
    technique

20
Backup Slides
21
Threats
  • Representativeness
  • Program
  • Changes
  • Tests and process
  • APFD as a test efficiency measure
  • Tools correctness

22
Experiment Subjects
23
FEP Computation
  • Probability that a fault causes a failure
  • Works with mutation analysis
  • Insert mutants
  • Determine how many mutant are exposed by a test
    case

FEP(t,s)
of mutants of s exposed by t
of mutants of s
24
FI Computation
  • Fault likelihood
  • Associated with measurable software attributes
  • Complexity metrics
  • Size, Control Flow, and Coupling
  • Generated fault index
  • principal component analysis

25
Overall
Group Technique Value
A Optimal 94.24
B St-fep-addtl 78.88
C St-fep-total 76.99
D C Fn-fi-fep-addtl 76.34
D C St-total 76.30
D E Fn-fi-fep-total 75.92
D E Fn-fi-total 75.63
D E Fn-fep-addtl 75.59
D E Fn-fep-total 75.48
F E Fn-total 75.09
F St-addtl 74.44
G Fn-fi-addtl 72.62
G Fn-addtl 71.66
H Random 59.73
Write a Comment
User Comments (0)
About PowerShow.com