A Statistical Analysis of the PrecisionRecall Graph - PowerPoint PPT Presentation

1 / 15
About This Presentation
Title:

A Statistical Analysis of the PrecisionRecall Graph

Description:

Set n= 2m and call the two m-halfes Z1 and Z2. Define gi (Z):=A(f,Zi) ... effective sample size is only the number of positive examples, in fact, only 2m. ... – PowerPoint PPT presentation

Number of Views:57
Avg rating:3.0/5.0
Slides: 16
Provided by: ral121
Category:

less

Transcript and Presenter's Notes

Title: A Statistical Analysis of the PrecisionRecall Graph


1
A Statistical Analysis of the Precision-Recall
Graph
  • Ralf Herbrich
  • Microsoft Research
  • UK

Joint work with Hugo Zaragoza and Simon Hill
2
Overview
  • The Precision-Recall Graph
  • A Stability Analysis
  • Main Result
  • Discussion and Applications
  • Conclusions

3
Features of Ranking Learning
  • We cannot take differences of ranks.
  • We cannot ignore the order of ranks.
  • Point-wise loss functions do not capture the
    ranking performance!
  • ROC or precision-recall curves do capture the
    ranking performance.
  • We need generalisation error bounds for ROC and
    precision-recall curves!

4
Precision and Recall
  • Given
  • Sample z((x1,y1),...,(xm,ym)) 2 (X 0,1)m
    with k positive yi together with a function fX
    ! R.
  • Ranking the sample
  • Re-order the sample f(x(1)) f(x(m))
  • Record the indices i1,, ik of the positive y(j).
  • Precision pi and ri recall

5
Precision-Recall An Example
After reordering
f(x(i))
6
Break-Even Point
1
0.9
Break-Even point
0.8
0.7
0.6
0.5
Precision
0.4
0.3
0.2
0.1
0
0
0.2
0.4
0.6
0.8
1
Recall
7
Average Precision
8
A Stability Analysis Questions
  • How much does A(f,z) change if we can alter one
    sample (xi,yi)?
  • How much does A(f,) change if we can alter z?
  • We will assume that the number of positive
    examples, k, has to remain constant.
  • We can only alter xi, rotate one y(i).

9
Stability Analysis
  • Case 1 yi0
  • Case 2 yi1

10
Proof
  • Case 1 yi0
  • Case 2 yi1

11
Main Result
  • Theorem For all probability measures, for all
    gt1/m, for all fX ! R, with probability at least
    1- over the IID draw of a training and test
    sample both of size m, if both training sample z
    and test sample z contain at least dme positive
    examples then

12
Proof
  • McDiarmids inequality For any function gZn ! R
    with stability c, for all probability measures P
    with probability at least 1- over the IID draw
    of Z
  • Set n 2m and call the two m-halfes Z1 and Z2.
    Define gi (Z)A(f,Zi). Then, by IID

13
Discussions
  • First bound which shows that asymptotically (m!1)
    training and test set performance (in terms of
    average precision) converge!
  • The effective sample size is only the number of
    positive examples, in fact, only 2m .
  • The proof can be generalised to arbitrary test
    sample sizes.
  • The constants can be improved.

14
Applications
  • Cardinality bounds
  • Compression Bounds
  • (TREC 2002)
  • No VC bounds!
  • No Margin bounds!
  • Union bound

15
Conclusions
  • Ranking learning requires to consider
    non-point-wise loss functions.
  • In order to study the complexity of algorithms we
    need to have large deviation inequalities for
    ranking performance measures.
  • McDiarmids inequality is a powerful tool.
  • Future work is focused on ROC curves.
Write a Comment
User Comments (0)
About PowerShow.com