You are on page 1of 15

A Statistical Analysis of the Precision-Recall Graph

Ralf Herbrich Microsoft Research UK

Joint work with Hugo Zaragoza and Simon Hill

Overview

The Precision-Recall Graph A Stability Analysis Main Result Discussion and Applications Conclusions

Features of Ranking Learning


We cannot take differences of ranks. We cannot ignore the order of ranks. Point-wise loss functions do not capture the ranking performance! ROC or precision-recall curves do capture the ranking performance. We need generalisation error bounds for ROC and precision-recall curves!
3

Precision and Recall

Given:

Sample z=((x1,y1),...,(xm,ym)) 2 (X {0,1})m with k positive yi together with a function f:X ! R. Re-order the sample: f(x(1)) f(x(m)) Record the indices i1,, ik of the positive y(j).

Ranking the sample:


Precision pi and ri recall:

Precision-Recall: An Example
After reordering:

f(x(i))

Break-Even Point
1 0.9
0.8 0.7

Break-Even point

Precision

0.6

0.5
0.4 0.3 0.2 0.1 0 0 0.2 0.4 0.6 0.8 1

Recall

Average Precision
1 0.9 0.8 0.7

Precision

0.6 0.5

0.4
0.3 0.2 0.1 0 0 0.2 0.4 0.6 0.8 1
7

Recall

A Stability Analysis: Questions


1.

2.

How much does A(f,z) change if we can alter one sample (xi,yi)? How much does A(f,) change if we can alter z? We will assume that the number of positive examples, k, has to remain constant. We can only alter xi, rotate one y(i).

Stability Analysis

Case 1: yi=0

Case 2: yi=1

Proof

Case 1: yi=0

Case 2: yi=1

10

Main Result
Theorem: For all probability measures, for all >1/m, for all f:X ! R, with probability at least 1- over the IID draw of a training and test sample both of size m, if both training sample z and test sample z contain at least dme positive examples then

11

Proof
1.

McDiarmids inequality: For any function g:Zn ! R with stability c, for all probability measures P with probability at least 1- over the IID draw of Z

2.

Set n= 2m and call the two m-halfes Z1 and Z2. Define gi (Z):=A(f,Zi). Then, by IID
12

Discussions

First bound which shows that asymptotically (m!1) training and test set performance (in terms of average precision) converge! The effective sample size is only the number of positive examples, in fact, only 2m . The proof can be generalised to arbitrary test sample sizes. The constants can be improved.
13

Applications

Cardinality bounds Compression Bounds


(TREC 2002)

Union bound:

No VC bounds! No Margin bounds!

14

Conclusions

Ranking learning requires to consider nonpoint-wise loss functions. In order to study the complexity of algorithms we need to have large deviation inequalities for ranking performance measures. McDiarmids inequality is a powerful tool. Future work is focused on ROC curves.

15

You might also like