Learning to Rank for Information Retrieval

Transcription

Learning to Rank for Information Retrieval
Learning to Rank
for Information Retrieval
Tie-Yan Liu
Lead Researcher
Microsoft Research Asia
Speaker
• Tie-Yan Liu
– Lead Researcher, Microsoft Research Asia
– Co-author of 70+ papers in SIGIR, WWW, NIPS, ICML, KDD, etc.
– Co-author of SIGIR Best Student Paper (2008) and JVCIR
Most Cited Paper Award (2004~2006).
– Author of two books on learning to rank.
– Associate editor, ACM Transactions on Information System;
Editorial board member, IR Journal.
– PC Chair of RIAO 2010, Area Chair of SIGIR (2008~2011),
Track Chair of WWW (2011).
2011/5/9
Tie-Yan Liu, Copyright reserved
2
Outline
•
•
•
•
•
•
Introduction
Major Approaches to Learning to Rank
Analysis of the Approaches
Benchmark Evaluation of the Approaches
Statistical Learning Theory for Ranking
Summary and Outlook
2011/5/9
Tie-Yan Liu, Copyright reserved
3
Ranking in Information
Retrieval
Overwhelmed by Flood of Information
2011/5/9
Tie-Yan Liu, Copyright reserved
5
Facts about the Web
• According to www.worldwidewebsize.com, major
search engines index at least tens of billions of web
pages.
• CUIL.com claimed that they index 127 billion pages.
• Google claimed that they know 1 trillion pages.
2011/5/9
Tie-Yan Liu, Copyright reserved
6
Ranking is Essential
Indexed
Document Repository
D  d j 
Ranking
model
Query
2011/5/9
Tie-Yan Liu, Copyright reserved
7
Conventional Ranking Models
• Query-dependent
–
–
–
–
Boolean model, extended Boolean model, etc.
Vector space model, latent semantic indexing (LSI), etc.
BM25 model, statistical language model, etc.
Span based model, distance aggregation model, etc.
• Query-independent
– PageRank, TrustRank, BrowseRank, Toolbar Clicks, etc.
2011/5/9
Tie-Yan Liu, Copyright reserved
8
Evaluation of Ranking Models
• Construct a test set containing a large number of
(randomly sampled) queries and their associated
documents, and relevance judgment for each querydocument pair.
• Evaluate the ranking result given by a model for a
particular query in the test set with an evaluation
measure.
• Use the average measure over the entire test set to
represent the overall ranking performance.
2011/5/9
Tie-Yan Liu, Copyright reserved
9
Collecting Queries and Documents
• Queries are usually randomly sampled from
query logs.
• TREC’s Pooling strategy for sampling documents
– A pool of possibly relevant documents is created by
taking a sample of documents selected by the various
participating systems.
– Top 100 documents retrieved in each submitted run
for a given query are selected and merged into the
pool for human assessment.
– On average, an assessor judges the relevance of
approximately 1500 documents per query.
2011/5/9
Tie-Yan Liu, Copyright reserved
10
Relevance Judgment
• Degree of relevance lk
– Binary: relevant vs. irrelevant
– Multiple ordered categories: Perfect > Excellent > Good >
Fair > Bad
• Pairwise preference lu ,v
– Document A is more relevant than
document B
• Total order
l
• Documents are ranked as {A,B,C,..}
according to their relevance
2011/5/9
Tie-Yan Liu, Copyright reserved
11
Evaluation Measure - MAP
• Precision at position k for query q:
#{relevant documents in top k results}
k
• Average precision for query q:
k P @ k  lk
AP 
# {relevant documents}
P@k 
1 1 2 3 
AP        0.76
3 1 3 5 
• MAP: averaged over all queries.
2011/5/9
Tie-Yan Liu, Copyright reserved
12
Evaluation Measure - NDCG
• NDCG at position n for query q:
NDCG @ k  Z k  G  1 ( j )  ( j )
k
j 1
Normalization
Cumulating
Gain
G ( j )  2
l
1
Position discount
 1 ( j )
 1,  ( j )  1 / log( j  1)
• Averaged over all queries.
 1 ( j ) The document ranked at
the j-th position in π.
2011/5/9
 ( j)
Tie-Yan Liu, Copyright reserved
The rank position of
document j in π.
13
Evaluation Measure - Summary
• Query-level: every query contributes equally to the
measure.
– Computed on documents associated with the same query.
– Bounded for each query.
– Averaged over all test queries.
• Position-based: rank position is explicitly used.
– Top-ranked objects are more important.
– Relative order vs. relevance score of each document.
– Non-continuous and non-differentiable w.r.t. scores
2011/5/9
Tie-Yan Liu, Copyright reserved
14
Learning to Rank
Challenges of Conventional Models
• Manual parameter tuning is usually difficult,
especially when there are many parameters and the
evaluation measures are non-smooth.
• Manual parameter tuning sometimes leads to overfitting.
• It is non-trivial to combine the large number of
models proposed in the literature to obtain an even
more effective model.
2011/5/9
Tie-Yan Liu, Copyright reserved
16
Machine Learning Can Help
• Machine learning is an effective tool
– To automatically tune parameters.
– To combine multiple evidences.
– To avoid over-fitting (by means of regularization, etc.)
• “Learning to Rank”
– In general, those methods that use
machine learning technologies to solve
the problem of ranking can be named as
“learning to rank” methods.
2011/5/9
Tie-Yan Liu, Copyright reserved
17
Learning to Rank
• In most recent works, learning to rank is
defined as having the following two properties:
– Feature based;
– Discriminative training.
2011/5/9
Tie-Yan Liu, Copyright reserved
18
Feature Based
• Documents represented by feature vectors.
– Features are extracted for each query-document pair.
– Even if a feature is the output of an existing retrieval
model, one assumes that the parameter in the model is
fixed, and only learns the optimal way of combining these
features.
• The capability of combining a large number of
features is very promising.
– It can easily incorporate any new progress on retrieval
model, by including the output of the model as a feature.
2011/5/9
Tie-Yan Liu, Copyright reserved
19
Discriminative Training
• An automatic learning process based on the
training data,
• Contains
the objects
under
– With the four pillars of
discriminative
learning
•
•
•
•
Input space,
Output space,
Hypothesis space,
Loss function.
investigation.
• Usually objects are
represented by feature vectors,
extracted according to different
applications
– Not even necessary to have a probabilistic
explanation.
2011/5/9
Tie-Yan Liu, Copyright reserved
20
Discriminative Training
• An automatic learning process based on the
training data,
are two relatedlearning
but different
– With the four pillars of• There
discriminative
•
•
•
•
Input space,
Output space,
Hypothesis space,
Loss function.
– Not even necessary
explanation.
2011/5/9
definitions of the output space.
- Output space of the task.
- Output space to facilitate learning.
• Example: using regression to solve
classification
- Output space of the task: {+1, -1}
- Output space to facilitate learning:
toreal
have
a probabilistic
numbers
Tie-Yan Liu, Copyright reserved
21
Discriminative Training
• An automatic learning process based on the
training data,
– With the four pillars of discriminative learning
•
•
•
•
Input space,
Output space,
Hypothesis space,
Loss function.
– Not even necessary
explanation.
2011/5/9
• Defines the class of functions
mapping the input space to the output
space.
• The functions operate on the feature
vectors of the input objects, and make
predictions according to the format of
tothe
have
a probabilistic
output
space.
Tie-Yan Liu, Copyright reserved
22
Discriminative Training
• An automatic learning process based on the
• The loss function measures to what
training data,
degree the prediction generated by the
is in accordance with the
– With the four pillars ofhypothesis
discriminative
learning
ground truth label.
•
•
•
•
Input space,
Output space,
Hypothesis space,
Loss function.
– Not even necessary to
explanation.
2011/5/9
- For example, widely used loss
functions for classification include the
exponential loss, the hinge loss, and
the logistic loss.
- With the loss function, an empirical
risk can be defined on the training set,
have
probabilistic
and theaoptimal
hypothesis is usually
(but not always) learned by means of
empirical risk minimization.
Tie-Yan Liu, Copyright reserved
23
Discriminative Training
• An automatic learning process based on the
training data,
• This is highly demanding for real search
engines,
– Everyday these search engines will receive a lot of
user feedback and usage logs
– It is very important to automatically learn from
the feedback and constantly improve the ranking
mechanism.
2011/5/9
Tie-Yan Liu, Copyright reserved
24
Learning to Rank Framework
2011/5/9
Tie-Yan Liu, Copyright reserved
25
Wide Applications of Learning to Rank
•
•
•
•
•
•
•
Document retrieval
Question answering
Multimedia retrieval
Text summarization
Online advertising
Collaborative filtering
Machine translation
2011/5/9
Tie-Yan Liu, Copyright reserved
26
Recent Trends on Learning to Rank
• Successfully applied to web search
• Hot topic in IR and ML
– Hundreds of publications at SIGIR, ICML, NIPS, etc.
– Multiple sessions at SIGIR every year
– Several workshops at SIGIR, ICML, NIPS, etc.
– Several tutorials at SIGIR, WWW, ACL, etc.
– Special issue at IRJ and JMLR
– Yahoo! Learning to rank challenges
– ……
2011/5/9
Tie-Yan Liu, Copyright reserved
27
Outline
•
•
•
•
•
•
Introduction
Major Approaches to Learning to Rank
Analysis of the Approaches
Benchmark Evaluation of the Approaches
Statistical Learning Theory for Ranking
Summary and Outlook
2011/5/9
Tie-Yan Liu, Copyright reserved
28
Learning to Rank Algorithms
Least Square Retrieval Function Query refinement (WWW 2008)
(TOIS 1989)
Nested Ranker (SIGIR 2006)
SVM-MAP (SIGIR 2007)
ListNet (ICML 2007)
Pranking (NIPS 2002)
MPRank (ICML 2007)
LambdaRank (NIPS 2006)
Frank (SIGIR 2007)
Learning to retrieval info (SCC 1995)
MHR (SIGIR 2007) RankBoost (JMLR 2003)
LDM (SIGIR 2005)
Large margin ranker (NIPS 2002)
IRSVM (SIGIR 2006)
RankNet (ICML 2005) Ranking SVM (ICANN 1999)
Discriminative model for IR (SIGIR 2004)
SVM Structure (JMLR 2005)
OAP-BPM (ICML 2003)
Subset Ranking (COLT 2006)
GPRank (LR4IR 2007) QBRank (NIPS 2007) GBRank (SIGIR 2007)
Constraint Ordinal Regression (ICML 2005)McRank (NIPS 2007)
AdaRank (SIGIR 2007)
RankCosine (IP&M 2007)
Relational ranking (WWW 2008)
2011/5/9
CCA (SIGIR 2007)
SoftRank (LR4IR 2007)
ListMLE (ICML 2008)
Supervised Rank Aggregation (WWW 2007)
Learning to order things (NIPS 1998)
Round robin ranking (ECML
2003)
Tie-Yan Liu, Copyright reserved
29
Key Questions to Answer
• To what respect are these learning to rank algorithms similar
and in which aspects do they differ? What are the strengths
and weaknesses of each algorithm?
• Empirically speaking, which of those many learning to rank
algorithms perform the best?
• What are the unique theoretical issues for ranking that should
be investigated?
• Are there many remaining issues regarding learning to rank to
study in the future?
2011/5/9
Tie-Yan Liu, Copyright reserved
30
Categorization of the Algorithms
Category
Algorithms
Pointwise
Approach
Regression: Least Square Retrieval Function (TOIS 1989), Regression Tree for Ordinal
Class Prediction (Fundamenta Informaticae, 2000), Subset Ranking using Regression
(COLT 2006), …
Classification: Discriminative model for IR (SIGIR 2004), McRank (NIPS 2007), …
Ordinal regression: Pranking (NIPS 2002), OAP-BPM (EMCL 2003), Ranking with
Large Margin Principles (NIPS 2002), Constraint Ordinal Regression (ICML 2005), …
Pairwise
Approach
Learning to Retrieve Information (SCC 1995), Learning to Order Things (NIPS 1998),
Ranking SVM (ICANN 1999), RankBoost (JMLR 2003), LDM (SIGIR 2005), RankNet
(ICML 2005), Frank (SIGIR 2007), MHR(SIGIR 2007), GBRank (SIGIR 2007), QBRank
(NIPS 2007), MPRank (ICML 2007), IRSVM (SIGIR 2006), LambdaRank (NIPS 2006),…
Listwise
Approach
Non-measure specific: ListNet (ICML 2007), ListMLE (ICML 2008), BoltzRank (ICML
2009) …
Measure-specific: AdaRank (SIGIR 2007), SVM-MAP (SIGIR 2007), SoftRank (LR4IR
2007), RankGP (LR4IR 2007), …
2011/5/9
Tie-Yan Liu, Copyright reserved
31
The Poinitwise Approach
The Pointwise Approach
• Reduce ranking to
– Regression
• Subset Ranking
– Classification
• Discriminative model for IR
• MCRank
– Ordinal regression
q

• PRanking
• Ranking with large margin principle
 x1 
 
 x2 
 
  
x 
 m
( x , y ), ( x , y ),..., ( x
1
2011/5/9
Tie-Yan Liu, Copyright reserved
1
2
2
m

, ym )
33
Subset Ranking using Regression
(D. Cossock and T. Zhang, COLT 2006)
• Regard relevance degree as real number, and use
regression to learn the ranking function.
L( f ; x j , y j )   f ( x j )  y j 
2
Regression-based
2011/5/9
Tie-Yan Liu, Copyright reserved
34
McRank
(P. Li, et al. NIPS 2007)
• Multi-class classification is used to learn the ranking
function.
– For document x j , the output of the classifier is ŷ j .
– Loss function: surrogate function of I{ yˆ j  y j }
• Ranking is produced by combining the outputs of the
classifiers.
K
pˆ j ,k  P( yˆ j  k ), f ( x j )   pˆ j ,k k
k 1
Classification-based
2011/5/9
Tie-Yan Liu, Copyright reserved
35
Pranking with Ranking
(K. Krammer and Y. Singer, NIPS 2002)
• Ranking is solved by ordinal regression
x  w x
Ordinal regression based
2011/5/9
Tie-Yan Liu, Copyright reserved
36
Problem with Pointwise Approach
• Properties of IR evaluation measures have not been
well considered.
– The fact is ignored that some documents are associated
with the same query and some are not.
• When the number of associated documents varies largely for
different queries, the overall loss function will be dominated by
those queries with a large number of documents.
– The position of documents in the ranked list is invisible to
the loss functions.
• The pointwise loss function may unconsciously emphasize too
much those unimportant documents (which are ranked low in the
final results).
2011/5/9
Tie-Yan Liu, Copyright reserved
37
The Pairwise Approach
The Pairwise Approach
• Reduce ranking to pairwise classification
–
–
–
–
–
RankNet and Frank
RankBoost
Ranking SVM
MHR
IR-SVM
q

 x1

 x2

 
x
 m








x1 , x2 ,1, x2 , x1 ,1,...,




x 2 , xm ,1, xm , x2 ,1 

2011/5/9
Tie-Yan Liu, Copyright reserved
39
RankNet
(C. Burges, et al. ICML 2005)
• Target probability:
Pu ,v  1, if yu ,v  1; Pu ,v  0, otherwise
• Modeled probability:
Pu ,v ( f ) 
exp( f ( xu )  f ( xv ))
1  exp( f ( xu )  f ( xv ))
• Cross entropy as the loss function
l ( f ; xu , xv , yu ,v )
  Pu ,v log Pu ,v ( f )  (1  Pu ,v ) log(1  Pu ,v ( f ))
• Use neural network as model, and gradient descent as
algorithm, to optimize the cross-entropy loss.
2011/5/9
Tie-Yan Liu, Copyright reserved
40
FRank
(M. Tsai, et al. SIGIR 2007)
• Similar setting to that of RankNet
• New loss function: fidelity
L( f ; xu , xv , yu ,v )
 1  Pu ,v  Pu ,v ( f )  (1  Pu ,v )  (1  Pu ,v ( f ))
2011/5/9
Tie-Yan Liu, Copyright reserved
41
RankBoost
(Y. Freund, et al. JMLR 2003)
• Exponential loss
L( f ; xu , xv , yu ,v )  exp(  yu,v ( f ( xu )  f ( xv )))
2011/5/9
Tie-Yan Liu, Copyright reserved
42
RankBoost
• Due to the analogy to AdaBoost, RankBoost
inherits
Weak
learnermany
that can
nice properties, such as the ability in feature
selection,
classify
important pairs
more correctly
will have
convergence in training, and certain generalization
abilities.
larger weight α
Hypothesis is the
combination of weak
learners
Emphasize those hard
pairs by setting their
distributions.
2011/5/9
Tie-Yan Liu, Copyright reserved
43
Ranking SVM
(R. Herbrich, et al. , Advances in Large Margin Classifiers, 2000;
T. Joachims, KDD 2002)
• Hinge loss
L( f ; xu , xv , yu,v )  1  yu,v  f ( xu )  f ( xv )
2011/5/9
Tie-Yan Liu, Copyright reserved
44
Ranking SVM
n
1
min || w || 2 C    u(,iv)
2
i 1 u ,v: yu( i,v) 1


wT xu(i )  xv(i )  1   u(,iv) , if yu(i,)v  1.
 uv(i )  0, i  1,..., n.
xu-xv as positive instance of
learning
Use SVM to perform binary
classification on these instances,
to learn model parameter w
• Since Ranking SVM is well rooted in the framework of SVM, it
inherits nice properties of SVM. For example, with the help of
margin maximization, Ranking SVM can have a good
generalization ability. Kernel tricks can also be applied to
Ranking SVM, so as to handle complex non-linear problems.
2011/5/9
Tie-Yan Liu, Copyright reserved
45
Problems with Pairwise Approach
• When we are given the relevance judgment in terms of
multiple ordered categories, converting it to pairwise
preference will lead to the missing of the information
about the finer granularity in the relevance judgment
• The distribution of document pair number is more
skewed than the distribution of document number,
with respect to different queries.
• The pairwise approach is more sensitive to noisy label
than the pointwise approach.
• Most of the pairwise ranking algorithms have not
considered the position in the final ranking results.
2011/5/9
Tie-Yan Liu, Copyright reserved
46
Improved Algorithms
• Multi-hyperplane ranker (MHR)
• IR-SVM
• LambdaRank
2011/5/9
Tie-Yan Liu, Copyright reserved
47
Multi-Hyperplane Ranker
(T. Qin, et al. SIGIR 2007)
• If we have K categories, then will have K(K-1)/2 pairwise
preferences between two labels.
– Learn a model fk,l for the pair of categories k and l, using any previous
method (for example Ranking SVM).
• Testing
– Rank Aggregation
f ( x )    k ,l f k , l ( x )
k ,l
2011/5/9
Tie-Yan Liu, Copyright reserved
48
IR-SVM
(J. Xu, et al. WWW 2007)
• Introduce query-level normalization to the pairwise loss
function.
– Use the number of document pairs to normalize the pairwise loss.
• IRSVM (Y. Cao, J. Xu, et al. SIGIR 2006; T. Qin, T.-Y. Liu, et al. IPM 2007)
N
1
1
min || w ||2 C  ~ ( i ) uv( i )
2
i 1 m
u ,v
Query-level normalizer
Significant improvement has been observed by using the query-level normalizer.
2011/5/9
Tie-Yan Liu, Copyright reserved
49
LambdaRank
(C. Burges, NIPS 2006)
• Basic idea
– The evaluation measures (which are position based) are
directly used to define the gradient with respect to each
document pair in the training process.
• Why is it feasible to directly define the gradient?
2011/5/9
Tie-Yan Liu, Copyright reserved
50
LambdaRank
• Lambda function
– The gradient is directly defined as a lambda function,
without caring what the corresponding loss function really
looks like.
– For each pair of documents xu and xv, in each round of
optimization, their scores are updated by +λ and − λ.
2011/5/9
Tie-Yan Liu, Copyright reserved
51
The Listwise Approach
Listwise Approach
• Measure-specific loss
– Optimize the IR evaluation measures directly
• Non measure-specific loss
– Optimize a loss function defined on all documents
associated with a query, according to the unique
properties of ranking
2011/5/9
Tie-Yan Liu, Copyright reserved
53
Measure-specific Listwise Ranking
• Also known as “Direct Optimization of IR Measures”.
• It is natural to directly optimize what is used to evaluate the
ranking results.
• However, it is non-trivial.
– Evaluation measures such as NDCG are non-continuous and nondifferentiable since they depend on the rank positions.
(S. Robertson and H. Zaragoza, Information Retrieval 2007).
– It is challenging to optimize such objective functions, since most
optimization techniques in the literature were developed to handle
continuous and differentiable cases.
2011/5/9
Tie-Yan Liu, Copyright reserved
54
Tackle the Challenges
• Approximate the objective
– Soften (approximate) the evaluation measure so as to make it smooth
and differentiable (SoftRank)
• Bound the objective
– Optimize a smooth and differentiable upper bound of the evaluation
measure (SVM-MAP)
• Optimize the non-smooth objective directly
– Use IR measure to update the distribution in Boosting (AdaRank)
– Use genetic programming (RankGP)
2011/5/9
Tie-Yan Liu, Copyright reserved
55
SoftRank
(M. Taylor, et al. LR4IR 2007/WSDM 2008)
• Key Idea:
– Avoid sorting by treating scores as random variables
• Steps
– Construct score distribution
– Map score distribution to rank distribution
– Compute expected NDCG (SoftNDCG) which is smooth and
differentiable.
– Optimize SoftNDCG by gradient descent.
2011/5/9
Tie-Yan Liu, Copyright reserved
56
Construct Score Distribution
• Considering the score for each document sj as random
variable, by using Gaussian.
p(s j )  Ν (s j | f ( x j ),  s2 )
2011/5/9
Tie-Yan Liu, Copyright reserved
57
From Scores to Rank Distribution
• Probability of du being ranked before dv

 uv   Ν (s | f ( xu )  f ( xv ),2 s2 )ds
0
• Rank Distribution
p (ju ) (r )  p (ju 1) (r  1) uj  p (ju 1) (r )(1   uj )
Define the probability of a document being ranked at a particular position
2011/5/9
Tie-Yan Liu, Copyright reserved
58
Soft NDCG
• Expected NDCG as the objective, or in the language of loss
function,
m
m 1
L( f ; x, y )  1  E[ NDCG]  1  Z m  (2  1)r p j ( r )
j 1
yj
r 0
– Only Pj(r) contain the model parameter w, and it is differential with
respect to w.
• A neural network is used as the model, and gradient descent
is used as the algorithm to optimize the above loss function.
2011/5/9
Tie-Yan Liu, Copyright reserved
59
SVM-MAP
(Y. Yue, et al. SIGIR 2007)
• Use the framework of Structured SVM for optimization
– I. Tsochantaridis, et al. ICML 2004; T. Joachims, ICML 2005.
• Sum of slacks ∑ξ upper bounds ∑ (1-AP).
1 2 C
min w 
2
N
N
(i )


i 1
y '  y (i ) : wT  (y (i ) , x (i ) )  wT  (y ' , x (i ) )  (y (i ) , y ' )   (i )
 ( y ' , x) 
 ( y'  y'
u
u ,v: yu 1, yv 0
( y, x ) 
2011/5/9
( x
v
u
u ,v: yu 1, yv 0
)( xu  xv )
(y (i ) , y' )  1  AP(y' )
 xv )
Tie-Yan Liu, Copyright reserved
60
Challenges in SVM-MAP
• For AP, the true ranking is a ranking where the relevant
documents are all ranked in the front, e.g.,
• An incorrect ranking would be any other ranking, e.g.,
• Exponential number of rankings, thus an exponential number
of constraints!
•  Solution: use a working set only containing the most
violated constraints.
2011/5/9
Tie-Yan Liu, Copyright reserved
61
Finding the Most Violated Constraint
• Most violated y’:
arg max ( y, y' )  wT ( y' , x)
y'
• If the relevance at each position is fixed, Δ(y,y’) will be the same.
• Strategy (with a complexity of O(mlogm))
– Sort the relevant and irrelevant documents and form a perfect ranking.
– Start with the perfect ranking and swap two adjacent relevant and
irrelevant documents, so as to find the optimal interleaving of the two
sorted lists.
2011/5/9
Tie-Yan Liu, Copyright reserved
62
AdaRank
(J. Xu and H. Li. SIGIR 2007)
Weak learner that can
• Embed IR evaluation measure in the distribution
rank important queries
update of Boosting
more correctly will have
larger weight α
Emphasize those hard
queries by setting their
distributions.
2011/5/9
Tie-Yan Liu, Copyright reserved
63
RankGP
(J. Yeh, et al. LR4IR 2007)
• Define ranking function as a tree
• Learning
– Single-population GP, which has been widely used to
optimize non-smoothing non-differentiable objectives.
• Evolution mechanism
– Crossover, mutation, reproduction, tournament selection.
• Fitness function
– IR evaluation measure (MAP).
2011/5/9
Tie-Yan Liu, Copyright reserved
64
Non Measure-specific Listwise Ranking
• Defining listwise loss functions based on the
understanding on the unique properties of ranking
for IR.
• Representative Algorithms
– ListNet
– ListMLE
– BoltzRank
2011/5/9
Tie-Yan Liu, Copyright reserved
65
Ranking Loss is Non-trivial!
• An example:
– function f:
f(A)=3, f(B)=0, f(C)=1
– function h:
h(A)=4, h(B)=6, h(C)=3
– ground truth g: g(A)=6, g(B)=4, g(C)=3
ACB
BAC
ABC
• Question: which function is closer to ground truth?
– Based on pointwise similarity: sim(f,g) < sim(g,h).
– Based on pairwise similarity: sim(f,g) = sim(g,h)
– Based on cosine similarity between score vectors?
f: <3,0,1>
g:<6,4,3>
h:<4,6,3>
Closer!
sim(f,g)=0.85
sim(g,h)=0.93
• However, according to NDCG, f should be closer to g!
2011/5/9
Tie-Yan Liu, Copyright reserved
66
ListNet
(Z. Cao, et al. ICML 2007)
• Ranked list ↔ Permutation probability distribution
• More informative representation for ranked list:
permutation and ranked list has 1-1 correspondence.
2011/5/9
Tie-Yan Liu, Copyright reserved
67
Defining Permutation Probability
• Probability of a permutation is defined with Luce Model
m
P( | f )  
j 1
 ( f ( x ( j ) ))

m
k j
 ( f ( x ( k ) ))
• Example:
PABC | f  
  f ( A) 
  f (B) 
  f (C) 


  f (A)     f (B)     f (A)    f (B)    f (C)    f (C) 
P(A ranked No.1)
P(B ranked No.2 | A ranked No.1)
= P(B ranked No.1)/(1- P(A ranked No.1))
P(C ranked No.3 | A ranked No.1, B ranked No.2)
2011/5/9
Tie-Yan Liu, Copyright reserved
68
Properties of Luce Model
• Continuous, differentiable, and concave w.r.t. score s.
• Suppose f(A) = 3, f(B)=0, f(C)=1, P(ACB) is largest and P(BCA) is
smallest; for the ranking based on f: ACB, swap any two,
probability will decrease, e.g., P(ACB) > P(ABC)
• Translation invariant, when  (s)  exp( s)
• Scale invariant, when  (s)  a  s
2011/5/9
Tie-Yan Liu, Copyright reserved
69
Distance between Ranked Lists
φ = exp
Using KL-divergence
to measure difference
between distributions
dis(f,g) = 0.46
dis(g,h) = 2.56
2011/5/9
Tie-Yan Liu, Copyright reserved
70
K-L Divergence Loss
• Loss function = KL-divergence between two permutation
probability distributions (φ = exp)
L( f ; x,  y )  DPy   || P |   f (x) 
Probability distribution defined
by the ground truth
Probability distribution defined
by the model output
• Model = Neural Network
• Algorithm = Gradient Descent
2011/5/9
Tie-Yan Liu, Copyright reserved
71
Limitations of ListNet
• Too complex to use in practice: O(m⋅m!)
• By using the top-k Luce model, the complexity of ListNet can
be reduced to polynomial order of m.
• However,
– When k is large, the complexity is still high!
– When k is small, the information loss in the top-k Luce model will
affect the effectiveness of ListNet.
2011/5/9
Tie-Yan Liu, Copyright reserved
72
Solution
• Given the scores outputted by the ranking model, compute
the likelihood of a ground truth permutation, and maximize it
to learn the model parameter.
• The complexity is reduced from O(m⋅m!) to O(m).
2011/5/9
Tie-Yan Liu, Copyright reserved
73
ListMLE
(F. Xia, et al. ICML 2008)
• Likelihood Loss
L( f ; x,  y )   log P( y |  ( f (x)))
• Model = Neural Network
• Algorithm = Stochastic Gradient Descent
2011/5/9
Tie-Yan Liu, Copyright reserved
74
Extension of ListMLE
• Dealing with other types of ground truth labels using
the concept of “equivalent permutation set”.
– For relevance degree
 y  { y | u  v, if l 1 (u )  l 1 (u ) }
y
y
– For pairwise preference
 y  { y | u  v, if l 1 (u ), 1 (u )  1}
y
y
L( f , x,  y )  min  log P y |   f (w, x)
 y  y
2011/5/9
Tie-Yan Liu, Copyright reserved
75
BoltzRank
(M. Volkovs and R. Zemel, ICML 2009)
• BoltzRank is also based on permutation probability
models.
• BoltzRank utilizes a scoring function composed of
individual and pairwise potentials to define a
conditional probability distribution, in the form of a
Boltzmann distribution, over all permutations of
documents retrieved for a given query.
2011/5/9
Tie-Yan Liu, Copyright reserved
76
BoltzRank
• E(π|s) thus represents the lack of compatibility
between the relative document orderings given by p
and those given by s, with more positive energy
indicating less compatibility.
• Using the energy function, we can now define the
conditional Boltzmann distribution over document
permutations by exponentiating and normalizing
2011/5/9
Tie-Yan Liu, Copyright reserved
77
BoltzRank
• After defining the permutation probability as above,
one can follow the idea in ListMLE or ListNet to
define the loss function.
• It is also possible to extend the above idea to
optimize evaluation measures.
– For example, suppose the evaluation measure is NDCG,
then its expected value with respect to all the possible
permutations can be computed as follows.
2011/5/9
Tie-Yan Liu, Copyright reserved
78
Discussions
• Advantages
– Take all the documents associated with the same
query as the learning instance
– Rank position is visible to the loss function
• Problems
– Complexity issue.
– The use of the position information is insufficient.
2011/5/9
Tie-Yan Liu, Copyright reserved
79
Outline
•
•
•
•
•
•
Introduction
Major Approaches to Learning to Rank
Analysis of the Approaches
Benchmark Evaluation of the Approaches
Statistical Learning Theory for Ranking
Summary and Outlook
2011/5/9
Tie-Yan Liu, Copyright reserved
80
Question
• Different loss functions are used in different
approaches, while the models learned with all
the approaches are evaluated by IR measures.
• What are the relationship between the loss
functions and IR evaluation measures?
2011/5/9
Tie-Yan Liu, Copyright reserved
81
The Pointwise Approach
• Many pointwise loss functions can upper
bound (1-NDCG).
• Minimization of the pointwise loss functions
can lead to the maximization of NDCG.
2011/5/9
Tie-Yan Liu, Copyright reserved
82
The Pointwise Approach
• For regression based algorithms (D. Cossock and T. Zhang,
COLT 2006)
1
1  NDCG( f ) 
Zm
1/ 


 2 j 


j

1


m
1/ 


   f ( x j )  y j  


j

1


m
• For multi-class classification based algorithms (P. Li, et al.
NIPS 2007)
15
1  NDCG( f ) 
Zm
2011/5/9
2
m
m
 m 2

2  j  m jm    I{ y j  yˆ j }
j 1
 j 1
 j 1
Tie-Yan Liu, Copyright reserved
83
The Pointwise Approach
• The minimization of the losses is only the sufficient
condition of minimizing (1-NDCG).
xi , yi
 x1 ,4 


x
,
3
 2 
 x ,2 
 3 
 x ,1 
 4 
xi , f ( xi )
Z m  21.4
1  NDCG( f )  0
m
I
j 1
{ y j  f ( x j )}
4
 x1 ,3 


x
,
2
 2 
 x ,1 
 3 
 x ,0 
 4 
2
2
 m

m
m


  m
15  
1
1
2 
  m 
    I{ y j  f ( x j )}  1.15  1
Z m  j 1  log( j  1) 
log(
j

1
)
j 1 
  j 1


2011/5/9
Tie-Yan Liu, Copyright reserved
84
Pairwise and Non-measure Specific Listwise
Approaches
(W. Chen, et al. NIPS 2009)
• Many pairwise and non-measure specific listwise
loss functions are surrogate functions, and also
upper bounds of a quantity named the “essential
loss”.
• The essential loss is an upper bound of (1-NDCG)
and (1-MAP).
• The minimization of these surrogate loss
functions can lead to the maximization of NDCG
and MAP.
2011/5/9
Tie-Yan Liu, Copyright reserved
85
Essential Loss for Ranking
• Model ranking as a sequence of classifications
Ground truth permutation:
Prediction of the ranking function:
y
 A
B
 
 
C 
 D 
2011/5/9

y
B
 A  B 
  
  C 
D D
C   
y  {A  B  C  D}
  {B  A  D  C}

y

B
   C   D 
D    
C   D  C 
 
Classifier
√
Tie-Yan Liu, Copyright reserved
86
Essential Loss vs. Ranking Measures
1) Both (1-NDCG) and (1-MAP) are upper bounded by the
essential loss.
2) The zero value of the essential loss is a necessary and
sufficient condition for the zero values of (1-NDCG) and
(1-MAP).
2011/5/9
Tie-Yan Liu, Copyright reserved
87
Essential Loss vs. Surrogate Losses
1) Many pairwise and listwise loss functions are upper
bounds of the essential loss.
2) Therefore, the pairwise and listwise loss functions are also
upper bounds of (1-NDCG) and (1-MAP).
2011/5/9
Tie-Yan Liu, Copyright reserved
88
Discussions
• Revealed that different loss functions are upper
bounds of the same thing.
• Explained why many different pairwise and
listwise algorithms perform similarly well.
• Suggest better algorithms
– The essential loss is a tighter upper bound of
measure-based ranking errors than existing surrogate
loss functions.
– Our experiments showed that higher ranking
performances can be achieved by optimizing such a
tighter bound.
2011/5/9
Tie-Yan Liu, Copyright reserved
89
Measure-specific Listwise Approach
• To what degree is a surrogate measure
correlated with the ranking measure?
• Can the maximization of a surrogate measure
lead to the same ranking model that
maximizes the ranking measure?
2011/5/9
Tie-Yan Liu, Copyright reserved
90
Tendency Correlation as a Tool
(Y. He and T.-Y. Liu, IR Journal 2010)
0
0.2
Finding the best linear
0.4 transformation, by
applying which to one of the functions,
the maximum difference between the
tendencies of the two functions in the
entire input space can be minimized.
2011/5/9
The “tendency” of a function with
respect to two inputs is defined as
the difference in the values of the
function at the two inputs.
Tie-Yan Liu, Copyright reserved
91
Theoretical Justification of Tendency Correlation
If the tendency correlation of a surrogate measure is
arbitrarily strong, its optimization will lead to the
ranking model that can optimize the corresponding
ranking measure, on both training and test sets.
2011/5/9
Tie-Yan Liu, Copyright reserved
92
Major Results - SoftRank
1) When a parameter in SoftRank is set to be small, the
tendency correlation of its surrogate measure will
become strong, regardless of the data distribution.
2) However, in this case, the surrogate measure will contain
many local optima, therefore robust global optimization
techniques are needed.
2011/5/9
Tie-Yan Liu, Copyright reserved
93
Major Results – SVM-MAP, etc.
1) The surrogate measures in SVM-MAP, SVM-NDCG, DORM,
and PermuRank are concave and easy to optimize.
2) No matter how the parameters in these algorithms are set,
for certain data distributions, the tendency correlation of
their surrogate measures can never be as strong as expected.
2011/5/9
Tie-Yan Liu, Copyright reserved
94
Summary
• With the tool of tendency correlation, we reveal
the relationship between surrogate measures and
original ranking measures.
• The theorems well explain previous empirical
observations
– Some methods can perform well on different datasets,
while some others cannot on some datasets.
• The theorems provide a guideline for trade-off
between effectiveness and ease of optimization.
2011/5/9
Tie-Yan Liu, Copyright reserved
95
Outline
•
•
•
•
•
•
Introduction
Major Approaches to Learning to Rank
Analysis of the Approaches
Benchmark Evaluation of the Approaches
Statistical Learning Theory for Ranking
Summary and Outlook
2011/5/9
Tie-Yan Liu, Copyright reserved
96
Benchmark Datasets for Learning
to Rank
Role of a Benchmark Dataset
• Enable researchers to focus on technologies
instead of experimental preparations.
• Enable fair comparison among algorithms.
• Examples
– Reuters and RCV-1 for text classification
– UCI for general machine learning
2011/5/9
Tie-Yan Liu, Copyright reserved
98
The LETOR Collection
(T.-Y. Liu, et al. LR4IR 2007)
• The first benchmark dataset for learning to
rank
– Document corpora
– Document sampling
– Feature extraction
– Meta information
2011/5/9
Tie-Yan Liu, Copyright reserved
99
Document Corpora
• The “.gov” corpus
• The OHSUMED corpus
– 106 queries
2011/5/9
Tie-Yan Liu, Copyright reserved
100
Document Sampling
• The “.gov” corpus
– Top 1000 documents per query returned by BM25
• The OHSUMED corpus
– All judged documents
2011/5/9
Tie-Yan Liu, Copyright reserved
101
Feature Extraction
• The “.gov” corpus
– 64 features
– TF, IDF, BM25, LMIR, PageRank, HITS, Relevance
Propagation, URL length, etc.
• The OHSUMED corpus
– 40 features
– TF, IDF, BM25, LMIR, etc.
2011/5/9
Tie-Yan Liu, Copyright reserved
102
Meta Information
• Statistical information about the corpus
– Number of documents, number of streams, and number of
(unique) terms in each stream.
• Raw information of the documents associated with
each query
– Term frequency, and document length.
• Relational information
– Hyperlink graph, sitemap information, and similarity
relationship matrix of the corpora.
2011/5/9
Tie-Yan Liu, Copyright reserved
103
LETOR Website
2011/5/9
Tie-Yan Liu, Copyright reserved
104
New Benchmark Datasets
• Yahoo! Challenge Datasets
• Microsoft Learning to Rank Datasets
• Both are large-scale datasets retired from
commercial search engines.
2011/5/9
Tie-Yan Liu, Copyright reserved
105
Yahoo! Challenge Datasets
• Released on March 2010.
• Accessible to participants of the challenge
• About 30K queries
• 700K documents
• 700 features for each document.
• http://learningtorankchallenge.yahoo.com/
2011/5/9
Tie-Yan Liu, Copyright reserved
106
Yahoo! Challenge Datasets
2011/5/9
Tie-Yan Liu, Copyright reserved
107
Microsoft Learning to Rank Datasets
• Released on June 16, 2010.
• Publicly available
• More than 30K queries
• 3.7M documents
• 136 features for each document.
• http://research.microsoft.com/en-us/projects/mslr/
2011/5/9
Tie-Yan Liu, Copyright reserved
108
Microsoft Learning to Rank Datasets
2011/5/9
Tie-Yan Liu, Copyright reserved
109
Evaluation Results on LETOR
Datasets
Algorithms under Investigation
• The pointwise approach
– Linear regression
• The pairwise approach
– Ranking SVM, RankBoost, Frank
• The listwise approach
– ListNet
– SVM-MAP, AdaRank
Linear ranking model (except RankBoost)
2011/5/9
Tie-Yan Liu, Copyright reserved
111
Experimental Results – TD2003
2011/5/9
Tie-Yan Liu, Copyright reserved
112
Experimental Results – TD2004
2011/5/9
Tie-Yan Liu, Copyright reserved
113
Experimental Results – NP2003
2011/5/9
Tie-Yan Liu, Copyright reserved
114
Experimental Results – NP2004
2011/5/9
Tie-Yan Liu, Copyright reserved
115
Experimental Results – HP2003
2011/5/9
Tie-Yan Liu, Copyright reserved
116
Experimental Results – HP2004
2011/5/9
Tie-Yan Liu, Copyright reserved
117
Experimental Results - OHSUMED
2011/5/9
Tie-Yan Liu, Copyright reserved
118
Experimental Results - Winning Numbers
2011/5/9
Tie-Yan Liu, Copyright reserved
119
Experimental Results - Winning Numbers
40
35
30
Regression
RankSVM
25
RankBoost
20
FRank
ListNet
15
AdaRank
10
SVMmap
5
0
N@1
2011/5/9
N@3
N@10
P@1
P@3
Tie-Yan Liu, Copyright reserved
P@10
MAP
120
Discussions
• NDCG@1
– {ListNet, AdaRank} > {SVM-MAP, RankSVM}
> {RankBoost, FRank} > {Regression}
• NDCG@3, @10
– {ListNet} > {AdaRank, SVM-MAP, RankSVM, RankBoost}
> {FRank} > {Regression}
• MAP
– {ListNet} > {AdaRank, SVM-MAP, RankSVM}
> {RankBoost, FRank} > {Regression}
The listwise approach has certain advantages, especially in
correctly predicting top-ranked documents.
2011/5/9
Tie-Yan Liu, Copyright reserved
121
Notes
• The experimental results can be found at the
LETOR website
– http://research.microsoft.com/~letor/
• The results are still primal
– The implementations of the these algorithms are
not perfect
– Almost every algorithm can be further improved.
– The datasets are relatively small.
2011/5/9
Tie-Yan Liu, Copyright reserved
122
Outline
•
•
•
•
•
•
Introduction
Major Approaches to Learning to Rank
Analysis of the Approaches
Benchmark Evaluation of the Approaches
Statistical Learning Theory for Ranking
Summary and Outlook
2011/5/9
Tie-Yan Liu, Copyright reserved
123
Why Theory?
• In practice, one can only observe experimental results on
relatively small datasets. To some extent, such empirical
results are not trustworthy, because
– Small training set cannot fully realize the potential of a learning
algorithm.
– Small test set cannot reflect the true performance of an
algorithm, since the query space is too huge to be represented
by a small number of samples.
• Statistical learning theory for ranking analyzes the
performance of an algorithm when the training data is infinite
and the test data is really randomly sampled.
2011/5/9
Tie-Yan Liu, Copyright reserved
124
Generalization Analysis
• More specifically, in the training phase, one
minimizes the empirical risk
on the training
data.
• In the test phase, we evaluate the expected risk
of the ranking model on any sample.
• Generalization analysis is concerned with the bound
of the difference between the expected and empirical
risks, when the number of training data approaches
infinity.
2011/5/9
Tie-Yan Liu, Copyright reserved
125
Generalization Analysis
?
• An algorithm is regarded as better than the other
algorithm if its empirical risk can converge to the
expected risk but that of the other cannot.
• An algorithm is regarded as better than the other if
its corresponding convergence rate is faster than that
of the other.
2011/5/9
Tie-Yan Liu, Copyright reserved
126
Generalization Analysis
• With different assumptions on the data, we
have different definitions of the empirical and
expected risks.
• As a consequence, we have different
generalization bounds.
2011/5/9
Tie-Yan Liu, Copyright reserved
127
Statistical Ranking Framework
• Document Ranking
– Assume that all the documents are i.i.d. no matter
they are associated with the same query or not
• Subset Ranking
– Regard documents as deterministically given, and only
regard queries as i.i.d. random variables
• Two-layer Ranking
– Assume that only the documents associated with the
same query are i.i.d. (the distribution depends on
queries), but queries are also i.i.d. in the query space
2011/5/9
Tie-Yan Liu, Copyright reserved
128
Document Ranking Framework
• Pointwise Approach
• Pairwise Approach
• Listwise Approach
– Cannot be explained using document ranking framework
2011/5/9
Tie-Yan Liu, Copyright reserved
129
Subset Ranking Framework
• Pointwise Approach
• Pairwise Approach
• Listwise Approach
2011/5/9
Tie-Yan Liu, Copyright reserved
130
Two-Layer Ranking Framework
• Pointwise Approach
• Pairwise Approach
• Listwise Approach
– Cannot be explained using document ranking framework
2011/5/9
Tie-Yan Liu, Copyright reserved
131
Two Kinds of Generalization Bounds
• Uniform generalization analysis
– To reveal a bound of the generalization ability for
any function in the function class under
investigation.
• Algorithm-dependent generalization analysis
– Only applicable to the model learned from the
training data using a specific algorithm.
2011/5/9
Tie-Yan Liu, Copyright reserved
132
Uniform Generalization Bound for
Document Ranking
(Y. Freund, et al. JMLR 2003)
• W.r.t. pairwise 0-1 loss, based on VC dimension.
Convergence Rate: log 𝑚 /𝑚
2011/5/9
Tie-Yan Liu, Copyright reserved
133
Uniform Generalization Bound for
Document Ranking
(S. Agarwal, et al. JMLR 2005)
• W.r.t. pairwise 0-1 loss, based on Rank Shatter
Coefficient.
Convergence Rate:
For linear scoring function in 1-D space:
1
𝑚
For linear scoring function in d-D space: log 𝑚 /𝑚
2011/5/9
Tie-Yan Liu, Copyright reserved
134
Uniform Generalization Bound for
Document Ranking
(P. Bartlett and S. Mendelson, JMLR 2002)
• W.r.t. pairwise 0-1 loss, for k-partite ranking, based
on properties of U-statistics.
1
Convergence Rate:
𝑚
2011/5/9
Tie-Yan Liu, Copyright reserved
135
Uniform Generalization Bound for
Document Ranking
(S. Clemencon, et al. Annual Stats 2008)
• W.r.t. pairwise surrogate loss, based on VC dimension.
1
Convergence Rate:
𝑚
2011/5/9
Tie-Yan Liu, Copyright reserved
136
Uniform Generalization Bound for
Subset Ranking
(Y. Lan, et al. ICML 2009)
• W.r.t. listwise surrogate loss, based on Rademacher
average.
• The generalization bound for ListMLE is much tighter than
that for ListNet, especially when the length of the list is large.
• The generalization bound for ListMLE decreases
monotonously, while that of ListNet increases monotonously,
with respect to the length of the list.
1
Convergence
• The linear transformation
functionRate:
is the best choice in terms
𝑛
of the generalization bound in most cases
2011/5/9
Tie-Yan Liu, Copyright reserved
137
Uniform Generalization Bound for
Two-Layer Ranking
(W. Chen, et al. NIPS 2010)
• W.r.t. pairwise 0-1/surrogate loss, based on
Rademacher average.
• The increasing number of either queries or documents per
query in the training data will enhance the two-layer
generalization ability.
• Only if m  ∞ and n  ∞ simultaneously, does the two-layer
generalization bound uniformly converge.
• If we only have the budget to label C documents in total, there
is an optimal tradeoff between the number of training queries
and that of training documents per query.
2011/5/9
Tie-Yan Liu, Copyright reserved
138
Algorithm-dependent Generalization
Bound for Document Ranking
(S. Agarwal, et al. ALT 2008 and COLT 2005)
• W.r.t. pairwise surrogate loss, based on stability.
For example, Ranking SVM is a stable algorithm,
16𝜅2
and its stability coefficient is β(𝑚) = λ𝑚 , where
K 𝑥, 𝑥 ≤ 𝜅 2 < ∞.
2011/5/9
Tie-Yan Liu, Copyright reserved
140
Algorithm-dependent Generalization
Bound for Subset Ranking
(Y. Lan, et al. ICML 2008)
• W.r.t. pairwise surrogate loss, based on query-level stability.
For Ranking SVM, 𝜏 𝑛 =
For IR-SVM, 𝜏 𝑛 =
4𝜅2
.
λ𝑛
4𝜅2
𝑚𝑖
max 1
𝑖
λ𝑛
𝑚
𝑛
.
The convergence rates for these two algorithms are both O(1/n). By
comparing the case with a finite number of training queries, the
bound for IR-SVM is much tighter than that for Ranking SVM.
2011/5/9
Tie-Yan Liu, Copyright reserved
141
Summary
2011/5/9
Tie-Yan Liu, Copyright reserved
142
Outline
•
•
•
•
•
•
Introduction
Major Approaches to Learning to Rank
Analysis of the Approaches
Benchmark Evaluation of the Approaches
Statistical Learning Theory for Ranking
Summary and Outlook
2011/5/9
Tie-Yan Liu, Copyright reserved
143
Listwise approach
- non-measure specific
- measure specific
• NIPS workshop on • ICML workshop on • 1st SIGIR workshop
learning to rank
structure output
on LR4IR
Data
Activity
Pointwise approach
Pairwise approach
Algorithms
Theory
• LETOR 1.0
• LETOR 2.0
•
•
•
•
Pranking
DMIR
Ranking SVM
RankBoost
< 2005
2011/5/9
• Gaussian process for
ordinal regression
• RankNet
2005
2006
WWW tutorial
•
SIGIR tutorial
•
2nd LR4IR workshop •
1st SIGIR workshop on •
•
diversity
• LETOR 3.0
• Generalization of
pairwise methods
• Consistency of
pointwise methods
•
•
•
• Subset Ranking •
• IR-SVM
•
• LambdaRank
•
• P-norm push
•
• Nested Ranker •
• RankCosine
•
•
•
•
•
• Consistency of
Listwise methods
MC-Rank
GBRank
MHR
Frank
ListNet
Supervised RA
RankGP
AdaRank
SVM-MAP
2007
Tie-Yan Liu, Copyright reserved
•
•
•
•
•
•
•
•
•
•
•
•
WWW tutorial
ACL tutorial
IRJ special issue
3rd LR4IR workshop
2nd Diversity workshop
• LETOR 4.0
• Generalization of
listwise methods
IsoRank
QBRank
•
CDN Ranker
•
ListMLE
•
Relational Ranking
•
GlobalRank
•
Query-dependent Ranker
AppRank
•
PermuRank
SoftRank
G-SoftRank
SVM-NDCG
2008
OWA Rank
BoltzmanRank
Bayes-Luce
SmoothRank
Robust sparse
ranker
……
2009
144
Answer to Question 1
• To what respect are these learning to rank algorithms
similar and in which aspects do they differ? What are
the strengths and weaknesses of each algorithm?
Approach
Modeling
Pointwise
Regression/
classification/
ordinal
regression
Pairwise
Pairwise
classification
Listwise
2011/5/9
Ranking
Pro
Con
Easy to leverage
existing theories
and algorithms
• Accurate ranking
≠ Accurate score or
category
•Effectively
• Position info is
invisible to the loss minimize (1NDCG).
query-level and
position based
• More complex
• New theory
needed
Tie-Yan Liu, Copyright reserved
Connection
145
Answer to Question 2
• Empirically speaking, which of those many learning
to rank algorithms perform the best?
– LETOR makes it possible to perform fair comparisons
among different learning to rank methods.
– Empirical studies on LETOR have shown that the listwise
ranking algorithms seem to have certain advantages over
other algorithms, especially in predicting the top-ranked
documents, and the pairwise ranking algorithms seem to
outperform the pointwise algorithms.
2011/5/9
Tie-Yan Liu, Copyright reserved
146
Answer to Question 3
• What are the unique theoretical issues for ranking
that should be investigated?
– Unique properties of ranking for IR: evaluation is
performed at query level and is position based.
– The risks should also be defined at the query level, and a
new theoretical framework is needed for conducting
analyses on the learning to rank methods.
– The “true loss” for ranking should consider the position
information in the ranking result, but not as simple as the
0−1 loss in classification.
2011/5/9
Tie-Yan Liu, Copyright reserved
147
Answer to Question 4
• Are there many remaining issues regarding learning
to rank to study in the future?
–
–
–
–
–
–
–
2011/5/9
Learning from logs
Feature engineering
Advanced ranking model
Ranking theory
Stability of ranking
Online learning
Large-scale learning to rank
Tie-Yan Liu, Copyright reserved
148
Learning from Logs
• Data scale matters a lot!
– Ground truth mining from logs is one of the possible
ways to construct large-scale training data.
• Ground truth mining may lose information .
– User sessions, frequency of clicking a certain
document, frequency of a certain click pattern, and
diversity in the intentions of different users.
• Directly learn from logs.
– E.g., Regard click-through logs as the “ground truth”,
and define its likelihood as the loss function.
2011/5/9
Tie-Yan Liu, Copyright reserved
149
Feature Engineering
• Ranking is deeply rooted in IR
– Not just new algorithms and theories.
• Feature engineering is very importance
– How to encode the knowledge on IR accumulated
in the past half a century in the features?
– Currently, these kinds of works have not been
given enough attention to.
2011/5/9
Tie-Yan Liu, Copyright reserved
150
Advanced Ranking Model
• Listwise ranking function
• Generative ranking model
• Semi-supervised ranking
• Relational ranking
• Query-dependent ranking
• ……
2011/5/9
Tie-Yan Liu, Copyright reserved
151
Ranking Theory
•
•
•
•
•
A more unified view on loss functions
True loss for ranking
Statistical consistency for ranking
Complexity of ranking function class
……
2011/5/9
Tie-Yan Liu, Copyright reserved
152
References
2011/5/9
Tie-Yan Liu, Copyright reserved
153
Acknowledgement
• I would like to thank
– My colleagues Tao Qin, Jun Xu, Hang Li, and Wei-Ying Ma
– Our interns Zhe Cao, Ming-Feng Tsai, Xiubo Geng, Yuting
Liu, Yanyan Lan, Fen Xia, Wei Chen, Jiang Bian, Yin He, Di
He, and Wenkui Ding.
– Our internal collaborators Chris Burges, Michael Taylor,
Stephen Robertson, Emine Yilmaz, and Ralf Herbrich.
– Our external collaborators Hsin-Hsi Chen, Hongyuan Zha,
Zhiming Ma, Liwei Wang, Cheng Xiang Zhai, Thorsten
Joachims, Olivier Chappell, and Yi Chang.
2011/5/9
Tie-Yan Liu, Copyright reserved
154
tyliu@microsoft.com
http://research.microsoft.com/people/tyliu/
http://weibo.com/tieyanliu/
2011/5/9
Tie-Yan Liu, Copyright reserved
155

Similar documents

Learning to Rank

Learning to Rank Learning to Rank Algorithms Least Square Retrieval Function Query refinement (WWW 2008) (TOIS 1989) Nested Ranker (SIGIR 2006) SVM-MAP (SIGIR 2007) ListNet (ICML 2007) Pranking (NIPS 2002) MPRank (...

More information