learning to rank
“Yahoo發起了一項學習排序競賽(Learning to Rank Challenge)作為ICML 2010大會的一部分,任何人可以以個人名義或組隊(最多10人)參賽。競賽3月1日開始,至5月31日結束,6月份公布獲獎名單。
競賽將公布兩個之前從未發布的真實數據形成的數據集。第一個數據集包括29921個請求,744692個URL地址,519個特征。第二個數據集包括6330個請求,172870個URL地址,596個特征。競賽的任務是根據訓練集中的數據構造一個排序函數,對驗證集和測試集中URL地址進行排序。
第1至4名優勝者將分別獲得8000,4000,2000,1000美元,并將被邀請參加ICML 2010大會。
更多詳細信息參見:Learning to Rank Challenge“
The task of learning to rank has recently drawn a lot of interest in
machine learning. As distinguished by [3] and [4], previous works
fall into three categories: (1) the point-wise approach, (2) the pairwise
approach, and (3) the list-wise approach.
In the point-wise approaches, each training instance is associated
with a rating. The learning is to find a model that can map
instances into ratings that are close to their true ones. A typical
example is PRank [5], which trains a Perceptron model to directly
maintain a totally-ordered set via projections. The pair-wise
approaches take pairs of objects and their relative preferences as
training instances and attempt learning to classify each object pair
into correctly-ranked or incorrectly-ranked. Indeed, most existing
methods are the pair-wise approaches, including Ranking SVM
[10], RankBoost [9], and RankNet [2]. Ranking SVM employs
support vector machine (SVM) to classify object pairs in
consideration of large margin rank boundaries. RankBoost
conducts Boosting to find a combined ranking which minimizes
the number of misordered pairs of objects. RankNet defines Cross
Entropy as a probabilistic cost function on object pairs and uses a
neural network model to optimize the cost function. Finally, the
list-wise approaches use a list of ranked objects as training
instances and learn to predict the list of objects. For example,
ListNet [3] introduces a probabilistic-based list-wise loss function
for learning. Neural network and gradient descent are employed
to train a list prediction model.
—–J. Yeh, J. Lin, H. Ke, and W. Yang. Learning to rank
for information retrieval using genetic programming.
In LR4IR, 2007.
When ranking problem is described as a machine
learning problem, proposing and minimizing the ranking
loss function becomes the key to learning to rank. There
are several popular approaches to constructing the
ranking loss function which are considered on different
instance level. One is building the loss function on
document instance level. SVOR [9] is proposed to
minimize the rank loss by aggregating the error on each
document instance. Another approach is pair-wise loss
function, which create the pair instance between two
documents with different relevance level, and denote
correct rank pair as positive (+1) instance while incorrect
as negative (-1). So the ranking problem is transformed
into a binary classification problem in RSVM [1],
RankBoost [2], RankNet [3]. And a recent approach is
list-wise, which define rank loss with the difference
between predicting document list and labeled list for each
query, in AdaRank [10], ListNet [11].
—–An Ensemble Approach to Learning to Rank
?
from":http://www.shamoxia.com/html/y2009/475.html
總結
以上是生活随笔為你收集整理的learning to rank的全部內容,希望文章能夠幫你解決所遇到的問題。
- 上一篇: IE6 葬礼
- 下一篇: Surface Computing