Ranking is a central topic in the field of search mainly because the performance of a search system can be evaluated by the accuracy of its ranking results. Traditionally, a ranking model is defined as a function of a query and a document, which represents the relevance of the document with respect to the query. In search, when given a query, the ranking model assigns a score to each of the documents retrieved with the query and produces the ranked list of documents sorted by the scores. The ranking model is typically a local model in the sense that the creation and utilization of the model only needs the involvement of a single document.
As search evolves, more and more useful information for ranking becomes available. This includes the content information of documents as well as the relation information between documents. The relations can be hyperlinks between documents, similarities between documents, etc. Ideally, the ranking model would be a function of the query and all the retrieved documents with the query. That is to say, ranking should be conducted on the basis of the contents of documents as well as the relations between documents.
In view of the above-described needs, the present invention provides an improved method for ranking documents using a ranking model. As described in detail below, one embodiment of the present invention employs Continuous Conditional Random Fields (CRF) as a model, which is a conditional probability distribution representing a mapping relationship from retrieved documents to their ranking scores. The model can naturally utilize features of the content information of documents as well as the relation information between documents for global ranking. The present invention also provides a learning algorithm for creating Continuous CRF. Also provided, the invention introduces Pseudo Relevance Feedback and Topic Distillation.
In one specific embodiment, the invention provides an improved method for ranking documents using a ranking model. The method includes the step of obtaining a number of documents and queries, where each query is associated with some documents. In addition, ground truth data, the document's features and relations, are also included. The labels include data that describes relationships between the documents. Using the relationship data, the method then maximizes a likelihood figure by employing a gradient ascent. The method then uses the likelihood to modify parameters of a ranking model. Once the parameters are established, the method ranks the documents by the use of the parameters. The method makes the ranking of the documents more consistent with the document label data.
In other embodiments, the present invention produces the likelihood figure by the use of specific functions. One specific function is configured to sort label data that describes content similarity relationships. Another specific function is configured to sort label data that describes parent-child relationships.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
The claimed subject matter is described with reference to the drawings, wherein like reference numerals are used to refer to like elements throughout. In the following description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the subject innovation. It may be evident, however, that the claimed subject matter may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to facilitate describing the subject innovation.
As utilized herein, terms “component,” “system,” “data store,” “evaluator,” “sensor,” “device,” “cloud,” “network,” “optimizer,” and the like are intended to refer to a computer-related entity, either hardware, software (e.g., in execution), and/or firmware. For example, a component can be a process running on a processor, a processor, an object, an executable, a program, a function, a library, a subroutine, and/or a computer or a combination of software and hardware. By way of illustration, both an application running on a server and the server can be a component. One or more components can reside within a process and a component can be localized on one computer and/or distributed between two or more computers.
Furthermore, the claimed subject matter may be implemented as a method, apparatus, or article of manufacture using standard programming and/or engineering techniques to produce software, firmware, hardware, or any combination thereof to control a computer to implement the disclosed subject matter. The term “article of manufacture” as used herein is intended to encompass a computer program accessible from any computer-readable device, carrier, or media. For example, computer readable media can include but are not limited to magnetic storage devices (e.g., hard disk, floppy disk, magnetic strips . . . ), optical disks (e.g., compact disk (CD), digital versatile disk (DVD) . . . ), smart cards, and flash memory devices (e.g., card, stick, key drive . . . ). Additionally it should be appreciated that a carrier wave can be employed to carry computer-readable electronic data such as those used in transmitting and receiving electronic mail or in accessing a network such as the Internet or a local area network (LAN). Of course, those skilled in the art will recognize many modifications may be made to this configuration without departing from the scope or spirit of the claimed subject matter. Moreover, the word “exemplary” is used herein to mean serving as an example, instance, or illustration. Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs.
The detailed description below first provides a description of embodiments of the present invention. The sections that follow disclose a formal definition of the problem of global ranking, then describes additional details of the methods that employ a Continuous CRF model for conducting global ranking.
The present invention provides an improved method for ranking documents using a ranking model. As described in detail below, one embodiment of the present invention employs Continuous Conditional Random Fields (CRF) as a model, which is a conditional probability distribution representing a mapping relationship from retrieved documents to their ranking scores. The model can naturally utilize features of the content information of documents as well as the relation information between documents for global ranking. The present invention also provides a learning algorithm for creating Continuous CRF. Also provided, the invention introduces Pseudo Relevance Feedback and Topic Distillation.
In one specific embodiment, the invention provides an improved method for ranking documents using a ranking model. With reference to
In other embodiments, the present invention produces the likelihood figure by the use of specific functions. One specific function is configured to sort label data that describes content similarity relationships. Another specific function is configured to sort label data that describes parent-child relationships. Details of each step of this embodiment follow a brief description of the background technology.
The Continuous CRF model is defined as a conditional probability distribution representing the mapping relationship from the retrieved documents to their ranking scores, where the ranking scores are represented by continuous variables. In Continuous CRF, we make use of both content information and relation information as features. The model is general in the sense that various types of relations for global ranking can be incorporated as features. We further propose a learning method for training Continuous CRF. Specifically, we use Maximum Likelihood Estimation and Gradient Ascent for parameter estimation. Our learning method turns out to be a novel learning-to-rank method for global ranking, in contrast to the exiting learning-to-rank methods which are by design local ranking methods. We apply Continuous CRF to Pseudo Relevance Feedback and Topic Distillation.
As a matter of background, this section briefly describes existing work in the field. Traditionally, document ranking was only conducted locally, in the sense that the ranking model is a function of a query and a single document. Although this makes the ranking model easy to create and use, its limitation is also clear. There is a large amount of in-formation which is useful for ranking, but is not local, for example, the relation information between documents.
Relation information between documents plays an important role in many information retrieval tasks. For example, ranking web pages on the basis of importance, improving relevance ranking by using similarity information, and diversifying search results.
Relation information has been used for importance ranking in web search. PageRank and HITS are well known algorithms for computing importance of web pages. They rank web pages based on the Markov chain model and authority-hub model respectively; both leverage the hyperlink (relation) information between web pages.
Topic Distillation is another example of using relation information in web search. Here, Topic Distillation refers to the search task in which one selects a page that can best represent the topic of the query from a web site by using structure (relation) information of the site. If both a page and its parent page are concerned with the topic, then the parent page is to be ranked higher. It is found that propagating the relevance of a web page to its neighborhood through the hyperlink graph can improve the accuracy of Topic Distillation. Furthermore, propagating the relevance of a web page to its parent page can also boost the accuracy.
Similarity between documents is useful information for relevance search. In Pseudo Relevance Feedback, we first conduct a round of relevance ranking, assuming that the top ranked documents are relevant; then conduct another round of ranking, using similarity information between the top ranked documents and the other documents, and boost some relevant documents dropped in the first round. Existing Pseudo Relevance Feedback methods can be clustered into two groups. In one group, the documents are ranked first based on relevance. Then, the top results are used to make an expansion of the query and the re-ranking with the expanded query is performed. In the other group, it is assumed that similar documents are likely to have similar ranking scores, and documents are re-ranked based on the similarities between documents.
In Subtopic Retrieval one also utilizes document similarity information. In the task, given a query, the returned documents should cover as many subtopics as possible. If there are multiple documents about the same subtopic, then only one document should be selected and ranked high. Although relation information between documents has been used in search, so far there has been no previous work which generalizes the specific ranking tasks.
Recently machine learning technologies called ‘learning to rank’ have been applied to information retrieval. In the approach, supervised learning is employed in ranking model construction. Previous work demonstrates that learning to rank has certain advantages when compared with the traditional non-learning approaches.
Previously people have tried to transform the problem of ranking into that of classification and apply existing classification techniques to the task. This is called the pairwise approach in the literature. For example, as classification techniques one can employ SVM and Neural Network, and derive the methods of RankSVM and RankNet. More recently, a number of authors have proposed directly defining a loss function on list of objects and optimizing the loss function in learning. This listwise approach formalizes the ranking problem in a more straightforward way and thus appears to be more effective.
Existing learning to rank methods do not consider using the relation information between objects (documents) in the models. As a result, they are not directly applicable to the cases in which relation information should be used. Making extensions of existing methods on the basis of heuristics would not work well, as will be seen in the experiment section.
Conditional Random Fields (CRF) is a discriminative model for sequence data prediction. It is defined as a conditional probability distribution of output label sequence given input observation sequence. The conditional probability distribution is an exponential model containing features based on both the input and output. It is assumed that there exists dependency between the adjacent labels in the output.
Document ranking in search is a problem as follows. When the user submits a query, the search system retrieves all the documents containing the query, calculates a ranking score for each of the documents using the ranking model, and sorts the documents according to the ranking scores. The scores determine the ranking orders of documents, and can represent relevance, importance, and/or diversity of documents.
Let q denote a query. Let d(q)={d1(q),d2(q), . . . ,dn
We call the ranking ‘local ranking,’ if the ranking model is defined as
yi(q)=f(q,di(q)),i=1, . . . ,n(q) (1)
Furthermore, we call the ranking ‘global ranking,’ if the ranking model is defined as
y(q)=F(q,d(q)) (2)
The major difference between the two is that F takes on all the documents as input, while f takes on individual documents as input. Note global ranking contains local ranking as its special case. Intuitively, in local ranking we look at the documents individually, while in global ranking we treat the documents as a whole. In global ranking, we use not only the content information of documents but also the relation information between documents. There are many specific application tasks that can be viewed as examples of global ranking. These include Pseudo Relevance Feedback, Topic Distillation, and Subtopic Retrieval.
The present invention provides a learning to rank method for global ranking, using Continuous CRF as model. Let x={x1,x2, . . . ,xn} denotes the input vectors of the documents retrieved with the query, y={y1,y2, . . . ,yn} denotes the ranking scores assigned to the documents, and R denotes the relation between the documents (and also the ranking scores of the documents). Here n stands for number of documents retrieved by the query and xiε,yiεi=1,2, . . . ,n and K1 is an integer. Note that in this paper we call x input vector, not feature vector, in order to distinguish it from the feature functions in the CRF model. Let θ={α,β},αε,βε be a vector of parameters respectively, where K2 is an integer. Let {fk(yi,x)}k=1 be a set of real-valued feature functions defined on x and yi(i=1, . . . ,n) and {gk(yi,yj,x)}k=1K2 be a set of real-valued feature functions defined on yi,yj, and x(i=1, . . . ,n,j=1, . . . ,n,i≠j).
Continuous Conditional Random Fields (CRF) is a conditional probability distribution with density function:
where Z(x) is a normalization function
Continuous CRF is a graphical model, as depicted in
In Continuous CRF, feature function fk represents the dependency between the ranking score of a document and the input vector of the document, and feature function gk represents the relationship between the ranking scores of two documents (e.g. similarity relation, parent-child relation). We call the feature functions fk vertex features, and the feature functions gk edge features. The edge feature functions are determined by the relation R. Different retrieval tasks have different definitions on R, as is explained below.
There are clear differences between the conventional CRF and the Continuous CRF proposed here. (1) The conventional CRF is usually defined on a chain while Continuous CRF is defined on a graph. (2) In the conventional CRF random variable y is discrete while in Continuous CRF y is continuous. This makes the inference of Continuous CRF is different from that of conventional CRF, as described below. The R in Continuous CRF defines the relations between ranking scores as well as specifies the corresponding feature function gk. An R is associated with a query and different queries may have different R's.
In learning, given training data we estimate the parameters θ={α,β} of Continuous CRF. Suppose that the training data {x(q),y(q)}q=1N is generated i.i.d. from a unknown probability distribution, where each x(q)={x1(q),x2(q), . . . ,xn(q)(q)} is a set of input vectors associated with the documents of query q, and each y(q)={y1(q),y2(q), . . . ,yn(q)(q)} is a set of ranking scores associated with the documents of query q. Note that we do not assume that the documents of a query are generated i.i.d. Further suppose that the relation data {R(q)}q=1N is given, where each R(q) is a set of relations on the documents (ranking scores) of query q.
We employ Maximum Likelihood Estimation to estimate the parameters. Specifically, we calculate the conditional log likelihood of the data with respect to the Continuous CRF model.
Specifically,
We take the parameter {circumflex over (θ)} that can maximize the log likelihood function as output. We know of no analytic solution to the optimization problem. We can employ the numerical method of Gradient Ascent to solve it.
In inference, given new data x and relation R we use the model with estimated parameter {circumflex over (θ)} to calculate the conditional probability Pr(y|x) and select the y satisfying
Equation (7) then becomes the ranking model.
In this section shows how to apply Continuous CRF to two global ranking tasks. Here we consider a method of using Continuous CRF for Pseudo Relevance Feedback, which belongs to the second group of existing methods. In fact, the method combines the two rounds of ranking of Pseudo Relevance Feedback into one by using the Continuous CRF model. In the method the similarities between any two documents are given and the input vectors are also defined. What we need to do is to specify the feature functions.
For each component in the input vector, we introduce a vertex feature function. Suppose that xi,k is the k-th component of input vector xi, we define the k-th feature function fk(yi,x) as
fk(yi,x)=−(yi−xi,k)2. (8)
Next, we introduce one (and only one) edge feature function.
where Si,j is similarity between documents di and dj. The larger Si,j is, the more similar the two documents are. Here the relation is represented by a matrix S, whose element in i-th row and j-th column is Si,j. Note that this specific feature function does not depend on x.
The Continuous CRF for Pseudo Relevance Feedback then be-comes
where Z(x) is defined as
To guarantee that
is integrable, we must have αk>0 and β>0.
The Continuous CRF can naturally model Pseudo Relevance Feed-back. First, if the value of Xi,k is high, then the value of yi is high with high probability. (For example, xi,k can be a feature representing tf−idf.) Second, if the value of Si,j is large, then ranking scores yi and yj are close with high probability.
With some derivation, we obtain
where A=αTeI+βD−βS,D is an n×n diagonal matrix with
is an n×n matrix, |A| is determinant of matrix
and X is a matrix whose i-th row k-th column element is xi,k.
In learning, we try to maximize the log likelihood. Note that maximization of L(θ) in Eq. (6) is a constrained optimization problem because we need to guarantee that ak>0 and β>0. Gradient Ascent cannot be directly applied to a constrained optimization problem. Here we adopt a technique similar to that in and then employ Gradient Ascent. Specifically, we maximize L(θ) with respect to log αk and log β instead of αk and β. As a result, the new optimization issue becomes unconstrained.
The gradients of L(θ) with respect to log αk and log β are computed as follows.
Now we show how to get the partial derivative
For simplicity, we omit the super script (q) in the equations hereafter. First
Furthermore,
where X: denotes the long column vector formed by concatenating the columns of matrix X, and X,k denotes the k-th column of matrix X. For example, if
then X:=[1,2,3,4]T.
Substituting Eq. (15)-(20) into Eq. (13) and (14), we obtain the gradient of the log likelihood function. Algorithm 1 in
In inference, we calculate the ranking scores of documents with respect to a new query in the following way.
Note that here the inference can be conducted with matrix computation, which is different from that in conventional CRF. The reason is that in Continuous CRF the output variable is continuous, while in conventional CRF it is discrete.
If we ignore the relation between documents and set β=0, then the ranking model degenerates to
ŷ=Xα
which is equivalent to a linear model used in conventional local ranking.
For n documents, the time complexity of straightforwardly computing the ranking model (21) is of order O(n3) and thus it is ex-pensive. The main cost of the computation comes from matrix inversion. In this paper we employ a fast computation technique to quickly perform the task. First, we make S a sparse matrix, which has at most k non-zero values in each row and each column. We can do so by only considering the similarity between each document and its
nearest neighbors. Next, we use the Gibbs-Poole-Stockmeyer algorithm to convert S to a banded matrix. Finally we solve the following system of linear equation and take the solution as ranking scores.
(αTeI+βD−βS)ŷ=Xα (22)
Let A=αTeI+β(D−S). A is a banded matrix when S is a banded matrix. Then, the scores ŷ in Eq. (22) can be computed with time complexity of O(n) when k<<n[6]. That is to say, the time complexity of testing a new query is comparable with those of existing local ranking methods.
We can also specify Continuous CRF to make it suitable for Topic Distillation. Here we assume that the parent-child relation between two pages is given. The input vectors are also defined. What we need to do is to specify the feature functions.
We define the vertex feature function fk(yi,x) in the same way as that in Eq. (8). Furthermore, we define the only edge feature function as
g(yi,yj,x)=Ri,j(yi−yj), (23)
where Ri,j denotes the parent-child relation: Ri,j=1 if document i is the parent of j, and Ri,j=0 for other cases.
The Continuous CRF for Topic Distillation then becomes
where Z(x) is defined as
To guarantee that
is integrable, we must have αk>0.
The Continuous CRF can naturally model Topic Distillation. First, if the value of xi,k is high, then the value of yi is high with high probability. Second, if the value of Ri,j is one, then the value of yi−yj is high with high probability.
With some derivation, we have
where n is the number of documents for this query, and a=αTe, b=2Xα+β(Dr−Dc)e, Dr and Dc are two diagonal matrixes with
In learning, we use Gradient Ascent to maximize the log likelihood. Again, we use the technique for optimization to guarantee αk>0. We compute the derivative of L(θ) with respect to β and the new optimization variable log αk as follows.
Now we show how to get the partial derivative
For simplicity, we omit the super script (q) hereafter.
where X,k denotes the k-th column of matrix X.
Substituting Eq. (29) and (30) into Eq. (27) and (28), we obtain the gradient of the log likelihood function. Due to space limitation, we omit the details of the learning algorithm, which is similar to Algorithm 1, shown in
In inference, we calculate the ranking scores of documents with respect to a new query in the following way.
Similarly to Pseudo Relevance Feedback, if we ignore the relation between documents and set β=0, the ranking model degenerates to a linear ranking model in conventional local ranking.
We can also conduct multiple global ranking tasks simultaneously. For example, we can combine Pseudo Relevance Feedback and Topic Distillation by using the following Continuous CRF model
In this case, the ranking scores of documents for a new query is calculated as follows.
Continuous CRF is a powerful model in the sense that various types of relations can be incorporated as edge feature functions.
6. Experiments
We applied Continuous CRF to Pseudo Relevance Feedback and Topic Distillation. We also compared the performances of Continuous CRF model against several baseline methods in the two tasks. As data, we used LETOR, which is a dataset for learning to rank research. We made use of OHSUMED in LETOR for Pseudo Relevance Feedback and TREC in LETOR for Topic Distillation. As evaluation measure, we utilized NDCG@n (Normalized Discounted Cumulative Gain).
The OHSUMED dataset in LETOR is derived from the OHSUMED data for research on relevance search. The document collection is a subset of MEDLINE, a database on medical publications.
There are 106 queries in OHSUMED data set, each associated with a number of documents. The relevance degrees of documents with respect to the queries are judged by humans, on three levels: definitely relevant, partially relevant, or not relevant. There are in total 16,140 query-document pairs with relevance judgments. Each query-document pair is represented by a 25 dimension feature vector, which means K1=25 in Eq. (10).
Similarity between documents is provided as relation information in the data, which is defined in the following way. First stop words are removed from the documents. Each document is represented as a term vector in the vector space model. Then the similarity Si,j between two documents i and j is calculated as co-sine between the term vectors of the two documents. (One should not confuse the term vector with the input vector in learning.)
6.1.2 Baseline Methods
As baseline methods, we used RankSVM and ListNet. RankSVM is a state-of-the-art algorithm of the pairwise approach to learning to rank, and ListNet is a state-of-the-art algorithm of the listwise approach.
The two learning to rank methods only use content information, but not relation information. To make fair comparisons, we added post processing to the two methods in which we incorporated relation information into the final ranking model creation. We refer to the methods as ‘RankSVM plus relation’ (RankSVM+R) and ‘List-Net plus relation’ (ListNet+R). Following prior work, we first calculated the scores of documents based on content information using a learning to rank method (RankSVM or ListNet), then propagated the scores using similarities between documents, and finally ranked the documents based on the propagated scores. Specifically, the final score list is calculated as
y+r=(I+β(D−S))−1y,
where y is the score list output by a learning algorithm (e.g. RankSVM or ListNet), and D and S are the same as those in Eq. (12). Here, is a parameter balancing the influence from content information and relation information.
For reference purposes, we also tried BM25 and Pseudo Relevance Feedback based on BM25. For the two baselines, we used the tools provided in Lemur toolkit. We conducted 5 fold cross validation for Continuous CRF and all the baseline methods, using the partition provided in LETOR.
Continuous CRF needs to use queries, their associated documents, and relevance scores as training data. Since LETOR only provides relevance labels, we mapped the labels to scores using heuristics. The rule is the ground truth score of a relevant document should be larger than that of an irrelevant document. We used validation set in LETOR to select the best mapping function.
For RankSVM+R and ListNet+R, we ran a number of experiments based on different values of parameter β. Here, we report the best performances of the methods. For RankSVM+R, β=0.2; for ListNet+R, β=0.1.
Table 1 shows the ranking accuracies of BM25, BM25 based Pseudo Relevance Feedback (PRF), RankSVM, ListNet, RankSVM+R ListNet+R, and Continuous CRF (CRF), in terms of NDCG averaged over five trials.
CRF's performance is superior to the performances of RankSVM and ListNet. This is particularly true for NDCG@ 1; CRF achieves about 5 points higher accuracy than RankSVM and more than 2 points higher accuracy than ListNet. The results indicate that learning with similarity information can indeed improve search relevance.
We can see that CRF performs much better than RankSVM+R and ListNet+R at all NDCG positions. This indicates that with the same information the proposed CRF can indeed perform better than the heuristic baseline methods.
RankSVM+R beats RankSVM largely at NDCG@ 1, while obtains similar results at NDCG@ 3-10, but a worse result at NDCG@2. ListNet+R works better than ListNet at NDCG@ 1, but does not as well as ListNet at the other positions. This seems to indicate that heuristically using relation in post-processing does not work well.
Continuous CRF also outperforms PRF (Pseudo Relevance Feed-back), the traditional method of using similarity information for ranking. The result suggests that it is better to leverage the machine learning techniques in Pseudo Relevance Feedback.
We made analysis on the results and found that CRF can in-deed improve relevance. Table 2 and 3 show the top 10 results of RankSVM and CRF for query “back pain-mri sensitivity etc in comparison to ct of lumbar spine” respectively. The documents in red are ‘definitely relevant’, documents in blue are ‘partially relevant’, and documents in black are ‘not relevant’.
It is obvious that CRF works better than RankSVM for this query. Document 262357 is a ‘definitely relevant’ document, but is ranked out of top 10 by RankSVM. Since this document is similar to documents 277424 and 169405 which are ranked at position 2 and 5, it is boosted to position 6 by CRF using similarity information.
In TREC 2004, there was a track for web search, called Topic Distillation, which is aimed at enhancing research on Topic Distillation, the task described in Section 2.
The TREC dataset in LETOR is derived from the TREC 2004 Topic Distillation data. There are 75 queries, and each query associated with about 1,000 documents. Each query document pair is associated with a label, representing whether the document is an entry page to the query (an answer) or not. There are 44 features defined over a query-document pair. It implies that K1=44 in Eq. (24). Furthermore, information on parent-child relation is also given. The element Ri,j equals 1 if page i is parent of page j in a website, and equals 0 otherwise.
As baseline methods, we used RankSVM and ListNet. Since RankSVM and ListNet do not use relation information, we tried two modifications of them, in which we used relation in-formation in post processing. We refer to them as ‘RankSVM plus relation’ (RankSVM+R) and ‘ListNet plus relation’ (ListNet+R). In RSVM+R (or ListNet+R), we use the ranking score of a child page output by RankSVM (or ListNet) to enhance the ranking score of its parent also output by RankSVM (or ListNet). The idea is similar to that of sitemap based score propagation.
We also tested non-learning methods of BM25 and sitemap based relevance propagation. The basic idea of sitemap based relevance propagation is to use the relevance of a child page to enhance the relevance of its parent page. There are two variants of the method: sitemap based term propagation (‘ST’ for short) and sitemap based score propagation (‘SS’ for short).
The following section describes the experimental results of the runs. We conduced 5-fold cross validation on our method and the base-line methods, using the partitions in LETOR. For RankSVM and ListNet, we refer to the results in LETOR.
Continuous CRF needs to use queries, their associated documents, and relevance scores as training data, while LETOR only provides ranking labels. Again, we used heuristics to map ranking labels to ranking score. The rule is that the score of an answer document should be larger than that of a non-answer document. We used the validation set in LETOR to select the best mapping function.
Table 4 in
CRF outperforms RankSVM and ListNet at all NDCG positions. This is particularly true for NDCG@1. CRF achieves 8 points higher accuracy than RankSVM and ListNet, which is a more than 15% relative improvement. Overall, learning using relation information can achieve better results than learning without using relation information. The result indicates that our method can effectively use the information in training of a Topic Distillation model.
CRF performs much better than RankSVM+R and ListNet+R at all NDCG positions. This indicates that with the same information the proposed CRF can indeed perform better than the heuristic methods.
RankSVM+R beats RankSVM largely at NDCG@1, while obtains slightly better results at NDCG@3-10 but a slightly worse result at NDCG@2. ListNet+R works better than ListNet at NDCG@2-10, but does not at NDCG@1. The results seem to indicate that simply using relation information as post-processing does not work very well.
Continuous CRF also outperforms SS and ST, the traditional method of using parent-child information for Topic Distillation. The result suggests that it is better to leverage the machine learning techniques using both content information and relation information in ranking.
We investigated the reason that CRF can achieve better results than other algorithms and concluded that it is because CRF can successfully leverage the relation information in ranking. Without loss of generality, we make a comparison between CRF and RankSVM.
Table 5 show top 10 results of RankSVM and CRF for query “HIV/AID.” The answer pages for this query are in red color. The answer page 643908 is not ranked in top 10 by RankSVM, because its content feature is not strong. Since the content features of its child pages (such as 220602, 887722, and other pages) are very strong, CRF can effectively use the parent-child relation information and boost it to position 4.
Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims. Accordingly, the invention is not limited except as by the appended claims.
Number | Name | Date | Kind |
---|---|---|---|
6112203 | Bharat et al. | Aug 2000 | A |
6738764 | Mao et al. | May 2004 | B2 |
7216123 | Kamvar et al. | May 2007 | B2 |
7231399 | Bem et al. | Jun 2007 | B1 |
7376649 | Yang et al. | May 2008 | B2 |
20060129550 | Zha et al. | Jun 2006 | A1 |
20080010043 | Thiesson et al. | Jan 2008 | A1 |
20080016050 | Stensmo | Jan 2008 | A1 |
20080028010 | Ramsey | Jan 2008 | A1 |
20080147641 | Leffingwell et al. | Jun 2008 | A1 |
Number | Date | Country |
---|---|---|
1462950 | Aug 2007 | EP |
Number | Date | Country | |
---|---|---|---|
20100082613 A1 | Apr 2010 | US |