This specification relates to generating a related group of search result documents from an initial group of search result documents.
Internet search engines provide information about Internet accessible search result documents (e.g., Web pages, images, text documents, multimedia content) that are responsive to a user's search query by returning a set of search results for the query. A search result includes, for example, a Uniform Resource Locator (URL) and a snippet of information for resources responsive to a query. Users are presented with many search results and view a subset of search result documents corresponding to the search results over time. Users may also browse various documents available on the Internet.
This specification describes technologies relating to generating a related group of search result documents from an initial group of search result documents. In general, one aspect of the subject matter described in this specification can be embodied in methods for identifying one or more second documents related to one or more first documents that include the actions of determining a respective strength of relationship score between each candidate document in a group of one or more candidate documents and each first document by aggregating user selection data for users. The first documents and the candidate documents are in a corpus of web documents, and the user selection data indicates, for each user, whether the user viewed the candidate document during a window of time after the first document is presented to the user on a search results web page in response to a query. The method further includes calculating an aggregate strength of relationship score for each candidate document from the respective strength of relationship scores for the candidate document and selecting the second documents from the candidate documents according to the aggregate strength of relationship scores for the candidate documents. Other embodiments of this aspect include corresponding systems, apparatus, and computer programs recorded on computer storage devices, each configured to perform the operations of the methods.
These and other embodiments can each optionally include one or more of the following features. The user selection data can further indicate whether each user viewed the candidate document for a threshold period of time. Aggregating user selection data can further comprise scaling the user selection data for one of the users by a scoring factor when the user views the candidate document during the window of time after the first document is selected by the user from the search results web page. Determining a respective strength of relationship score between each candidate document and each first document can further comprise using a popularity of the candidate document to normalize the respective strength of relationship score. The one or more second documents can be associated with a natural language, and determining a respective strength of relationship score between each candidate document and each first document can further include scaling the strength of relationship score by a percentage of users who viewed the candidate document and are associated with the natural language.
The embodiments can further include identifying one or more documents responsive to a query as the first documents, and generating an augmented set of documents responsive to the query by including one or more of the second documents in the first documents. The embodiments can further include receiving the query and presenting the augmented set of documents in response to the query.
The embodiments can further include selecting the one or more first documents from documents a first user has viewed for a second period of time and presenting the one or more second documents as suggested documents. Presenting the one or more second documents as suggested documents can include presenting the one or more second documents in a toolbar.
The embodiments can further include receiving input from a second user indicating that one or more of the first documents are preferred documents and calculating a respective document weight for each preferred document. Calculating the aggregate strength of relationship score for each candidate document can include weighting the strength of relationship scores for the candidate document and each preferred document by the respective document weight for the preferred document. The embodiments can further include receiving input from the second user indicating that one or more of the first documents are disliked documents and calculating a respective document weight for each disliked document. Calculating the aggregate strength of relationship score for each candidate document can include weighting the strength of relationship scores for the candidate document and each disliked document by the respective document weight for the disliked document. The embodiments can further include presenting one or more of the second documents as suggested documents.
The embodiments can further include selecting the one or more first documents based on one or more first queries issued during a session, where each of the first documents is responsive to one of the first queries, identifying one or more second queries corresponding to the one or more second documents from data associating queries and documents, and presenting the one or more second queries as suggested queries.
Particular embodiments of the subject matter described in this specification can be implemented so as to realize one or more of the following advantages. A related set of “better” documents can be generated from an initial set of documents. The initial documents can be documents that a user has viewed during a search session, documents that a user has browsed (i.e., viewed), or documents that a user has bookmarked. The related set of documents can be used to generate suggested documents for a user. The related set of documents can be used to return more relevant search results in response to users, for example, through session-based ranking, where search results are ranked in the overall context of a session of user search queries, or by augmenting a set of search results responsive to a search query with other related documents. The related set of documents can be used to generate suggested queries.
The details of one or more embodiments of the subject matter described in this specification are set forth in the accompanying drawings and the description below. Other features, aspects, and advantages of the subject matter will become apparent from the description, the drawings, and the claims.
Like reference numbers and designations in the various drawings indicate like elements.
First, the user issues a query 110 for “San Francisco” through the search engine user interface 102. A search engine generates relevant search results 112 and 114 and presents them to the user. The user clicks (e.g., selects with a mouse or through interaction with another input device such as, for example, a keyboard, a microphone, a camera, or a tactile input device) a search result for “Cheap San Francisco Hotels” 112, and the user interface changes to 104 and presents the document 116 corresponding to the search result selected by the user. The user views the search result document 116 for some period of time. While
At a later time, the user interface changes to 106 where the user issues a new query 118 for “San Francisco Vacation.” The search engine generates relevant search results 120 and 122 and presents them to the user. The user clicks search result 122 for “Delicious Eats by the Bay,” and the user interface changes to 108 and displays the document 124 corresponding to the search result selected by the user. The user views the search result document 124 for some period of time.
The directed edges from the source nodes to the sink nodes represent a relationship between the documents corresponding to the nodes. In
Various other metrics can be used to relate source nodes to sink nodes. For example, in some implementations, a given search result is considered to have been presented to a user only if the user clicks on the given search result or a search result below the given search result in the order of search results. For example, in
In some implementations, source nodes are related to sink nodes only when the time during which the user viewed the sink document satisfies a threshold. For example, a sink document could have to be viewed for a minimum amount of time. The minimum amount of time can be determined, for example, empirically. In other implementations, source nodes are related to sink nodes only when the time between when the user was presented with and/or clicked the source node and viewed the sink document satisfies a threshold. For example, the sink document could have to be viewed within a given time period (e.g., two hours) of visiting the source document.
In some implementations, the edges are weighted, for example, by the time the user spent viewing the sink document, or by whether the user viewed the sink document after clicking on the source document or merely being presented with the source document. For example, longer viewing times can be given a higher weight, and edges from source nodes whose source results were clicked on rather than just presented can be given a higher weight.
In some implementations the system maintains additional state data to avoid double counting documents, for example, in the situation where a user is presented with a search result for document A, views document B, is presented with a search result for document A, and then views document C. Document C would be counted only one time for document A, even though it was viewed after document A was presented two times.
The weighted edges shown in
A user 302 interacts with the search system 314 through a client device 304. For example, the client device 304 can be a computer (e.g., a personal computer, a mobile phone, etc.) coupled to the search system 314 through a wired or wireless local area network (LAN) or wide area network (WAN), e.g., the Internet. In some implementations, the search system 314 and the client device 304 are one machine. For example, a user can install a desktop search application on the client device 304. The client device 304 will generally include a random access memory (RAM) 306 and a processor 308.
A user 302 submits a query 310 to a search engine 330 within a search system 314. When the user 302 submits a query 310, the query 310 is transmitted through a network to the search system 314. The search system 314 can be implemented as, for example, computer programs running on one or more computers in one or more locations that are coupled to each other through a network. The search system 314 includes an index database 322 and a search engine 330. The search system 314 responds to the query 310 by generating search results 328, which are transmitted through the network to the client device 304 in a form that can be presented to the user 302 (e.g., as a search results web page to be displayed in a web browser running on the client device 304).
When the query 310 is received by the search engine 330, the search engine 330 identifies documents that match the query 310. The search engine 330 will generally include an indexing engine 320 that indexes web documents (e.g., web pages, images, multimedia content, or news articles on the Internet) found in a corpus (e.g., a collection or repository of content), an index database 322 that stores the index information, and a ranking engine 352 (or other software) to rank the documents that match the query 310. The indexing and ranking of the documents can be performed, for example, using conventional techniques. The search engine 330 transmits the search results 328 through the network to the client device 304 for presentation to the user 302.
In some implementations, the search system further includes a set improver engine 360 which generates related sets of search result documents from initial sets of documents. In some implementations, the search engine uses the sets generated by the set improver engine 360 to rank documents responsive to the user's query or to identify documents and queries to present to the user. The set improver engine 360 is described in more detail below with reference to
As shown in
In response to the user selecting the search button 408 or issuing a search command, the search system 314 returns a ranking or result list 410 which is an ordered list of search results for documents that are responsive to the query 404. The result list 410 includes two search results, URL A 410a and URL B 410b. The result list 410 can contain the text of the URL itself, a short description of the information found within each document, a snapshot of the portion of the document which contains the query, other suitable information, or a combination of these. A user then continues to issue search queries or view documents, and, at a later period in time the user views the document corresponding to URL D 412 in user interface 414.
In various implementations, the model database 402 is built as users interact with search engines and view documents. The search system 314, or another system, tracks which search results are presented to a user and which documents a user views within a period of time after being presented with the search result. In some implementations, the period of time can be a fixed period of time, for example, 30 minutes, 2 hours, 10 hours, or 24 hours from when the search result is presented. In other implementations, the period of time is a fixed window of time (e.g., a 24 hour window that resets every 24 hours, regardless of when the search result is presented). In still other implementations, the period of time is a user session. In general, a user session is a period during which a user submits queries. A user session can be measured in a number of ways including, for example, by a specified period of time (for example, thirty minutes), by a specified number of queries (for example, fifteen queries), until a specified period of inactivity (for example, ten minutes without submitting a query), while a user is logged-in to a system, or while a user submits queries that relate to similar topics. In some implementations, the system also tracks whether the second document was viewed after being presented in response to a query that overlaps with the query the first document was presented in response to, and only relates the two documents in the model when the queries overlap. Overlapping queries are queries that have one or more words in common.
For example, in
In some implementations, the search system or another system only includes entries for a pair of documents when a user clicks on the first document and then later views the second document (and not when the first document is merely presented as a search result). In other implementations, the search system or another system maintains multiple models, for example, one storing data for documents that were presented, and one storing data for documents that were clicked on.
In various implementations, model data is specific to a geographic location (e.g., a city, metropolitan region, state, country, or continent), specific to a language preference of users (e.g., as specified by the users in a profile or as determined from user search behavior, for example, the language of the queries issued by the user), or specific to a locale (e.g., specific to a geographic location or a set of geographic locations, and optionally a language preference). Other extensions of model data are possible. For example, in some implementations, the model can alternatively, or additionally, track what documents a user views after a user views a given document, independent of what, if any, search results are presented.
The set improver engine 502 includes a strength of relationship score generator 510 and a selection engine 512. The boxes shown in
The individual document score generator 604 generates a strength of relationship score between the candidate document 602 and each document in the first set of documents 504 from the model data 402. For example, in
The score aggregator 606 then aggregates the individual scores for the candidate document and each document in the set of first documents 504 resulting in an aggregate score 608 for the candidate document 602. In the examples given below, the conditional probability P(g i) is used as the individual score for a candidate document g and a document i in the first set of documents. However, in some implementations other individual scores are used instead of the conditional probability, for example, the joint probability P(g, i), or the scores described above with reference to
The score aggregator can aggregate the scores in various ways. For example, in some implementations, the score aggregator 606 sums the individual scores, e.g.:
S(g)=ΣiP(g|i),
where S(g) is the aggregate score for candidate document g over all documents i in the set of first documents.
In other implementations, the score aggregator 606 calculates a weighted sum of the individual scores, e.g.:
S(g)=ΣiwiP(g|i),
where wi is the weight corresponding to document i. In some implementations, the weights can be positive or negative. A positive weight for a document i indicates that the document is representative of the documents the user is interested in (for example, because a user for whom the related set of documents is being generated has indicated that the user found that document relevant). A negative weight for a document i indicates that the document is not representative of the documents the user is interested in (for example, because a user for whom the related set of documents is being generated has indicated (implicitly or explicitly) that the user did not find the document relevant).
In some further implementations, the score aggregator scales the score, for example, by taking the logarithm of the sum of the individual scores, e.g.:
S(g)=ln(ΣiwiP(g|i)).
In some implementations, the score aggregator normalizes the score for a candidate document by an overall popularity of the document, e.g., how often a population of users views the document compared to how often the population views other documents. For example, the score aggregator can divide by the score by the overall popularity of the candidate document (optionally weighted by the weights of the first documents), e.g.:
where P(g) is the overall popularity of the candidate document g. P(g) can be derived from user selection data, for example, by dividing the number of times that users viewed the document g by the number of times users viewed any document.
Alternatively, the score aggregator can normalize the score by summing logarithms of the individual probabilities and then subtracting a logarithm corresponding to the popularity of the candidate document, e.g.:
S(g)=(Σiwi ln(P(g|i)))−(ln(P(g))Σi|wi|).
Other techniques of aggregation are also possible, for example, averaging the individual scores instead of summing them.
The system determines a respective strength of relationship score between candidate documents and first documents (step 702). The system determines each strength of relationship score from aggregate user selection data for a population of users. The user selection data indicates, for each user, whether the user viewed the candidate document during a window of time after the first document was presented to the user on a search results web page in response to a query. For example, in some implementations, the user selection data is the data stored in the model 402 illustrated in
In some implementations, the system identifies related documents that are tailored to a particular natural language, geographic location, or locale. For example, in some implementations, the system maintains separate models for different languages, locations, and locales, and uses the appropriate data to generate the related documents. In other implementations, the system does not maintain separate models but instead tracks what fraction of users viewing a candidate document are associated with the language, location, or locale, and then scales the model data according to the fractions. For example, if 90% of the time, people who view a candidate document are associated with the language French, and 5% of the time, people who view the candidate document are associated with the language English, then if the system is considering the candidate document for a user associated with the language English, then the scores for the candidate document would be scaled by 0.05.
The system calculates an aggregate strength of relationship score for each candidate document (step 704). The system calculates the aggregate strength of relationship score from the individual strength of relationship scores for the candidate document, for example, as described above with reference to
The system identifies a set of one or more second documents from the set of candidate documents according to the aggregate strength of relationship scores (step 706), for example, as described above with reference to
The identified set of second documents can be used in various applications, some exemplary examples of which are described below.
The system selects a first set of one or more documents responsive to a search query (step 802). The query can be received from a user, or can be identified by the system, for example, by parsing past records of what queries users have submitted. The system identifies the set of documents responsive to the query, for example, as described above with reference to
The system obtains a second set of documents related to the first set of documents (step 804), for example, as described above with reference to
The system generates an augmented set of documents from the documents responsive to the query and the second set of documents (step 806). In various implementations, the system generates the augmented set of documents by inserting one or more of the documents from the second set of documents into the set of documents responsive to the query. In some implementations, the set of documents responsive to the query is ordered according to an order, and the system determines locations in the order at which to insert each of the second set of documents. The system can determine the locations in various ways, for example, by randomly selecting locations, or selecting locations based on the aggregated strength of relationship score for the document (e.g., by selecting higher locations when the aggregated strength of relationship score is higher). In some implementations, the system uses a framework that tests different locations for the second set of documents over time.
In some implementations, the system generates the augmented set of documents in response to receiving the search query from the user, and then presents the augmented set of documents in response to the query, for example, by transmitting the search results to a user computer for presentation to the user, as described above with reference to
The system receives a query (step 902), for example, as described above with reference to
The system selects a first set of documents from documents responsive to queries issued during a session (step 904). For example, during a user search session the system can track what queries a user submits (and optionally, what documents are responsive to those queries). The system then selects a number of the queries and a number of documents responsive to each query. The system can use various techniques to select the documents. In some implementations, the system selects all documents presented in response to a given number of past queries. The number of queries can be fixed (e.g., the past five queries) or can be all queries occurring within a set time window from when the query is issued (e.g., all queries issued in the last fifteen minutes). In other implementations, the system selects all documents responsive to the query received in step 902 (or alternatively, the last query issued before that) and all documents that a user selected after they were presented in response to a given number of past queries (e.g., the past five queries). In some implementations, the system only includes documents presented in response to queries that have some query overlap with the query.
The system obtains a second set of documents related to the first set of documents (step 906), for example, as described above with reference to
The system ranks a set of documents responsive to the query using the second set of documents (step 908). The set of documents responsive to the query can be generated, for example, as described above with reference to
The system selects a first set of documents from a user's history (step 1002). In some implementations, the history is a search history that is maintained, for example, by the search system 314. The search history tracks which queries a user submits and which documents a user views during a search session. The search history can correspond to a current user session or a given period of time (e.g., the last fifteen minutes, the last week, or the last few months).
The system can select the first set of documents in various ways. In some implementations, the system selects documents that a user has viewed in the current session as the first set of documents (e.g., all of the documents viewed during the session, a number of the documents the user has viewed recently, such as the last five documents the user viewed, or all documents the user viewed in response to the last query the user submitted). In other implementations, the system selects documents that are responsive to one or more queries the user submitted during the current session. Other techniques for selecting the first set of documents are also possible, for example, in some implementations, the system selects the first set of documents from a set of bookmarked documents that the user has saved for future reference, or from a set of documents.
The system obtains a second set of documents related to the first set of documents (step 1004), for example, as described above with reference to
The system presents the second set of documents to the user as suggested documents (step 1006). In some implementations, the system presents the first set of documents by presenting them in (or transmitting them to the user's computer for presentation in), for example, a search engine interface or a toolbar window. In other implementations, the system presents an “I'm Feeling Lucky” button to the user. When the user clicks on the button, the system causes the user's web browser to load the first document in the set of second documents.
The system receives data indicating one or more preferred documents in a first set of documents (step 1202). The first set of documents can be, for example, a set of documents responsive to a search query, or the first set of documents described above with reference to
The system then determines weights for the preferred documents (step 1204). In some implementations, the weights are derived from the time the user viewed the document or the rating the user gave the document (e.g., the weights can be the time, the rating, or a factor derived from the time or the rating). In other implementations, the weights are fixed, e.g., each document that is preferred (e.g., because a user said it was preferred, viewed it for a threshold period of time, or gave it a rating above a threshold) is given the same weight.
The system then obtains a second set of documents related to the first set of documents using the weight for the preferred documents (step 1206). For example, the system can obtain the set of documents as described above with reference to
In some implementations, the system then presents the related documents to the user, for example, as described above with reference to
In some implementations, the system also receives data indicating that a user dislikes a document, determines weights for the disliked documents (e.g., negative weights, or weights less than 1), and obtains the second set of documents based on the weights for the preferred and disliked documents.
The system selects a first set of documents from documents responsive to queries issued during a session (step 1302), for example, as described above with reference to
The system obtains a second set of documents related to the first set of documents, for example, as described above with reference to
The system generates a set of queries related to the second set of documents (step 1306). The system generates the set of related queries by mapping the second set of documents to a set of related queries, and then selecting the queries most often associated with the documents. For example, in some implementations, the system generates a list of queries for each document, where the document is responsive to the queries, counts the number of times each query is associated with a document, and then selects a number of the queries with the highest count (or all queries whose count exceeds a threshold). The number and the threshold can be determined, for example, empirically. In other implementations, the system uses a document-query database to identify queries that are related to documents from user preference data, where the user preference data indicates how often a user selects a document after viewing a query. The user preference data can be stored, for example, in a database that stores a tuple for each document and query pair. Each tuple identifies the document, the query, and includes aggregated click data for all users that select the document in response to the query. The aggregated click data can be, for example, a tuple score. In general, the tuple score estimates how responsive users found a given document to be as a search result for a given query. For example, a tuple score can be an average of the weighted long clicks for a given document and query pair, a sum of the weighted clicks for a given document and query pair, or a click fraction for a given document and query pair. Various weights are also possible, for example, the clicks can be weighted by the length of the clicks (e.g., how long the user viewed the document), by how quickly a user clicked on a result, and whether a user clicked on a result first, second, third, etc. Various types of click fractions are possible. For example, in some implementations, the click fraction is calculated by dividing the number of clicks (or the number of weighted clicks) by the number of impressions (e.g., the number of times the document was presented as a search result for the query). In other implementations, the click fraction is calculated by dividing the sum of the weighted clicks by the total number of clicks. In still other implementations, the click fraction is the sum of weighted clicks for the document presented in response to the query divided by sum of weighted clicks for all documents presented in response to the query. Other click fractions are also possible. In various implementations, the data is specific to a geographic location (e.g., a city, metropolitan region, state, country, or continent), specific to a language preference of users (e.g., as specified by the users in a profile or as determined from user search behavior, for example, the language of the queries issued by the user), or specific to a locale (e.g., specific to a geographic location or a set of geographic locations, and optionally a language preference). The user preference data can be generated, for example, by a search engine or a user's browser.
In some implementations, rather than choosing the queries that overlap with the most documents, the system chooses a set of queries that covers the second set of documents (or a subset corresponding to the top scoring documents in the second set of documents) with minimal overlap. For example, if the top set of documents includes documents 1-9, query A is associated with documents 1, 2, 5, and 6, query B is associated with documents 3, 4, 7, 8, and 9, and query C is associated with documents 1, 2, 7, 8, and 9, the system would select query A and query B and ignore query C, because queries A and B have no overlap in documents, and every document associated with query B is also associated with either query A or query B. Standard techniques can be used to select the queries such that they have minimum overlap in documents.
The system presents the set of queries as suggested queries (step 1308). The system presents the first set of documents by presenting them in (or transmitting them to the user's computer for presentation in), for example, a search engine interface or a toolbar window.
The set improver engine 502 includes a strength of relationship score generator 510 and a selection engine 512, and generates a set of related documents from a set of initial documents, for example, as described above with reference to
The server 1502 can also have hardware or firmware devices including one or more processors 1512, one or more additional devices 1514, computer readable medium 1516, a communication interface 1518, and one or more user interface devices 1520. Each processor 1512 is capable of processing instructions for execution within the server 1502. In some implementations, the processor 1512 is a single or multi-threaded microprocessor, for example. Each processor 1512 is capable of processing instructions stored on the computer readable medium 1516 or on another storage device. The server 1502 uses its communication interface 1518 to communicate with one or more computers, for example, over a network. Examples of user interface devices 1520 include a display, a camera, a speaker, a microphone, a tactile feedback device, a keyboard, and a mouse.
The server 1502 can store instructions that implement operations associated with the modules described above, for example, on the computer readable medium 1516 or one or more additional devices 1514, for example, one or more of a floppy disk device, a hard disk device, an optical disk device, or a tape device. In some implementations, the server 1502 also stores additional data, for example, model data, a document-query database, or other data, on the computer readable medium 1516 or one or more additional devices 1514.
User computers such as user computer A 1504 are connected to the server 1502 through the network 1506. Users submit search queries to the server 1502 and select search results received from the server 1502 using user computers such as user computer A 1504, for example, using a web-browser running on the computer. Users may also view other documents using the web-browser. Users can also use user computers such as user computer A 1504 to view recommendations from the server 1502, for example, recommendations for related documents or queries.
Embodiments of the subject matter and the functional operations described in this specification can be implemented in digital electronic circuitry, or in computer software, firmware, or hardware, including the structures disclosed in this specification and their structural equivalents, or in combinations of one or more of them. Embodiments of the subject matter described in this specification can be implemented as one or more computer programs, i.e., one or more modules of computer program instructions encoded on a computer storage medium for execution by, or to control the operation of, data processing apparatus. Alternatively or in addition, the program instructions can be encoded on a propagated signal that is an artificially generated signal, e.g., a machine-generated electrical, optical, or electromagnetic signal, that is generated to encode information for transmission to suitable receiver apparatus for execution by a data processing apparatus. The computer storage medium can be a machine-readable storage device, a machine-readable storage substrate, a random or serial access memory device, or a combination of one or more of them.
The term “data processing apparatus” encompasses all kinds of apparatus, devices, and machines for processing data, including by way of example a programmable processor, a computer, or multiple processors or computers. The apparatus can include special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit). The apparatus can also include, in addition to hardware, code that creates an execution environment for the computer program in question, e.g., code that constitutes processor firmware, a protocol stack, a database management system, an operating system, or a combination of one or more of them.
A computer program (also known as a program, software, software application, script, or code) can be written in any form of programming language, including compiled or interpreted languages, or declarative or procedural languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program may, but need not, correspond to a file in a file system. A program can be stored in a portion of a file that holds other programs or data (e.g., one or more scripts stored in a markup language document), in a single file dedicated to the program in question, or in multiple coordinated files (e.g., files that store one or more modules, sub-programs, or portions of code). A computer program can be deployed to be executed on one computer or on multiple computers that are located at one site or distributed across multiple sites and interconnected by a communication network.
The processes and logic flows described in this specification can be performed by one or more programmable processors executing one or more computer programs to perform functions by operating on input data and generating output. The processes and logic flows can also be performed by, and apparatus can also be implemented as, special purpose logic circuitry, e.g., an FPGA (field programmable gate array) or an ASIC (application-specific integrated circuit).
Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. The essential elements of a computer are a processor for performing or executing instructions and one or more memory devices for storing instructions and data. Generally, a computer will also include, or be operatively coupled to receive data from or transfer data to, or both, one or more mass storage devices for storing data, e.g., magnetic, magneto-optical disks, or optical disks. However, a computer need not have such devices. Moreover, a computer can be embedded in another device, e.g., a mobile telephone, a personal digital assistant (PDA), a mobile audio or video player, a game console, a Global Positioning System (GPS) receiver, or a portable storage device (e.g., a universal serial bus (USB) flash drive), to name just a few.
Computer-readable media suitable for storing computer program instructions and data include all forms of non-volatile memory, media and memory devices, including by way of example semiconductor memory devices, e.g., EPROM, EEPROM, and flash memory devices; magnetic disks, e.g., internal hard disks or removable disks; magneto-optical disks; and CD-ROM and DVD-ROM disks. The processor and the memory can be supplemented by, or incorporated in, special purpose logic circuitry.
To provide for interaction with a user, embodiments of the subject matter described in this specification can be implemented on a computer having a display device, e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor, for displaying information to the user and a keyboard and a pointing device, e.g., a mouse or a trackball, by which the user can provide input to the computer. Other kinds of devices can be used to provide for interaction with a user as well; for example, feedback provided to the user can be any form of sensory feedback, e.g., visual feedback, auditory feedback, or tactile feedback; and input from the user can be received in any form, including acoustic, speech, or tactile input. In addition, a computer can interact with a user by sending documents to and receiving documents from a device that is used by the user; for example, by sending web pages to a web browser on a user's client device in response to requests received from the web browser.
Embodiments of the subject matter described in this specification can be implemented in a computing system that includes a back-end component, e.g., as a data server, or that includes a middleware component, e.g., an application server, or that includes a front-end component, e.g., a client computer having a graphical user interface or a Web browser through which a user can interact with an implementation of the subject matter described in this specification, or any combination of one or more such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication, e.g., a communication network. Examples of communication networks include a local area network (“LAN”) and a wide area network (“WAN”), e.g., the Internet.
The computing system can include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
While this specification contains many specific implementation details, these should not be construed as limitations on the scope of any invention or of what may be claimed, but rather as descriptions of features that may be specific to particular embodiments of particular inventions. Certain features that are described in this specification in the context of separate embodiments can also be implemented in combination in a single embodiment. Conversely, various features that are described in the context of a single embodiment can also be implemented in multiple embodiments separately or in any suitable subcombination. Moreover, although features may be described above as acting in certain combinations and even initially claimed as such, one or more features from a claimed combination can in some cases be excised from the combination, and the claimed combination may be directed to a subcombination or variation of a subcombination.
Similarly, while operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in sequential order, or that all illustrated operations be performed, to achieve desirable results. In certain circumstances, multitasking and parallel processing may be advantageous. Moreover, the separation of various system components in the embodiments described above should not be understood as requiring such separation in all embodiments, and it should be understood that the described program components and systems can generally be integrated together in a single software product or packaged into multiple software products.
Particular embodiments of the subject matter have been described. Other embodiments are within the scope of the following claims. For example, the actions recited in the claims can be performed in a different order and still achieve desirable results. As one example, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In certain implementations, multitasking and parallel processing may be advantageous.
As another example, the strength of relationship score between each candidate document and each first document described above, for example, with reference to
This application is a continuation application of, and claims priority to, pending U.S. patent application Ser. No. 12/506,203 filed on Jul. 20, 2009, entitled “Generating a Related Set of Documents for an Initial Set of Documents”, which is incorporated by reference herein in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
5265065 | Turtle | Nov 1993 | A |
5488725 | Turtle | Jan 1996 | A |
5696962 | Kupiec | Dec 1997 | A |
5920854 | Kirsch et al. | Jul 1999 | A |
5963940 | Liddy et al. | Oct 1999 | A |
6006222 | Culliss | Dec 1999 | A |
6006225 | Bowman et al. | Dec 1999 | A |
6014665 | Culliss | Jan 2000 | A |
6026388 | Liddy et al. | Feb 2000 | A |
6067565 | Horvitz | May 2000 | A |
6078916 | Culliss | Jun 2000 | A |
6078917 | Paulsen et al. | Jun 2000 | A |
6088692 | Driscoll | Jul 2000 | A |
6134532 | Lazarus et al. | Oct 2000 | A |
6182066 | Marques et al. | Jan 2001 | B1 |
6182068 | Culiss | Jan 2001 | B1 |
6185559 | Brin et al. | Feb 2001 | B1 |
6249252 | Dupray | Jun 2001 | B1 |
6269368 | Diamond | Jul 2001 | B1 |
6285999 | Page | Sep 2001 | B1 |
6321228 | Crandall et al. | Nov 2001 | B1 |
6327590 | Chidlovskii et al. | Dec 2001 | B1 |
6341283 | Yamakawa et al. | Jan 2002 | B1 |
6353849 | Linsk | Mar 2002 | B1 |
6363378 | Conklin et al. | Mar 2002 | B1 |
6370526 | Agrawal et al. | Apr 2002 | B1 |
6421675 | Ryan et al. | Jul 2002 | B1 |
6473752 | Fleming, III | Oct 2002 | B1 |
6480843 | Li | Nov 2002 | B2 |
6490575 | Berstis | Dec 2002 | B1 |
6526440 | Bharat | Feb 2003 | B1 |
6529903 | Smith et al. | Mar 2003 | B2 |
6539377 | Culliss | Mar 2003 | B1 |
6560590 | Shwe et al. | May 2003 | B1 |
6567103 | Chaudhry | May 2003 | B1 |
6587848 | Aggarwal et al. | Jul 2003 | B1 |
6615209 | Gomes | Sep 2003 | B1 |
6623529 | Lakritz | Sep 2003 | B1 |
6640218 | Golding et al. | Oct 2003 | B1 |
6658423 | Pugh et al. | Dec 2003 | B1 |
6671681 | Emens et al. | Dec 2003 | B1 |
6678681 | Brin et al. | Jan 2004 | B1 |
6701309 | Beeferman et al. | Mar 2004 | B1 |
6725259 | Bharat | Apr 2004 | B1 |
6738764 | Mao et al. | May 2004 | B2 |
6754873 | Law et al. | Jun 2004 | B1 |
6792416 | Soetarman et al. | Sep 2004 | B2 |
6795820 | Barnett | Sep 2004 | B2 |
6816850 | Culliss | Nov 2004 | B2 |
6853993 | Ortega et al. | Feb 2005 | B2 |
6873982 | Bates et al. | Mar 2005 | B1 |
6877002 | Prince | Apr 2005 | B2 |
6882999 | Cohen et al. | Apr 2005 | B2 |
6901402 | Corston-Oliver et al. | May 2005 | B1 |
6912505 | Linden et al. | Jun 2005 | B2 |
6944611 | Flank et al. | Sep 2005 | B2 |
6944612 | Roustant et al. | Sep 2005 | B2 |
6954750 | Bradford | Oct 2005 | B2 |
6963867 | Ford et al. | Nov 2005 | B2 |
6990453 | Wang et al. | Jan 2006 | B2 |
7016939 | Rothwell et al. | Mar 2006 | B1 |
7028027 | Zha et al. | Apr 2006 | B1 |
7072886 | Salmenkaita et al. | Jul 2006 | B2 |
7085761 | Shibata | Aug 2006 | B2 |
7113939 | Chou et al. | Sep 2006 | B2 |
7117206 | Bharat et al. | Oct 2006 | B1 |
7136849 | Patrick | Nov 2006 | B2 |
7146361 | Broder et al. | Dec 2006 | B2 |
7222127 | Bem et al. | May 2007 | B1 |
7231399 | Bem et al. | Jun 2007 | B1 |
7243102 | Naam et al. | Jul 2007 | B1 |
7249126 | Ginsburg et al. | Jul 2007 | B1 |
7266765 | Golovchinsky et al. | Sep 2007 | B2 |
7293016 | Shakib et al. | Nov 2007 | B1 |
7379951 | Chkodrov et al. | May 2008 | B2 |
7382358 | Kushler et al. | Jun 2008 | B2 |
7395222 | Sotos | Jul 2008 | B1 |
7426507 | Patterson | Sep 2008 | B1 |
7451487 | Oliver et al. | Nov 2008 | B2 |
7499919 | Meyerzon et al. | Mar 2009 | B2 |
7505964 | Tong et al. | Mar 2009 | B2 |
7516146 | Robertson et al. | Apr 2009 | B2 |
7526470 | Karnawat et al. | Apr 2009 | B1 |
7533092 | Berkhin et al. | May 2009 | B2 |
7533130 | Narayana et al. | May 2009 | B2 |
7552112 | Jhala et al. | Jun 2009 | B2 |
7565363 | Anwar | Jul 2009 | B2 |
7565367 | Barrett et al. | Jul 2009 | B2 |
7566363 | Starling et al. | Jul 2009 | B2 |
7574530 | Wang et al. | Aug 2009 | B2 |
7584181 | Zeng et al. | Sep 2009 | B2 |
7603350 | Guha | Oct 2009 | B1 |
7610282 | Datar et al. | Oct 2009 | B1 |
7657626 | Zwicky | Feb 2010 | B1 |
7676507 | Maim | Mar 2010 | B2 |
7680775 | Levin et al. | Mar 2010 | B2 |
7693818 | Majumder | Apr 2010 | B2 |
7716225 | Dean et al. | May 2010 | B1 |
7747612 | Thun et al. | Jun 2010 | B2 |
7756887 | Haveliwala | Jul 2010 | B1 |
7783632 | Richardson et al. | Aug 2010 | B2 |
7801885 | Verma | Sep 2010 | B1 |
7809716 | Wang et al. | Oct 2010 | B2 |
7818315 | Cucerzan et al. | Oct 2010 | B2 |
7818320 | Makeev | Oct 2010 | B2 |
7836058 | Chellapilla | Nov 2010 | B2 |
7844589 | Wang et al. | Nov 2010 | B2 |
7849089 | Zhang et al. | Dec 2010 | B2 |
7853557 | Schneider et al. | Dec 2010 | B2 |
7856446 | Brave et al. | Dec 2010 | B2 |
7860886 | Loftesness | Dec 2010 | B2 |
7877404 | Achan et al. | Jan 2011 | B2 |
7895177 | Wu | Feb 2011 | B2 |
7925649 | Jeh et al. | Apr 2011 | B2 |
7953740 | Vadon et al. | May 2011 | B1 |
7974974 | Tankovich et al. | Jul 2011 | B2 |
7987185 | Mysen et al. | Jul 2011 | B1 |
8019650 | Donsbach et al. | Sep 2011 | B2 |
8024326 | Tong et al. | Sep 2011 | B2 |
8024330 | Franco et al. | Sep 2011 | B1 |
8027439 | Zoldi et al. | Sep 2011 | B2 |
8037042 | Anderson et al. | Oct 2011 | B2 |
8037043 | Zoeter et al. | Oct 2011 | B2 |
8037086 | Upstill et al. | Oct 2011 | B1 |
8051061 | Niu et al. | Nov 2011 | B2 |
8060456 | Gao et al. | Nov 2011 | B2 |
8060497 | Zatsman et al. | Nov 2011 | B1 |
8065296 | Franz et al. | Nov 2011 | B1 |
8069182 | Pieper | Nov 2011 | B2 |
8073263 | Hull et al. | Dec 2011 | B2 |
8073772 | Bishop et al. | Dec 2011 | B2 |
8073867 | Chowdhury | Dec 2011 | B2 |
8082242 | Mysen et al. | Dec 2011 | B1 |
8086282 | Hellberg | Dec 2011 | B2 |
8086599 | Heymans | Dec 2011 | B1 |
8090717 | Bharat et al. | Jan 2012 | B1 |
8156111 | Jones et al. | Apr 2012 | B2 |
8224827 | Dean et al. | Jul 2012 | B2 |
8239370 | Wong et al. | Aug 2012 | B2 |
8396865 | Ie et al. | Mar 2013 | B1 |
8412699 | Mukherjee et al. | Apr 2013 | B1 |
8447760 | Tong et al. | May 2013 | B1 |
8458165 | Liao et al. | Jun 2013 | B2 |
8498974 | Kim et al. | Jul 2013 | B1 |
8521725 | Pearson et al. | Aug 2013 | B1 |
8583636 | Franz et al. | Nov 2013 | B1 |
8661029 | Kim et al. | Feb 2014 | B1 |
8738596 | Kim et al. | May 2014 | B1 |
20010000356 | Woods | Apr 2001 | A1 |
20020034292 | Tuoriniemi et al. | Mar 2002 | A1 |
20020042791 | Smith et al. | Apr 2002 | A1 |
20020049752 | Bowman et al. | Apr 2002 | A1 |
20020103790 | Wang et al. | Aug 2002 | A1 |
20020123988 | Dean et al. | Sep 2002 | A1 |
20020133481 | Smith et al. | Sep 2002 | A1 |
20020165849 | Singh et al. | Nov 2002 | A1 |
20030009399 | Boerner | Jan 2003 | A1 |
20030018707 | Flocken | Jan 2003 | A1 |
20030028529 | Cheung et al. | Feb 2003 | A1 |
20030037074 | Dwork et al. | Feb 2003 | A1 |
20030078914 | Witbrock | Apr 2003 | A1 |
20030120654 | Edlund et al. | Jun 2003 | A1 |
20030135490 | Barrett et al. | Jul 2003 | A1 |
20030149704 | Yayoi et al. | Aug 2003 | A1 |
20030167252 | Odom et al. | Sep 2003 | A1 |
20030195877 | Ford et al. | Oct 2003 | A1 |
20030204495 | Lehnert | Oct 2003 | A1 |
20030220913 | Doganata et al. | Nov 2003 | A1 |
20030229640 | Carlson et al. | Dec 2003 | A1 |
20040006456 | Loofbourrow | Jan 2004 | A1 |
20040006740 | Krohn et al. | Jan 2004 | A1 |
20040034632 | Carmel et al. | Feb 2004 | A1 |
20040049486 | Scanlon et al. | Mar 2004 | A1 |
20040059708 | Dean et al. | Mar 2004 | A1 |
20040083205 | Yeager | Apr 2004 | A1 |
20040093325 | Banerjee et al. | May 2004 | A1 |
20040119740 | Chang et al. | Jun 2004 | A1 |
20040122811 | Page | Jun 2004 | A1 |
20040153472 | Rieffanaugh, Jr. | Aug 2004 | A1 |
20040158560 | Wen et al. | Aug 2004 | A1 |
20040186828 | Yadav | Sep 2004 | A1 |
20040186996 | Gibbs et al. | Sep 2004 | A1 |
20040199419 | Kim et al. | Oct 2004 | A1 |
20040215607 | Travis, Jr. | Oct 2004 | A1 |
20050015366 | Carrasco et al. | Jan 2005 | A1 |
20050021397 | Cui et al. | Jan 2005 | A1 |
20050027691 | Brin et al. | Feb 2005 | A1 |
20050033803 | Vleet et al. | Feb 2005 | A1 |
20050050014 | Gosse et al. | Mar 2005 | A1 |
20050055342 | Bharat et al. | Mar 2005 | A1 |
20050055345 | Ripley | Mar 2005 | A1 |
20050060290 | Herscovici et al. | Mar 2005 | A1 |
20050060310 | Tong et al. | Mar 2005 | A1 |
20050060311 | Tong et al. | Mar 2005 | A1 |
20050071741 | Acharya et al. | Mar 2005 | A1 |
20050102282 | Linden | May 2005 | A1 |
20050125376 | Curtis et al. | Jun 2005 | A1 |
20050160083 | Robinson | Jul 2005 | A1 |
20050192946 | Lu et al. | Sep 2005 | A1 |
20050198026 | Dehlinger et al. | Sep 2005 | A1 |
20050222987 | Vadon | Oct 2005 | A1 |
20050222998 | Driessen et al. | Oct 2005 | A1 |
20050240576 | Piscitello et al. | Oct 2005 | A1 |
20050240580 | Zamir et al. | Oct 2005 | A1 |
20050256848 | Alpert et al. | Nov 2005 | A1 |
20060036593 | Dean et al. | Feb 2006 | A1 |
20060047643 | Chaman | Mar 2006 | A1 |
20060069667 | Manasse et al. | Mar 2006 | A1 |
20060074903 | Meyerzon et al. | Apr 2006 | A1 |
20060089926 | Knepper et al. | Apr 2006 | A1 |
20060095421 | Nagai et al. | May 2006 | A1 |
20060106793 | Liang | May 2006 | A1 |
20060123014 | Ng | Jun 2006 | A1 |
20060173830 | Smyth et al. | Aug 2006 | A1 |
20060195443 | Franklin et al. | Aug 2006 | A1 |
20060200476 | Gottumukkala et al. | Sep 2006 | A1 |
20060200556 | Brave et al. | Sep 2006 | A1 |
20060227992 | Rathus et al. | Oct 2006 | A1 |
20060230040 | Curtis et al. | Oct 2006 | A1 |
20060259467 | Westphal | Nov 2006 | A1 |
20060259476 | Kadayam et al. | Nov 2006 | A1 |
20060293950 | Meek et al. | Dec 2006 | A1 |
20070005575 | Dai et al. | Jan 2007 | A1 |
20070005588 | Zhang et al. | Jan 2007 | A1 |
20070038659 | Datar et al. | Feb 2007 | A1 |
20070050339 | Kasperski et al. | Mar 2007 | A1 |
20070061195 | Liu et al. | Mar 2007 | A1 |
20070061211 | Ramer et al. | Mar 2007 | A1 |
20070081197 | Omoigui | Apr 2007 | A1 |
20070106659 | Lu et al. | May 2007 | A1 |
20070112730 | Gulli et al. | May 2007 | A1 |
20070130370 | Akaezuwa | Jun 2007 | A1 |
20070156677 | Szabo | Jul 2007 | A1 |
20070172155 | Guckenberger | Jul 2007 | A1 |
20070180355 | McCall et al. | Aug 2007 | A1 |
20070192190 | Granville | Aug 2007 | A1 |
20070208730 | Agichtein et al. | Sep 2007 | A1 |
20070214131 | Cucerzan et al. | Sep 2007 | A1 |
20070233653 | Biggs et al. | Oct 2007 | A1 |
20070255689 | Sun et al. | Nov 2007 | A1 |
20070260596 | Koran et al. | Nov 2007 | A1 |
20070260597 | Cramer et al. | Nov 2007 | A1 |
20070266021 | Aravamudan et al. | Nov 2007 | A1 |
20070266439 | Kraft | Nov 2007 | A1 |
20070288450 | Datta et al. | Dec 2007 | A1 |
20080010143 | Kniaz et al. | Jan 2008 | A1 |
20080027913 | Chang et al. | Jan 2008 | A1 |
20080052219 | Sandholm et al. | Feb 2008 | A1 |
20080052273 | Pickens | Feb 2008 | A1 |
20080059453 | Laderman | Mar 2008 | A1 |
20080077570 | Tang et al. | Mar 2008 | A1 |
20080082518 | Loftesness | Apr 2008 | A1 |
20080091650 | Fontoura et al. | Apr 2008 | A1 |
20080104043 | Garg et al. | May 2008 | A1 |
20080114624 | Kitts | May 2008 | A1 |
20080114729 | Raman et al. | May 2008 | A1 |
20080114750 | Saxena et al. | May 2008 | A1 |
20080140699 | Jones et al. | Jun 2008 | A1 |
20080162475 | Meggs et al. | Jul 2008 | A1 |
20080183660 | Szulcewski | Jul 2008 | A1 |
20080189269 | Olsen | Aug 2008 | A1 |
20080208825 | Curtis et al. | Aug 2008 | A1 |
20080228442 | Lippincott et al. | Sep 2008 | A1 |
20080256050 | Zhang et al. | Oct 2008 | A1 |
20080313168 | Liu et al. | Dec 2008 | A1 |
20080313247 | Galvin | Dec 2008 | A1 |
20090006438 | Tunkelang et al. | Jan 2009 | A1 |
20090012969 | Rail et al. | Jan 2009 | A1 |
20090055392 | Gupta et al. | Feb 2009 | A1 |
20090070194 | Song | Mar 2009 | A1 |
20090157643 | Gollapudi et al. | Jun 2009 | A1 |
20090182723 | Shnitko et al. | Jul 2009 | A1 |
20090187557 | Hansen et al. | Jul 2009 | A1 |
20090228442 | Adams et al. | Sep 2009 | A1 |
20090287656 | Bennett | Nov 2009 | A1 |
20090313242 | Kodama et al. | Dec 2009 | A1 |
20100106706 | Rorex et al. | Apr 2010 | A1 |
20100131563 | Yin | May 2010 | A1 |
20100205541 | Rapaport et al. | Aug 2010 | A1 |
20100228738 | Mehta et al. | Sep 2010 | A1 |
20100241472 | Hernandez | Sep 2010 | A1 |
20100299317 | Uy | Nov 2010 | A1 |
20100325131 | Dumais et al. | Dec 2010 | A1 |
20110179093 | Pike et al. | Jul 2011 | A1 |
20110219025 | Lipson et al. | Sep 2011 | A1 |
20110264670 | Banerjee et al. | Oct 2011 | A1 |
20110282906 | Wong | Nov 2011 | A1 |
20110295844 | Sun et al. | Dec 2011 | A1 |
20110295879 | Logis et al. | Dec 2011 | A1 |
20120011148 | Rathus et al. | Jan 2012 | A1 |
20120191705 | Tong et al. | Jul 2012 | A1 |
Number | Date | Country |
---|---|---|
WO 0077689 | Dec 2000 | WO |
WO 0116807 | Mar 2001 | WO |
WO 0167297 | Sep 2001 | WO |
WO 2004059514 | Jul 2004 | WO |
Entry |
---|
Joachims, “Evaluating Search Engines Using Clickthrough Data”, Cornell University, Department of Computer Science, Draft, Feb. 19, 2002, 13 pages. |
Jansen et al., “An Analysis of Web Documents Retrieved and Viewed”, School of Information Sciences and Technology, The Pennsylvania State University, the 4th International Conference on Internet Computing, Las Vegas, Nevada, pp. 65-69, Jun. 23-26, 2003, 5 pages. |
Diligenti, et al., Users, Queries and Documents: A Unified Representation for Web Mining, wi-iat, vol. 1, 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology, 2009, pp. 238-244. |
Agichtein, et al; Improving Web Search Ranking by Incorporating User Behavior Information; Aug. 2006; Proceedings of the Twenty-Ninth Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, p. 19-26. |
Agichtein, et al; Learning User Interaction Models for Predicting Web Search Rersult Performances; Aug. 2006; Proceedings of the Twenty-Ninth Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, p. 3-10. |
Linden, Greg et. al., Amazon.com Recommedations: Item-to-Item Collaborative Filtering, [online], http://computer.org/internet/, IEEE Internet Computing, Jan.-Feb. 2003, IEEE Computer Society, pp. 76-80. |
Boyan et al.; A Machine Learning Architecture for Optimizing Web Search Engines; Aug. 1996; Internet-based information systems-Workshop Technichal Report—American Association for Artificial Intelligence, pp. 1-8. |
Cutrell, et al.; Eye tracking in MSN Search: Investigating snippet length, target position and task types; 2007; Conference on Human Factors in Computing Systems—Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. |
Gr{hacek over (c)}ar, Miha, User Profiling: Collaborative Filtering, SIKDD 2004, Oct. 12-15, 2004, Ljubljana, Slovenia, 4 pages. |
Nicole, Kristen, Heeii is StumbleUpon Plus Google Suggestions, [online], Retrieved from the Internet http://mashale.com/2007/05/15/heeii/, 11 pages. |
Hofmann, Thomas, Latent Semantic Models for Collaborative Filtering, ACM Transactions on Information Systems, vol. 22, No. 1, Jan. 2004, pp. 89-115. |
Kelly et al.; Implicit Feedback for Inferring User Preference: A Bibliography; SIGIR Forum, vol. 37, No. 2 (2003), pp. 18-28. |
Lemire, Daniel, Scale and Translation Invariant Collaborative Filtering Systems, Published in Information Retrieval, 8(1), pp. 129-150, 2005. |
Radlinski et al., Query Chains: Learning to Rank from Implicit Feedback, KDD '05, Aug. 21-24, 2005, Chicago, Illinois, USA, 10 pages. |
Schwab, et al., Adaptivity through Unobstrusive Learning, 2002, 16(3), pp. 5-9. |
Stoilova, Lubomira et al., GiveALink: Mining a Semantic Network of Bookmarks for Web Search and Recommendation, LinkKDD '05, Aug. 21, 2005, Chicago, IL, USA, 8 pages. |
Xiao, et al., Measuring Similarity of Interests for Clustering Web-Users, ADC, 2001, p. 107-114. |
Xie et al., Web User Clustering from Access Log Using Belief Function, K-CAP '01 Oct. 22-23, 2001, Victoria, British Columbia, Canada, pp. 202-208. |
Yu et al., Selecting Relevant Instances for Efficient and Accurate Collaborative Filtering, CIKM '01, Nov. 5-10, 2001, Atlanta, Georgia, pp. 239-246. |
Zeng et al., Similarity Measure and Instance Selection for Collaborative Filtering, WWW '03, May 20-24, Budapest, Hungary, pp. 652-658. |
Soumen Chakrabarti, et al. “Enhanced Topic Distillation using Text, Markup tags, and Hyperlinks”. ACM, Sep. 9-12, 2001, pp. 206-216. |
Gabriel Somlo et al., “Using Web Helper Agent Profiles in Query Generation”, ACM, Jul. 2003, pp. 812-818. |
Bar-Llan et al., “Presentation Bias is Significant in Determining User Preference for Search Results—A User Study”; Journal of the American Society for Information Science and Technology, vol. 60, Issue 1 (p. 135-149), Sep. 2008, 15 pages. |
Bar-Llan et al.; “Methods for comparing rankings of search engine results”; Computer Networks: The International Journal of Computer and Telecommunications Networking, Jul. 2006, vol. 50, Issue 10, 19 pages. |
Boldi, et al.; The Query-flow Graph Model and Applications; CKIM '08, Oct. 26-30, Napa Valley, California, USA, pp. 609-617. |
Burke, Robin, Integrating Knowledge-based and Collaborative-filtering Recommender Systems, AAAI Technical Report WS-99-01. Compilation copyright © 1999, AAAI (www.aaai.org), pp. 69-72. |
Craswell, et al.; Random Walks on the Click Graph; Jul. 2007; SIGIR '07, Amsterdam, the Netherlands, 8 pages. |
Google News Archive, Jul. 8, 2003, Webmasterworld.com, [online] Retrieved from the Internet http://www.webmasterwolrd.com/forum3/15085.htm [retrieved on Nov. 20, 2009] 3 pages. |
Joachims et al., “Search Engines that Learn from Implicit Feedback”; Aug. 2007, IEEE Computer Society. |
Joachims, T., Evaluating retrieval performance using clickthrough data. Proceedings of the SIGIR Workshop on Mathematical/Formal Methods in Information Retrieval; Aug. 12-15, 2002; Tampere, Finland, 18 pages. |
Joachims; Optimizing search engines using clickthrough data; 2002; Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 133-142. |
U.S. Appl. No. 11/556,143, filed Nov. 2, 2006, in Office Action mailed Apr. 20, 2011, 18 pages. |
U.S. Appl. No. 11/556,143, filed Nov. 2, 2006, in Office Action mailed Jan. 25, 2010, 14 pages. |
U.S. Appl. No. 11/556,143, filed Nov. 2, 2006, in Office Action mailed Jul. 6, 2010, 20 pages. |
U.S. Appl. No. 11/685,095, filed Mar. 12, 2007, in Office Action mailed Apr. 13, 2011, 31 pages. |
U.S. Appl. No. 11/685,095, filed Mar. 12, 2007, in Office Action mailed Feb. 25, 2009, 21 pages. |
U.S. Appl. No. 11/685,095, filed Mar. 12, 2007, in Office Action mailed Feb. 8, 2010, 31 pages. |
U.S. Appl. No. 11/685,095, filed Mar. 12, 2007, in Office Action mailed Sep. 10, 2009, 23 pages. |
U.S. Appl. No. 11/556,086, filed Nov. 2, 2006, in Office Action mailed Jun. 23, 2010, 21 pages. |
W3C, URIs, URLs and URNs: Classification and Recommendations 1.0, Report from the joint W3C/IETF URI Planning Interest Group, Sep. 21, 2001, 8 pages. |
Xiao et al., Measuring Similarity of Interests for Clustering Web-Users, ADC, 2001, pp. 107-114. |
Zeng, et al., “Learning to Cluster Web Search Results” SIGIR '04, Proceesdings of the 27th Annual International ACM SIGIR conference on research and development in information retrieval, 2004. |
Australian Patent Office Non-Final Office ACtion in AU App. Ser. No. 2004275274, mailed Feb. 3, 2010, 2 pages. |
Dan Olsen et al., “Query-by-critique: Spoken Language Access to Large Lists”, ACM, Oct. 2002, pp. 131-140. |
Susan Gauch et al., “A Corpus Analysis Approach for Automatic Query Expansion and its Extenstion to Multiple Databases”, ACM, Jul. 1999, pp. 250-269. |
Nicolas Bruno et al., “Top-K Selection Queries over Relational Databases: Mapping Strategies and Performance Evaluation”, ACM, Jun. 2002, pp. 153-187. |
Ji-Rong Wen et al., “Query Clustering using User Logs”, ACM, Jan. 2002, pp. 59-81. |
Brin, S. and L. Page, The Anatomy of Large-Scale Hypertextual; Web Search Engine, Computer Science Department, Apr. 1998. |
International Search Report and Written Opinion for Application No. PCT/US2004/029615, dated Jan. 19, 2005, 8 pages. |
Hungarian Patent Office, International Search Report and Written Opinion for Application No. 200806756-3 dated Nov. 19, 2010 12 pages. |
Authorized Officer Athina Nickitas-Etienne, International Preliminary Report and Written Opinion for Application No. PCT/US2004/029615, mailed Mar. 23, 2006. |
Indian Office Action in Indian Application No. 686/KOLNP/2006, mailed Jun. 3, 2008, 2 pages. |
Danish Search Report and Written Opinion for Application No. 20060130-7, dated Jun. 21, 2007, 15 pages. |
Jones et al., “Pictures of Relevance: A Geometric Analysis of Similarity Measures”, Journal of the American Society for Information Science, Nov. 1987, 23 pages. |
Kaplan et al., “Adaptive Hypertext Navigation Based on User Goals and Context”, User Modeling and User-Adapted Interaction 2, Sep. 1, 1993; pp. 193-220, 28 pages. |
Liddy et al., “A Natural Language Text Retrieval System With Relevance Feedback”, 16th National Online, May 2-6, 1995, 3 pages. |
“Personalizing Search via Automated Analysis of Interests and Activities,” by Teevan et al. IN: SIGIR'05 (2005), 8 pages. |
Number | Date | Country | |
---|---|---|---|
Parent | 12506203 | Jul 2009 | US |
Child | 13617019 | US |