The present invention is directed to the field of query processing.
Many World Wide Web sites permit users to perform searches to identify a small number of interesting items among a much larger domain of items. As an example, several web index sites permit users to search for particular web sites among most of the known web sites. Similarly, many online merchants, such as booksellers, permit users to search for particular products among all of the products that can be purchased from a merchant. In many cases, users perform searches in order to ultimately find a single item within an entire domain of items.
In order to perform a search, a user submits a query containing one or more query terms. The query also explicitly or implicitly identifies a domain of items to search. For example, a user may submit a query to an online bookseller containing terms that the user believes are words in the title of a book. A query server program processes the query to identify within the domain items matching the terms of the query. The items identified by the query server program are collectively known as a query result. In the example, the query result is a list of books whose titles contain some or all of the query terms. The query result is typically displayed to the user as a list of items. This list may be ordered in various ways. For example, the list may be ordered alphabetically or numerically based on a property of each item, such as the title, author, or release date of each book. As another example, the list may be ordered based on the extent to which each identified item matches the terms of the query.
It is fairly common for a user to specify a query containing a term that doesn't match any items. This may happen, for example, where a user mistypes the non-matching term, or when the user types a term that, though it does not match any items, is a synonym of a word that matches one or more items. In such cases, conventional techniques, which present only items that satisfy the query, present no items to the user. When no items are presented to a user in response to issuing a query, the user can become frustrated with the search engine, and may even discontinue its use.
Accordingly, an approach to processing search queries that does not rely on the terms of the query matching text associated with items would have significant utility.
A software facility (the “facility”) is described that (1) discerns from user behavior associations between query terms and items relevant to such query terms, and (2) exploits such associations, such as (a) to add to a query result items not included in the query result produced for the query by a search engine but nonetheless likely to be relevant to the query, or (b) to generate a set of item recommendations for a user based upon the contents of the queries recently submitted by the user.
For a group of users, the facility monitors queries submitted and item selection actions taken, such as viewing an item's item detail page (e.g., as in an online catalog), adding an item to a shopping cart, adding an item to a gift registry, ordering an item, sampling an item, or rating or reviewing an item. For each item selection action by a user, the facility (1) identifies queries submitted by that same user in a short period of time before the selection action, and (2) correlates the item that is the subject of the selection action with the query terms in the identified queries. The facility uses the formed correlations to generate, for each of a number of combinations of a query term and an item, a score reflecting the tendency of users submitting queries containing the query term to go on to select the item. The facility can use a variety of approaches described below to perform the monitoring and determine the scores.
In order to exploit these term/item scores to add to a query result items not included by the search engine in the query result produced for the query but nonetheless likely to be relevant to the query, when a new query is received, in addition to being provided to the search index to generate a search result listing items that satisfy the query, it is provided to the facility to identify items that, when paired with the query term(s) in the query, have the highest scores. The facility provides the identified items for inclusion in or display with the query result generated by the search engine.
In this way, the facility makes query results potentially more valuable in a variety of different situations. Where a user has submitted a query containing a misspelled query term that, if properly spelled, would match a particular item, and the misspelling is somewhat common among users tracked by the facility, the facility can cause the item to be included in the augmented search result. For example, where a user submits the query “Morrissy” in an attempt to locate an album by the musician Morrissey, the facility may add one or more Morrissey albums to the produced query result where other users have previously submitted a query containing “Morrissy,” and gone on to do any of the following: discovered their spelling error, resubmitted the search with the correct spelling, and selected the album; or used a way other than search to locate and select the album, such as using a browse tree, best seller list, recommendation, wish list, reference to an external source such as an online or paper music magazine, etc. Similarly, when a user has submitted a query containing a synonym for a search term that would match an item that is commonly used by other users to search for the item, the facility can cause the item to be included in the augmented search result. For example, where a user submits the query “Samsung mobile” in an attempt to find wireless telephones manufactured by Samsung, none of which has the word “mobile” associated with them, the facility can cause one or more Samsung cellular wireless telephone items to be added to the query result.
Where a user has submitted a query containing a query for a first item that is commonly purchased—or otherwise selected—together with a second, “complement” item, the facility can add the complement item to the search result produced for the query. For example, where a user submits a query that identifies a particular printer item that is often purchased together with a particular printer cable item, the facility can add the printer cable item to the produced search result. Additionally, where a number of users use the same query terms to search for a first item, then follow a “related items” link on the first item's item detail page to a second item, the facility can add the second item to the search result. For example, where users commonly (1) submit the query “Grinch” to locate the DVD Dr. Seuss—How the Grinch Stole Christmas, then (2) follow a “related items” link on that item's item detail page to the item detail page for the DVD Frosty the Snowman, the facility can add the Frosty the Snowman DVD item to the search results generated for subsequent “Grinch” queries.
The facility can also exploit the term/item scores to generate a set of item recommendations for a user based upon the contents of the queries recently submitted by the user. When an opportunity arises to present item recommendations to a user—such as when the user displays a special recommendations page, or when the user displays a page of another type that has space reserved for user recommendations—the facility retrieves a history of the user's recent queries, and for each item, combines scores for combinations of that item with each query term appearing in the user's query history. The facility then selects a set of items to recommend that have the highest combined score. For example, where an opportunity arises to provide recommendation to a user who has recently submitted the queries “detective” and “mystery,” and where users that submit queries containing either of these query terms frequently buy or otherwise select the book The Adventures of Sherlock Holmes, the facility can provide recommendations to the user that include this book item. Recommendations generated in this way are often “fresher” than conventional recommendations based upon the user's purchases, which are sometimes criticized as stagnant. In some embodiments, recommendations generated for a user based upon the contents of the queries recently submitted by the user are combined with conventional recommendations for presentation to the user.
By compiling and/or exploiting the term/item scores in some or all of the ways described above, the facility can assist users by leveraging the earlier experiences and actions of all other users. In particular, where other users have had to expend a larger amount of effort to select an item not initially included with the query result that they received, a user that receives an augmented search result or recommendation produced by the facility that contains the item can select the item directly from the augmented search result or recommendation and avoid any additional effort.
Those skilled in the art will appreciate that queries submitted as shown in
In steps 501-505, the facility loops through events in the web server log that have not yet been processed. In step 502, the facility branches on the event type of the event: if the event is an item selection event, then the facility continues in step 503; if the event is a search query event, then the facility continues in step 504; if the event is an event of another type, then the facility continues in step 505. In step 503, the facility adds a row for the event to an item selection event table.
Returning to
Returning to
In steps 506-508, the facility loops through each row of the search query event table. In step 507, the facility copies the query terms from the current row of the search query event table to the preceding query terms field of each row of the item selection event table that both (1) contains the same user ID, and (2) has a time that is within a specified period after the time of the row of the search query event table. The size of the specified period of time is configurable, and should be based on what is understood to be the typical amount of time after performing a search that a user may select an item that is related to the search. Typical periods of time may be ten minutes, one hour, five hours, one day, one week, etc. In step 508, if additional rows of the search query event table remain to be processed, then the facility continues at step 506 to process the next row, else the facility continues in step 509.
Returning to
Embodiments of the facility generate term/item scores in a variety of ways. Most of the approaches to scoring used by the facility involve, at a minimum, counting the number of users that both (1) submitted a query for the query term, and (2) within the prescribed amount of time, performed a selection action with respect to the item. In one approach to scoring, the facility may simply sum the number of users that both submitted a query for the term and selected the item. In some embodiments, the facility weights differently the types of selection actions. For example, the facility may weight a purchase selection action more heavily than a display item detail selection action. Also, the facility may weight more heavily selection actions that require a greater level of effort to perform, as is discussed in greater detail in U.S. patent application Ser. No. 09/847,954, entitled “Identifying the Items Most Relevant to a Current Query Based on Items Selection in Connection with Similar Queries,” filed on May 2, 2001 (now U.S. Pat. No. 7,124,129, issued on Oct. 17, 2006), which is hereby incorporated by reference in its entirety.
Additional approaches used by embodiments of the facility include:
score1(term, item)=C/SQRT(A·B) (1)
The above mentioned scores can be combined in a variety of ways to produce a weighted average of multiple scores. For example:
score5(term, item)=ΣWiscorei(term, item) (5)
where W is a weighting function for each correlation score, scorei is the correlation score itself, and ΣWi=1. For example, the faculty could combine approaches (2) and (4) above as follows:
CS=w*(score2)=(1−w)*(score4) (6)
where w is a weighting factor such as 0.20.
In some embodiments, the facility uses a sliding window approach to incorporate the most recent information derived from weblogs in the relatedness scores determined by the facility. Suitable approaches of this type are described in the aforementioned U.S. patent application Ser. No. 09/847,954 (now U.S. Pat. No. 7,124,129), and in U.S. patent application Ser. No. 10/817,554, entitled “Automated Detection of Associations Between Search Criteria and Item Categories Based on Collective Analysis of User Activity Data,” filed on Apr. 2, 2004 (published on Oct. 6, 2005 as 2005/0222987 A1), which is hereby incorporated by reference in its entirety. In some embodiments, the facility uses an approach of maintaining separate sets of relatedness scores for different user populations of various types. A suitable approach of this type is described in the aforementioned U.S. patent application Ser. No. 09/847,954.
In steps 1102-1104, the facility loops through each item. In step 1103, the facility combines the term/item scores for the combinations of the current item and each of the terms read in step 1101 to obtain a combined score for the item. In step 1104, if additional items remain to be processed, then the facility continues in step 1102 to process the next item, else the facility continues in step 1105. In some embodiments (not shown), the facility limits the number of items for which it performs the processing in step 1103, such as by performing the processing only for items that have at least one term/item score with one of the red terms that exceeds a score threshold.
In step 1105, the facility recommends one or more items having the highest combined scores. After step 1105, these steps conclude.
The facility may perform the steps of
It will be appreciated by those skilled in the art that the above-described facility may be straightforwardly adapted or extended in various ways. While the foregoing description makes reference to particular embodiments, the scope of the invention is defined solely by the claims that follow and the elements recited therein.
| Number | Name | Date | Kind |
|---|---|---|---|
| 5278980 | Pedersen et al. | Jan 1994 | A |
| 5694592 | Driscoll | Dec 1997 | A |
| 6006225 | Bowman et al. | Dec 1999 | A |
| 6098065 | Skillen et al. | Aug 2000 | A |
| 6185558 | Bowman et al. | Feb 2001 | B1 |
| 6195654 | Wachtel | Feb 2001 | B1 |
| 6363378 | Conklin et al. | Mar 2002 | B1 |
| 6366910 | Rajaraman et al. | Apr 2002 | B1 |
| 6421675 | Ryan et al. | Jul 2002 | B1 |
| 6430558 | Delano | Aug 2002 | B1 |
| 6502091 | Chundi et al. | Dec 2002 | B1 |
| 6584462 | Neal et al. | Jun 2003 | B2 |
| 6606102 | Odom | Aug 2003 | B1 |
| 6665655 | Warner et al. | Dec 2003 | B1 |
| 6671681 | Emens et al. | Dec 2003 | B1 |
| 6732088 | Glance | May 2004 | B1 |
| 6732090 | Shanahan et al. | May 2004 | B2 |
| 6778979 | Grefenstette et al. | Aug 2004 | B2 |
| 6785671 | Bailey et al. | Aug 2004 | B1 |
| 7089237 | Turnbull et al. | Aug 2006 | B2 |
| 7152061 | Curtis et al. | Dec 2006 | B2 |
| 7162454 | Donner et al. | Jan 2007 | B1 |
| 20020002502 | Maes et al. | Jan 2002 | A1 |
| 20020072974 | Pugliese et al. | Jun 2002 | A1 |
| 20020078045 | Dutta | Jun 2002 | A1 |
| 20020103789 | Turnball et al. | Aug 2002 | A1 |
| 20020188694 | Yu | Dec 2002 | A1 |
| 20020194081 | Perkowski | Dec 2002 | A1 |
| 20020198882 | Linden et al. | Dec 2002 | A1 |
| 20030078928 | Dorosario et al. | Apr 2003 | A1 |
| 20030220909 | Farrett | Nov 2003 | A1 |
| 20040260677 | Malpani et al. | Dec 2004 | A1 |
| Number | Date | Country |
|---|---|---|
| WO 0116806 | Mar 2001 | WO |
| WO 0146870 | Jun 2001 | WO |