Alerting users to items of current interest

Information

  • Patent Grant
  • 9317560
  • Patent Number
    9,317,560
  • Date Filed
    Monday, April 22, 2013
    11 years ago
  • Date Issued
    Tuesday, April 19, 2016
    8 years ago
Abstract
Disseminating to a participant an indication that an item accessible by the participant via a network is of current interest is disclosed. An indication that the item is of current interest is received in real time. The indication is processed. The participant is informed that the item is of current interest.
Description
TECHNICAL FIELD

The present invention relates generally to communications and computer networks. More specifically, alerting users to dynamic content accessible via a communications or computer network that is of interest at the time of the alert is disclosed.


BACKGROUND

The use of the Internet, and in particular the World Wide Web, and other communication and computer networks has grown dramatically in recent years. The emergence of technologies for broader bandwidth communications, better compression technology, and new and less expensive digital recording and imaging technology, have all contributed to explosive growth in the volume and diversity of content available via communication and/or computer networks, such as the World Wide Web.


However, this proliferation of content, such as audio, image, and video content, presents certain challenges from the perspective of users seeking content of current interest. First, the shear volume of content available makes it difficult for users to find the content in which they are most interested in accessing at any given time. Apart from having to sort through the enormous volume of content available, much of the content of potentially greatest interest, at least to many users, is dynamic. At certain times, a file or other electronic resource may be of great interest while at other times, or perhaps even most of the time, it is not of great interest or not interesting at all.


For example, thousands of and perhaps in excess of a hundred thousand web cameras, or “webcams”, are in use. Webcams are cameras used to provide images of a target of interest via a site on the World Wide Web. Images are updated in varying manners and at varying intervals, depending on the site. A webcam might be used, for example, to provide images of a watering hole in Africa. Typically, users would access a website associated with the webcam to view activity at the watering hole. However, there would be many periods during which nothing of particular interest (e.g., no animals, etc.) would be happening at the watering hole. Conversely, there would be occasional periods when activity of great interest would be occurring, such as the presence of a rare or endangered animal at the watering hole. Users would have no way of knowing when such activity would be occurring and might miss the most interesting images if they did not happen to check the website at the right time. The same problems arise with respect to files or other electronic resources other than webcam content provided via the World Wide Web, including other media such as audio.


As a result, there is a need for a way to alert users to web content or other electronic resources available via a communications or computer network that are of interest at a particular time. To meet this latter need, there is a need to provide a way to become aware that dynamic web content or an electronic resource other than web content is of interest at a given time, and to quantify the degree or level of current interest. In addition, there is a need to consider the interests of a user when determining which web content or other electronic resources likely will be of the greatest interest to the user.


There is also a need to ensure that interested users receive alerts with respect to web content or other electronic resources that are of interest only to a relatively small community of users, or that are of interest on only relatively rare or infrequent occasions. There is a risk, otherwise, that indications of current interest regarding such files and other electronic resources would be masked by more voluminous or frequent activity with respect to more widely popular or pervasive resources or types of resources (such as pornography sites on the World Wide Web).


SUMMARY

Accordingly, alerting users of items of current interest is disclosed. The level of current interest of a particular file or other electronic resource is determined based on indications received from alerting users. One or more users receive an alert that the item is of current interest. Normalization of the level of current interest of a file or other resource, such as to adjust for items of current interest to a small community or for items of current interest only infrequently, also is described.


It should be appreciated that the present invention can be implemented in numerous ways, including as a process, an apparatus, a system, a device, a method, or a computer readable medium such as a computer readable storage medium or a computer network wherein program instructions are sent over optical or electronic communication links. Several inventive embodiments of the present invention are described below.


Disseminating to a participant an indication that an item accessible by the participant via a network is of current interest is disclosed. In one embodiment, an indication that the item is of current interest is received in real time. The indication is processed. The participant is informed that the item is of current interest.


In one embodiment, a computer is configured to receive in real time an indication that an item is of current interest; process the indication; and inform a participant that the item is of current interest. A database, associated with the computer, is configured to store data relating to the item.


In one embodiment, a computer program product for disseminating to a participant an indication that an item accessible by the participant via a network is of current interest comprises computer instructions for receiving in real time an indication that the item is of current interest; processing the indication; and informing the participant that the item is of current interest.


These and other features and advantages of the present invention will be presented in more detail in the following detailed description and the accompanying figures, which illustrate by way of example the principles of the invention.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will be readily understood by the following detailed description in conjunction with the accompanying drawings, wherein like reference numerals designate like structural elements, and in which:



FIG. 1 is a schematic diagram illustrating a system used in one embodiment to alert users to dynamic content of interest at the time of the alert (also referred to herein as an “item of current interest”).



FIG. 2A is a series of three screen shots showing three different states of an alert submission display 200 used in one embodiment.



FIG. 2B is an illustration of the data structure used in one embodiment for alerts submitted by an alerting user.



FIG. 3 is a flow chart illustrating a process used in one embodiment to alert users of items of current interest.



FIG. 4 is a flow chart illustrating a process used in one embodiment to receive an alert, as in step 302 of FIG. 3.



FIG. 5 is an illustration of the data structure used in one embodiment for the alert object.



FIG. 6 is a flowchart illustrating a process used in one embodiment to process an alert, as in step 304 of FIG. 3.



FIG. 7 is an illustration of six database tables 700 used in one embodiment to store data concerning alerts received with respect to items of current interest associated with URLs.



FIG. 8A is a flowchart illustrating a process used in one embodiment to update the intensity sum for a URL, as in step 606 of FIG. 6.



FIG. 8B is a flowchart illustrating a process used in one embodiment to update the intensity rank for a URL to reflect the intensity of the current alert.



FIG. 8C is a flowchart illustrating a process used in one embodiment to update the interest category weight for a URL with respect to the interest category indicated in an alert.



FIG. 9 is a flowchart illustrating a process used in one embodiment to purge records for URLs that are determined to be no longer of current interest by calculating a time decayed intensity rank at intervals, even if no new alert has been received, and purging from the database the records for a URL if the time decayed intensity rank is below a prescribed threshold.



FIG. 10 is a flowchart illustrating a process used in one embodiment to disseminate an alert to a participant, as in step 306 of FIG. 3.



FIG. 11 shows an exemplary participant display 1100 used in one embodiment to disseminate alert information to a participant.



FIG. 12 is a flowchart illustrating a process used in one embodiment to build a list of hot URLs responsive to a request, as in step 1008 of FIG. 10.





DETAILED DESCRIPTION

A detailed description of a preferred embodiment of the invention is provided below. While the invention is described in conjunction with that preferred embodiment, it should be understood that the invention is not limited to any one embodiment. On the contrary, the scope of the invention is limited only by the appended claims and the invention encompasses numerous alternatives, modifications and equivalents. For the purpose of example, numerous specific details are set forth in the following description in order to provide a thorough understanding of the present invention. The present invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the present invention is not unnecessarily obscured.



FIG. 1 is a schematic diagram illustrating a system used in one embodiment to alert users to dynamic content of interest at the time of the alert (also referred to herein as an “item of current interest”). The system 100 includes at least one alerting user 102 who accesses dynamic content associated with a uniform resource locator (URL), determines the content is of current interest, and sends an alert indicating that the URL is of current interest, as described more fully below. The system 100 also includes at least one participant 104. In one embodiment, participant 104 provides an indication of the participant's interests and receives a list of URLs providing the location of dynamic content, such as web content on the World Wide Web, that may be of interest to the participant at the time of the alert, as described more fully below. Both the alerting user 102 and the participant 104 are connected to a web server 105 via the Internet. Web server 105 is a computer system configured to present web pages and other web browser readable files, and to receive data from users, via the World Wide Web. Web server 105 is connected to an application server 106 and is configured to provide data to and receive data and instructions from application server 106. Application server 106 is configured to perform the application logic functions described more fully below. In one embodiment, the functions performed by the application server, as described more fully below, are divided among two or more computers so as to optimize the distribution of work load among the computers and to minimize the time the system takes to respond to inputs and queries from users.


When an alert has been received and is being processed, as described more fully below, the application server 106 comprises an alert software object 108 used to store data relating to and perform certain processing with respect to an alert, as described more fully below. The alert software object 108 uses data provided in an alert sent by alerting user 102, along with data retrieved from database 110 associated with the application server 106, to process the alert. Certain of the data that results from the processing performed by alert software object 108 is then stored in database 110. In one embodiment, database 110 is stored in memory in application server 106. In one embodiment, database 110 is stored in a separate structure, such as a database server, connected, either directly or through a communication link, with application server 106.


In one embodiment, when a request from a participant for a list of URLs for items of current interest is received, the application server 106 comprises a hot list software object 112 used to store certain data concerning and perform certain operations with respect to the request from the participant and the response thereto. In one embodiment, the hot list object 112 comprises an interest category array 114. In one embodiment, the interest category array 114 is comprised of one or more interest category objects, each of which stores data relating to one interest category identified in the participant's request as being of interest to the participant. In one embodiment, the hot list object 112 comprises a hot token array 116. The hot token array 116 is comprised of a hot token object for each URL of current interest in the database for the category or categories indicated in the participant's request.


As indicated in FIG. 1, an alert sent by an alerting user includes, in one embodiment, at least the URL of the web content considered by the alerting user to be of current interest. In one embodiment an alert may also include an interest selection, meaning a category or subject area to which the alerting user believes the web content relates, and/or a caption in which the alerting user may provide text indicating what the alerting user believes to be of current interest in the web content.



FIG. 2A is a series of three screen shots showing three different states of an alert submission display 200 used in one embodiment. One view is comprised of blank alert submission display 202. Blank alert submission display 202 includes a submission button 204 used to submit an alert with respect to the URL of the web content currently being accessed by the alerting user. Blank alert submission display 202 also includes an interest category selection area 206. In one embodiment, as illustrated in FIG. 2A, the interest category selection area 206 is configured as a pull down menu activated by selecting the downward arrow on the right side of interest category selection area 206. Blank alert submission display 202 also includes a caption area 208 in which an alerting user may enter text associated with the alert, such as text indicating why the alerting user believes the URL to be of current interest. As shown in interest category selection display 212, when the downward arrow button on the right side of interest category selection area 206 is selected, a pull down menu 214 is presented, and an alerting user may select one of the interest categories listed in the pull down menu 214 in the manner well known in the art. As shown in the completed alert submission display 222 of FIG. 2A, the interest category selected by the alerting user is shown in the interest category selection area 206. In the example shown in FIG. 2A, the category selected is “NATURE”. In addition, the caption entered by the alerting user, the comment “rhino!” in the example shown in FIG. 2A, appears in the caption area 208 of the alert submission display. As noted above, the alerting party posts the alert to the application server via the Internet and the web server by selecting the submission button 204.



FIG. 2B is an illustration of the data structure used in one embodiment for alerts submitted by an alerting user. The alert includes an ALERTER_ID field 240 in which data identifying the alerting user is provided. The alert also includes a URL field 242 in which the URL of the web content or other electronic resource being accessed by the alerting user when the alert was sent is stored. The alert also includes an INTEREST SELECTION field 244 in which the interest category selected by the alerting user, if any, is provided. Finally, the alert includes a CAPTION field 246 in which the caption entered by the alerting user, if any, is provided.



FIG. 3 is a flow chart illustrating a process used in one embodiment to alert users of items of current interest. The process begins in step 302 in which an alert indicating that an item is of current interest is received. Next, in step 304, the alert is processed. Finally, in step 306, the alert is disseminated to one or more participants, as described more fully below.



FIG. 4 is a flow chart illustrating a process used in one embodiment to receive an alert, as in step 302 of FIG. 3. The process begins with step 402 in which a transmission comprising an alert is received from an alerting user. As noted above, in one embodiment an alert includes at least the URL of the web content being accessed by the alerting user at the time the alert was sent. In one embodiment, as described above, the alert also includes data indicating the identity of the alerting user. In addition, as noted above, the alert may include, at the option of the alerting user, an interest selection and/or a caption for the alert. The process shown in FIG. 4 continues with step 404 in which a new alert software object is created at the application server, such as application server 106 of FIG. 1. Next, in step 406, the data provided in the alert is stored in the alert object. In step 408, a time stamp indicating the time when the alert was received is stored in the alert object. Finally, in step 410, an ALERT_ID, which uniquely identifies the alert and distinguishes the alert and its associated object from other alerts and their associated objects, is obtained and stored in the alert object.



FIG. 5 is an illustration of the data structure used in one embodiment for the alert object. Data field 502 is used to store the ALERT_ID described above. Data field 504 is used to store the time stamp described above. Data fields 506-512 are used to store the ALERTER_ID, URL, INTEREST SELECTION, and CAPTION described above, respectively. ALERT INTENSITY field 514 is used to store a number indicating the intensity or weight to be afforded to the incoming alert. The ALERT INTENSITY is determined as described below. The alert object also stores properties retrieved from various database tables, described more fully below. For example, the alert object includes a LAST_TIME field 516 used to store data retrieved from the database indicating the time of the most recent prior alert. The alert object also includes a LAST_RANK field 518 used to store a numerical ranking retrieved from the database that indicates the overall level or degree of current interest of an item as indicated by all of the alerts that have been submitted with respect to a URL during the current period of activity with respect to the URL through the most recent prior alert. The alert object also includes a LAST_WEIGHT field 520 used to store data retrieved from a database table, as described below, that represents the number of prior alerts received for the URL in the interest category indicated by the current alert, as described more fully below. The alert object also includes a LAST_INTENSITY_SUM field 522 in which the sum of the intensities of all prior alerts for the URL during the current period of activity with respect to the URL, which sum is retrieved from a database table described more fully below, is stored. Finally, the alert object includes a LAST_NORMAL_TIME field 524 used to store the time, retrieved from a database table as described more fully below, when the last normalization calculation was performed.



FIG. 6 is a flowchart illustrating a process used in one embodiment to process an alert, as in step 304 of FIG. 3. The process begins with step 602 in which the intensity of the alert is determined. The term intensity as used herein refers to the weight or value to be assigned to a particular alert regarding an item. In one embodiment, the intensity is a value between 0 and 1. In one embodiment, the value assigned for the intensity is higher if the alerting user selects an interest category for the alert than it would have been if the same alerting party had not selected an interest category. In one embodiment, the intensity value is higher if the alerting party provides a caption for the alert than it would have been if the alerting party had not provided a caption. In one embodiment, the intensity of an alert is increased if it is determined that the alerting party is a party that has provided particularly relevant or helpful alerts in the past, or is trusted for some other reason, such as expertise, academic credentials, or reputation within a particular community of interest. In one embodiment, the intensity of an alert is decreased if it is determined that the alerting party has provided unhelpful or erroneous alerts in the past, or if it is determined that the alerting party cannot be trusted as much as other alerting parties for other reasons, such as reputation in the relevant community. In one embodiment, it is possible to provide both an active alert by selecting an alert button and to provide a passive alert by merely accessing a URL with respect to which an alerting party previously submitted an active alert. In one embodiment, an active alert is assigned a higher intensity value than a passive alert.


For example, a passive alert may be arbitrarily assigned a baseline intensity value of 0.3 and an active alert a baseline intensity value of 0.5. For an active alert, 0.1 could be added for each of the following conditions that is satisfied by the alert: an interest category selection was included in the alert; a caption was included in the alert; and/or the source of the alert is particularly trusted. Conversely, 0.1 could be subtracted from the intensity of an alert from a source known to be unreliable. Alternatively, alerts from sources known to be unreliable may be blocked and not assigned any intensity value.


The process illustrated in FIG. 6 continues with step 604 in which data values for the alert object data fields described above that are not included in the alert transmission received from the alerting party are retrieved from the database.


Next, in step 606, the intensity sum for the URL, which is the sum of the intensity values for all of the alerts with respect to the URL, is updated. Next, in step 608, the intensity rank for the URL is updated to reflect the new alert. In step 610, the interest weight value, which represents the number of alerts for a particular URL in which a particular category of interest was indicated, is updated. Finally, in step 612, the updated data values are stored to the database.



FIG. 7 is an illustration of six database tables 700 used in one embodiment to store data concerning alerts received with respect to items of current interest associated with URLs. The database tables 700 include an INTEREST_ID table 702 used to provide a unique identifier, labeled INTEREST_ID in FIG. 7, for each interest category, denominated INTEREST_CAT in FIG. 7. Database tables 700 also include a URL_ID table 704 used to provide a unique identifier, labeled URL_ID in FIG. 7, for each URL.


Database tables 700 also include an INTERESTS table 706 used to store the interest weight, denominated WEIGHT in FIG. 7, for each interest category with respect to which an alert has been submitted for a URL. As noted above, in one embodiment, the weight is the total number of alerts received within a given interest category for a URL. For example, if five alerts indicating the interest category People and three alerts indicating the interest category Nature have been submitted for a URL, there will be two entries for the URL in the interest table, one for each interest category. The weight in the entry for the category People would be “5” and the weight for the URL in the category Nature would be “3”.


The database tables 700 also include a RANK table 708 used to store a rank value for each URL associated with an item of current interest, a time stamp when the rank was last calculated, and a data entity denominated NUM_ALERT in FIG. 7, which represents the total number of alerts submitted for the URL.


The database tables 700 also include a COMMENTS table 710 used to store any comment submitted with an alert and to associate each comment with the corresponding URL. Finally, the database tables 700 include a NORMALIZE table 712 used to store the sum of the intensities of the alerts submitted for a URL (INTENSITY_SUM) and a time stamp indicating when the last normalization was performed.



FIG. 8A is a flowchart illustrating a process used in one embodiment to update the intensity sum for a URL, as in step 606 of FIG. 6. The process begins with step 802 in which the current intensity sum is retrieved from the database, as in step 604 of FIG. 6. If there is no existing record for the URL in the NORMALIZE table (i.e., the alert being processed is the first alert for the URL), a URL_ID is assigned for the URL, a record for the URL is created in the NORMALIZE table, and the retrieved current intensity sum is set to zero. Next, in step 804, the intensity sum is incremented by the amount of the intensity of the current alert. For example, if the previous intensity sum was 4.7 and the intensity for the current alert was 0.5, the intensity sum would be incremented to the value of 4.7+0.5=5.2. Finally, in step 806, the intensity sum time stamp stored in NORMALIZE table 712 shown in FIG. 7 (which is the same as the LAST_NORMAL_TIME stored in field 524 of FIG. 5) is updated to the time stamp of the current alert. In one embodiment, the intensity sum is updated, and a normalization is performed as described more fully below, each time a new alert is received for a URL. In such an embodiment, the time stamp stored in the NORMALIZE table 712 of FIG. 7 will be the same as the time stamp stored in the RANK table 708 of FIG. 7, as both the rank and the intensity sum are updated each time an alert is received.



FIG. 8B is a flowchart illustrating a process used in one embodiment to update the intensity rank for a URL to reflect the intensity of the current alert. The process begins with step 822 in which the current intensity rank is retrieved from the database, as in step 604 of FIG. 6. As shown in FIG. 7, in one embodiment, this value is retrieved from the RANK table 708. If there is no entry in the RANK table for the URL, i.e., the alert being processed is the first alert for the URL, a record in the RANK table is created for the URL (identified by the URL_ID assigned to the URL) and the current intensity rank is set to zero. Next, in step 824, the intensity rank is updated to reflect the intensity of the current alert. In one embodiment, if the current alert has been received within a predetermined time interval τ after the last alert for the URL, the updated intensity rank is a function of the last rank and the intensity of the current alert in accordance with the following formula:

r′=(k−r)*Ialert+r

Where k is the maximum intensity value, which as noted above is one in one embodiment, r is the last rank, r′ is the updated rank, and Ialert is the intensity value for the current alert. Restating the formula to reflect the fact that in one embodiment, the maximum intensity level k=1, the formula becomes:

r′=(1−r)*Ialert+r


If an alert is the first alert received for a URL, the last rank is considered to be zero (r=0) and the above formula results in the new rank being equal to the intensity value for the current alert. For example, if the intensity value for the current alert is 0.5, the updated heat rank r′=(1−0)*0.5+0=0.5. If a subsequent alert of intensity 0.6 is received, the formula results in the updated intensity rank being calculated as follows:

r′=(1−0.5)*0.6+0.5=0.8


As the example illustrates, so long as additional alerts are received within the time interval each incoming alert will cause the intensity rank for the URL to increase until the intensity rank approaches the maximum intensity value k (in the example, the rank would approach k=1). The speed with which the intensity rank for a particular URL approaches the maximum value k depends on the intensity value of the incoming alerts and the frequency with which alerts are received.


In one embodiment, if the predetermined time interval r referred to above has expired between the last alert and the current alert, the updated intensity rank is calculated by a modified formula which reduces the updated intensity rank in accordance with an exponential decay function that effectively adjusts the updated intensity rank downward to account for the passage of time between the last alert and the current alert. All other things being equal, this adjustment would result in a site that received alerts more frequently to have a higher rank than a site that received alerts separated by more than the predetermined time interval. To determine the updated intensity rank as adjusted for the passage of time, the following formula is used in one embodiment:

r′=[(k−4)*Ialert+r]*e−a(Δt-τ)


In this formula, k, r, and Ialert are the same as above, a is the weight assigned to the decay function (a higher value for a will result in a greater amount of decay per unit time), Δt is the amount of time that has elapsed between the current alert and the previous alert, and τ is the predetermined time interval referred to above.


In one embodiment, the updated intensity rank is normalized by multiplying the updated intensity rank by two factors. The first factor is a low frequency enhancement factor designed to enhance the intensity rank of URLs with respect to which alerts are received relatively less frequently relative to the intensity rank of URLs regarding which alerts are received more frequently. The purpose of this enhancement factor is to ensure that sites that are of current interest only from time to time are not masked by the intensity ranking calculated for sites that are of current interest more frequently. In one embodiment, the low frequency enhancement factor is the time of the current alert minus the time of the last update to the intensity rank.


The second factor by which the updated intensity rank is multiplied is a low volume enhancement factor The purpose of this factor is to ensure that the intensity rank of URLs that are of current interest only to a smaller community of users will not be overshadowed by the intensity rank of URLs that are of current interest to a large community. In one embodiment, the low volume enhancement factor is the inverse of the intensity sum for the URL. Accordingly, in one embodiment, the normalized intensity rank is determined by the following formula:

r″=r′*(tcurrent−tfirst)*1/n

Where


r″=normalized intensity rank


r′=updated intensity rank before normalization


tcurrent=timestamp of current alert


tfirst=timestamp of first alert for URL


n=intensity sum=sum of all alert intensities for URL.


Once the intensity rank has been updated and normalized, the process shown in FIG. 8B continues with step 826 in which the time stamps for the normalization and intensity rank tables are updated to the time stamp of the current alert.



FIG. 8C is a flowchart illustrating a process used in one embodiment to update 20 the interest category weight for a URL with respect to the interest category indicated in an alert. The process begins with step 842 in which the database is queried to determine if a record exists for the URL for the interest category indicated in the alert. In step 844, it is determined whether the query performed in step 842 identified an existing database table entry for the URL for the interest category indicated in the alert (i.e., whether a prior alert indicated the same interest category for the URL). If it is determined in step 844 that a database entry does not exist for the interest category with respect to the URL, the process proceeds to step 846 in which a record in the INTEREST table is created for the URL with respect to the interest category of the alert. The process then proceeds to step 850 in which the weight value is incremented for the URL with respect to the interest category by increasing the value from zero to one for the new record.


If it is determined in step 844 that there is an existing record for the interest category for the alert with respect to the alert URL, the process proceeds to step 848 in which the weight value stored in the record is retrieved. The process then continues to step 850 in which the retrieved weight is incremented by one to reflect the current alert. For example, if the retrieve weight were 7, the weight would be incremented to 8 in step 850 to reflect the current alert.



FIG. 9 is a flowchart illustrating a process used in one embodiment to purge records for URLs that are determined to be no longer of current interest by calculating a time decayed intensity rank at intervals, even if no new alert has been received, and purging from the database the records for a URL if the time decayed intensity rank is below a prescribed threshold. The process shown in FIG. 9 begins with step 902 in which the intensity rank for a URL is retrieved. In one embodiment, the intensity rank is retrieved and the process shown in FIG. 9 is performed, at a predetermined arbitrary time interval τ.


The process shown in FIG. 9 continues with step 904 in which an intensity rank adjusted for time decay is calculated for the URL. In one embodiment, the time decayed 5 intensity rank is determined by the following formula:

rt=e−a(Δt-τ)*r

Where


rt=time decayed intensity rank


a=weight of decay function


Δt=time elapsed since last alert


τ=predetermined time interval referred to above


r=stored intensity rank.


As can be seen from the above formula, the time decayed intensity rank decays exponentially over time if no new alerts are received. If it is determined in step 906 of the process shown in FIG. 9 that the time decayed intensity rank is below the intensity rank threshold, the process proceeds to step 908 in which the record for the URL is deleted. If it is determined in step 906 that the time decayed intensity rank is not below the intensity rank threshold, the process proceeds to step 910 in which the intensity rank as stored in the database is left unchanged.



FIG. 10 is a flowchart illustrating a process used in one embodiment to disseminate an alert to a participant, as in step 306 of FIG. 3. The process begins with step 1002 in which a request containing interest category filter selections made by the participant is received. Next, in step 1004, a hot list software object is created at the application server, as shown in FIG. 1 and described above. Then, in step 1006, an array of interest categories, such as the interest category array 114 described above with respect to FIG. 1, is created within the hot list object. Next, in step 1008, a list of hot URLs responsive to the request is built. Finally, in step 1010, the list of hot URLs responsive to the request is sent to the participant.



FIG. 11 shows an exemplary participant display 1100 used in one embodiment to disseminate alert information to a participant. The display 1100 includes a URL entry and display area 1102. The URL for the web content or other electronic resource currently being accessed by the participant is displayed in the URL entry and display area 1102, and the participant may enter the URL for the web content or other electronic resource the participant wishes to access manually in the URL entry and display area 1102, as in the URL or address field for a World Wide Web browser. The display 1100 also includes a content display area 1104 in which the web or other content for the URL listed in URL entry and display area 1102 is displayed. For example, if the URL is the URL of web content accessed via the Internet, the web content associated with the URL will be displayed in URL display area 1104.


The display 1100 also includes an interest category filter selection area 1106 in which interest categories are listed along with a check box for each category listed. The participant selects the check box for each interest category for which the participant would like URLs of current interest to be included in the participant's hot list.


In one embodiment, filter selection area 1106 includes for each category a sensitivity entry area (not shown in FIG. 11) to be used to provide an indication of the participant's degree or level of interest. For example, in one embodiment a participant may enter a whole number from 1 to 5, with 1 indicating the lowest level of sensitivity (e.g., the participant does not want to receive a notification regarding a URL in the category unless a significant number of alerts have been received regarding the URL, or only when the intensity rank for the URL exceeds a predetermined, relatively high threshold) and 5 representing the highest level of sensitivity (e.g., the participant wants to receive a notification even if there has only been one or relatively few alerts concerning a URL, or if one or more alerts have been received but the intensity rank for the URL is relatively low).


In one embodiment, a request is sent to the application server automatically at predetermined intervals. The request contains the interest categories that are in the selected state at the time the request is sent. In one embodiment, the display 1100 includes a submit button (not shown in FIG. 11) that, when selected, causes a request containing the interest categories selected by the participant at the time to be posted to the application server via the Internet.


The display 1100 also includes a hot list display area 1108 in which the hot list of URLs returned by the system to the participant in response to a request is presented. As shown in FIG. 11, in one embodiment, each URL is represented by a hypertext link that, when selected, causes the URL of the listed cite to appear in the URL entry and display area 1102 and the content associated with the URL to be displayed in the URL display area 1104.


In one embodiment, the display 1100 is modified to include an alert submission display area such as the alert submission display shown in FIG. 2A. This would permit a participant to send an active alert to the application server if the participant encounters a URL of current interest.



FIG. 12 is a flowchart illustrating a process used in one embodiment to build a list of hot URLs responsive to a request, as in step 1008 of FIG. 10. The process begins with step 1202 in which all URLs of current interest within the categories indicated in the request are found.


Next, in step 1204, a “hot token” object is created in a hot token array within the hot list object for each URL found in step 1202, as described above with respect to hot token array 116 shown in FIG. 1. Each hot token object holds the URL_ID, the WEIGHT for the URL with respect to the interest category indicated in the request, the sum of the WEIGHT values for each category associated with the URL in the database, and the intensity rank (RANK) for the URL.


Next, in step 1206, a list rank is determined for each URL retrieved in response to the request. In one embodiment, a list rank value is calculated for each URL and is used to determine the list rank (or the order in which the responsive URLs will be placed to determine which URLs will be provided). In one embodiment, an initial list rank value is calculated for each URL based on the interest category weight(s) for the URL with respect to the interest category or categories in the request, along with the interest weight for any interest category or categories that are associated with the URL in the database but which are not among the categories indicated in the request. In one embodiment, the initial list rank value “v” of a URL number “n” (vn) is calculated according to the following formula:







v
n

=





f
k







f
m









Where


vn=initial list rank value of URL “n”


fk=interest weight for URL for each request category


fm=interest weight for each category associated with URL in database


For example, if at the time of the request there had been ten alerts submitted for a particular URL and three of the alerts were associated with a first category, two with a second category and five with a third category, and if a request were received that included among the request categories the first and third categories, the initial list rank value “v” for URL number “n” calculated in accordance with the above formula would be as follows:







v
n

=




3

+

5




3

+

2

+

5




0.74





It should be noted that the use of the square root of the weight for each category tends to give relatively greater effect to the weight of interest categories associated with the URL by a minority of alerting users because using the square root reduces the net effect of the greater weight value associated with interest categories indicated by the majority of alerting users. As with the normalization of the intensity rank described above, this has the effect of giving more visibility to matters of interest to a relatively smaller community.


In an embodiment in which the participant indicates a level of sensitivity with respect to each selected interest category, as described above, the formula for the initial list rank value is modified to take into consideration the sensitivity “s” indicated for each category of interest. In one embodiment, the initial list rank value formula is modified as follows:







v
n

=






s
k

*

f
k









s
m

*

f
m










Where

    • vn=initial list rank value of URL “n”
    • fk=interest weight for URL for each request category
    • fm=interest weight for each category associated with URL in database
    • sk=sensitivity indicated for request category “k”
    • sm=sensitivity indicated for request category corresponding to interest category “m”, if any (sm=1 for interest categories not in request).


For example, in the example described above, assume the participant indicated a sensitivity level of 1 with respect to the first category and 5 with respect to the third category, the initial list rank value would be calculated as follows:







v
n

=





1
*
3


+


5
*
5






1
*
3


+


1
*
2


+


5
*
5





0.83





(As noted above, the sensitivity level sn, used for the second category, having weight “2” in the denominator, is set at “1” because in the example the participant did not select that category.)


The initial list rank value determined by this calculation (0.83) is greater than the initial list rank value found in the above calculation of an initial list rank value in an embodiment in which sensitivity levels are not assigned or considered (0.74). This illustrates the effect of assigning sensitivity levels. The initial list rank value determined in the second calculation, which takes into account a sensitivity level for each category, is higher than it would have been found to be without regard to sensitivity because the participant indicated a higher sensitivity for one of the categories with respect to which alerts had been received for the URL.


In this way, high-sensitivity users are more likely to become aware of and access a URL with respect to which one or more alerts have been received in a category for which the user has indicated a high sensitivity. If such a high-sensitivity user chose to send alerts of their own with respect to the URL, such activity would increase the intensity rank for the URL (as described above) and would tend to propagate the original alert or alerts to lower-sensitivity users (because the intensity rank is factored into the final list rank used to identify the final list of URLs to be provided to a participant, as described below). If such lower-sensitivity users were to send even more alerts, the original alerts would be further propagated to even lower-sensitivity users, and so on.


In one embodiment, the initial list rank value determined by the interest category weights, as described above, is used along with the intensity rank for the URL to calculate a final list rank value for the URL. In one embodiment, the final list rank value for URL number “n” is calculated in accordance with the following formula:

vn′=rn(α+(1−α)vn)

Where


vn′=final list rank value


rn=intensity rank for URL


α=weight factor (0≦α≦1)


vn=initial list rank value


In the above equation, the weight factor α determines the relative weight afforded to the intensity rank for the URL and the initial list rank calculated based on the interest category weights as described above. If the value for α is selected to be 1, the final list rank would be equal to the intensity rank for the URL and the initial list rank would not factor into the final list rank at all. Therefore, a higher weight factor will tend to increase the influence of the intensity rank for the URL and decrease the effect of the initial list rank. Stated another way, a low weight factor tends to give more effect to the extent to which the interest categories associated with the URL in the database match the interest categories indicated in the request from the participant. Conversely, a higher weight factor tends to give greater effect to the overall popularity of the URL as measured by the intensity rank.


Once the list rank for each retrieved URL has been calculated in step 1206, in step 1208 the retrieved URLs are sorted by list rank. Then, in step 1210, the top ten URLs by list rank are selected as the hot list of URLs to be sent to the participant in response to the request. The number ten is an arbitrary number and either a fewer number or greater number of URLs may be included.


Although the foregoing invention has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications may be practiced within the scope of the appended claims. It should be noted that there are many alternative ways of implementing both the process and apparatus of the present invention. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.

Claims
  • 1. A computer-implemented method of informing a participant user that an item accessible by the participant user via a network is of current interest, comprising: receiving via the network indications from source users other than the participant user that the item is of current interest;processing the indications using one or more computer processors including: determining an intensity value for a first indication of the indications from a first user of the source users;blocking a second indication of the indications from a second user of the source users from being used in ranking the item based on data regarding identity of the second source user; andcalculating a rank for the item based on the intensity value of the first indication and not based on an intensity value for the second indication, the rank indicating a level of current interest of the item relative to other items accessible by the participant user via the network; andinforming the participant user via the network that the item is of current interest based on the rank for the item.
  • 2. The method of claim 1, wherein the first and second indications are received in real time from the first and second users while the first and second users are viewing the item.
  • 3. The method of claim 1, wherein the data regarding identity of the second source user comprises data regarding unwanted indications from the second source user.
  • 4. The method of claim 3, wherein the data regarding unwanted indications from the second source user comprises data regarding unreliableness of the second source user.
  • 5. The method of claim 1, further comprising: determining a time decay for the first indication;wherein the rank is calculated based on the time decay for the first indication.
  • 6. The method of claim 5, wherein the time decay is determined based on elapsed time from when the first indication was received.
  • 7. The method of claim 6, wherein the time decay causes the rank to decrease as the elapsed time from when the first indication was received increases.
  • 8. The method of claim 7, wherein the rank decreases exponentially as the elapsed time from when the first indication was received increases.
  • 9. The method of claim 5, wherein the rank is calculated based on the time decay after expiration of a predetermined time interval of elapsed time from when the first indication was received.
  • 10. The method of claim 1, wherein the determining the intensity value for the first indication includes determining the intensity value based on data regarding a type of activity taken by the first user in relation to the item.
  • 11. A computer-implemented method of informing a participant user that an item accessible by the participant user via a network is of current interest, comprising: receiving via the network an indication from a source user other than the participant user that the item is of current interest;processing the indication using one or more computer processors including: determining an intensity value for the indication based on data provided with the indication regarding a type of activity taken by the source user in relation to the item;determining a time decay for the indication; andcalculating a rank for the item based on the intensity value and the time decay for the indication, the rank indicating a level of current interest of the item relative to other items accessible by the participant user via the network; andinforming the participant user via the network that the item is of current interest based on the rank for the item.
  • 12. The method of claim 11, wherein the indication is received in real time from the source user while the source user is viewing the item.
  • 13. The method of claim 11, wherein the time decay is determined based on elapsed time from when the indication was received.
  • 14. The method of claim 13, wherein the time decay causes the rank to decrease as the elapsed time from when the indication was received increases.
  • 15. The method of claim 14, wherein the rank decreases exponentially as the elapsed time from when the indication was received increases.
  • 16. The method of claim 11, wherein the rank is calculated based on the time decay after expiration of a predetermined time interval of elapsed time from when the indication was received.
  • 17. The method of claim 16, wherein the time decay is determined based on subtracting the predetermined time interval from the elapsed time from when the indication was received.
  • 18. The method of claim 11, wherein data regarding the type of activity taken by the source user is received with the indication.
  • 19. The method of claim 18, wherein a higher intensity value is determined for the indication in response to the activity taken by the source user being of an active type and a lower intensity value is determined for the indication in response to the activity taken by the source user being of a passive type.
  • 20. The method of claim 11, wherein: the determining the intensity value for the indication includes adjusting the intensity value based on data regarding identity of the source user; andwherein the rank is calculated based on the adjusted intensity value.
  • 21. A computer-implemented method of informing a participant user that a particular item accessible by the participant user via a network is of current interest, comprising: receiving via the network a particular indication from a source user other than the participant user that the particular item is of current interest;processing the indication using one or more computer processors including: determining an intensity value for the particular indication including adjusting the intensity value based on data regarding additional indications previously received from the source user for additional items accessed by the source user via the network;determining a time decay for the particular indication; andcalculating a rank for the particular item based on the intensity value and the time decay for the particular indication, the rank indicating a level of current interest of the particular item relative to other items accessible by the participant user via the network; andinforming the participant user via the network that the particular item is of current interest based on the rank for the particular item.
  • 22. The method of claim 21, wherein the particular and additional indications are received in real time from the source user while the source user is viewing the particular and additional items.
  • 23. The method of claim 21, wherein the data regarding the additional indications previously received from the source user comprises data regarding quality of the additional indications previously received from the source user.
  • 24. The method of claim 23, wherein the quality of the additional indications previously received from the source user comprises relevance of the additional indications previously received from the source user.
  • 25. The method of claim 23, wherein the quality of the additional indications previously received from the source user comprises helpfulness of the additional indications previously received from the source user.
  • 26. The method of claim 21, wherein the time decay is determined based on elapsed time from when the particular indication was received.
  • 27. The method of claim 26, wherein the time decay causes the rank to decrease as the elapsed time from when the particular indication was received increases.
  • 28. The method of claim 27, wherein the rank decreases exponentially as the elapsed time from when the particular indication was received increases.
  • 29. The method of claim 28, wherein the rank is calculated based on the time decay after expiration of a predetermined time interval of elapsed time from when the particular indication was received.
  • 30. The method of claim 21, wherein the determining the intensity value for the particular indication includes determining the intensity value based on data regarding a type of activity taken by the first user in relation to the item.
  • 31. The method of claim 21, wherein adjusting the intensity value reflects the participant user and the source user being in a relatively smaller community of users within a larger community of users.
  • 32. The method of claim 21, wherein the data regarding the additional indications previously received from the source user comprises data indicating the source user is associated with a relatively smaller community of users within a larger community of users.
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 12/411,500 filed Mar. 26, 2009, entitled “Alerting Users to Items of Current Interest,” which is a continuation of U.S. patent application Ser. No. 10/800,393 filed Mar. 12, 2004, entitled “Alerting Users to Items of Current Interest,” which is a continuation of U.S. patent application Ser. No. 09/656,638 filed Sep. 7, 2000, entitled “Alerting Users to Items of Current Interest” (now U.S. Pat. No. 6,757,682), which claims priority under 35 U.S.C. §119(e) from U.S. Provisional Patent Application No. 60/178,627 filed Jan. 28, 2000, entitled “Alerting Users to Web Sites of Current Interest and Handling Large Increases in User Traffic,” all of which are incorporated herein by reference in their entireties. This application is related to U.S. patent application Ser. No. 09/656,518 filed Sep. 7, 2000, entitled “Quantifying the Level of Interest of an Item of Current Interest” (now U.S. Pat. No. 6,556,989), which is incorporated herein by reference in its entirety; and U.S. patent application Ser. No. 09/658,346 filed Sep. 7, 2000, entitled “Normalizing a Measure of the Level of Current Interest of an Item Accessible Via a Network” (now U.S. Pat. No. 6,721,744), which is incorporated herein by reference in its entirety.

US Referenced Citations (365)
Number Name Date Kind
3868675 Firmin Feb 1975 A
3884403 Brewer May 1975 A
3919475 Dukich et al. Nov 1975 A
4033335 Nickles Jul 1977 A
4051534 Dukich et al. Sep 1977 A
4131919 Lloyd et al. Dec 1978 A
4217609 Hatori et al. Aug 1980 A
4260229 Bloomstein Apr 1981 A
4283735 Jagger Aug 1981 A
4319286 Hanpachern Mar 1982 A
4390904 Johnston et al. Jun 1983 A
4446997 Himberg May 1984 A
4520404 Von Kohorn May 1985 A
4524381 Konishi Jun 1985 A
4526308 Dovey Jul 1985 A
4527201 Cappels Jul 1985 A
4536066 Bauer Aug 1985 A
4554418 Toy Nov 1985 A
4574354 Mihalik et al. Mar 1986 A
4602297 Reese Jul 1986 A
4605964 Chard Aug 1986 A
4605973 Von Kohorn Aug 1986 A
4618895 Wright Oct 1986 A
4714184 Young et al. Dec 1987 A
4739398 Thomas et al. Apr 1988 A
4750052 Poppy et al. Jun 1988 A
4750053 Allen Jun 1988 A
4777537 Ueno et al. Oct 1988 A
4782401 Faerber et al. Nov 1988 A
4814876 Horio et al. Mar 1989 A
4827532 Bloomstein May 1989 A
4837817 Maemori Jun 1989 A
4841575 Welsh et al. Jun 1989 A
4843484 Kanamaru et al. Jun 1989 A
4845658 Gifford Jul 1989 A
4847543 Fellinger Jul 1989 A
4913539 Lewis Apr 1990 A
4930160 Vogel May 1990 A
4934821 Morton Jun 1990 A
4965825 Harvey et al. Oct 1990 A
4989104 Schulein et al. Jan 1991 A
5012334 Etra Apr 1991 A
5012335 Cohodar Apr 1991 A
5025394 Parke Jun 1991 A
5038217 Hayashi et al. Aug 1991 A
5040081 McCutchen Aug 1991 A
5105184 Pirani et al. Apr 1992 A
5109482 Bohrman Apr 1992 A
5136655 Bronson Aug 1992 A
5146353 Isoguchi et al. Sep 1992 A
5157742 Niihara Oct 1992 A
5172281 Ardis et al. Dec 1992 A
5175769 Hejna, Jr. et al. Dec 1992 A
5177796 Feig et al. Jan 1993 A
5179449 Doi Jan 1993 A
5182641 Diner et al. Jan 1993 A
5185667 Zimmermann Feb 1993 A
5187571 Braun et al. Feb 1993 A
5226093 Iwase Jul 1993 A
5233654 Harvey et al. Aug 1993 A
5239428 Nishida et al. Aug 1993 A
5241428 Goldwasser et al. Aug 1993 A
5249289 Thamm et al. Sep 1993 A
5253061 Takahama et al. Oct 1993 A
5253066 Vogel Oct 1993 A
5253275 Yurt et al. Oct 1993 A
5262856 Lippman et al. Nov 1993 A
5265180 Golin Nov 1993 A
5267042 Tsuchiya et al. Nov 1993 A
5295064 Malec et al. Mar 1994 A
5295089 Ambasz Mar 1994 A
5299019 Pack et al. Mar 1994 A
5305195 Murphy Apr 1994 A
5305400 Butera Apr 1994 A
5309234 Kranawetter et al. May 1994 A
5317730 Moore et al. May 1994 A
5329320 Yifrach Jul 1994 A
5333091 Iggulden et al. Jul 1994 A
5343251 Nafeh Aug 1994 A
5347632 Filepp et al. Sep 1994 A
5377051 Lane et al. Dec 1994 A
5384703 Withgott et al. Jan 1995 A
5396287 Cho Mar 1995 A
5396583 Chen et al. Mar 1995 A
5404316 Klingler et al. Apr 1995 A
5406626 Ryan Apr 1995 A
5416310 Little May 1995 A
5420801 Dockter et al. May 1995 A
5421031 De Bey May 1995 A
5428774 Takahashi et al. Jun 1995 A
5428778 Brookes Jun 1995 A
5436542 Petelin et al. Jul 1995 A
5436637 Gayraud et al. Jul 1995 A
5436653 Ellis et al. Jul 1995 A
5438357 McNelley Aug 1995 A
5438362 Tabuchi Aug 1995 A
5438423 Lynch et al. Aug 1995 A
5444478 Lelong et al. Aug 1995 A
5467271 Abel et al. Nov 1995 A
5467288 Fasciano et al. Nov 1995 A
5473379 Horne Dec 1995 A
5477331 Yamaguchi et al. Dec 1995 A
5485611 Astle Jan 1996 A
5486852 Arai Jan 1996 A
5488409 Yuen et al. Jan 1996 A
5498002 Gechter Mar 1996 A
5510830 Ohia et al. Apr 1996 A
5514861 Swartz et al. May 1996 A
5524051 Ryan Jun 1996 A
5524193 Covington et al. Jun 1996 A
5528282 Voeten et al. Jun 1996 A
5535382 Ogawa Jul 1996 A
5537151 Orr et al. Jul 1996 A
5537530 Edgar et al. Jul 1996 A
5539483 Nalwa Jul 1996 A
5546145 Bernardi et al. Aug 1996 A
5550754 McNelley et al. Aug 1996 A
5553221 Reimer et al. Sep 1996 A
5555463 Staron Sep 1996 A
5572261 Cooper Nov 1996 A
5572643 Judson Nov 1996 A
5575443 Honeycutt Nov 1996 A
5583560 Florin et al. Dec 1996 A
5583565 Cortjens et al. Dec 1996 A
5583571 Friedland Dec 1996 A
5583652 Ware Dec 1996 A
5589892 Knee et al. Dec 1996 A
5590195 Ryan Dec 1996 A
5592626 Papadimitriou et al. Jan 1997 A
5594498 Fraley Jan 1997 A
5598352 Rosenau et al. Jan 1997 A
5604551 Choi et al. Feb 1997 A
5606359 Youden et al. Feb 1997 A
5608839 Chen Mar 1997 A
5612742 Krause et al. Mar 1997 A
5613032 Cruz et al. Mar 1997 A
5613909 Stelovsky Mar 1997 A
5614940 Cobbley et al. Mar 1997 A
5617526 Oran et al. Apr 1997 A
5623173 Fasullo et al. Apr 1997 A
5635982 Zhang et al. Jun 1997 A
5664227 Mauldin et al. Sep 1997 A
5666159 Parulski et al. Sep 1997 A
5671014 Ito et al. Sep 1997 A
5678793 Hill Oct 1997 A
5682597 Ganek et al. Oct 1997 A
5684514 Branscomb Nov 1997 A
5687095 Haskell et al. Nov 1997 A
5689300 Shibata et al. Nov 1997 A
5689648 Diaz et al. Nov 1997 A
5692213 Goldberg et al. Nov 1997 A
5692661 Kellerman Dec 1997 A
5694474 Ngo et al. Dec 1997 A
5701582 DeBey Dec 1997 A
5703655 Corey et al. Dec 1997 A
5703795 Mankovitz Dec 1997 A
5713021 Kondo et al. Jan 1998 A
5717814 Abecassis Feb 1998 A
5717869 Moran et al. Feb 1998 A
5721823 Chen et al. Feb 1998 A
5724424 Gifford Mar 1998 A
5724567 Rose et al. Mar 1998 A
5724646 Ganek et al. Mar 1998 A
5726660 Purdy et al. Mar 1998 A
5726717 Peters et al. Mar 1998 A
5729108 Steele Mar 1998 A
5729741 Liaguno et al. Mar 1998 A
5737009 Payton Apr 1998 A
5740037 McCann et al. Apr 1998 A
5740549 Reilly et al. Apr 1998 A
5742339 Wakui Apr 1998 A
5742517 Van Den Bosch Apr 1998 A
5748190 Kjorsvik May 1998 A
5749010 McCumber May 1998 A
5749081 Whiteis May 1998 A
5751336 Aggarwal et al. May 1998 A
5751806 Ryan May 1998 A
5752113 Borden May 1998 A
5754230 Tsuruta May 1998 A
5754938 Herz et al. May 1998 A
5754939 Herz et al. May 1998 A
5758181 Becker May 1998 A
5765133 Antoniol et al. Jun 1998 A
5768528 Stumm Jun 1998 A
5768640 Takahashi et al. Jun 1998 A
5768648 Skipp et al. Jun 1998 A
5774498 Oya et al. Jun 1998 A
5774664 Hidary et al. Jun 1998 A
5778181 Hidary et al. Jul 1998 A
5781894 Petrecca et al. Jul 1998 A
5791907 Ramshaw et al. Aug 1998 A
5793367 Taguchi Aug 1998 A
5793971 Fujita et al. Aug 1998 A
5796426 Gullichsen et al. Aug 1998 A
5796945 Tarabella Aug 1998 A
5796967 Filepp et al. Aug 1998 A
5805156 Richmond et al. Sep 1998 A
5809161 Auty et al. Sep 1998 A
5818439 Nagasaka et al. Oct 1998 A
5818510 Cobbley et al. Oct 1998 A
5819284 Farber et al. Oct 1998 A
5822493 Uchara et al. Oct 1998 A
5826206 Nemeth Oct 1998 A
5828994 Covell et al. Oct 1998 A
5831662 Payton Nov 1998 A
5835667 Wactlar et al. Nov 1998 A
5844693 Miyata Dec 1998 A
5850520 Griebenow et al. Dec 1998 A
5864871 Kitain et al. Jan 1999 A
5870143 Suzuki Feb 1999 A
5870754 Dimitrova et al. Feb 1999 A
5870770 Wolfe Feb 1999 A
5873076 Barr et al. Feb 1999 A
5880788 Bregler Mar 1999 A
5884141 Inoue et al. Mar 1999 A
5886739 Winningstad Mar 1999 A
5892536 Logan et al. Apr 1999 A
5892909 Grasso et al. Apr 1999 A
5893062 Bhadkamkar et al. Apr 1999 A
5893091 Hunt et al. Apr 1999 A
5907322 Kelly et al. May 1999 A
5913040 Rakavy et al. Jun 1999 A
5917542 Moghadam et al. Jun 1999 A
5926205 Krause et al. Jul 1999 A
5929904 Uchida Jul 1999 A
5936659 Viswanathan et al. Aug 1999 A
5940004 Fulton Aug 1999 A
5946050 Wolff Aug 1999 A
5959623 van Hoff et al. Sep 1999 A
5963916 Kaplan Oct 1999 A
5968175 Morishita et al. Oct 1999 A
5974235 Nunally et al. Oct 1999 A
5978828 Greer et al. Nov 1999 A
5982979 Omata et al. Nov 1999 A
5999173 Ubillos Dec 1999 A
6005564 Ahmad et al. Dec 1999 A
6009204 Ahmad Dec 1999 A
6018359 Kermode et al. Jan 2000 A
6020883 Herz et al. Feb 2000 A
6020884 MacNaughton et al. Feb 2000 A
6025837 Matthews, III et al. Feb 2000 A
6034652 Freiberger et al. Mar 2000 A
6035336 Lu et al. Mar 2000 A
6041142 Rao et al. Mar 2000 A
6049777 Sheena et al. Apr 2000 A
6061055 Marks May 2000 A
6061056 Menard et al. May 2000 A
6064980 Jacobi et al. May 2000 A
6069621 Schupak May 2000 A
6072542 Wilcox et al. Jun 2000 A
6078740 DeTreville Jun 2000 A
6078916 Culliss Jun 2000 A
6081551 Etoh Jun 2000 A
6098064 Pirolli et al. Aug 2000 A
6118925 Murata et al. Sep 2000 A
6141693 Perlman et al. Oct 2000 A
6160950 Shimazaki et al. Dec 2000 A
6172675 Ahmad et al. Jan 2001 B1
6182067 Presnell et al. Jan 2001 B1
6195657 Rucker et al. Feb 2001 B1
6199080 Nielsen Mar 2001 B1
6208989 Dockter et al. Mar 2001 B1
6212657 Wang et al. Apr 2001 B1
6236395 Sezan et al. May 2001 B1
6243725 Hempleman et al. Jun 2001 B1
6259817 Ahmad Jul 2001 B1
6263507 Ahmad et al. Jul 2001 B1
6266085 Kato et al. Jul 2001 B1
6266649 Linden et al. Jul 2001 B1
6272231 Maurer et al. Aug 2001 B1
6282362 Murphy et al. Aug 2001 B1
6297845 Kuhn et al. Oct 2001 B1
6351599 Komeno Feb 2002 B1
6360202 Bhadkamkar et al. Mar 2002 B1
6360234 Jain et al. Mar 2002 B2
6360261 Boyd et al. Mar 2002 B1
6366296 Boreczky et al. Apr 2002 B1
6377519 Lee et al. Apr 2002 B1
6385386 Aotake May 2002 B1
6385619 Eichstaedt et al. May 2002 B1
6404811 Cvetko et al. Jun 2002 B1
6434621 Pezzillo et al. Aug 2002 B1
6452969 Yim Sep 2002 B1
6460036 Herz Oct 2002 B1
6466918 Spiegel et al. Oct 2002 B1
6496568 Nelson Dec 2002 B1
6502139 Birk et al. Dec 2002 B1
6556989 Naimark et al. Apr 2003 B1
6563532 Strub et al. May 2003 B1
6574630 Augustine et al. Jun 2003 B1
6591245 Klug Jul 2003 B1
6594682 Peterson et al. Jul 2003 B2
6596030 Ball et al. Jul 2003 B2
6622305 Willard Sep 2003 B1
6624846 Lassiter Sep 2003 B1
6633910 Rajan et al. Oct 2003 B1
6681369 Meunier et al. Jan 2004 B2
6690273 Thomason Feb 2004 B2
6691106 Sathyanarayan Feb 2004 B1
6701528 Arsenault et al. Mar 2004 B1
6704750 Asazu Mar 2004 B2
6721744 Naimark et al. Apr 2004 B1
6728678 Bhadkamkar et al. Apr 2004 B2
6748455 Hinson et al. Jun 2004 B1
6750880 Freiberger et al. Jun 2004 B2
6757682 Naimark et al. Jun 2004 B1
6757866 Dey et al. Jun 2004 B1
6782186 Covell et al. Aug 2004 B1
6788314 Freiberger et al. Sep 2004 B1
6803945 Needham Oct 2004 B1
6804675 Knight et al. Oct 2004 B1
6807558 Hassett et al. Oct 2004 B1
6825875 Strub et al. Nov 2004 B1
6868452 Eager et al. Mar 2005 B1
6880171 Ahmad et al. Apr 2005 B1
6901209 Cooper et al. May 2005 B1
6934461 Strub et al. Aug 2005 B1
6961954 Maybury et al. Nov 2005 B1
6963867 Ford et al. Nov 2005 B2
6986156 Rodriguez et al. Jan 2006 B1
6993787 Kamel et al. Jan 2006 B1
7062484 Pass Jun 2006 B2
7072575 Kang Jul 2006 B2
7082407 Bezos et al. Jul 2006 B1
7130528 Blair et al. Oct 2006 B2
7143431 Eager et al. Nov 2006 B1
7155735 Ngo et al. Dec 2006 B1
7181417 Langseth et al. Feb 2007 B1
7194186 Strub et al. Mar 2007 B1
7209942 Hori et al. Apr 2007 B1
7266771 Tow et al. Sep 2007 B1
7310615 Lewis Dec 2007 B2
7340760 Wachfogel et al. Mar 2008 B2
7348935 Freiberger et al. Mar 2008 B1
7519271 Strub et al. Apr 2009 B2
7565681 Ngo et al. Jul 2009 B2
7849475 Covell et al. Dec 2010 B2
20020006266 Yoon et al. Jan 2002 A1
20020013949 Hejna, Jr. Jan 2002 A1
20020031331 Kwoh et al. Mar 2002 A1
20020073105 Noguchi et al. Jun 2002 A1
20020120925 Logan Aug 2002 A1
20020157103 Song et al. Oct 2002 A1
20020159750 Jasinschi et al. Oct 2002 A1
20020170068 Rafey et al. Nov 2002 A1
20030014326 Ben-Meir et al. Jan 2003 A1
20030043194 Lif Mar 2003 A1
20030093790 Logan et al. May 2003 A1
20030149574 Rudman Aug 2003 A1
20030163824 Gordon et al. Aug 2003 A1
20040022313 Kim Feb 2004 A1
20040078812 Calvert Apr 2004 A1
20040123324 Sazzad et al. Jun 2004 A1
20040162772 Lewis Aug 2004 A1
20040170385 Bhadkamkar et al. Sep 2004 A1
20040210555 Naimark et al. Oct 2004 A1
20050055306 Miller et al. Mar 2005 A1
20050183127 Ngo et al. Aug 2005 A1
20060031879 Colter et al. Feb 2006 A1
20060031885 Colter et al. Feb 2006 A1
20060031916 Colter et al. Feb 2006 A1
20080127240 Covell et al. May 2008 A1
20090097823 Bhadkamkar et al. Apr 2009 A1
20090276817 Colter et al. Nov 2009 A1
20110055861 Covell et al. Mar 2011 A1
Foreign Referenced Citations (28)
Number Date Country
4408131 Jul 1995 DE
179701 Apr 1986 EP
513601 Nov 1992 EP
674315 Sep 1995 EP
689362 Dec 1995 EP
2231246 Nov 1990 GB
1174192 Jul 1989 JP
2054331 Feb 1990 JP
4051628 Feb 1992 JP
4061628 Feb 1992 JP
5284437 Oct 1993 JP
6162166 Jun 1994 JP
07114572 May 1995 JP
7507169 Aug 1995 JP
7295531 Nov 1995 JP
7319691 Dec 1995 JP
08106543 Apr 1996 JP
9269923 Oct 1997 JP
8902628 Mar 1989 WO
9319427 Sep 1993 WO
9430000 Dec 1994 WO
9515801 Jun 1995 WO
9612240 Apr 1996 WO
9627983 Sep 1996 WO
9630864 Oct 1996 WO
9715926 May 1997 WO
9735296 Sep 1997 WO
9827497 Jun 1998 WO
Non-Patent Literature Citations (113)
Entry
“About Gitzo,” http/lwww.gitzo.com/en/company.htm, Gitzo SA Ball Head Camera Mounts, 1996 and 1997, 4 pages.
“About Upoc,” http://www.upoc.com/about.jsp, Sep. 25, 2000, 1 page.
“About Upoc,” http://www.upoc.com/about.jsp, Feb. 7, 2001, 1 page.
Arons, B., “Techniques, Perception, and Applications of Time-Compressed Speech,” Proceedings of 1992 Conference American Voice 110 Society, Sep. 1992, 19 pages.
Brooke et al., “Computer Graphics Animations of Talking Faces Based on Stochastic Models,” ISSIPNN '94, 1994 International Symposium on Speech, Image Processing and Neural Networks, Proceedings, vol. 1, Apr. 1994, pp. 73-76.
Brown, Martin G. et al., “Automatic Content-Based Retrieval of Broadcast News,” Proceedings of the third ACM international conference on Multimedia, Association for Computing Machinery (ACM) 1995, 29 pages.
Buckley et al., “The Effect of Adding Relevance Information in a Relevance Feedback Environment,” Proceedings of the 17th Annual Conference on Research and Development in Information Retrieval, Jul. 3-6, 1994, pp. 292-300.
Buckley, Chris et al., “The Effect of Adding Relevance Information Relevance Feedback Environment,” SIGIR '94 Proceedings of the 17th annual international ACM SIGIR conference on Research and development in information retrieval, Association for Computing Machinery (ACM) 1994, 9 pages.
Carter, S. et al., “Improving Video-on Demand Server Efficiency Through Stream Tapping,” 1997 IEEE, pp. 200-207.
Carter, S. et al., “Video-on-Demand Broadcasting Protocols,” pp. 1-25.
Cavnar, William B. et al., “N-Gram-Based Text Categorization,” Proceedings of SDAIR 94, 3rd Annual Symposium on Document Analysis and Information Retrieval, Information Science Research Institute (ISRI) 1994, 14 pages.
Chen, T. et al., “Lip synchronization in talking head video utilizing speech information,” SPIE, Visual Communications and Image Processing '95, vol. 2501, Jul. 1995, pp. 1690-1701.
Chen F. et al., “The Use of Emphasis to Automatically Summarize a Spoken Discourse,” 1992 IEEE International Conference on Acoustics, Speech and Signal Processing, Mar. 23-26, 1992, pp. 1-229 and 1-232.
Chesnais, Pascal R. et al, “The Fishwrap Personalized News System,” Proceedings of the Second International Workshop on Community Networking 1995: Integrated Multimedia Services to the Home, IEEE 1995, 8 pages.
Clouthier, Ron et al., “Imaging Applications in the Work World,” SPSE's International Symposium and Exposition on Electronic Imaging Devices and Systems '88, The International Society for Optical Engineering (SPIE) 1988, 10 pages.
CNN at Work White Paper, 1994, 22 pages.
“Correlating Audio and Moving-Image Tracks,” IBM Technical Disclosure Bulletin, vol. 33, No. 10A, Mar. 1991, pp. 295-296.
Covell et al., “Spanning the Gap between Motion Estimation and Morphing,” Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Apr. 19-22, 1994, Adelaide, Australia, http://www.mangolassi.org/covell/1994-0101, 8 pages [Internet Accessed on Jul. 11, 2008].
“Covert Video/Audio Sunglasses,” From Rock House Products, http://www.rock2000.com/nanny/glasses.htm, 1219198, 2 pages.
Dellafera, Anthony C. et al, “The Zephyr Notification Service,” Proceedings of Winter 1988 Usenix Technical Conference, 1988, 9 pages.
Douglis et al., “Rate of Change and other Metrics: a Live Study of the World Wide Web,” USENIX symposium on Internet Technologies and Systems, Dec. 1997, 14 pages, Monterey, CA.
Drummond, Chris, et al. “Intelligent Browsing for Multimedia Applications,” Proceedings of the Third IEEE International Conference on Multimedia Computing and Systems 1996, IEEE 1996, 4 pages.
Elliott, E., “Multiple Views of Digital Video,” MIT Media Laboratory, Interactive Cinema Group, Mar. 23, 1992, 4 pages.
Elliott, E., “Watch, Grab, Arrange, See: Thinking with Motion Images via Streams and Collages,” Masters Thesis, School of Architecture and Planning, Massachusetts Institute of Technology, Feb. 1993, 105 pages.
eNow-Company Background, http://www.enow.com/html/compl.html, Feb. 7, 2001, 2 pages.
“Evolution of Steve Mann's ‘existential computer’,” ‘Personal Imaging’, IEEE Computer Society, http://computer.org/pubs/computer/19971029710c.htm, 1997, 1 page.
Flanagan et al., “Computer-steered microphone arrays for sound transduction in large rooms,” J. Acoust. Soc. Am. 78 (5), Nov. 1985, pp. 1508-1518.
Forsyth, David et al., “Searching for Digital Pictures,” Scientific American, Jun. 1997, pp. 88-93.
Froba, Bernhard et al., “Multi-Sensor Biometric Person Recognition in an Access Control System,” Proceedings Second International Conference on Audio- and Video-Based Biometric Person Authentication, Mar. 22-23, 1999, Washington, D.C., pp. 55-59.
Frost III, Otis, “An Algorithm for Linearly Constrained Adaptive Array Processing,” Proceedings of the IEEE, vol. 60, No. 8, Aug. 1972, pp. 926-935.
Fujikawa, “Subrogation of User Account by Advertisement,” Nikei Communications, Mar. 4, 1996, No. 217.
Gemperle, Francine et al., “Design for Wearability,” IEEE, 1998, pp. 116-122.
Glance, Natalie et al., “Knowledge Pump: Community-Centered Collaborative Filtering,” Proceedings of the DELOS Workshop on Collaborative Filtering, 1997, 5 pages.
Glance, Natalie et al., “Making Recommender Systems Work for Organizations,” Proceedings of PAAM'99, London, UK, Apr. 19-21, 1999, 21 pages.
Gomes, Lee, “Upstart's Internet ‘TV’ has Microsoft Tuned in,” Wall Street Journal, Jul. 29, 1996, 1 page.
Greening, Dan R., “Building Consumer Trust with Accurate Product Recommendations,” A White Paper on LikeMinds WebSell 2.1, LikeMinds, Inc. 1997, 15 pages.
Grenier, Yves, “A microphone array for car environments,” Speech Communication, vol. 12, No. 1, Mar. 1993, pp. 25-39.
Guernsey, L., “Mining the ‘Deep Web’ with Specialized Drills,” Jan. 25, 2001, The New York Times, 3 pages.
Haeberli, Paul et al., “Automatic Panoramic Image Merging,” World Wide Web, Grafica Obscura, 1998, 2 pages.
Hall, Allan, “A new spin on power dressing,” Associated Newspapers Ltd., Sep. 14, 1999, 1 page.
Hauptmann et al., “Speech for Multimedia Information Retrieval,” UIST '95 (ACM), Nov. 14, 1995, pp. 79-80.
Hauptmann et al., “Text, Speech, and Vision for Video Segmentation: The Informedia Project,” AAAI Fall Symposium, Computational Models for Integrating Language and Vision, Nov. 10-12, 1995, 12 pages.
Healey, Jennifer et al., “Quantifying Driver Stress: Developing a System for Collecting and Processing Bio-Metric Signals in Natural Situations,” Proceedings of the Rocky Mountain Bio-Engineering Symposium, Apr. 16-18, 1999, pp. 1-6.
Healey, Jennifer et al., “StartleCam: A Cybernetic Wearable Camera,” Second International Symposium on Wearable Computers, Oct. 19-20, 1998, pp. 42-49.
Henton, C. et al., “Saying and Seeing It With Feeling: Techniques for Synthesizing Visible, Emotional Speech,” SSW2-1994, pp. 73-76.
Hjetsvold, Rune et al., “Integrated, Video Archive Tools,” Multimedia '95: Proceedings of the Third ACM International Conference on Multimedia, Association for Computing Machinery (ACM) 1995, 11 pages.
Hollands, P., “Keeping Track of Changes to Web-based Resources,” eLib, URL-Minder, May 20, 1996, 8 pages.
Horner, C., “NewsTime: A Graphical User Interface to Audio News,” Masters Thesis, School of Architecture and Planning, Massachusetts Institute of Technology, Jun. 1993, pp. 1-84.
Huttenlocher et al., “Tracking Non-Rigid Objects in Complex Scenes,” IEEE International Conference on Computer Vision, 1993, pp. 93-101.
“Informedia: Experience on Demand,” DARPA/ITO Joint IC and V/IM Principal Investigator Meeting, Oct. 15-17, 1997, 2 pages.
Iwayama, Makoto et al., “Cluster-Based Text Categorization: A Comparison of Category Search Strategies,” SIGIR '95: Proceedings of the 18th annual international ACM SIGIR Conference on Research and Development in Information retrieval, Association for Computing Machinery (ACM) 1995, 8 pages.
Joachims, Thorsten et al., “WebWatcher: Machine Learning and Hypertext,” Fachgruppentreffen Maschinelles Lernen, 1995, 7 pages.
Kelly, Patrick H. et al., “An Architecture for Multiple Perspective Interactive Video,” Multimedia 1995, pp. 201-212.
Kindermann, C. et al., “The MIHMA Demonstrator Application: bmt line,” Technical Report, Non Standard Logics S. A. and TU Berlin 1996, 39 pages.
La Lunette Camera—Alain Mikli, http://www.mikli.fr/eng/Actualites/LcameralLcamera.htm?Nav=1, La Lunette Camera Sunglasses, Jul. 27, 1998, 1 page.
Lichty, Tom, “The Official America Online for Windows Tour Guide” (2nd Edition) Version 2.5, Ventana Press, 1994, 54 pages.
Lindblad et al., “ViewStation Applications: Implications for Network Traffic,” IEEE Journal on Selected Areas in Communications, vol. 13, No. 5, Jun. 1995, pp. 768-777, New York, NY, US.
Mann, S. et al., Video Orbits of the Projective Group: A simple approach to featureless estimation of parameters, IEEE Trans., Image Proc., Jul. 1997, pp. 1-31.
Mann, S. et al., Virtual Bellows: Constructing High Quality Stills from Video, IEEE, 1994, pp. 363-367.
Masand, Brij et al., “Classifying News Stories using Memory Based Reasoning,” SIGIR '92 Proceedings of the 15th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, Association for Computing Machinery (ACM) 1992, 7 pages.
MicroStrategy Launches Strategy.com, the World's First Personal Intelligence Network, Jun. 28, 1999.
Moulines et al., “Non-parametric techniques for pitch-scale and time-scale modification of speech,” Speech Communication, vol. 16, 1995, pp. 175-205.
Nesbitt, Steve, “WebCobra: An Automated Collaborative Filtering Agent System for the World Wide Web,” Degree Thesis, The Department of Computer Science at James Cook University of North Queensland, 1997, 114 pages.
Nielsen, Jakob, “Hypertext & Hypermedia,” Academic Press, Inc. 1990, 276 pages.
1998 Everest Expedition: Technology, http://www.everest.org/EveresUtech.html, Datacam Technology, Jul. 31, 1998, 2 pages.
Novell GroupWise Reference for Mcintosh Version 4.1, Novell, Inc. (1993, 1994), 210 pages.
Paris et al., “A Low Bandwidth Broadcasting Protocol for Video on Demand,” IC3N '98, Oct. 1998, pp. 640-647.
PCT International Search Report for corresponding International Application PCT/US97/04610 (WO 1997/035296), 3 pages.
Phillips, P. Jonathon et al., “The FERET Evaluation,” Face Recognition: From Theory to Applications 1998, pp. 244-261.
Pentland A. et al., “Photobook: Content-Based Manipulation of Image Databases,” International Journal of Computer Vision, vol. 18, No. 3, pp. 233-254, Kluwer Academic Publishers 1996, 22 pages.
Pentland A. et al., “Photobook: Content-Based Manipulation of Image Databases,” Proceedings of the Society of Photo-Optical Instrumentation Engineers (SPIE), 2185, 34, SPIE 1994, 24 pages.
Product description for SMPTE Time Code Products for GPS Video Position Logging, Horita Co., Inc., Viejo, CA, 1998, 1 page.
Product description for Studio Master Time Code and Sync Generators, Horita Co., Inc., Viejo, California, 1998, 1. about.e.
Rabiner, A Tutorial on Hidden Markov Models and Selected Applications in Speech Recognition, Proceedings of the IEEE, vol. 77, No. 2, Feb. 1989, pp. 257-285.
Reynolds, D., “A Gaussian Mixture Modeling Approach to Text-Independent Speaker Identification,” PhD. Thesis, Dept. of Electrical Engineering, Georgia Institute of Technology, Aug. 1992, pp. 1-154.
Rigdon, Joan E., “Screen Savers Go Beyond Fish, Flying Toasters,” Wall Street Journal, Feb. 13, 1996, 1 page.
Rosenberg, Aaron et al., “Recent Research in Automatic Speaker Recognition,” Advances in Speech Signal Processing, 1992, pp. 701-738.
Roucos et al., “High Quality Time-Scale Modification for Speech,” Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, IEEE, 1985, pp. 493-496.
Rucker, James et al. “Personalized Navigation for the Web,” Communications of the ACM, vol. 40 Issue 3, Mar. 1997, 3 pages.
Sakauchi et al., “Multimedia Database Systems for the Contents Mediator,” IEICE Trans. Inf. and Syst., vol. E79-D, No. 6, Jun. 1996, pp. 641-646.
Salton et al., “Improving Retrieval Performance by Relevance Feedback,” Journal of the American Society for Information Science, vol. 41, No. 4, Jun. 1990, pp. 288-297.
Salton, Gerard et al., “Improving Retrieval Performance by Relevance Feedback,” Journal of the American Society for Information Science, John Wiley & Sons, Inc. 1990, 10 pages.
Salton, Gerard et al., “Introduction to Modern Information Retrieval,” McGraw-Hill Book Company 1983, pp. 120-134, 137-145, 151-152, 236-244, and 424.
Scheirer et al., “Construction and Evaluation of a Robust Multifeature Speech/Music Discriminator,” Proc. ICASSP, Apr. 21-24, 1997, pp. 1-4.
Schwab et al., “Learning to Recommend From Positive Evidence,” pp. 241-247.
Scott et al., “Synthesis of Speaker Facial Movement to Match Selected Speech Sequences,” Speech Science and Technology '94, Dec. 5-8, Perth, Australia, 6 pages.
Search Report PCT/US97/22145 (WO1998027497), Apr. 29, 1998.
Senior, Andrew, “Face and feature finding for a face recognition system,” Proceedings Second International Conference on Audio- and Video-based Biometric Person Authentication, Mar. 22-23, 1999, Washington, D.C., pp. 154-159.
Shahraray et al., “Automatic generation of pictoral transcripts of video programs,” SPIE, vol. 2417, Jan. 1995, pp. 512-518.
Shibata M., “A Description Model of Video Content and Its Application for Video Structuring,” Systems and Computers in Japan, vol. 27, No. 7, Jun. 1996, pp. 70-83.
Slaney, M. et al., “Baby Ears: A Recognition System for Affective Vocalizations,” Proceeding of the 1998 International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Seattle, WA, May 12-15, 1998, 4 pages.
Sodergard et al., “Integrated Multimedia Publishing-Combining TV and Newspaper Content on Personal Channels,” W8-papers, Jun. 18, 1999, http://www8.org/w8-papers/1b-multimedia/integrated/integrated.html, 24 pages [accessed on Jan. 23, 2007].
Sony ClipLink Guide, 1996, pp. 9-14.
Sony PowerHAD digital camcorder operating instructions, 1998, pp. 65-71.
Staff Reporter, “PointCast Inc. is Testing New Screen-Saver Product,” Wall Street Journal, May 1996, 1 page.
Supercircuits Microvideo Catalog, GC1 Professional Grade Color Covert Glasses Cam, 1998, 1 page.
Szeliski, R. et al., “Creating Full View Panoramic Image Mosaics and Environment Maps,” Computer Graphics Proceedings, Annual Conference Series, 1997, pp. 251-258.
Szeliski, Richard, Video Mosaics for Virtual Environments, IEEE Computer Graphics and Application, Mar. 1996, pp. 22-30.
Tennenhouse et al., “The ViewStation: a software-intensive approach to media processing and distribution,” Proceedings of the 17th Annual Conference on Research and Development in Information Retrieval, Jul. 3-6, 1994, pp. 104-115.
Terlaga, Amy, “The URL Minder,” Conntech 1995 Archives, Jul. 29, 1995, 2 pages.
Tokuda, “To make the Internet Free—Mr. Y. Itakera (President, Hyper Net),” Nikkei Business Publications, Mar. 11, 1996, No. 831.
Velbon.com, “CX Series Photo Tripods,” and “OF-Series ‘Dual Function’ PhotoNideo Tripods,” 3 pages [Internet accessed on Apr. 13, 1999].
VideoLogger.RTM.: Automate Video Encoding and Indexing, Datasheet, Virage, Inc., San Mateo, California, 2003, 3 pages.
VideoLogger.RTM. SDK: Virage Open Architecture, Datasheet, Virage, Inc., San Mateo, California, 2003, 2 pages.
Virtual Tech, “Thousands of NEW Products are introduced every year by hundreds of companies . . . ,” Toyoda MachineJY USA, believed to be in use in Japan in 1998, 2 pages.
Wactlar et al., “Intelligent Access to Digital Video: Informedia Project,” Computer, vol. 29, No. 5, May 1996, pp. 46-52.
Welcome to MyZDnet Alerts, http://my.zdnet.com/alerthelp, Sep. 25, 2000, 1 page.
Welsh et al., “Facial-Feature Image Coding Using Principal Components,” Electronics Letters, vol. 28, No. 22, Oct. 1992, pp. 2066-2067.
Widyantoro, Ioerget, Yen, “An Adaptive Algorithm for Learning Changes in User Interests,” Department of Computer Sciences, Texas A&M University, pp. 405-412.
“Worldwide Webcam,” http://www.worldwide-top100.com/topsites1/, Sep. 25, 2000, 4 pages.
Yeung, M. M. et al., “Efficient Matching and Clustering of Video Shots,” IEEE '95 (ICIP), vol. 1, Oct. 1995, pp. 338-341.
Zabih et al., “A Real-Time System for Automatically Annotating Unstructured Image Sequences,” IEEE International Conference on Systems, Man, and Cybernetics, 1993, 6 pages.
Zhang et al., “Automatic Parsing of News Video,” IEEE Conference on Multimedia Computing and Systems, 1994, pp. 45-54.
Related Publications (1)
Number Date Country
20140046957 A1 Feb 2014 US
Provisional Applications (1)
Number Date Country
60178627 Jan 2000 US
Continuations (3)
Number Date Country
Parent 12411500 Mar 2009 US
Child 13867792 US
Parent 10800393 Mar 2004 US
Child 12411500 US
Parent 09656638 Sep 2000 US
Child 10800393 US