This invention pertains to multimedia authoring methods, systems, software, and product distribution media. In particular, this invention automatically generates a single-media or multimedia presentation based on a user's stored media files, thereby automatically producing a customized story for the user requiring minimal user effort.
It is widely acknowledged that viewing images in the form of a single-media or multi-media thematic presentation, referred to herein as a “story,” or a hardcopy thematic album is much more compelling than browsing through a number of random hard-copy prints, or looking at a random series of static images presented sequentially using a slide projector, computer, or television. The selective addition of other elements to the presentation such as a sound track appropriate to the content of the images, the insertion of interesting transitions between the images, the addition of video or of various video-style special effects including fades and dissolves, image-collaging, backgrounds and borders, and colorization makes the presentation much more interesting to the viewer and can greatly enhance the emotional content of the images being presented. The proliferation in the home of new television-based viewing platforms able to accommodate multimedia, including DVD, Video CD players, home media servers, and high definition digital displays also increases the demand for this type of story or presentation.
For the ordinary photographic consumer, the creation of a multimedia presentation or album of still images is not presently very easy or convenient. The selection and layout of digital image assets can be a significant and time-consuming process. Even if the images are available in digital form, a consumer must have facility with multimedia authoring software tools such as Macromedia Director™ or Adobe Premier™ in order to create such a presentation. These software tools, while very flexible, are aimed more at the professional presentation creator, have multiple functional options, and require a great deal of time and experience to develop the skill needed to use them to advantage. More recently, template-based multimedia presentation applications such as Photojam™, offered by Shockwave.com™, or PC-based “movie making” applications such as Apple's i-Movie™ have become available. While these applications can simplify the creation of multimedia presentations for a consumer, they do not help to automate many of the story making options. Current applications often require the user to select a presentation theme and to select the assets, such as pictures, video, and music; that are used to automatically generate an image product. In addition, these applications offer no way to automatically generate an image product such as for special occasions, holidays, anniversaries, or for selected other events or calendar dates.
Thus, there remains a need for an automated authoring system where an inexperienced user can receive an automatically generated single-media or multimedia story and obtain copies of the presentation over a variety of channels and in a variety of formats suitable for various types of presentation devices.
In answer to these and other needs, and in accordance with one preferred embodiment of the present invention, there is provided a method for automatically generating a customized story, image product, or presentation on a digital storage device of a set of digital media files provided by a user, comprising the steps of analyzing the digital media files for semantic information, including metadata, and organizing the digital image assets in association with a selected presentation format and on a medium that can be viewed by the user, the format automatically chosen in accordance with the semantic and metadata information, or pre-selected by the user or by the computer system.
Another preferred embodiment of the present invention is a method, software, and a programmed computer system for automatic story-creation from a collection of assets (still images, video, music, public content) utilizing prescribed template rules applied to the collection. The template rules rely on metadata associated with the assets, personal profile and/or user preference data acquired from the user. Metadata can be in the form of EXIF data, index values from image understanding and classification algorithms, GPS data, and/or personal profile/preferences. These rules, or a subset of them, when automatically applied to a collection within the system, will produce a story for rendering via a multimedia output engine. The story can be delivered to the user on a variety of storage media such as CDs, DVDs, magnetic discs, and portable flash memory media. The story can be transmitted via cellular networks, by satellite providers, or over local and wired area networks. The story can be received and viewed by the user on a variety of hand held display devices such as PDAs and cell phones. The story can be received at a home and displayed on a computer, television, or over theater style projection systems.
Another preferred embodiment of the invention comprises a method for automatically creating an image product comprising the steps of obtaining a plurality of digital media files associated with an event such as a birthday, holiday, anniversary, or other occasion. Classifying the event is accomplished based on analyzing the digital media files and automatically determining a format of an output product based upon the analysis, and then selecting which ones of the digital media files will be included in accordance with requirements of said output image product.
Another preferred embodiment of the invention comprises a program storage device storing a computer program for execution on a computer system. The program is capable of automatically generating an image product utilizing a number of digital media files that are resident in the computer system. The program is designed to first detect an image product trigger which might be a calendar date, a user request for an image product, or an upload to the computer system of a plurality of digital media files such as images, sound files, video, etc. The program locates a plurality of digital media files associated with an event if it is a calendar event, for example, or if the trigger is an upload of media files, the program will determine if the media files satisfy an output product format type. The program automatically classifies the plurality of digital media files based on analyzing metadata associated therewith and automatically selects those files, based on the classifying step, that satisfy an output product format type. The selected media files are ranked based on one or more of a variety of metrics, such as an image value index, and some or all of the ranked files are included in an appropriate image product format that is related to the event.
Other embodiments that are contemplated by the present invention include computer readable media and program storage devices tangibly embodying or carrying a program of instructions readable by machine or a processor, for having the machine or computer processor execute instructions or data structures stored thereon. Such computer readable media can be any available media which can be accessed by a general purpose or special purpose computer. Such computer-readable media can comprise physical computer-readable media such as RAM, ROM, EEPROM, CD-ROM, or other optical disk storage, magnetic disk storage, or other magnetic storage devices, for example. Any other media which can be used to carry or store software programs which can be accessed by a general purpose or special purpose computer are considered within the scope of the present invention.
These, and other, aspects and objects of the present invention will be better appreciated and understood when considered in conjunction with the following description and the accompanying drawings. It should be understood, however, that the following description, while indicating preferred embodiments of the present invention and numerous specific details thereof, is given by way of illustration and not of limitation. Many changes and modifications may be made within the scope of the present invention without departing from the spirit thereof, and the invention includes all such modifications.
Overview:
With respect to
The story viewer 102 can be used for viewing an image product or other media assets on a monitor or other display apparatus coupled to the computer system, and may include capabilities for editing media assets. Numerous computer implemented image-editing applications are well known in the art, and are not described further herein. Although it is referred to as a story viewer, audio playback of digital audio assets can also be included in the story viewer, wherein the playback can occur simultaneously with, or separately from, viewing image assets.
The story viewer can utilize a computer system connection to a network, such as an Internet connection, for sending or receiving completed multimedia stories to and from another computer system. It can also utilize a network connection for sending completed stories or other media asset collections over a cell network to a hand held device such as a multimedia capable cell phone or PDA, or to a printer for printing a story. One of the distinctive features of the present invention is a selection algorithm, which automatically selects and sequences media assets.
A story notifier 103 is used for automatically notifying a user that an image product has been generated by the system. The story notifier can send notices over a network to another computer, utilize a computer system connection to an RSS feed for sending notifications that a story has been generated, or over a cell network to a hand held device such as a cell phone or PDA. In the latter instance, an SMS (Short Messaging System) protocol can be implemented on the computer system. The computer system can also be programmed to notify a user via a display screen message or by an audio signal. A user can also access the story notifier to forward notifications to another device using any of the above means.
Although
The bottom portion of
With reference to the front-end user interface, the user introduces selected assets into the system database by activating the asset uploader 101. This component then communicates with the asset import 104 component. The asset import functions to store copies of the assets into the asset store 112 and informs the system manager 107 that it has completed the upload. The asset import component can be located on the computer system or it may be located on a server connected to a computer locally or over a network. In one preferred embodiment, communication between the asset import and system manager occurs via the database 113, however, each of the back-end components can be implemented to communicate directly with the system manager 107. For ease of illustration,
The semantic indexers 110 include metadata extraction mechanisms for extracting metadata already included in the digital asset, such as embedded by a digital camera, as explained above, and recording it in the database. Other examples of such metadata include the capture date and time, among many other examples as described herein. The indexers can also include complex algorithms that analyze stored assets to generate more complex metadata. A preferred embodiment of the present invention sequences the operation of various semantic indexers that organize a set of media assets because some of the indexers may rely on the output of other indexers for their operation. Such ordering would be managed by the System Manager 107 or it may be sequenced via a table lookup for a strict ordering scheme, or it may be stored in a dependency tree or other suitable data structures. All of the generated metadata are recorded in the database 113 and are appropriately associated with their corresponding media asset. In a preferred embodiment, any of the metadata may be stored in the triplestore 115, a type of database optimized for storing large quantities of unstructured data.
When the last semantic indexer has completed, or at least a sufficient number of pre-selected indexers have completed, the system manager 107 will activate the story suggester 106 to determine if one or more appropriate stories should be created, which, if so determined, will cause the generation of an image product or story. The story suggester, in turn, will activate the inference engine 111 to evaluate the various rules stored in the rule base 114 to determine if any of the story rules stored therein can be satisfied by a current collection of media assets. This is referred to as an event based trigger for story generation. Other types of programmable triggers might include monitoring for an upload of assets by a user. For example, if a user has uploaded a number of assets the story suggester will begin to analyze the assets to generate a possible story if the inference engine determines that a sufficient number of story rules have been satisfied. One preferred embodiment of the inference engine is the Prolog inference engine having the rule base 114 represented as a set of Prolog clauses organized into sets of named rules stored in an XML file and evaluated by the Prolog engine as requested. Prolog is a declarative logic programming language used as explained in detail below.
When the story suggester is searching for stories to create based upon a date based trigger, such as an anniversary, holiday, birthday, or other event, the story suggester 106 requests that the inference engine 111 evaluate the Prolog clause suggestStoryByEvent, looking for valid bindings for several free variables including, but not necessarily limited to, the user, the story type, the intended recipient, and the product type. If a valid set of variable bindings is identified (such as by the Prolog inferencing engine), the story suggester will then obtain from the smart asset selector 109 the appropriate set of assets to go with the suggested story, and then request that the product generator 108 create the desired product representation which may include a photo album, or rendering an image on a mug or on a T-shirt. The product generator will create one or more files of the appropriate format representing the image product, which may include instructions to be sent over a network to a product maker if the image product so requires. The product generator may store the resulting file(s) in the asset store 112 enabling the resulting product to be treated as another asset in the user's collection. If there are only a small number of assets that satisfy a story rule, it may be determined to make a mug or other product containing a single image or small number of images. The system manager 107 is notified by the product generator when the image product has been generated, at which point the system manager alerts the story notifier service 105, which in turn causes the story notifier 103 to inform the user that a new product, or product preview, has been created. In addition to the notification methods described earlier, the notification may be in the form of a pop-up window on a display containing text and graphics information indicating that an image product has been created and is ready for viewing. The user may then view the product using the story viewer 102. The story viewer may be implemented as a browser such as Internet Explorer, or a video playback device such as Windows Media Player. In a preferred embodiment, the user has the option to request from the story viewer that the product be sent to a printer for a hard-copy rendition of the product, such as a bound photo album, if appropriate. The user may also request that a mug, for example, be produced and delivered. Such an implementation requires that an order screen be presentable to the user to provide contact information for a fulfillment provider, which may include a direct link to a provider website, and to obtain a user's delivery request information. To display the product, the story viewer requests and obtains the necessary assets from the asset store 112.
The system manager may also launch the story suggester on a periodic basis, such as nightly, or monthly, or some other period, to determine if a calendar event driven story can be created from digital media files stored on the computer system. This can optionally be driven by an upcoming event based on a time window selected by the user. The reader will appreciate that alternative architectures may result in fundamentally the same behavior. For example, the story suggester 106 and smart asset selector 109 components may be combined into a single component, or the story suggester may directly invoke the smart asset selector to determine that the appropriate set of assets are available for a particular story. In suggesting and creating stories for a particular user, the story suggestor and smart asset selector may consider only assets and metadata owned by that user, or they may consider all assets in the system to which the user has access, including assets that other system users may have shared with the user.
Metadata:
Metadata encompasses data that is stored and associated with a media asset. In general, by way of example and not by limitation, there are three sources of metadata: capture device metadata, such as time, date, and location provided by a digital camera; user provided metadata such as via a capture device user interface or an image editing application interface; and derived metadata such as by face recognition or scene classification applications. Derived metadata also includes metadata deduced from existing metadata of any type. Metadata can be generated at the time of storing files on a computer of captured image data. Metadata can be generated automatically by a capture device or entered manually into storage by a user at the time of capturing an image, for example. It can also be generated automatically without a user's knowledge by programmed operation of image recognition software. Such software may be capable of generating many levels of metadata based on extrapolating existing metadata information. For example, a family tree may be inferred given sufficient existing metadata for known family members who are depicted in stored image media collections.
With reference to
Face clustering uses data generated from detection and feature extraction algorithms to group faces that appear to be similar. As explained in detail below, this selection may be triggered based on a numeric confidence value. Location-based data 207, as described in U.S. Patent Publication No. US2006/0126944; entitled: “Variance-Based Event Clustering”; filed on Nov. 17, 2004; can include cell tower locations, GPS coordinates, and network router locations. A capture device may or may not include metadata archiving with an image or video file; however, these are typically stored with the image as metadata by the recording device, which captures an image, video or sound. Location-based metadata can be very powerful when used in concert with other attributes for media clustering. For example, the U.S. Geological Survey's Board on Geographical Names maintains the Geographic Names Information System, which provides a means to map latitude and longitude coordinates to commonly recognized feature names and types, including types such as church, park or school. Item 208 exemplifies identification or classification of a detected event into a semantic category such as birthday, wedding, etc. as described in detail in U.S. Patent Publication No. US2007/0008321; entitled: “Identifying Collection Images With Special Events”; filed on Jul. 11, 2005. Media assets classified as an event can be so associated because of the same location, setting, or activity per a unit of time, and are intended to be related, to the subjective intent of the user or group of users. Within each event, media assets can also be clustered into separate groups of relevant content called sub-events. Media in an event are associated with same setting or activity, while media in a sub-event have similar content within an event. An image value index (“IVI”) 209 is defined as a measure of the degree of importance (significance, attractiveness, usefulness, or utility) that an individual user might associate with a particular asset (and can be a stored rating entered by the user as metadata), and is described in detail in U.S. patent application Ser. No. 11/403,686; filed on Apr. 13, 2006; entitled: “Value Index From Incomplete Data”; and in U.S. patent application Ser. No. 11/403,583; filed on Apr. 13, 2006; entitled: “Camera User Input Based Image Value Index”). Automatic IVI algorithms can utilize image features such as sharpness, lighting, and other indications of quality. Camera-related metadata (exposure, time, date), image understanding (skin or face detection and size of skin/face area), or behavioral measures (viewing time, magnification, editing, printing, or sharing) can also be used to calculate an IVI for any particular media asset. The prior art references listed in this paragraph are hereby incorporated by reference in their entirety.
Video key frame extraction 210 is the process of extracting key-frames and/or a salient shot, scene, or event, and the associated audio to provide a summary or highlight of a video sequence, and is described in detail in U.S. patent application Ser. No. 11/346,708; entitled: “Extracting Key Frame Candidates From Video Clip”. EXIF data 211 (Exchangeable Image File format for digital still cameras: EXIF Version 2.2, JEITA CP-3451, Japan Electronics and Information Technology Industries Association, April 2002) is data generated by a recording device and is stored with the captured media file. For example, a digital camera might include various camera settings associated with an image such as f-stop, speed, and flash information. These camera-generated data may also include GPS data indicating geographic location related to where an image was captured. All metadata, whether input by a user, provided by a recording apparatus, or inferred by a computer system can be used by the programmed computer system to generate additional metadata based on inferences that can be determined from existing metadata. The prior art references listed in this paragraph are hereby incorporated by reference in their entirety.
With reference to
This user profile also includes information about people related to “Peter Jones” such as family and friends, which will also be associated by the program with the profiled person. The depiction of user input information related to a person should not be limited only to the examples illustrated in
For familial relationships, the system does not require that the user enter all family relationships—one does not need to say, for example, that Jane is Ann's daughter, that Jane is Mary's grandchild, that Jane is Bill's niece, etc. Instead, the system requires only that the canonical relationships of spouse, and parent/child be entered; all other familial relationships can be automatically inferred by the system. Relationships by marriage can likewise be inferred, such as mother-in-law. The system can provide a way for the user to specify that such a relationship has terminated as a consequence of divorce, for example.
With reference to
Some metadata may be obtained from third-party sources, such as weather or calendar services, as performed by external data accessors 118. For example, it may be useful in constructing a story to know what the weather was like a given day in a given location. The date and time information, combined with the GPS information recorded in an EXIF file may be provided as input to an external web service providing historical weather information. Location information together with date information can identify an event or a likely special interest. Alternatively, such information may be provided to a service describing events, enabling the system to know, for example, what happened at a stadium on a particular date.
General Algorithm:
A preferred embodiment of the present invention includes an algorithm that automatically checks whether required components for a story exist in a user's media database. Any programmable event may be used to trigger an evaluation of a media database for the possibility of a story creation. An upload of one or more media files can initiate a check for automatic story creation. A periodic calendar date or an approaching anniversary or holiday can trigger a check of media assets for use in a story product. The significance of a particular calendar date can be inferred based on the frequency and type of user activity during time periods surrounding the calendar date. A user's preferred activities can also be inferred from analyzing his or her media assets, or by tracking a frequency of how often particular types of assets are accessed. The ways in which a user's media assets, and the user's interaction with those assets, can be analyzed is virtually unlimited. The embodiments described herein are not meant to restrict the present invention to any specific embodiment.
With respect to a recurring date, such as a holiday, an example algorithm for suggesting a Mother's Day story for a particular user in a preferred embodiment can be expressed in the Prolog computer language. It has an English equivalent as follows:
The above rule depicts a high level rule used to determine whether a particular story product should be created for a particular user. It should be noted at this point that those skilled in the art will recognize the virtually unlimited number of rules that can be programmed to implement the present invention. If the rule above is satisfied, then a Mother's Day story type is used for the next step in story creation.
The story type defines a set of rules used to pick the assets to use to make a particular story product. The smart asset selector 109 executes the rule set requested by the story suggester 106 to determine the appropriate set of assets for the story product being created. In the preferred embodiment, the rules making up a rule set are expressed in Prolog, using a version of Prolog where clauses are written in a parenthesized prefix form known as S-expressions.
“Best” may be defined according to a variety of programmed metrics, or a combination thereof, including various image value index (IVI) metrics. These criteria can be extended to other types of dates besides holidays. The above rules are merely exemplary; the Prolog language enables an arbitrary set of constraints to be defined. In a preferred embodiment, the exact definition of best is defined using additional Prolog clauses.
The story suggester 106 requests that the smart asset selector compute the set of assets matching the rule set “Mother's Day Album.” The smart asset selector 109 in turn requests that the inference engine 111 execute the associated rules stored in Rulebase 114, determining which assets satisfy the constraints specified by the rules. Continuing the previous example, given the rule set 2 as the rule set “Mother's Day Album”, which is shown in part in
A satisfied rule set will specify a number of assets, if they exist. It's possible that an asset store will contain no assets satisfying a rule set. In that event, a default selection algorithm can be programmed to select available assets if a story product must be generated. A rule set may require the rule set, or its constituent rules, to match a minimum number of assets; if insufficient assets are present, then the story product is not created. A rule set may also specify further constraints on the assets that must be adhered to by the product generator. For example, a rule set may specify a sequence that the assets must follow in the final product and/or how the assets are to be grouped. The scope of the present invention includes all such embodiments.
Another preferred embodiment of the present invention is in the form of an event driven story type. This story type is triggered based upon an event. For example, an upload of assets to the computer system can be a triggering event. In one embodiment, the system, upon receipt of a set of assets, attempts to classify those assets as belonging to one or more event types. The system combines this event classification with additional information about the user to suggest a particular story type. In general, the programmed computer system includes the following routines for generating a story product for this event type:
The interest and activity ontology defines an extensible list of possible activities, interests and hobbies. For example, a subset of the ontology may include the following classes:
A full ontology class can be scaled to contain an arbitrary amount of information. The computer system, upon uploading of a set of assets, for example, a series of photos from a digital camera, attempts to first group those assets into events using the event classifier, described above, and then classify the events according to the interest and activity ontology. In one preferred embodiment, the programmed computer system classifies assets belonging to one of the following example high-level event types 208:
These event types are selected because images can be categorized into these four categories using metadata analysis. These categories can be mapped to one or more classes from the previous activity and interest ontology. For example, the event type Outdoor Sports is mapped to the item 1.b Outdoor Sports in the ontology.
The product catalog likewise contains a set of possible product types, along with the activities/interests those products may be associated with:
Given the above, the system can suggest a themed story based upon an upload of a set of digital media assets. For example, suppose a father uploads a set of pictures from his daughter Jane's recent little league game, and the system knows the following information:
The specific algorithm for picking a story based on automatically selecting a theme associated with a set of pictures is as follows, in one preferred embodiment:
This rule, along with many other such rules, is stored in the rule repository 114 and is executed by the inference engine 111 when requested by the story suggestor 106.
With reference to
The previously described inference engine 111 of
Event E1 513 is owned by user Alex 501, as shown by link 514, so Alex satisfies rule clause 4.1. Event E1 contains pictures P1 518 through Pn. Moreover, Event E1 has activity type Outdoor Sports, shown by nodes 513 and 510 and “classifiedAs” link 512. Consequently, rule clause 4.2 is satisfied by binding the variable EventType to Outdoor Sports.
A set of pictures making up an event is considered to feature a particular person if that person is portrayed in the pictures. More complex definitions of what it means for a set of pictures to feature a person may be defined to require that the person be predominantly portrayed in those pictures, for example, appearing in a majority of the pictures, etc. Using the simple definition that an event features a person if the person appears in a picture belonging to the event, the rule clause 4.3 is satisfied by binding the variable Person to Jane, in light of the statement represented by 518, 515 and 503. Clause 4.4 is satisfied by binding User to Alex, supported by the statement represented by 501, 502 and 503, that Alex is a parent of Jane. Clause 4.5 is satisfied by binding ActivityType to the class baseball, supported by the statement represented by 503, 504 and 506, that Jane likes baseball. Given the binding of ActivityType to baseball, clause 4.6 is satisfied by binding Product to the baseball album, using 519, 520 and 506. Given that baseball is a subclass of Outdoor Sports (506, 505, 507), the variable binding of Activity to baseball and EventType to Outdoor Sports satisfies clause 4.7 and so, by this example, the entire rule 4 is satisfied given the variable binding of User to Alex and Product to baseball album. More precisely, the variables are bound to the corresponding universal resource identifiers representing Alex and the baseball album product.
As noted previously, the preferred embodiment uses a Prolog inferencing engine to search for solutions to rules, where the rules are represented using Prolog clauses, but other mechanisms for describing constraints may also be used.
Referring to
Story Creation:
With reference to
With reference to
Another preferred embodiment of the present invention includes the option of providing the user with edit capability for editing an image product that is presented to the user for his or her approval. With reference to
The methods, systems, software, and product distribution media described herein, illustrate embodiments of the present invention wherein a computer program automatically creates a composite image product. Part of the power of the present invention is that it allows automatic asset selection whereby the computer system selects a subset of images in an intelligent fashion so that, for example, all the pictures in a collection need not be included in the image product. The number of assets selected may be determined by the output product desired. As another example, if a two-minute multimedia presentation is selected at a transition rate of four seconds per slide, this would require thirty images. This constraint may be specified as part of a programmed rule set.
The computer system may generate image products based on calendar entries that identify significant dates. The dates may be personally significant, such as anniversaries or birthdays, or they may be holidays such as Mother's Day or New Years Day. The data for these calendar dates may be input to the system by users or they may be inferred by the programmed computer system. One method for the system to infer dates that are significant to the user is to track dates when images are captured or uploaded to the system and the categories of those images. The system can then infer significant dates if particular types of images are captured or uploaded at the same time on an annual, or some other, basis.
With reference to
It will be understood that, although specific embodiments of the invention have been described herein for purposes of illustration and explained in detail with particular reference to certain preferred embodiments thereof, numerous modifications and all sorts of variations may be made and can be effected within the spirit of the invention and without departing from the scope of the invention. Accordingly, the scope of protection for this invention is limited only by the following claims and their equivalents.
This application claims priority under 35 U.S.C. §120 as a continuation-in-part application of commonly owned related U.S. patent application Ser. No. 11/758,358; filed on Jun. 5, 2007 now abandoned; entitled: “Automatic Story Creation Using Semantic Classifiers For Images And Associated Metadata”, by Catherine D. Newell et al.
Number | Name | Date | Kind |
---|---|---|---|
5499366 | Rosenberg et al. | Mar 1996 | A |
6230312 | Hunt | May 2001 | B1 |
6282317 | Luo et al. | Aug 2001 | B1 |
6351556 | Loui et al. | Feb 2002 | B1 |
6389181 | Shaffer et al. | May 2002 | B2 |
6396963 | Shaffer et al. | May 2002 | B2 |
6480840 | Zhu et al. | Nov 2002 | B2 |
6492998 | Kim et al. | Dec 2002 | B1 |
6504951 | Luo et al. | Jan 2003 | B1 |
6587839 | McIntyre et al. | Jul 2003 | B1 |
6606411 | Loui et al. | Aug 2003 | B1 |
6629104 | Parulski et al. | Sep 2003 | B1 |
6636648 | Loui et al. | Oct 2003 | B2 |
6697502 | Luo | Feb 2004 | B2 |
6766363 | Rothschild | Jul 2004 | B1 |
6771801 | Fisher et al. | Aug 2004 | B1 |
6782397 | Tanaka | Aug 2004 | B2 |
6865297 | Loui et al. | Mar 2005 | B2 |
6940545 | Ray et al. | Sep 2005 | B1 |
7035467 | Nicponski | Apr 2006 | B2 |
7062454 | Giannini et al. | Jun 2006 | B1 |
7092966 | McIntyre | Aug 2006 | B2 |
7110575 | Chen et al. | Sep 2006 | B2 |
7233684 | Fedorovskaya et al. | Jun 2007 | B2 |
7272788 | Anderson et al. | Sep 2007 | B2 |
7286723 | Taugher et al. | Oct 2007 | B2 |
7333967 | Bringsjord et al. | Feb 2008 | B1 |
7362919 | Das et al. | Apr 2008 | B2 |
7487072 | Semple et al. | Feb 2009 | B2 |
7586524 | Tsue et al. | Sep 2009 | B2 |
7711145 | Gallagher | May 2010 | B2 |
7889794 | Luo et al. | Feb 2011 | B2 |
7970240 | Chao et al. | Jun 2011 | B1 |
8135684 | Fedorovskaya et al. | Mar 2012 | B2 |
8201073 | Canora et al. | Jun 2012 | B2 |
8315463 | Gallagher et al. | Nov 2012 | B2 |
8330830 | Manico et al. | Dec 2012 | B2 |
8401334 | Neville et al. | Mar 2013 | B2 |
8570586 | Calhoon et al. | Oct 2013 | B2 |
20010032151 | Paul et al. | Oct 2001 | A1 |
20010041020 | Shaffer et al. | Nov 2001 | A1 |
20020016971 | Berezowski et al. | Feb 2002 | A1 |
20020154158 | Fukuda et al. | Oct 2002 | A1 |
20020161635 | Berg et al. | Oct 2002 | A1 |
20020199191 | Thursfield et al. | Dec 2002 | A1 |
20030128877 | Nicponski | Jul 2003 | A1 |
20030144923 | Capozzi et al. | Jul 2003 | A1 |
20030160824 | Szumla | Aug 2003 | A1 |
20030174217 | Kito et al. | Sep 2003 | A1 |
20040008180 | Appling, III | Jan 2004 | A1 |
20040044725 | Bell et al. | Mar 2004 | A1 |
20040081338 | Takenaka | Apr 2004 | A1 |
20040119875 | Hayaishi et al. | Jun 2004 | A1 |
20040172405 | Farran | Sep 2004 | A1 |
20040172419 | Morris et al. | Sep 2004 | A1 |
20040179719 | Chen et al. | Sep 2004 | A1 |
20040208377 | Loui et al. | Oct 2004 | A1 |
20040254859 | Aslanian, Jr. | Dec 2004 | A1 |
20050063613 | Casey et al. | Mar 2005 | A1 |
20050084232 | Herberger et al. | Apr 2005 | A1 |
20050105775 | Luo et al. | May 2005 | A1 |
20050105776 | Luo et al. | May 2005 | A1 |
20050111737 | Das et al. | May 2005 | A1 |
20050116027 | Algiene et al. | Jun 2005 | A1 |
20050131767 | Heins | Jun 2005 | A1 |
20050149970 | Fairhurst et al. | Jul 2005 | A1 |
20050206751 | Manico et al. | Sep 2005 | A1 |
20050216283 | Squilla et al. | Sep 2005 | A1 |
20060036949 | Moore et al. | Feb 2006 | A1 |
20060055992 | Koo | Mar 2006 | A1 |
20060061804 | Manico et al. | Mar 2006 | A1 |
20060080314 | Hubert et al. | Apr 2006 | A1 |
20060088207 | Schneiderman | Apr 2006 | A1 |
20060094462 | Nguyen et al. | May 2006 | A1 |
20060098941 | Abe et al. | May 2006 | A1 |
20060126944 | Loui et al. | Jun 2006 | A1 |
20060129606 | Horvitz et al. | Jun 2006 | A1 |
20060242178 | Butterfield et al. | Oct 2006 | A1 |
20060251292 | Gokturk et al. | Nov 2006 | A1 |
20060251382 | Vronay et al. | Nov 2006 | A1 |
20060253783 | Vronay et al. | Nov 2006 | A1 |
20060271486 | Cross et al. | Nov 2006 | A1 |
20070008321 | Gallagher et al. | Jan 2007 | A1 |
20070061759 | Klein, Jr. | Mar 2007 | A1 |
20070098303 | Gallagher et al. | May 2007 | A1 |
20070101271 | Hua et al. | May 2007 | A1 |
20070238082 | Ingrassia | Oct 2007 | A1 |
20070300158 | Kasperkiewicz et al. | Dec 2007 | A1 |
20080021894 | Styles | Jan 2008 | A1 |
20080034374 | Davis | Feb 2008 | A1 |
20080205789 | Ten Kate et al. | Aug 2008 | A1 |
20090029771 | Donahue | Jan 2009 | A1 |
20090052736 | Kacker | Feb 2009 | A1 |
20120185538 | Lee et al. | Jul 2012 | A1 |
Number | Date | Country |
---|---|---|
1 377 045 | Jan 2004 | EP |
1 443 428 | Aug 2004 | EP |
2006-510113 | Mar 2006 | JP |
0177776 | Oct 2001 | WO |
2006065223 | Jun 2006 | WO |
Entry |
---|
Office Action on Japanese Application 2010-511182, mailed Oct. 15, 2013. (English translation not available). |
D. E. Millard, C. Bailey, T. Brody, D. Dupplaw, W. Hall, S. Harris, K. R. Page, G. Power, and M. J. Weal, “Hyperdoc: An Adaptive Narrative System for Dynamic Multimedia Presentations,” University of Southampton ECSTR-IAM02-006, 2003. |
A. Gallagher, M. Das, and A. Loui, “User-Assisted People Search in Consumer Image Collections,” in IEEE Intern. Conf. on Multimedia and Expo (ICME), Beijing, China: IEEE, Jul. 2-5, 2007. |
J. Lou and M. Boutell, “Natural Scene Classification Using Overcomplete ICA,” Pattern Recognition, vol. 38, pp. 1507-1519, 2005. |
A. Savakis, S. Etz, and A. Loui, “Evaluation of Image Appeal in Consumer Photography,” in SPIE Human Vision and Electronic Imaging, Jan. 2000. |
A. Loui and A. Savakis, “Automated Event Clustering and Quality Screening of Consumer Pictures for Digital Alburning,” IEEE Transaction on Multimedia, vol. 5, pp. 390-402, 2003. |
J. Grant and D. Beckett, Eds., “RDF Test Cases,” World Wide Web Consortium, Feb. 2004. http://www.w3.org/TR/testcases/. |
International Preliminary Report on Patentability for PCT/US2008/007017, issued Dec. 7, 2009. |
International Search Report for PCT/US2008/007017, mailed Oct. 23, 2008. |
Non-Final Office Action on U.S. Appl. No. 11/758,358, mailed Sep. 8, 2010. |
Number | Date | Country | |
---|---|---|---|
20080304808 A1 | Dec 2008 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11758358 | Jun 2007 | US |
Child | 11935737 | US |