This application claims the benefit of the Spanish Patent Application No. ES P200930270, filed on Jun. 5, 2009, which is hereby incorporated by reference in its entirety.
Embodiments of the present invention relate to the field of digital photography, and more specifically in providing an automatic photo recommender according to objective and subjective parameters.
The amount of multimedia information generated by our users is dramatically increasing nowadays. In particular, the number of photographs that are taken, uploaded to web servers, needed to be reviewed and processed, or simply shared with others (e.g. after a holiday trip) is unmanageable by individuals. Then, there is a clear need to automate the process of photo selection.
For professional photographers the problem is basically the same: a big amount of photographs taken during sessions (e.g. fashion week events) which need to be reviewed, selected and post-processed. A tool which simplifies their workflow and facilitates their job in the darkroom is really a must for most of them. Take into account that most of professional photographers are freelance, and then they appreciate very much to dedicate more time to take photos and to deal with potential customers.
The selection of photos is currently done based on user's rating, tagging or any kind of metadata (e.g. camera model, shutter speed, f-stop, ISO and so on) but not considering other parameters regarding user's behaviour.
For instance, patent document US2004126038 discloses an external device that tags photographs from cameras. However the present invention makes use of tagging in a first step (then the better the tagging is, the better the results are) but another ones are required as prefiltering, filtering and tuning to sort them properly.
Document US2006173746 discloses an efficient methods for temporal event clustering of digital photographs. It deals with the clustering of photographs considering their time-stamps. The present invention is not related to the clustering of photographs but to find the optimal number of photos and its order to be shown them to each user.
Embodiments of the present invention relate to the automation of the very tedious process of photo selection. Disclosed methods take into account the user preferences, and maximizing the quality of the photos and the amount of information shown.
Disclosed embodiments provide a number of advantages, including but not limited to:
To provide these and other advantages, disclosed embodiments relate to methods for recommending photographs. In an example method, the recommended photographs are chosen from a set of digital photographs, and said recommendation is provided to at least one user. An example method comprises:
In the relevant function of the tuning step tuning weighs weighT are preferably assigned to each objective and subjective parameter, such that:
being:
k the number of the determined objective and subjective parameters,
weighTi the weigh assigned to the parameter i,
M the maximum score obtainable in the tuning step;
The score obtained for each photograph in the tuning step, tuning score scoret, can then be obtained by applying the following equation:
being fpi, equal to 0 in case parameter i is not present in said photograph, and 1 in case parameter i is present in said photograph.
The method can also comprise a prefiltering phase, before the tuning step, for discarding photographs according to objective and/or subjective parameters. The prefiltering phase can be carried out by configuring a threshold for at least one objective parameter and comparing said threshold with the value of the corresponding objective parameters obtained in the tagging step.
The method can also comprise a filtering step, after the tuning step. The filtering step preferably comprises:
The filtering weighs weighF can be assigned to each item such that:
being:
j the number of items identified,
weighFi the weigh assigned to the item i,
N the maximum score obtainable in the filtering step;
The score obtained for each photograph in the filtering step, filtering score scoref, can be obtained by applying the following equation:
being fi equal to 0 in case item i is not present in said photograph, and 1 in case item i is present in said photograph.
The items photographed can be, at least, of the following type: people, objects and places.
The total score scoreTOTAL of each photograph can be equal to its tuning score scoret, in case there is no filtering step. If there is a filtering step, the total score scoreTOTAL of each photograph is preferably such that scoreTOTAL=α·scoret+β·scoref, being α, β configurable parameters.
The recommended photographs finally showed can be a determined number T of photographs, with T<K, being K the total number of photographs of the set.
The relevant function in the tuning step can be automatically tuned in an iterative and learning process.
The objective parameters can be selected, at least, from the following:
The subjective parameters can be selected, at least, from the following:
Disclosed embodiments are also directed to a system for recommending photographs. The disclosed system comprises an electronic device which in turn comprises storing means in which the set of digital photographs are stored and data processing means configured to execute the method for recommending photographs previously discussed.
The electronic device can be a computer with display means configured to show the at least one user the photographs recommendation.
A series of drawings which aid in better understanding the invention and which are expressly related with an embodiment of said invention, presented as a non-limiting example thereof, are very briefly described below.
Embodiments of the present invention are directed to an advanced method to sort a set of photographs which maximizes the amount of information contained in this set and matches with the user's likes. The example method is based on the following steps:
1. Tagging.
2. Prefiltering (optional).
3. Tuning.
4. Filtering (optional).
5. Show results to the user.
1. Tagging
Based on objective parameters (as people and objects photographed, framing, main subject is well focused, inside/outside, closed eyes, smiling . . . ) and subjective parameters collected from the user's behaviour (as number of times it has been displayed, time you've spent watching it, it's been shared or not and the number of times, stars, more comments, explicitly selected by the user . . . ). The tagging phase can be automatically done, because the image recognition technology is mature enough to implement it.
Objective Parameters
There are other composition rules like:
Then, the photograph can be tagged attending to each one of these composition rules indicating if each one has been followed or not.
An example of a histogram can be the one represented in
Attending to this, a photograph can be tagged as:
The subjective parameters are collected from the users' behaviour. In that case two different scenarios can be distinguished: (a) when the set of photographs is locally stored, and (b) when the set of photographs is stored in a server (e.g. Flickr). The parameters change in each case:
(a) Local photos: photos stored in the local device, e.g. mobile phone or computer. For instance:
(b) Photos in a remote server like Flickr and accessed by the user using a device like a mobile phone or computer. For instance:
The tagging phase is automatically done, because the image recognition technology is mature enough to implement it.
2. Optional Prefiltering
Discard photos according to objective and/or subjective parameters to enhance the quality of the final photo selection. This can be done by configuring a threshold for each objective parameter and comparing these ones with those generated in the tagging step. For instance, if the main subject is out of focus (the main subject will be the biggest one and will be located in the proximity of one of the emphasis points according to the rule of thirds), or if the photo is overexposed or subexposed, etc. This can also be done by choosing only the photographs previously watched or marked by a user, discarding the rest of photographs.
3. Tuning
Apply over the previous ones the rest of criteria (objective and subjective parameters) weighing them according to a relevant function. This function will provide a score for each photo to arrange the set of photos. Optionally, the function could be automatically tuned in an iterative process, e.g. using a learning mechanism based on neural networks.
The relevant function can be defined in a similar way as the table shown in
4. Optional Filtering
The purpose of this step, which is optional, is to select the minimum number of photos that include the people, the objects and places photographed in a set of photos. It will follow the next steps:
If the final user indicates to the system that only a set of T photographs (T<K, K=total number of photos) should be selected, then the system will use the final scores to make the selection of the T photos (the T photos with higher score).
5. Show Results to the User
The photographs recommendation is shown to the user. The user can choose to iterate the procedure in step 3, tuning.
| Number | Date | Country | Kind |
|---|---|---|---|
| 200930270 | Jun 2009 | ES | national |
| Number | Name | Date | Kind |
|---|---|---|---|
| 6999624 | Nelson | Feb 2006 | B1 |
| 6999625 | Nelson | Feb 2006 | B1 |
| 7076097 | Kondo et al. | Jul 2006 | B2 |
| 7099860 | Liu et al. | Aug 2006 | B1 |
| 7529732 | Liu et al. | May 2009 | B2 |
| 7809192 | Gokturk et al. | Oct 2010 | B2 |
| 8165378 | Fritz et al. | Apr 2012 | B2 |
| 8194934 | Eura et al. | Jun 2012 | B2 |
| 8194986 | Conwell | Jun 2012 | B2 |
| 8311291 | Yonezawa et al. | Nov 2012 | B2 |
| 8351665 | Tearney et al. | Jan 2013 | B2 |
| 8385685 | Bedros et al. | Feb 2013 | B2 |
| 20040126038 | Aublant et al. | Jul 2004 | A1 |
| 20050058350 | Dugan et al. | Mar 2005 | A1 |
| 20060173746 | Cooper et al. | Aug 2006 | A1 |
| 20060210121 | Nakano et al. | Sep 2006 | A1 |
| 20060262976 | Hart et al. | Nov 2006 | A1 |
| 20070047819 | Hull et al. | Mar 2007 | A1 |
| 20070177805 | Gallagher | Aug 2007 | A1 |
| 20070297675 | Lee et al. | Dec 2007 | A1 |
| 20080181508 | Kaneda et al. | Jul 2008 | A1 |
| 20080310725 | Kurata et al. | Dec 2008 | A1 |
| 20080313214 | Duhig et al. | Dec 2008 | A1 |
| 20090003707 | Fukuda et al. | Jan 2009 | A1 |
| 20090024580 | Obrador | Jan 2009 | A1 |
| 20090064008 | Liu et al. | Mar 2009 | A1 |
| 20090297010 | Fritz et al. | Dec 2009 | A1 |
| 20090324049 | Kontos et al. | Dec 2009 | A1 |
| 20100049740 | Iwase et al. | Feb 2010 | A1 |
| 20100086176 | Yokono et al. | Apr 2010 | A1 |
| 20100172555 | Hasezawa et al. | Jul 2010 | A1 |
| 20100312609 | Epshtein et al. | Dec 2010 | A1 |
| 20110002543 | Almodovar Herraiz et al. | Jan 2011 | A1 |
| 20110208716 | Liu et al. | Aug 2011 | A1 |
| 20120177294 | Ke et al. | Jul 2012 | A1 |
| 20120207360 | Mehanian et al. | Aug 2012 | A1 |
| 20120314913 | Bedros et al. | Dec 2012 | A1 |
| Number | Date | Country |
|---|---|---|
| 1202187 | May 2002 | EP |
| Entry |
|---|
| Kim et al Relevance Feedback Using Adaptive Clustering for Content Based Image Retrieval, In Proc. of the ACM SIGMOD Int. Conf. on Management of Data, 2003. |
| Number | Date | Country | |
|---|---|---|---|
| 20110002543 A1 | Jan 2011 | US |