Method and apparatus for photograph finding

Information

  • Patent Grant
  • 11625441
  • Patent Number
    11,625,441
  • Date Filed
    Friday, July 9, 2021
    3 years ago
  • Date Issued
    Tuesday, April 11, 2023
    a year ago
Abstract
Digital image data including discrete photographic images of a variety of different subjects, times, and so forth, are collected and analyzed to identify specific features in the photographs. In an embodiment of the invention, distinctive markers are distributed to aid in the identification of particular subject matter. Facial recognition may also be employed. The digital image data is maintained in a database and quarried in response to search requests. The search requests include criteria specifying any feature category or other identifying information, such as date, time, and location that each photograph was taken, associated with each photograph. Candidate images are provided for review by requesters, who may select desired images for purchase or downloading.
Description
BACKGROUND OF THE INVENTION
Field of the Invention

Photography has transformed how people conceive of the world. Photographs allow people to see all sorts of things that are actually many miles away and/or years preceding. Photography lets people capture moments in time and preserve them for years to come.


Often people at a public place notice that a stranger has taken a photograph of which they would love to have a copy, Alternatively, after going somewhere, a person may bemoan the fact that he did not have a photograph of the event (in the present context, photograph also includes video, audio, or other representation).


A need exists, therefore, to provide a method and apparatus for identifying and connecting people with photographs they want. In addition, there is a need to provide a method and apparatus for characterizing errant photographs stored on computer databases that makes use of a variety of attributes to reliably characterize photographs in such a way as to reduce the amount of manual review necessary to identify and connect people with the photographs they want.


SUMMARY OF THE INVENTION

The present invention provides a method and apparatus that matches people with photographs in which they accidentally (or purposely) appear or with photographs of events they have attended.


Specifically, in one embodiment, a web site is created with a database backend. The database is seeded with information provided by (1) the photographer; (2) recovering metadata from the photograph; (3) reading devices such as a Global Positioning System (GPS) device; (4) referencing the account data of the photographer (i.e., account number, photographer's zip code or area code, etc.); (5) analyzing the photograph (i.e., computer recognizes eye color, optical character recognizes any text found in the photograph, recognizes the number of persons, the gender of persons, the hair color, the time of day by optical character recognizing any clocks in the photograph or analyzing the lighting conditions, the weather, etc.); (6) photograph quality information; and/or (7) any other information.


A user looking for a photograph would visit the web site and search for certain criteria. The user is then provided with a gallery of thumbnails that match the criteria. When the user identifies a photograph he wants to own, he can then download the full quality version, or order print(s). In a preferred implementation, the user is charged some amount of money that is split between the site owner and the photographer. Alternatively, the user may be charged in some other way, such as by viewing advertisements or by exchanging credits for downloads or by some other payment or a combination thereof.


A more complete understanding of the present invention will be afforded to those skilled in the art, as well as a realization of additional advantages and objects thereof, by a consideration of the following detailed description of the preferred embodiments.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a flow diagram showing exemplary steps of a method according to the invention.



FIG. 2 is a diagram showing an exemplary distinctive marker for photographic data.





DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS

The present invention provides a method and apparatus that matches people with photographs in which they accidentally (or purposely) appear or with photographs of events they have attended.



FIG. 1 illustrates exemplary steps of a method 100 according to the invention. At optional step 102, distinctive markers may be distributed to persons desiring to contribute photographic images to a database. The markers may comprise, for example, distinctive bins, badges, or stickers for placing on objects to be photographed. The markers should be designed so as to be easily recognized using automatic recognition algorithms, but should not be too conspicuous.


At step 104, image data is collected from a variety of sources. It may be desirable to accept material from as many sources as possible, to increase the number of images available for browsing. Optionally, images may be accepted from qualified sources only.


At step 104, source information regarding each photograph is collected. For example, at step 108, the image provider may be asked for information, such as the time and date of the photograph, the subject matter, location, photographer, etc. Image metadata may also be read to obtain such information. Other ways of associating images to identifying information may include, for example, reading devices such as a Global Positioning System (GPS) device attached to a camera or other image capture device, or by referencing account data of the image contributor (e.g., account number, photographer's zip code or area code, etc.).


At step 110, image data is analyzed to identify any characteristics that may be of interest to users. Such characteristics may include, for example, eye color, words and sentences, a number or gender of persons, the hair color, time of day, lighting conditions, and so forth. For further example, at step 112, a facial recognition program as known in the art may be used to analyze any faces appearing in the photos at a sufficiently high resolution. At step 114, the images may be analyzed for the presence of any known markers. And at step 116, other features and qualities of the image may be classified, for example, whether it is taken indoors or outdoors, whether it contains people, dogs, cats, or other animals, whether it contains automobiles, airplanes, or other objects, and so forth. At step 118, selected feature information and other source information is associated with each image and provided to any suitable relational database.


At step 120, requests specifying search criteria for photographic images are received. For example, a user looking for a photograph may visit a web site hosted by the system and fill out a search form specifying search criteria of interest. The criteria may include specific subject matter, times, dates, and locations. For example, “Disneyland AND Matterhorn AND blue eye AND child AND Jan. 1, 2004 AND morning” would search for a photograph or photographs taken at Disneyland's Matterhorn with a child who has blue eyes on the morning of Jan. 1, 2004.


At step 122, the image database is queried as known in the art, to identify images that at least partially match the search criteria. Such images may be presented, at step 124, to the user. For example, the user may be provided with a gallery of “thumbnail” (reduced-size) images generated from images that match the criteria. When the user identifies a photograph he wants to own, he can then download the full quality version, or order print(s). In a preferred implementation, the user is charged some amount of money that is split between the site owner and the photographer. Alternatively, the user may be charged in some other ways such as by viewing advertisements or by exchanging credits for downloads or by some other payment or a combination thereof. The price can be on a sliding scale depending on the quality of the photograph that the user downloads or the size or quality of the print. For example, a photograph may cost $1.00 for 1024×768 resolution or $2.00 for 1600×1200 resolution. Similarly, a print may cost $1.00 for 3×5 or $5 for 8×10. For downloads, an “upgrade” may be possible by charging the difference between the resolutions. An automated process may be used to reduce the number of pixels for purposes of having a lower quality version to sell.


In addition, a surcharge may be applied (even if no surcharge is required) for various enhancements to the photograph, such as “upconverting” to a higher resolution, eliminating red-eye, enhancing shadow, color, or brightness, etc.


Moreover, when a photographer takes photographs, he can be provided with printed cards bearing a Uniform Resource Locator (URL) and a unique code in order that the user would be able to enter into the web site to find the photograph or the series of photographs then being taken. The photographer can also distribute cards (the printed cards bearing the URL and the unique code or any other cards known to those skilled in the art) to people whom he photographs, whether intentionally or inadvertently. The photographer can further advertise the same (e.g., the URL and the unique code) via a mark on his camera, a T-shirt, or other means.


Fixed-place cameras can also serve this function (e.g., the of photographer). For example, a camera set up at an intersection in Hollywood might take and upload one photograph every 10 seconds.


Photographers can also be given accounts and be allowed to upload photographs to the site. The database is populated during this process, although additional database information can be added later by web site users. In addition, the number of times the photograph has been purchased and/or viewed can be a part of the database.


In one embodiment, the method and apparatus of the present invention should be capable of face recognition. It should assign values to various factors (i.e., ratio of distance between pupils to distance to tip of nose, etc.). It would add this information to the database for uploaded photographs. A user can then upload a photograph of the target person and the system would then generate the same data from that photograph and use it to limit the possible search targets.


A provider of the present method and apparatus or a photographer can also hand out pins, clothing, or other materials that are marked in a way that allows a computer to later recognize them in a photograph. FIG. 2 shows an exemplary distinctive marker 200 having an optical code 202, such as a bar code. The marker may have a color combination, distinctive shape, lettering, bar code, or other optical pattern, or some combination of the foregoing, that is unique to the marker. The marker may be computer generated, for example, and produced using an end-user's laser or ink-jet printer. The marker may be associated with specific information, for example, a particular user account, photographer, subject matter type, person, event, or location. Users can later search for photographs containing an image of the marker.


Numerous distribution mechanisms exist whereby photographs may be distributed from a source over a wide area network, such as the Internet. In some cases, the photographs are distributed using a centralized server system (such as Napster 2.0, eBay, or from a web site). In other cases, the photographs are distributed using a decentralized system (such as Gnutella). In a preferred implementation, the photographs are distributed to a person using the centralized server system or using a central hub.


Embodiments of the present invention operate in accordance with at least one web-hosting mechanism and a plurality of user mechanisms communicating over a wide area network, such as the Internet. Specifically, a web-hosting mechanism includes a database, an interface application and a server, wherein the server is adapted to communicate with a plurality of user mechanisms over a wide area network. It should be appreciated that the mechanisms described can include, but are not limited to, personal computers, mainframe computers, personal digital assistances, wireless communication devices and all other physical and wireless connected network devices generally known to those skilled in the art. It should further be understood that the database depicted can include, but is not limited, to RAM, cache memory, flash memory, magnetic disks, optical disks, removable disks, SCSI disks, IDE hard drives, tape drives, and all other types of data storage devices (and combinations thereof, such as RAID devices) generally known to those skilled in the art. In addition, the mechanisms described above are for purposes of example only and the invention is not limited thereby.


Having thus described several embodiments for photograph finding, it should be apparent to those skilled in the art that certain advantages of the system have been achieved. It should also be appreciated that various modifications, adaptations, and alternative embodiments thereof may be made within the scope and spirit of the present invention. For example, in the context of the present invention a photograph can include video, audio, and/or other representation of how people conceive of the world. The invention is defined by the following claims.

Claims
  • 1. A method, comprising: a computer system receiving a search request that includes a digital image;the computer system determining a set of multiple characteristics of the digital image, including: one or more objects appearing in the digital image;one or more colors appearing in the digital image; andlighting conditions appearing in the digital image;the computer system comparing the set of multiple characteristics with corresponding characteristics for respective ones of a plurality of stored digital images; andthe computer system transmitting a response to the search request, wherein the response indicates one or more stored digital images that meet one or more matching criteria for the comparing.
  • 2. The method of claim 1, wherein the set of multiple characteristics further includes one or more characteristics of the following characteristics: faces recognized in the digital image;time of day represented in the digital image;a device used to capture the digital image; anda geographic location at which the digital image was captured.
  • 3. The method of claim 1, wherein the one or more objects include visual facial characteristics.
  • 4. The method of claim 1, wherein the one or more objects include a known marker obj ect.
  • 5. The method of claim 1, wherein at least one of the multiple characteristics is determined based on image metadata and at least one of the multiple characteristics is determined based on analyzing image data of the digital image.
  • 6. The method of claim 1, wherein the search request is received via a wide area network, the digital image is uploaded via the wide area network, and the response is transmitted via the wide area network.
  • 7. The method of claim 6, wherein the transmitting the response includes causing reduced-size versions, of the one or more stored digital images that meet one or more matching criteria, to be displayed via a web site.
  • 8. The method of claim 1, further comprising: the computer system receiving one or more search parameters based on user input;wherein the comparing includes comparing characteristics for stored digital images with the received one or more search parameters.
  • 9. A system, comprising: one or more processors; andone or more memories having program instructions stored thereon that are executable by the one or more processors to: receive a search request that includes a digital image;determine a set of multiple characteristics of the digital image, including: one or more objects appearing in the digital image;one or more colors appearing in the digital image; andlighting conditions appearing in the digital image;compare the set of multiple characteristics with corresponding characteristics for respective ones of a plurality of stored digital images; andtransmit a response to the search request, wherein the response indicates one or more stored digital images that meet one or more matching criteria for the comparison.
  • 10. The system of claim 9, wherein the set of multiple characteristics further includes the following characteristics: faces recognized in the digital image; andtime of day represented in the digital image.
  • 11. The system of claim 10, wherein the set of multiple characteristics further includes the following characteristics: a device used to capture the digital image; anda geographic location at which the digital image was captured.
  • 12. The system of claim 9, wherein at least one of the multiple characteristics is determined based on image metadata and at least one of the multiple characteristics is determined based on an analysis of pixel data of the digital image.
  • 13. The system of claim 9, wherein the search request is received via a wide area network, the digital image is uploaded via the wide area network, and the response is transmitted via the wide area network.
  • 14. A non-transitory computer-readable medium having instructions stored thereon that are executable by a computing device to perform operations comprising: receiving a search request that includes a digital image;determining a set of multiple characteristics of the digital image, including: one or more objects appearing in the digital image;one or more colors appearing in the digital image; andlighting conditions appearing in the digital image;comparing the set of multiple characteristics with corresponding characteristics for respective ones of a plurality of stored digital images; andtransmitting a response to the search request, wherein the response indicates one or more stored digital images that meet one or more matching criteria for the comparing.
  • 15. The non-transitory computer-readable medium of claim 14, wherein the set of multiple characteristics further includes one or more characteristics of the following characteristics: faces recognized in the digital image;time of day represented in the digital image;a device used to capture the digital image; anda geographic location at which the digital image was captured.
  • 16. The non-transitory computer-readable medium of claim 14, wherein at least one of the multiple characteristics is determined based on image metadata and at least one of the multiple characteristics is determined based on analyzing image data of the digital image.
  • 17. The non-transitory computer-readable medium of claim 14, wherein the search request is received via a wide area network, the digital image is uploaded via the wide area network, and the response is transmitted via the wide area network.
  • 18. The non-transitory computer-readable medium of claim 14, wherein the transmitting the response includes causing reduced-size versions, of the one or more stored digital images that meet one or more matching criteria, to be displayed via a web site.
  • 19. The non-transitory computer-readable medium of claim 14, wherein the operations further comprise: receiving one or more search parameters based on user input;wherein the comparing includes comparing characteristics for stored digital images with the received one or more search parameters.
  • 20. The non-transitory computer-readable medium of claim 14, wherein the one or more objects include visual facial characteristics.
CROSS-REFERENCE TO RELATED APPLICATIONS

The present application is a continuation of U.S. application Ser. No. 17/020,671, filed Sep. 14, 2020 (now U.S. Pat. No. 11,061,971), which is a continuation of U.S. application Ser. No. 16/105,876, filed Aug. 20, 2018 (now U.S. Pat. No. 10,776,430), which is a continuation of U.S. application Ser. No. 15/481,595, filed Apr. 7, 2017 (now U.S. Pat. No. 10,055,497), which is a continuation of U.S. application Ser. No. 14/518,655, filed Oct. 20, 2014 (now U.S. Pat. No. 9,619,486), which is a continuation of applicaton Ser. No. 13/776,463, filed Feb. 25, 2013 (now U.S. Pat. No. 8,867,798), which is a continuation of U.S. application Ser. No. 13/090,026, filed Apr. 19, 2011 (now U.S. Pat. No. 8,385,691), which is a continuation of U.S. application Ser. No. 12/874,929, filed Sep. 2, 2010 now U.S. Pat. No. 7,929,810), which is a continuation of U.S. application Ser. No. 12/325,589 filed Dec. 1, 2008 (now U.S. Pat. No. 7,844,141), which is a continuation of U.S. application Ser. No. 11/056,699, filed Feb. 10, 2005 (now Pat. No. 7,460,7371 which claims priority to U.S. Provisional Appl. No. 60/544,570, filed Feb. 12, 2004 the disclosures of each of the above-referenced applications are incorporated by reference herein in their entireties.

US Referenced Citations (99)
Number Name Date Kind
5280530 Trew et al. Jan 1994 A
5748904 Huang et al. May 1998 A
5793371 Deering Aug 1998 A
5835244 Bestmann Nov 1998 A
5848410 Walls et al. Dec 1998 A
5912980 Hunke Jun 1999 A
5926647 Adams et al. Jul 1999 A
5926812 Hilsenrath et al. Jul 1999 A
6094649 Bowen et al. Jul 2000 A
6353840 Saito et al. Mar 2002 B2
6397213 Cullen et al. May 2002 B1
6421675 Ryan et al. Jul 2002 B1
6424587 Hosoda Jul 2002 B1
6530944 West et al. Mar 2003 B2
6564214 Bhide May 2003 B1
6608923 Zhang et al. Aug 2003 B1
6654758 Teague Nov 2003 B1
6658408 Yano et al. Dec 2003 B2
6690828 Meyers Feb 2004 B2
6738653 Sfez et al. May 2004 B1
6772150 Whitman et al. Aug 2004 B1
6778697 Shin et al. Aug 2004 B1
6792434 Moghaddam et al. Sep 2004 B2
6873806 Kobayashi et al. Mar 2005 B2
6894494 Stergiopoulos et al. May 2005 B2
6898582 Lange May 2005 B2
6919892 Cheiky Jul 2005 B1
6963663 Yoshida Nov 2005 B1
6975418 Ohta et al. Dec 2005 B1
6982811 Sato Jan 2006 B2
7006881 Hoffberg Feb 2006 B1
7027054 Cheiky et al. Apr 2006 B1
7110996 Kawamura Sep 2006 B2
7112806 Lussier Sep 2006 B2
7140535 Sussman Nov 2006 B2
7171058 Luo Jan 2007 B2
7174332 Baxter et al. Feb 2007 B2
7181480 Nikiel et al. Feb 2007 B1
7184814 Lang et al. Feb 2007 B2
7209653 Umeda Apr 2007 B2
7231243 Tearney et al. Jun 2007 B2
7233684 Fedorovskaya et al. Jun 2007 B2
7236264 Kanno Jun 2007 B2
7259785 Stavely et al. Aug 2007 B2
7266544 Dodge et al. Sep 2007 B1
7286723 Taugher et al. Oct 2007 B2
7286871 Cohen Oct 2007 B2
7287253 Yamamura et al. Oct 2007 B2
7298931 Kim et al. Nov 2007 B2
7302057 Rotholtz et al. Nov 2007 B2
7308550 Cornett Dec 2007 B2
7319780 Fedorovskaya et al. Jan 2008 B2
7333963 Widrow et al. Feb 2008 B2
7346631 Amirghodsi Mar 2008 B2
7356512 Okita et al. Apr 2008 B2
7359535 Salla et al. Apr 2008 B2
7366994 Loui Apr 2008 B2
7379627 Li et al. May 2008 B2
7382903 Ray Jun 2008 B2
7388693 Ahmed et al. Jun 2008 B2
7418116 Fedorovskaya et al. Aug 2008 B2
7428744 Ritter Sep 2008 B1
7440594 Takenaka Oct 2008 B2
7474808 Ozaki et al. Jan 2009 B2
7499916 Liu et al. Mar 2009 B2
7555148 Steinberg et al. Jun 2009 B1
7562299 Millar et al. Jul 2009 B2
7564994 Steinberg et al. Jul 2009 B1
7583294 Ray et al. Sep 2009 B2
7596247 Ioffe Sep 2009 B2
7609864 Nagasaka et al. Oct 2009 B2
7639898 Chan Dec 2009 B1
7660445 Ray Feb 2010 B2
7706633 Chefd'hotel et al. Apr 2010 B2
7711145 Gallagher May 2010 B2
7715659 Zhao et al. May 2010 B2
7813822 Hoffberg Oct 2010 B1
7843495 Aas et al. Nov 2010 B2
7860320 Luo et al. Dec 2010 B2
7881506 Nagasaka et al. Feb 2011 B2
7921020 Kalies Apr 2011 B2
7941197 Jain et al. May 2011 B2
7941442 Li May 2011 B2
8005272 Grim, III et al. Aug 2011 B2
8064650 Webb Nov 2011 B2
8065611 Chan et al. Nov 2011 B1
8150584 Mackin et al. Apr 2012 B2
8199979 Steinberg et al. Jun 2012 B2
8224849 Li Jul 2012 B2
8316237 Felsher Nov 2012 B1
8374466 Webb Feb 2013 B2
8460103 Mattice et al. Jun 2013 B2
8513513 Komano Aug 2013 B2
8531474 Chang et al. Sep 2013 B2
8543573 MacPherson Sep 2013 B2
8583263 Hoffberg Nov 2013 B2
9311619 Mosley Apr 2016 B2
9547883 Stepanenko Jan 2017 B1
9560415 Good Jan 2017 B2
Related Publications (1)
Number Date Country
20210406323 A1 Dec 2021 US
Provisional Applications (1)
Number Date Country
60544570 Feb 2004 US
Continuations (9)
Number Date Country
Parent 17020671 Sep 2020 US
Child 17371595 US
Parent 16105876 Aug 2018 US
Child 17020671 US
Parent 15481595 Apr 2017 US
Child 16105876 US
Parent 14518655 Oct 2014 US
Child 15481595 US
Parent 13776463 Feb 2013 US
Child 14518655 US
Parent 13090026 Apr 2011 US
Child 13776463 US
Parent 12874929 Sep 2010 US
Child 13090026 US
Parent 12325589 Dec 2008 US
Child 12874929 US
Parent 11056699 Feb 2005 US
Child 12325589 US