Not Applicable
Not Applicable
Not Applicable
The present invention relates to modifying video images, in particular varying features of a person on live video.
Software has existed for some time to modify images, such as a person's face. U.S. Pat. No. 5,825,941 is an example of software allowing a patient to see the effects of proposed plastic surgery before it is done. Other software allows users to modify static pictures that can be saved and printed, such as by varying the features of the face or morphing two faces together. US Published Application No. 20050026685 shows a system for morphing video game characters. US Published Application No. 20040201666 shows using user voice or other inputs to modify a virtual character.
A variety of applications are directed to modifying avatars for use with instant messaging. These applications allow a user to personalize the avatar image that is the representation of the user. For example, U.S. Published Application No. 20030043153 shows determining facial expressions from a webcam video feed and mapping those expressions to an avatar. An animation vector is mapped to a target mix vector. Each morph target may represent a facial expression, such as a neutral facial expression, a surprised facial expression, an angry facial expression, and a smiling facial expression.
The present invention provides a method for modifying, or morphing, a video image. A video feed is received which include an image of a person's face. A number of facial features are detected from the video feed. One or more of those facial features are selected and modified, with the modification being applied to the video feed to produce a morphed image. For example, the morphing can produce fun effects such as oversized or too small eyes, lips, etc.
In one embodiment, the user can choose to modify particular features directly. For example, the user can thicken the lips, widen the mouth, etc. The user can then do similar modifications separately to the eyes, nose, ears, etc. Additionally, the user can select to have a hat or other accessories superimposed over the face.
In another embodiment, the user can select from multiple expressions. The selected expression would then automatically change the particular features required to achieve that expression. For example, the expression for a smile can cause the lips and mouth to raise in the smile expression, and additionally can cause smile lines or dimples and a twinkling appearance in the eyes. Alternately another “expression” such as a younger age could cause the smoothing of the skin to eliminate wrinkles, modification of the facial features to produce a more defined chin, eyes, etc.
In one embodiment, a video feed from a web cam, such as a USB video feed is provided through a camera driver to feature recognition software. That software then provides the location of the identified features to a separate morphing software module. The morphing software module then responds to user inputs to select and apply modifications to the various facial features. These modifications are then applied to the video feed, to produce an output video feed which is only slightly delayed in real time.
System
The modified video pipe can be used by any application that uses live video input. This includes instant messaging applications (MSN Messenger, Yahoo Messenger, AOL messenger, etc.) as well as video capture applications (when—for example—an artist is capturing the video to the disk). This means that the morphing can be applied before the video is captured to the disk and stored or before it is sent to the interlocutor on the other side of the communication channel. This is because the morphing engine is integrated into the system-level part of the software, which makes it transparent to the applications using the video stream that lay above the system layer at the user layer. In this way the present invention is generic—i.e. it can coexist with any video application and modify its input according to the settings on quick assistant.
Computer 14 includes a camera driver 16 which receives the live video and provides it to feature recognition software module 18. Feature recognition module 18 could be any software which extracts information regarding facial features from a video image, such as the software produced by Neven Vision, a machine vision technology company with offices in Santa Monica, Calif.
The detected information regarding facial features is provided to a morphing module 20. The morphing module responds to user inputs to select facial features, determine the type of morphing, and apply the morphing. The user input can be provided through any peripheral, such as keyboard 22 or mouse 24 with associated keyboard driver 26 and mouse driver 28. After the morphing has been applied, the video feed is provided to application software 30. The application software could be an instant messenger application, a video conferencing application, or any other application which uses video. In one embodiment, the delay caused by the use of the feature recognition and morphing software is less than one frame of video.
Morphing Quick Assistant
Options
The morphing module allows a virtual makeover to be applied to a person. The size of the eyes can be changed, and their location can be modified slightly. The iris color can be changed. The lips could be made wider or narrower, or could be accentuated in other ways. A variety of other modifications are available.
Manual Selection
When Lips Enhancement is selected the box 58 of
Expression Selection
Preview Screen
Variations
In one embodiment, the deformation applied to a face (e.g., ‘fuller lips’) would move with the subject. In some prior art applications, the deformation is applied to frontal face. In an embodiment of the present invention, the 3D position of the head and the features is localized and the deformation is adjusted from a frontal position to the actual position of the user. Thus, even if the face is not frontal, the deformation (e.g. full lips) is maintained when the subject rotates his head so the lips or other feature would still look natural, but they would be ‘fuller’. This is accomplished with a 3D image being stored and modified for each feature, or with algorithms for adjustment in 3D on the fly being used.
In one embodiment, the present invention is able to apply the modifications to live video on the fly with a delay of less than one frame. This is possible because only certain features of the face are morphed or modified, not the entire face. In one embodiment, less than 20% or less than 10% of the video is modified. This limits the amount of processing necessary and increases the speed at which the morphing can be done.
In another embodiment, the user selects the morphing to be applied before the video is fed to the application software program. In another embodiment, the user can also do the adjustments on the fly, morphing or changing the video image in the middle of an instant messenger conversation, for example.
In one embodiment, the invention is integrated into the video pipe so that it is transparent to any application program. The application program need not be modified to accommodate the present invention, which is applied at the driver level of the video feed.
In another embodiment, popular morphing combinations are stored as a macro or particular expression. These are provided on a website which is accessible by users, and users can create their own and upload them to the website for sharing with other users. Morphing to produce features similar to those of celebrities can be stored as a macro on the web site for downloading.
In one embodiment, such as for a conference call with multiple people present, the morphing can be applied independently to different facial features. Thus, one person could be made to have wider eyes, another can be made smile all the time, another could be made to be anger, and another could be made to look younger.
In addition, an entire image can be morphed with, for example, effects similar to those found in a fun house mirror room. By keeping the effects simple, they can be applied in real time without undue delay. Delay is also cut down by the ability to bypass the facial feature detection software, since this is not needed when the entire image is morphed. In one embodiment, the morphing causes a delay of less than 20 milliseconds.
As will be understood as those of skill in the art, the present invention could be embodied in other specific forms without departing from the essential characteristics of the invention. For example, the morphing could be applied to features of the face of a pet, such as a dog, or any other image. The morphing module could also be made to appear upon initiation of an application program which uses still pictures, such as the sending of snapshots from a cell phone. The morphing engine could be made to appear anytime a video feed is requested by an application program, or only for certain types of application programs. Accordingly, reference should be made to the appended claims which set forth the scope of the claimed invention.
Number | Name | Date | Kind |
---|---|---|---|
2297844 | Shoemaker | Oct 1942 | A |
3941925 | Busch et al. | Mar 1976 | A |
4021847 | Van Atta | May 1977 | A |
4160263 | Christy et al. | Jul 1979 | A |
4206965 | McGrew | Jun 1980 | A |
4234894 | Tokumaru et al. | Nov 1980 | A |
4255764 | Howe | Mar 1981 | A |
4267561 | Karpinsky et al. | May 1981 | A |
4306252 | Fernside | Dec 1981 | A |
4337482 | Coutta | Jun 1982 | A |
4340905 | Balding | Jul 1982 | A |
4371893 | Rabeisen | Feb 1983 | A |
4383279 | Kenney, II | May 1983 | A |
4398211 | Young | Aug 1983 | A |
4411489 | McGrew | Oct 1983 | A |
4445140 | Netzer | Apr 1984 | A |
4470816 | Marshall et al. | Sep 1984 | A |
4491868 | Berridgem, Jr. et al. | Jan 1985 | A |
4602280 | Maloomian | Jul 1986 | A |
4658297 | Nomura et al. | Apr 1987 | A |
4668095 | Maeda | May 1987 | A |
4682220 | Beurskens | Jul 1987 | A |
4727416 | Cooper et al. | Feb 1988 | A |
4731666 | Csesznegi | Mar 1988 | A |
4731743 | Blancato | Mar 1988 | A |
4757387 | Saito | Jul 1988 | A |
4763146 | Niikura | Aug 1988 | A |
4776796 | Nossal | Oct 1988 | A |
4777527 | Camps et al. | Oct 1988 | A |
4795253 | Sandridge et al. | Jan 1989 | A |
4795261 | Nakata et al. | Jan 1989 | A |
4800432 | Barnett et al. | Jan 1989 | A |
4804983 | Thayer, Jr. | Feb 1989 | A |
4809247 | Elliott | Feb 1989 | A |
4815845 | Colbaugh et al. | Mar 1989 | A |
4823285 | Blancato | Apr 1989 | A |
4827532 | Bloomstein | May 1989 | A |
4844475 | Saffer et al. | Jul 1989 | A |
4845641 | Ninomiya et al. | Jul 1989 | A |
4864410 | Andrews et al. | Sep 1989 | A |
4868661 | Takahashi | Sep 1989 | A |
4872056 | Hicks et al. | Oct 1989 | A |
4888668 | Roll | Dec 1989 | A |
4890159 | Ogin | Dec 1989 | A |
4896175 | Thayer, Jr. | Jan 1990 | A |
4897827 | Raetzer et al. | Jan 1990 | A |
4899921 | Bendat et al. | Feb 1990 | A |
4903068 | Shiota | Feb 1990 | A |
4908700 | Ishii et al. | Mar 1990 | A |
RE33224 | Speigelstein | May 1990 | E |
4933773 | Shiota et al. | Jun 1990 | A |
4934773 | Becker | Jun 1990 | A |
4937665 | Schiffman | Jun 1990 | A |
4945379 | Date et al. | Jul 1990 | A |
4959670 | Thayer, Jr. | Sep 1990 | A |
4965819 | Kannes | Oct 1990 | A |
4969040 | Gharaui | Nov 1990 | A |
4971312 | Weinreich | Nov 1990 | A |
4973890 | Desjardins | Nov 1990 | A |
4979815 | Tsikos | Dec 1990 | A |
4987432 | Landwehr | Jan 1991 | A |
4991005 | Smith | Feb 1991 | A |
4995068 | Chou et al. | Feb 1991 | A |
4999586 | Meyer et al. | Mar 1991 | A |
5005148 | Behensky et al. | Apr 1991 | A |
5012274 | Dolgoff | Apr 1991 | A |
5025320 | Kyhl | Jun 1991 | A |
5027138 | Gandrud | Jun 1991 | A |
5027219 | Stuttler et al. | Jun 1991 | A |
5032863 | Date et al. | Jul 1991 | A |
5034618 | Akeel et al. | Jul 1991 | A |
5034822 | Stevens | Jul 1991 | A |
5035475 | Lee et al. | Jul 1991 | A |
5051872 | Anderson | Sep 1991 | A |
5060063 | Freeman | Oct 1991 | A |
5060171 | Steir et al. | Oct 1991 | A |
5061996 | Schiffman | Oct 1991 | A |
5075769 | Allen et al. | Dec 1991 | A |
5076275 | Bechor et al. | Dec 1991 | A |
5086223 | Hidaka | Feb 1992 | A |
5086354 | Bass et al. | Feb 1992 | A |
5096287 | Kakunami et al. | Mar 1992 | A |
5117283 | Kroos et al. | May 1992 | A |
5117285 | Nelson et al. | May 1992 | A |
5127032 | Lam et al. | Jun 1992 | A |
5129716 | Holakovszky et al. | Jul 1992 | A |
5132804 | Takahashi | Jul 1992 | A |
5137450 | Thomas | Aug 1992 | A |
5138647 | Nguyen et al. | Aug 1992 | A |
5140428 | Park | Aug 1992 | A |
5142414 | Koehler | Aug 1992 | A |
5150215 | Shi | Sep 1992 | A |
5153752 | Kurematsu et al. | Oct 1992 | A |
5160265 | Kim | Nov 1992 | A |
5160948 | Bae | Nov 1992 | A |
5162787 | Thompson et al. | Nov 1992 | A |
5164834 | Fukuda et al. | Nov 1992 | A |
5168354 | Martinez et al. | Dec 1992 | A |
5184228 | Kobayashi et al. | Feb 1993 | A |
5189490 | Shetty et al. | Feb 1993 | A |
5190286 | Watanebe et al. | Mar 1993 | A |
5196949 | Swanberg | Mar 1993 | A |
5198305 | Wada et al. | Mar 1993 | A |
5206629 | Demond et al. | Apr 1993 | A |
5218441 | Karcher | Jun 1993 | A |
5218626 | Mckee | Jun 1993 | A |
5223927 | Kageyama et al. | Jun 1993 | A |
5231432 | Glenn | Jul 1993 | A |
5235420 | Gharaui | Aug 1993 | A |
5249967 | O'Leary et al. | Oct 1993 | A |
5257091 | Caicedo, Jr. et al. | Oct 1993 | A |
5274464 | Strolle et al. | Dec 1993 | A |
5274498 | Rios-Rivera et al. | Dec 1993 | A |
5277499 | Kameyama | Jan 1994 | A |
5283433 | Tsien | Feb 1994 | A |
5296924 | de Saint Blancard et al. | Mar 1994 | A |
5299274 | Wysocki et al. | Mar 1994 | A |
5300942 | Dolgoff | Apr 1994 | A |
5301030 | Ko | Apr 1994 | A |
5303043 | Glenn | Apr 1994 | A |
5309496 | Winsor | May 1994 | A |
5311568 | McKee, Jr. et al. | May 1994 | A |
5317405 | Kuriki et al. | May 1994 | A |
5321499 | Yu et al. | Jun 1994 | A |
5358407 | Lainer | Oct 1994 | A |
5369433 | Baldwin et al. | Nov 1994 | A |
5376978 | Bae | Dec 1994 | A |
5380206 | Asprey | Jan 1995 | A |
5392080 | Galt et al. | Feb 1995 | A |
5394198 | Janow | Feb 1995 | A |
5396301 | Sasaki et al. | Mar 1995 | A |
5398082 | Henderson et al. | Mar 1995 | A |
5414521 | Ansley | May 1995 | A |
5424838 | Siu | Jun 1995 | A |
5451163 | Black | Sep 1995 | A |
5455694 | Ariki et al. | Oct 1995 | A |
5465132 | Mangelsdorf | Nov 1995 | A |
5467104 | Furness, III et al. | Nov 1995 | A |
5513116 | Buckley et al. | Apr 1996 | A |
5546316 | Buckley et al. | Aug 1996 | A |
5559714 | Banks et al. | Sep 1996 | A |
5603617 | Light | Feb 1997 | A |
5612716 | Chida et al. | Mar 1997 | A |
5681223 | Weinreich | Oct 1997 | A |
5717454 | Adolphi et al. | Feb 1998 | A |
5724519 | Kato et al. | Mar 1998 | A |
5825941 | Linford et al. | Oct 1998 | A |
5850463 | Horii | Dec 1998 | A |
5986670 | Dries et al. | Nov 1999 | A |
5993048 | Banks et al. | Nov 1999 | A |
6121953 | Walker | Sep 2000 | A |
6148148 | Wain et al. | Nov 2000 | A |
6206750 | Barad et al. | Mar 2001 | B1 |
6229904 | Huang et al. | May 2001 | B1 |
6298197 | Wain et al. | Oct 2001 | B1 |
6333985 | Ueda et al. | Dec 2001 | B1 |
6425825 | Sitrick | Jul 2002 | B1 |
6580811 | Maurer et al. | Jun 2003 | B2 |
6606096 | Wang | Aug 2003 | B2 |
6614466 | Thomas | Sep 2003 | B2 |
6824387 | Sakai et al. | Nov 2004 | B2 |
6829432 | Misumi et al. | Dec 2004 | B2 |
6879712 | Tuncay et al. | Apr 2005 | B2 |
6903756 | Giannini | Jun 2005 | B1 |
7039222 | Simon et al. | May 2006 | B2 |
7062454 | Giannini et al. | Jun 2006 | B1 |
20010036860 | Yonezawa | Nov 2001 | A1 |
20030043153 | Buddemeier et al. | Mar 2003 | A1 |
20030198384 | Vrhel | Oct 2003 | A1 |
20040085324 | Yao | May 2004 | A1 |
20040201666 | Matsuo et al. | Oct 2004 | A1 |
20040239689 | Fertig et al. | Dec 2004 | A1 |
20050026685 | Ruark et al. | Feb 2005 | A1 |
20050041867 | Loy et al. | Feb 2005 | A1 |
20050286799 | Huang et al. | Dec 2005 | A1 |
20060075448 | McAlpine et al. | Apr 2006 | A1 |
Number | Date | Country |
---|---|---|
09160575 | Jun 1997 | JP |
11-175061 | Feb 1999 | JP |
2000-209500 | Jul 2000 | JP |
2002-042168 | Aug 2002 | JP |
WO 9703517 | Jan 1997 | WO |
WO 0203708 | Jan 2002 | WO |
Number | Date | Country | |
---|---|---|---|
20070014485 A1 | Jan 2007 | US |