The present invention relates to the field of processing a video image, including methods of processing a composite image which includes a visual indication of at least one feature in the composite image.
The ready availability and affordability of broadband internet access globally has led to an exponential increase in the adoption of communication methods like video messaging and video conferencing. The variety and availability of free software enabling such communication methods from companies like Microsoft, Yahoo, Skype, and America Online also aid in increasing the popularity of such communication methods. Currently, there are also an increasing number of devices which are used for video conferencing. These devices may or may not employ the aforementioned communication software for video conferencing.
Generally, during video conferencing, parties who are communicating with each other only see one another. There are certain inconveniences for a first party to highlight/point out to the second party features relating to objects which are outside a field of view of an image capture device in a video conferencing device, while still being able to see one another. This limitation may cause some problems during certain instances of communication between the parties.
In a first aspect, there is provided a method for processing a composite video image. The method includes obtaining a first image of a scene through a first image capturing device with the first image including a visual representation for directing a receiver's attention to at least one portion of the first image. A second image of an appearance of a user of a second image capturing device is also obtained through the second image capturing device. The second image is preferably obtained simultaneously when obtaining the first image. The second image may be either a face or an upper torso of the user of the second image capturing device.
A composite image that includes both the first image and the second image is subsequently generated, the first image making up a substantial proportion of the composite image with the second image being located at a position to create a perception that the user is making the visual representation to the receiver. The composite image is then transmitted to the receiver. The composite image may undergo perspective gaze correction before transmission to the receiver, with the gaze corrected image being akin to the user of the second image capturing device looking upon the scene of the first image.
It is preferable that the first image capturing device and second image capturing device include a lens of a type such as, for example, standard (50 mm), wide-angle or fish eye. The at least one of the first and second image capturing devices have either a fixed or variable field of view. A first field of view of the first image capturing device and a second field of view of the second image capturing device may or may not overlap.
Preferably, the visual representation may be an arm, a hand, a finger, or an object used for pointing. The visual representation may be visually enhanced using wavefront coding to enable it to appear as an optically focused object.
It is preferable that the first image capturing device and the second image capturing device are integrated in a single apparatus, the single apparatus being, for example, a handheld media player, a mobile phone, a handheld digital camera, a webcam and the like.
There is provided an alternative method for processing a composite video image. The method includes obtaining a first image of a scene through a first lens of a first image capturing device; using a preview image of the first image and overlaying a visual representation for directing a receiver's attention to at least one portion of the scene on the preview image; the first image capturing device focus-locking on at least one portion of the first image, with the focus-lock enabling movement of the visual representation when a field of view of the first image capturing device is varied. A composite image that includes the visual representation is then generated and transmitted to the receiver.
The first lens may be a type such as, for example, standard (50 mm), wide-angle, fish eye and the like. The first image capturing device may have either a fixed or variable field of view. It is preferable that the visual representation is of a form such as, for example, an arrow, a cross hair, a pointer, a cross and so forth. The focus-lock may preferably maintain a position of at least one portion of the first image in the preview image.
The variation of the field of view of the first image capturing device may be detected using techniques for detecting movement of the first image capturing device such as, for example, an electronic compass for detecting horizontal panning using bearings, an inertial meter for detecting vertical tilting, a tilt sensor for detecting vertical tilting, optical flow analysis for detecting specific pixel movements and the like.
In order that the present invention may be fully understood and readily put into practical effect, there shall now be described by way of non-limitative example only preferred embodiments of the present invention, the description being with reference to the accompanying illustrative drawings.
There is shown in
A second image 106 of an appearance of a user of a second image capturing device may be obtained through a second image capturing device (24). It should be noted that the first image 102 may be obtained after obtaining the second image 106. The order of obtaining the first image 102 and the second image 106 is variable. The second image 106 may also be obtained simultaneously when obtaining the first image 102. The first and second image capturing devices may include lenses of a type such as, for example, standard (50 mm), wide-angle, fish eye and the like. At least one of the first and second image capturing devices may have a fixed field of view. Similarly, at least one of the first and second image capturing devices may have a variable field of view. Further discussion of the first and second image capturing devices will be provided in a subsequent portion of the description. Both the first image capturing device and the second image capturing device may be integrated in a single apparatus, the single apparatus may be, for example, a handheld media player, a mobile phone, a handheld digital camera, a webcam and the like.
Subsequently, a composite image 98 is generated (26) that includes both the first image 102 and the second image 106. The first image 102 may make up a substantial proportion of the composite image 98 with the second image 106 being located at a position to create a perception that the user of the second image capturing device is making the visual representation 104 to the receiver.
Referring to
Referring to
Finally, the composite image 98 may be transmitted to the receiver (28). The receiver is thus able to view the composite image 98 in which the user of the second image capturing device is able to indicate features within the first image 102 which aid in communication with the receiver.
Referring to
Referring to FIGS. 2 and 9A-9C, there is provided an alternative method 60 for processing a composite video image. The method 60 includes obtaining a first image 202 of a scene (62) through a first lens 203 of a first image capturing device 204. The first image capturing device 204 may be similar to the apparatus 40 as mentioned earlier. This is depicted in window 200 of
A visual representation 206 for directing a receiver's attention to at least one portion of the scene is overlaid on a preview image 208 (64). The preview image 208 may be cropped 210 from the first image 202. The visual representation 206 is shown to be a cross hair, but may also be, for example, an arrow, a cross, a pointer and the like. The visual representation 206 may be moved about in the preview image 208 using controls on the first image capturing device 204. Alternatively, the preview image 208 may be shown on a touch-sensitive display and the visual representation 206 is moved about using a sense of touch on the touch-sensitive display.
The first image capturing device 204 is focus-locked (66) on at least one portion of the first image 202 (a flower pot as shown in
Based on a comparison of window 200 and window 250, it can be seen that the field of view of the first image capturing device 204 has shifted left-wards (panning). Consequently, the flower pot in the first image 202 is moved to the right, but due to the focus-lock, the flower pot in the first image 202 appears to be stationary. This is shown in
The user of the first image capturing device 204 may pan the first image capturing device 204 leftwards to cause the visual representation 206 to move leftwards. It can be seen that such a process is intuitive in nature. While the example shown relates to panning of the first image capturing device 204, it should be noted that the method 60 also covers tilting of the first image capturing device 204.
Subsequently, a composite image identical to the preview image 208 is generated (70) that includes both the first image 202 and the visual representation 206. Finally, the composite image may be transmitted to the receiver (72). The receiver is thus able to view the composite image 98 in which the user of the first image capturing device 204 is able to indicate features within the first image 202 which aid in communication with the receiver.
Whilst there has been described in the foregoing description preferred embodiments of the present invention, it will be understood by those skilled in the technology concerned that many variations or modifications in details of design or construction may be made without departing from the present invention.
Number | Date | Country | Kind |
---|---|---|---|
200706524-6 | Sep 2007 | SG | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/SG2008/000331 | 9/5/2008 | WO | 00 | 3/5/2010 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2009/031978 | 3/12/2009 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
5025314 | Tang et al. | Jun 1991 | A |
5473369 | Abe | Dec 1995 | A |
5781174 | Uya et al. | Jul 1998 | A |
6771303 | Zhang et al. | Aug 2004 | B2 |
6774932 | Ewing et al. | Aug 2004 | B1 |
6806898 | Toyama et al. | Oct 2004 | B1 |
6816625 | Lewis et al. | Nov 2004 | B2 |
7184574 | Zahavi | Feb 2007 | B1 |
7502018 | Tamura | Mar 2009 | B2 |
7535486 | Motomura et al. | May 2009 | B2 |
7606476 | Toji | Oct 2009 | B2 |
20030016285 | Drost et al. | Jan 2003 | A1 |
20040114042 | Paolini et al. | Jun 2004 | A1 |
20040119851 | Kaku | Jun 2004 | A1 |
20040189804 | Borden et al. | Sep 2004 | A1 |
20050128323 | Choi | Jun 2005 | A1 |
20050280707 | Sablak et al. | Dec 2005 | A1 |
20060044396 | Miyashita et al. | Mar 2006 | A1 |
20060054787 | Olsen et al. | Mar 2006 | A1 |
20060098729 | Shen | May 2006 | A1 |
20070098380 | Spielberg | May 2007 | A1 |
20070109398 | Teo | May 2007 | A1 |
20070222859 | Chang et al. | Sep 2007 | A1 |
20070248348 | Lee et al. | Oct 2007 | A1 |
20080030592 | Border et al. | Feb 2008 | A1 |
20080084508 | Cole et al. | Apr 2008 | A1 |
20080129825 | DeAngelis et al. | Jun 2008 | A1 |
20080266380 | Gorzynski et al. | Oct 2008 | A1 |
20080297586 | Kurtz et al. | Dec 2008 | A1 |
20100134642 | Thorn | Jun 2010 | A1 |
Number | Date | Country | |
---|---|---|---|
20100309337 A1 | Dec 2010 | US |