The invention relates to a video processing apparatus and method, and in particular to a video compression apparatus and method.
Video compression techniques are commonly used for transmitting video signals more efficiently over communication channels having a limited bandwidth. In modern day video compression techniques, such as MPEG4, region based coding is proposed to enable different regions in the scene to be coded with different qualities. The main objective of this technique is to send important objects with high quality, while less important regions of the scene are transmitted with lower quality.
“Region based Video Coding using Mathematical Morphology”, Philippe Salembier et al, Proceedings of the IEEE, Vol. 83, No. 6, June 1995, discloses a region based coding in which regions in an image are segmented based on intensity, color and grey value. This has the disadvantage that it is not clear which is the significant object in the scene. Often, the significant object will be the moving object in the image.
The aim of the present invention is to provide an improved video processing.
The invention is defined by the independent claims. The dependent claims define advantageous embodiments.
According to a first aspect of the present invention, there is provided a video processing apparatus for processing an image signal having one or more regions of interest. The apparatus comprises depth estimation means for determining the depth of a region in the image signal, and providing a corresponding depth signal. A data compressor receives the image signal and the depth signal, and is configured to compress the image data in a particular region based on the corresponding depth signal received from the depth estimation means.
The invention has the advantage of being able to compress a region of the image signal, for example relating to a particular object, based on the depth of that region in the image signal, and hence the importance of the region within the overall image signal.
According to another aspect of the invention, there is provided a mobile communications device comprising a first imaging means for taking a first image signal, and a second imaging means for taking a second image signal. The first and second imaging means are arranged to point in substantially the same direction.
The communications device according to this aspect of the invention has the advantage of being able to determine depth information in the image signal being viewed, which can then be used to dynamically compress different regions in the image signal as described above.
According to another aspect of the invention, there is provided a method of processing an image signal having one or more regions of interest. The method comprises the steps of determining the depth of a region in the image, signal to provide a corresponding depth signal. The depth signal is used by a data compressor for compressing the image signal, such that the image data for a particular region is compressed based on the corresponding depth of that region in the image signal.
For a better understanding of the invention, and to show more clearly how it may be carried into effect, reference will now be made, by way of example only, to the following drawings in which:
A data compressor 7 receives an image signal from one of the cameras, for example the first camera 1, and compresses the video data in the image signal to produce a compressed image signal 14. The data compression level is based on the depth signal 13 received from the depth estimator 5.
For example, the apparatus can be configured to compress image data based on the assumption that the objects that are closer to the camera are more important than the objects in the background.
The depth signal 13 is determined based on the first image signal 9 and second image signal 11 received by the depth estimator 5. The first image signal 9 and the second image signal 11 are used to determine the disparity between corresponding pixels for the same object in the left and right images.
Preferably, the disparity is translated into a depth signal per pixel, which is used to control the degree of quantization in the data compressor 7 when compressing the normal image.
Thus, according to the invention, objects that are closer to the cameras are coded with high quality, i.e. high quantization, while objects that are further away from the cameras are subjected to lower coding, i.e. lower quantization resulting in a lower bandwidth requirement.
Optionally, a decision can be made to completely ignore pixels relating to the insignificant parts of the scene. In such pixels, the data compressor 7 can be configured to insert data that is more easily coded in place of the true background information. Alternatively, a flag or indicator can be inserted, which causes a receiver to insert pixel data at the receiver side.
Disparity of a specific object in a stereoscopic image is the difference in pixels between the position of the object at the left image and the position of the same object at the right image. In other words, for a given pixel relating to a particular object, the disparity between the images seen by the first and second cameras 1, 3 will be small if the pixel relates to an object that is far away from the cameras, while the disparity will be large if the pixel relates to an object that is close to the cameras. Thus, the pixel data will appear in almost the same position in both the image frames when that pixel data relates to an object that is far away from the cameras 1, 3. Conversely, the pixel data will appear in significantly different positions in the image frame when the pixel data relates to an object that is close to the cameras 1, 3.
For example, in
Various techniques for calculating the depth of an object from the images obtained from two cameras are known per se, and will not be described in greater detail in this application. These techniques include the steps of taking a specific pixel from a first image and finding the corresponding pixel in the second image. If the corresponding pixel is found, then the disparity is calculated, and a depth value assigned to that pixel.
From the above it will be seen that each pixel in the image signal is provided a depth signal, which is used to provide the quantization value for the data compressor when compressing the normal image.
The invention is particularly suited for applications in which video data must be compressed for transmission over a communication channel having a limited bandwidth. For example, the invention is particularly suited for use in a mobile telephone. According to this aspect of the invention, there is provided a mobile telephone having first and second cameras, the first and second cameras being arranged to point in substantially the same direction. The cameras can be used to determine depth information, for use in providing a depth signal for the data compression, as described above.
Alternatively, the video processing apparatus could be used to reduce the amount of video data to be stored, for example in a mobile telephone or video camera.
Although the preferred embodiment has been described in relation to the cameras providing “right” and “left” versions of the scene, it will be appreciated that any orientation will be possible, providing the two cameras are in a fixed position relative to one another. In addition, the depth value can also be measured using other means, such as “time of flight of light” from an object in a scene, or other focusing techniques for determining the depth of an object. Furthermore, when used with video cameras, the depth of an object in a scene can be determined from successive frames of the same scene, provided an object is moving between the respective frames. Although such an embodiment relies on knowledge about the size of the objects in a scene, it can nevertheless be useful for deteniniing which object is in front of the other objects, thereby enabling the closest object (and hence the most important object) to be determined.
In addition, although the preferred embodiment has been described on the basis that objects in the foreground are more important than objects in the background, it will be readily appreciated that the invention can also be used in reverse, whereby objects in the background are treated as the more important objects, for example in securing applications in which a background scene is being monitored. Alternatively, the invention could be used to provide the best quality at a predetermined depth from the cameras, for example if the cameras are used in a fixed location, and intended to monitor a scene that is at a predetermined distance away from the cameras.
The invention described in the embodiments above has the advantage of being able to compress a region of an image signal, for example relating to a particular object, based on the depth of that region in the image signal, and hence the importance of the region within the overall image signal.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative embodiments without departing from the scope of the appended claims. The word ‘comprising’ does not exclude the presence of elements or steps other than those listed in a claim.
In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word “a” or “an” preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the device claim enumerating several means, several of these means may be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
Number | Date | Country | Kind |
---|---|---|---|
04103122.0 | Jul 2004 | EP | regional |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/IB2005/052135 | 6/28/2005 | WO | 00 | 12/19/2006 |