Hereinafter, embodiments according to an embodiment of the invention will be described in detail with reference to the drawings.
In the conspicuous area detecting part 2, a frame memory (FM) 4 delays the video signal S1 by a period of a single frame and outputs it, and a motion vector computing part 5 subsequent thereto receives the video signal delayed by a single frame and the original video signal S1 and detects a motion vector v1 for every pixel. In addition here, the detection of the motion vector may be performed in units of blocks for detecting motion vectors. Moreover, only a brightness signal may be used to detect the motion vector v1. Furthermore, various schemes of detecting the motion vector can be widely applied including block matching, gradient, etc.
The conspicuous area detecting part 2 uses the motion vector v1 to detect a conspicuous degree that indicates the possibility of being a conspicuous area, and determines the conspicuous degree based on a predetermined threshold to detect the conspicuous area.
Here, as shown in
In contrast to this, in the case of zooming out, in the majority of cases, the conspicuous area having been zoomed in is turned to zoom out and shoot. Also in this case, it can be said that the possibility of being the conspicuous area is higher in the portion in which the motion vector is smaller.
In addition, as shown in
On this account, also in the case in which a camera is tilted, it can be said that the possibility of being the conspicuous area is higher in the portion in which the amount of movement is smaller, and the magnitude of the absolute value of the motion vector is used to express the conspicuous degree.
Accordingly, in the case in which no zooming operation is performed, the conspicuous degree can be detected from the magnitude of the relative motion with respect to the background, whereas for the zooming operation, the conspicuous degree can be expressed by the distance from the center of the screen.
However, in the actual imaging result, the pan, tilt and zooming operations are sometimes combined. Therefore, it is necessary to determine the conspicuous degree with no distinction depending on the existence of the zooming operation. Then, in the embodiment, attention is given to that the outer part of the screen is highly unlikely to be the conspicuous area, and the conspicuous degree is detected with reference to the motion vector v1 detected in the outer part of the screen.
Here, as shown in
Here, the relative motion vector with respect to the background is expressed, and then, as shown in
ro pan=|motion vector v1 for every pixel−vppan| (2)
In contrast to this, the same thing can be applied to the case of a tilt, and similarly, the motion vector v1 for every pixel is processed to detect the conspicuous degree. In addition, in this case, instead of the motion vectors v1L and v1R detected in a certain area at the left end of the screen and a certain area at the right end of the screen, respectively, motion vectors detected in a certain area at the upper end of the screen and in a certain area at the lower end of the screen, respectively, may be used.
In contrast to this, as shown in
vzoom=|vz| (3)
In this case, since the area is more probably the conspicuous area as closer to the center of the screen, the conspicuous degree ro zoom for every pixel caused by zooming is defined by interpolation in which a position on the screen is a variable. More specifically, in the embodiment, linear interpolation is applied to interpolation. As shown in
In addition, the conspicuous degree ro zoom is expressed by the following equation, if h width/2<h pos.
In addition here, for interpolation in which the conspicuous degree ro zoom is computed, various interpolations can be applied, not restricted to linear interpolation. For example, this scheme may be performed in which the distance from the center of the screen is set to a variable, and a curve of the second order is used for interpolation to compute the conspicuous degree ro zoom.
In practice, it is difficult to operate a camera as closely following the motion of a subject on the basis of the definitions for the conspicuous degrees ro pan and ro zoom from equations 1 to 5. In addition, in the embodiment, in the actual imaging result, as shown in
In other words, in the conspicuous area detecting part 2, the histogram computing part 6 receives the motion vector v1 detected in the motion vector computing part 5, and selectively uses the motion vector v1 in the outer part of the screen to compute the histogram in units of frames. Moreover, it detects the motion vector having the largest distribution from the computed histogram, sets the vector to the representative value of the motion vector v1, and notifies the conspicuous degree computing part 7 about the representative value. Thus, the histogram computing part 6 detects the motion vector in the outer part of the screen from the video signal S1, and notifies the conspicuous degree computing part 7.
More specifically, in the embodiment, as shown in
The conspicuous degree computing part 7 uses the representative value determined in the histogram computing part 6, and processes the motion vector v1 for every pixel to compute the conspicuous degree for every pixel. In addition, it determines the conspicuous degree based on the threshold suited for the process of the image processing part 3, generates a conspicuous area signal S3 that indicates whether to be a conspicuous area, and outputs the conspicuous area signal S3 as the detected result of the conspicuous area to the image processing part 3.
Here,
Here, the motion vectors va and vb detected from the background shown in
va=vp+vz+α1
vb=vp+vz+α2 (6)
In addition here, α1 and α2 are components caused by fluctuations and variations on the background, but α1 and α2 can be regarded as value 0 by determining the representative value in the histogram computing part 6. Then, α1 and α2 are regarded as nearly value 0, the equation 6 is solved for the vectors vp and vz, and then the relation between the following equations can be obtained.
The conspicuous degree computing part 7 uses the equation 7-1 to average the representative values of the motion vectors detected in each portion, and detects the in-phase component of the representative values to detect the motion vector on the background vp caused by making a pan and tilt. In addition, it uses the equation 7-1 to process the representative values of the motion vectors detected in each portion, and detects the antiphase component of the representative values to detect the motion vector vz in the end part caused by zooming.
Subsequently, the conspicuous degree computing part 7 goes to Step SP3, uses the motion vector vz caused by zooming at the end part detected in Step SP2, and computes the conspicuous degree ro zoom for every pixel caused by zooming. In addition here, r is the distance from the center of the screen for every pixel.
At the same time, the conspicuous degree computing part 7 computes the following equation, and detects the motion component of the motion vector v1 caused by zooming operations detected for every pixel. In addition, in the following equation, components in the X-direction and components in the Y-direction for each value are indicated by the suffix of the signs x and y. Therefore, the conspicuous degree computing part 7 linearly interpolates the motion caused by zooming operations detected at the right and left ends of the screen for each of the components in the X-direction and the components in the Y-direction in accordance with the distance from the center of the screen in the X-direction and in the Y-direction by computing Equation 9, and computes the motion caused by zooming operations for every pixel.
In addition, the computation of the equation 8 is a computation that combines the computation of the equation 9 with the computation in which the motion vector v1 zoom determined by the equation 9 is subtracted from the motion vector vz at the end part to obtain the absolute value. Therefore, in the process in Step SP3, this scheme may be performed in which instead of directly determining the conspicuous degree ro zoom caused by zooming operations through the computation of the Equation 8, the vector v1 zoom (v1 zooms, v1 zoomy) caused by zooming operations for every pixel determined through the computation of the Equation 9 is subtracted from the motion vector vz caused by zooming operations detected at the end part to obtain the absolute value, and the conspicuous degree ro zoom caused by zooming is computed for every pixel by computing the equation 8.
Subsequently, the conspicuous degree computing part 7 goes to Step SP4, and subtracts the vector v1 zoom (v1 zooms, v1 zoomy) for every pixel determined in Step SP3 from the corresponding motion vector v1 for every pixel through the computation of the following equation to compute the motion vector v0 for every pixel caused by making a pan and tilt. Accordingly, in the embodiment, the motion vector v1 detected in each portion is separated into the motion component caused by zooming operations and the motion component caused by making a pan and tilt to detect the conspicuous degree for the both components. In addition, in this case, this scheme may be performed in which the computation of the equation 8 is performed for each of the components in the X-direction and the components in the Y-direction, the vectors obtained from the computed components in the X-direction and the components in the Y-direction are added to the corresponding motion vector v1 for every pixel, and the motion vector v0 caused by making a pan and tilt is computed for every pixel. In addition, in the computation of the equation 9, instead of linear interpolation, an interpolation based on the curve of the second order may be applied. Moreover, this scheme may be performed in which the X-component and the Y-component are weighted to detect the conspicuous degree, and changes in the conspicuous degree with respect to the motion in each portion are varied by the pan operations and the tilt operations. In addition here, instead of a simple addition of the conspicuous degrees ro zoom and ro pan, the total conspicuous degree may be computed by weighted addition, and the weighting factor in weighted addition may be switched depending on the shooting mode.
v0=v1−v1 zoom (10)
In addition, the motion vector v0 computed for every pixel and the motion vector on the background vp caused by making a pan and tilt computed with the equation 7-1 are used to compute the following equation, and the conspicuous degree ro pan caused by making a pan and tilt is computed.
ro pan=|v0 vp| (11)
Subsequently, the conspicuous degree computing part 7 goes to Step SP5, adds the conspicuous degrees ro zoom and ro pan computed in Step SP3 and Step SP4 for every pixel to compute a total conspicuous degree R1, and goes to Step SP6 to end the process steps.
The conspicuous degree computing part 7 turns the computed conspicuous degree r1 to the absolute value, determines it based on a predetermined threshold, and then generates the conspicuous area signal S3.
In the configuration described above, in the image processing circuit 1 (
Then, in the embodiment, the motion component caused by zooming operations and the motion component caused by making a pan and tilt are detected from the representative values detected at the left end and the right end (Step SP2 in
Moreover, the motion component caused by zooming operations is used to detect the motion caused by zooming operations in each portion of the video signal S1, and the conspicuous degree caused by zooming operations is detected (Step SP3 in
In addition, the motion component caused by zooming operations is removed of the motion vector v1 detected in each portion based on the motion caused by zooming operations in each portion, and then the conspicuous degree caused by making a pan and tilt is detected based on the relative amount of movement to the motion component caused by making a pan and tilt detected from the representative value. Moreover, the conspicuous degree caused by making a pan and tilt is added to the conspicuous degree caused by zooming operations to detect the total conspicuous degree, and the total conspicuous degree is determined based on the threshold to generate the conspicuous area signal S3 that indicates whether to be a conspicuous area. In the image processing circuit 1, the video signal S1 is subjected to image processing based on the conspicuous area signal S3.
Accordingly, in the image processing circuit 1, even in the case in which it is difficult to detect a conspicuous area depending on color, shapes, texture, spatial frequencies and others included in two dimensional video, the motion vector is used to detect a conspicuous area, and a conspicuous area can be detected more adequately than before. In addition, at this time, the motion vector detected in each portion of the imaging result is separated into the component caused by zooming operations and the component caused by making a pan and tilt to detect the conspicuous degree for components in each area, and the conspicuous degree is detected in a comprehensive manner, whereby a conspicuous area can be detected more adequately than before.
More specifically, in the embodiment, a histogram is generated for each area in which the representative value is detected in the histogram computing part 6, and the motion vector having the largest distribution in the histogram is set to the representative value. Therefore, for example, as shown in
In addition, the antiphase component is detected from the representative value to detect the motion component caused by zooming operations, as well as the in-phase component is detected from the representative value to detect the motion component caused by making a pan and tilt, whereby the motion component caused by zooming operations and the motion component caused by making a pan and tilt are detected through a simple computation, and a conspicuous area can be detected more adequately than before with a simple configuration as a whole.
Moreover, the motion component caused by zooming operations in each portion is detected by interpolation in which the position on the screen is the variable with reference to the motion in the outer part of the screen detected from the representative value, whereby the motion caused by zooming operations in each portion is detected with simple operations, and a conspicuous area can be detected more adequately than before with a simple configuration as a whole.
According to the configuration above, the motion vector detected in each portion of the imaging result is separated into the component caused by zooming operations and the component caused by making a pan and tilt to detect the conspicuous degree that indicates the probability of being a conspicuous area for the components in each area, and the conspicuous degree is detected from the conspicuous degree in a comprehensive manner, whereby a conspicuous area can be detected more adequately than before.
In addition, a histogram is generated for each area in which the representative value is detected, and the motion vector having the largest distribution in the histogram is set to the representative value, whereby the motion on the background is reliably detected to improve the detection accuracy of the conspicuous area.
Moreover, the antiphase component is detected from the representative value to detect the motion component caused by zooming operations, and/or the in-phase component is detected from the representative value to detect the motion component caused by making a pan and tilt, whereby the motion component caused by zooming operations and the motion component caused by making a pan and tilt can be detected through a simple computation, and a conspicuous area can be correspondingly detected more adequately than before with a simple configuration as a whole.
Furthermore, the motion component caused by zooming operations in each portion is detected by interpolation in which the position on the screen is the variable with reference to the motion in the outer part of the screen detected from the representative value, whereby the motion caused by zooming operations in each portion is detected by simple operations, and a conspicuous area can be correspondingly detected more adequately than before with a simple configuration as a whole.
In the image processing circuit 11, a conspicuous area detecting part 12 detects a conspicuous degree S4 of a video signal S1, and the conspicuous degree S4 is determined based on a threshold on an image processing part 13 side to generate a conspicuous area signal. The image processing part 13 subjects the video signal S1 to image processing based on the conspicuous area signal as similar to the image processing part 3 described in the embodiment 1.
In addition, as similar to the description in the embodiment 1, the conspicuous area detecting part 12 uses a representative value determined in a histogram computing part 6 to process a motion vector v1, and detects a conspicuous degree S5. A conspicuous degree computing part 17 is configured the same as the conspicuous degree computing part 7 in the embodiment 1 except that the conspicuous degree S5 is outputted directly.
The conspicuous area detecting part 12 delays the conspicuous degree S4 that is to be outputted to the image processing part 13 by a period of a single frame through a frame memory 18. Moreover, it weights and adds the conspicuous degree S4 delayed by a single frame to the conspicuous degree S5 outputted from the conspicuous degree computing part 17 by means of multiplication circuits 19 and 20, and an addition circuit 21, and generates the conspicuous degree S4. Thus, the conspicuous area detecting part 12 smoothes and outputs the conspicuous degree to stabilize the detection of the conspicuous area. In addition, as depicted by a dotted line in
Moreover, for example, this scheme may be performed in which weighting factors 1-K and K in the conspicuous area detecting part 12 are switched depending on the shooting mode such as scenic mode, portrait mode, and sports mode to vary the time constant for smoothing. More specifically, for example, the time constant is made long in the scenic mode and portrait mode, whereas the time constant is made short in the sports mode.
According to the embodiment, the conspicuous degree is smoothed and outputted, whereby the detection of the conspicuous area can be stabilized.
In addition, in the embodiments described above, the case is described in which the motion vector is simply used to detect the conspicuous area, but an embodiment of the invention is not restricted thereto. This scheme may be performed in which such a configuration is provided that a change in a scene is detected to stop detecting the motion vector, and a change in a scene is detected to hold the conspicuous degree at a value a single frame before. With this scheme, the conspicuous degree can be detected stably with no influence of the absence of the motion vector caused by finishing the operation of a camera, for example.
Moreover, in the embodiments described above, the case is described in which the motion vector is simply used to detect the conspicuous area, but an embodiment of the invention is not restricted thereto. This scheme may be performed in which the physical properties such as color, shapes, texture, spatial frequencies and others included in two dimensional video are used to combine a technique of detecting the conspicuous area to detect the conspicuous area. In addition in this case, in the case in which the OR operation of the conspicuous area signal detected by each technique generates the overall conspicuous area signal, for example, various configurations can be applied for the combinations of these techniques. In addition, the detection technique of the conspicuous area may be switched depending on the shooting mode. Furthermore, this scheme may be performed in which the conspicuous degrees detected by the individual techniques are weighted and added to combine these techniques, and the weighting factors are switched depending on the shooting mode.
In addition, in the embodiments described above, the case is described in which the video signal is subjected to image processing, but an embodiment of the invention is not restricted thereto. It can be widely adapted to the case in which the conspicuous area is detected for focus control.
An embodiment of the invention can be used for adjusting image quality of video signals, for example.
It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Number | Date | Country | Kind |
---|---|---|---|
P2006-224911 | Aug 2006 | JP | national |