This invention belongs to digital image processing technology, especially involving with video digital image motion processing technology.
Currently, with respect to the video digital image motion processing, the processing is often focused on motion features and their changes at local area such as processed pixel points and/or some adjacent pixel points and etc., integration of processing results on local area motion features of all pixel points in an image composes a final processing result of the image. A motion adaptive algorithm, shown below, is an example to explain this common video image motion processing method.
Motion adaptive algorithm is a video digital image processing technology based on motion information, usually adopted in various image processing such as image interpolation, image de-interplacement, image de-noising, image enhancement and etc. The basic idea of the motion adaptive algorithm is to utilize multi-frame image for detecting motion status of pixel points and for adjusting if the pixel point trends to static or motion, which is then used as a foundation for further operation processing. If the pixel point trends to static status, thus, a pixel point at the same position on an adjacent frame will have similar features to the current pixel point, and can be used as relative accurate reference information, this method is called as inter-frame processing. But if the pixel point trends to motion status, thus information of a pixel point at the same position on an adjacent frame cannot be used as reference, therefore, only can a space-adjacent pixel point on the same frame be used as reference information, namely, so-called intra-frame processing.
In practical application, motion situation of each pixel point on the same frame is different with each other, in order to make up for problems caused by using single method, two processing algorithms of inter-frame and intra-frame mentioned above are combined together for obtaining the best image result. Motion adaptive algorithm weights and averages the results obtained via these two algorithms, its formula is as follows:
P
results
=a×P
intra−(1−a)×Pinter
Wherein, Presult is its finally processed result, Pintra is the intra-frame processed result and Pinter is the inter-frame processed result. That is, the larger the motion adaptive weight value a is, the stronger the motion will be, thus it trends intra-frame processing; otherwise, if the motion adaptive weight value a is smaller, thus it trends to inter-frame processing. The motion adaptive weight value is an absolute value of a differential value between pixel points relative to two adjacent frames. Its formula is as follows:
a=P(n,i,j)−P(n−1,i,j)
Wherein, P is a luminance value of the pixel point; n is a sequential number of the image frame based on time; i is a line number of the image, on which the pixel point is located; j is a row number of the image, on which the pixel point is located.
The above explanation shows: the object processed by this image motion processing method is the pixel point, and simultaneously, by using information of local area around the processing pixel point as auxiliary information. This image processing method, because of focusing the identified object only on a micro-local area, will result in error if in comparison with global image identification method by eye. Therefore, when the image is influenced by problems such as inter-frame delay, noise and etc., especially when under situation of motion and static both existed in the image, larger identification error may occur, also with truncation effect easily occurred on edge of the area truncation.
This invention, focused on the problem of larger error existed in current video image motion processing method and caused by judging at a limited local area, offers a video image motion processing method introduced with global feature classification.
Another purpose of this invention is to offer a device for implementing this video image motion processing method introduced with global feature classification.
The technical idea of this invention is to utilize the global feature information of a processing video image and the local feature information of its pixel points for classifying certain local motion feature information of the pixel points, assigning correction value to each classification, and then using the correction value to correct certain local motion feature information of the pixel points, and finally to obtain more accurate local motion feature information of the pixel points.
Technical scheme of this invention is as follows:
The video image motion processing method introduced with global feature classification includes steps as follows:
The local motion features obtained in the Step A include motion adaptive weight values of the pixel points; final motion adaptive weight values of the pixel points, obtained when the local motion features corrected as said in the Step E, are the motion adaptive weight values of the pixel points.
The local motion features as said in the step A also include motion feature values between pixel point fields, which show motion status between pixel point fields, the formula for obtaining the inter-field motion feature value is as follows:
Motionfield=|(P(n,i−1,j)+P(n,i,+1,j)/2−P(n+1,i,j)|;
Wherein, Motionfield is a motion feature value between pixel point fields; P is a luminance value of the pixel point; n is a sequential number of the image field based on time; i is a line number of the image, on which the pixel point is located; j is a row number of the image, on which the pixel point is located.
The said local features obtained in the Step A also include judgment value for judging if the pixel point is an edge point or not, which is obtained via edge detection.
The said edge detection includes steps as follows:
Obtaining the global features as said in the Step B includes steps as follows:
The selected pixel points as said in the Step (1) of obtaining the global feature are the edge pixel points.
The classification as said in the Step C refers to classifying to obtain several classifications and to sort out the pixel points into each classification in accordance to the obtained global features, the motion adaptive weight values, the edge point judgment values and the inter-field motion feature values, of which all are used as classification basis for the processing pixel point.
The classification method said in the Step C is a decision-tree classification method.
The correction formula adopted in the correction said in the Step D is as follows:
a′=Clip(f(a,k),m,n);
Wherein, a′ is the final motion adaptive value; a is the motion adaptive weight value obtained in the Step A; k is a classification parameter in the Step D; f(a, k) is a binary function of variables a and k; Clip ( ) is a truncation function, ensuring output value within a range of [m, n].
The device for implementing the video image motion processing method introduced with global feature classification includes units as follows: a local feature capture unit, a global feature capture unit, a classification unit and a correction unit; the local feature capture unit is respectively connected with the classification unit and the correction unit; the global feature capture unit is respectively connected with the local feature capture unit and the classification unit; the classification unit is also connected with the correction unit; the said local feature capture unit is used to extract the local feature of the pixel point in the processing video image, the said local feature includes the local motion feature; the said global feature capture unit is used to extract the global feature of the processing video image; the said classification unit is used to classify the pixel points in the processing video image in accordance with results of the global feature capture unit and the local feature capture unit, and assigning the correction parameters to the classifications obtained after classifying; the correction unit utilizes the correction parameters obtained by the classification unit to correct the certain local features obtained by the local feature capture unit.
The said local feature capture unit includes a motion detection unit, the said motion detection unit outputs its results to the said classification unit; the result obtained by the motion detection unit is the motion adaptive weight value and the inter-field motion feature value of the processing pixel points.
The said local feature capture also includes an edge detection unit, the said edge detection unit outputs its results to the said global feature capture unit; the result obtained by the edge detection unit is a judgment value for judging if the processing pixel point is the edge point or not.
Because of introducing the global feature of the processing video image to classify the local motion features of the pixel points, and to accurately correct according to different classification, final results of the local motion features obtained by using technical scheme of this invention are more accurate. Because human eyes recognize image's effect via judging in a global and macro-view, classifying the local motion features of the pixel points via introducing the global feature can correct errors on the local motion features of the pixel points in a global view, and can avoid distortion, caused by various interferences, on motion features obtained only locally, thus improving accuracy of the local motion features of the pixel points.
When conducting the motion detection, though the most direct method for global statistics is to process all of image pixel points, that is, making statistics on motion situation of each pixel point in the image, the motion status of different pixel points in the same frame of image are all different, and a large part of the pixel points for a general continuous video is at a static status (even if human eyes feel the image in moving), and the edge pixel points in a image can further represent image motion status, that is, namely if the edge pixel point is in motion, there is motion in the image; if the edge pixel point is not in motion, there is no motion in the image. Therefore, introducing motion information of the processing video image edge pixel point for classifying, judging and processing the motion feature of the pixel point can more accurately identify motion status of an image.
Under situation of interlaced image processing during the edge detection, not only is the information of adjacent pixel point at the same field of the pixel point used as foundation, but also the information of an adjacent pixel point corresponding to the pixel point at the front field is used, that is, the motion detection shall detect motion results between adjacent fields. Because there is inter-field time gap in the original motion information obtained via inter-frame differential value (namely inter-frame motion) of the pixel point motion feature, and if change frequency of the pixel point is just the same as field frequency, it is impossible to detect the field motion (for example, (n−1) field is in black, (n) field is in white, and (n+1) filed is also in black, thus it can be judged that there is no frame motion). So, the inter-field motion detection is introduced in order to avoid such problem.
The technical scheme of this invention is explained below in combination of these figures mentioned above.
As shown in the
Because the processing video image global feature is introduced for classifying the local motion features of the pixel points, and accurately correcting in accordance to different classifications, the results of the final local motion features obtained by adopting the technical scheme of this invention are more accurate. Because human eyes recognize image's effect via judging in a global and macro-view, classifying the local motion features of the pixel points via introducing the global feature can correct errors on the local motion features of the pixel points in a global view, and can avoid distortion, caused by various interferences, on the motion features obtained only locally, thus the accuracy of the local motion features of the pixel points is improved largely.
This invention will be further explained below via a video image motion detection method introduced with global feature classification (called as this motion detection method hereinafter). The processing video image signal in this Embodiment is an interlace signal, that is, one frame of an image includes two fields of image information in time sequence, the image of each field has respectively odd-line pixel information or even-line pixel information, wherein, processing focused on interlaced signal situation (such as introducing former field information in inter-field motion feature value algorithm and in edge judgment) can be omitted under situation of line-by-line signal.
In the phase for obtaining the local features, this motion detection method captures three local feature values: the pixel point's motion adaptive weight value, the inter-field motion feature value and the edge judgment value in the processing video image.
In the phase for obtaining the global feature, first conducting statistics of the motion adaptive weight value of the edge pixel points; then, conducting primary classification for the processing video image, in accordance with the statistic results of the edge pixel point's motion adaptive weight value and in comparison to the experience value, that is, whether the global image of the processing video image is in motion trend or in static trend.
In the classification phase, in accordance with judgment on the global image whether it is in motion trend or in static trend, and according to three local features such as the pixel point's motion adaptive value, the inter-field motion feature value and the edge judgment value said above, classifying the global pixel points to distribute each pixel point to its own classification finally, and then assigning the correction parameter to each classification belonged by the pixel point. Foundation of each classification is all the different sections divided on numerical interval on the basis of experience, and these sections are used as classification sorts, for example, a threshold can be determined for the motion adaptive weight value on the basis of experience, if the motion adaptive weight value of a pixel point is higher than this threshold, this pixel point is put into the motion pixel point's classification; the pixel points being lower than the threshold is put into the non-motion pixel point classification.
In the correction phase, by using the correction parameters obtained in the phase of the global pixel point classification, correcting the motion adaptive weight value of the processing video image pixel point to obtain motion adaptive weight value of the pixel point.
Technical measures of the steps are explained in details as follows:
There are several ways for capturing the motion adaptive weight value, for example, simply using absolute value of an inter-frame differential value to obtain it, its formula is as follows:
a(n,i,j)=|P(n+1,i,j)−P(n−1,i,j)|
Wherein, a(n, I, j) is the motion adaptive weight value of the pixel points; P is a luminance value of the pixel points; n is a number of the image frame in time sequence; i is a line number of the image on which the pixel point is located; j is a row number on which the pixel point is located. For simplifying the followed data calculation, normalizing the obtained a in an equal proportion into 1, that is, limiting the obtained a values in an equal proportion into an interval of [0, 1].
Capturing the inter-field motion feature value, that is, obtaining the motion results between adjacent fields, its significance is that: the motion adaptive weight value obtained in 1.1 is an inter-frame motion value, but under situation of interlaced processing, original motion information exists time gap between two fields, thus, if the change frequency of the pixel point is just in accordance to field frequency, the field motion cannot be detected (for example, (n−1) field is in black, (n) field is in white, and (n+1) is also in black, it will be judged as no frame motion). For remedying this problem, it is necessary to introduce the inter-field detection, its detection foundation is the differential value relationship between P(n, i−1, j) and P(n, i+1, j), and P(n+1, i, j) (or P(n−1, i, j). Its formula is as follows:
Motionfield=|(P(n,i−1,j)+P(n,i+1,j)/2−P(n+1,i,j)|
Wherein, Motionfield is the inter-field motion feature value; P is a luminance value of the pixel points; n is a number of the image field on time sequence; i is a line number of the image on which the pixel point is located; j is a row number of the image on which the pixel point is located.
Though the most direct method for conducting statistic and judgment of the global motion situation is to process all of the pixel points on full-frame image, in a frame of image, the motion status of different pixel points are different with each other and for a common continuous video image, most pixel points are in static status, therefore, accuracy is often influenced if conducting statistics and judgment on motion status of all pixel points globally. In practical situation, edges in an image can more accurately represent the motion status of an image, so making statistics and judgment on the motion status of the edge pixel points can improve accuracy.
The edge detection includes steps as follows:
For the pixel point belonging to the edge, its motion adaptive weight value is calculated into statistic data, and the pixel point at non-edge is emitted. Finally, after processing the full frame image, motion statistics result of the edge pixel can be obtained. Many statistics methods such as histogram statistics or probability density statistics can be used as a statistics method to conduct statistics for the motion adaptive weight value of the pixel points. The method adopted here is respectively to make statistics on non-motion (that is, the inter-frame motion adaptive weight value is 0) pixel point numbers Ns and on motion (that is, the inter-frame motion adaptive weight value is non-0) pixel point numbers Nm. The statistics target can also be the motion adaptive weight value of all pixel points, or the motion adaptive weight value of pixel points selected according to other rules.
Classifying the statistics results obtained in 2.1 according to the rules listed below, to obtain different image global motion status:
Nm/Ns>p, the image trends to motion status;
Nm/Ns<p, the image trends to static status;
q≦Nm/Ns≦p, the image is either in motion status or in static status.
Wherein, p and q is respectively adjustable threshold, and p>q. In this embodiment, p and q is the selected values as follows: p=5, q=⅕. These three statuses mentioned above correspond respectively to a value, for example, 0, 1, 2 that is called as motion status, for easily processing followed. The status value obtained above is used as the global feature applying in the followed steps. Because at the time obtaining information of this frame image status, processing of this frame image has been finished simultaneously, the obtained motion status is used in the next frame processing. In order to avoid mutation of the smooth image, the value obtained in this current image and corresponded by the motion status is averaged arithmetically with the values corresponded by the motion statuses of several former frame images (commonly, three frames) to decrease the mutation in critical status.
In order to conduct different motion correction for the different status pixel points in the processing video image, in this section, the global feature, the edge judgment value, the motion adaptive weight value and the inter-field motion feature value all obtained according to the above descriptions are used as classification foundations, and excluding special description in this embodiment, these classification foundations are all divided into sorts within their value ranges and according to the given thresholds. These classification foundations will be overlapped to build a several layer's classification structure, for example, overlapping the edge judgment value and the motion adaptive weight value, and using these two values respectively as a coordinate, to build a two-dimensional system as shown in the
It shall be specially explained that the non-edge non-motion pixel points are further divided here into the non-interfiled motion pixel C4 and the interfiled motion pixel C5. This is a treatment for high-frequency changing situation said above, that is, there is not motion in the inter-frame at that time, if there is an inter-field motion existed, error will occur in the judgment. In order to avoid such situation occurring, it is necessary to distinguish the situation of existed inter-field motion.
Each pixel point in the processing video image is classified. Common model classification methods include: decision tree, linear classification, Bayes classification, support vector classification and etc. Here, the decision tree classification method is adopted to classify pixel points.
As shown in the
On the basis of classification belonging to pixel point, respectively determining its corresponded correction parameter k; correcting, by using k value, the pixel point motion adaptive weight value obtained initially. Because of being able to more accurately correct the motion adaptive weight value obtained initially in a global view, it is possible to obtain a more accurate final motion adaptive weight value. The motion adaptive weight value is in a certain range, therefore the final motion adaptive weight value after corrected shall still be in this range, and the value higher than it is truncated. The correction formula is as follows:
a′−Clip(f(a,k),m,n);
Wherein, a′ is the final motion adaptive value; a is the motion adaptive weight value obtained in the Step A; k is the classification parameter in the Step D; f(a, k) is a binary function of the variables a and k; Clip ( ) is a truncation function, ensuring output value within the range of [m, n], that is, if higher than n, taking n as the value; if lower than m, taking m as the value. If a is normalized to 1 before, here a′ shall be within [0, 1] range.
The local feature capture unit is used to extract the local feature of the pixel point in the processing video image, the said local feature includes the local motion feature; the global feature capture unit is used to extract the global feature of the processing video image; the classification unit is used to classify the global pixel points in the processing video image in accordance with results of the local feature capture unit, and assigning the correction parameters to the classifications obtained after classified; the correction unit utilizes the correction parameters obtained by the classification unit to correct the certain local features obtained by the local feature capture unit.
In this embodiment of the device for implementing video image motion detection method introduced with global feature classification, the local feature capture unit includes a motion detection unit, the motion detection unit receives the processing video image information, the results obtained by the motion detection unit is the processing pixel point motion adaptive weight value and the inter-field motion feature value. The result of the motion detection unit outputs to the followed classification unit.
In this embodiment of the device for implementing video image motion detection method introduced with global feature classification, the local feature capture unit also includes an edge detection unit. The edge detection unit receives the processing video image information, and the obtained result is a judgment value to judge whether the processing pixel point is an edge point or not. The result of the edge detection unit outputs to the global feature capture unit.
In this embodiment of the device for implementing video image motion detection method introduced with global feature classification, its global feature capture unit also includes an edge pixel statistics unit that is used for conducting statistics of the local motion feature of the global edge pixel point (substantially referring to the motion adaptive weight value), and using its result to classifying in the classification unit. The classification unit judges the image-belonged classification according to the statistic results of the global edge pixel point motion feature, and this classification is used as a foundation for the followed classification.
Operation process of the device for implementing video image motion detection method introduced with global feature classification is as follows:
Information of the processing video image is first processed by the local feature capture unit, to obtain the pixel-point's motion adaptive weight value, the inter-field motion feature value and the judgment value for judging whether the pixel point is an edge point or not. After the global feature capture unit receives the judgment value for judging whether the pixel point obtained by the local feature capture unit is an edge point or not, statistics on the motion adaptive weight value of the edge pixel point is conducted, and the result obtained by comparing the statistic result with a pre-set value is delivered to the classification unit. The classification unit obtains information delivered by the local feature capture unit and the global feature capture unit (the pixel-point's motion adaptive weight value, the inter-field motion feature value, the judgment value for judging whether the pixel point is an edge point or not, and the result obtained by comparing the said statistic results), distributing the processing pixel points into a definite classification according to above information, and assigning the correction parameters to these classifications. The correction unit utilizes the correction parameter obtained in the classification unit to correct the pixel-point's motion adaptive weight value obtained by the local feature capture unit, to obtain the final motion adaptive weight value. Thus, the device for implementing the video image motion detection method introduced with global feature classification has finished its operation process.
It shall be pointed out that: the substantial operation ways mentioned above in the embodiment can let the technicians in this area comprehensively understand this invention, but does not limit this invention in any ways. Therefore, though the attached reference figures and the embodiment in this specification have explained this invention in details, the technicians in this area shall understand that: this invention can still be altered or equal-substituted; but any technical schemes and any modifications not breaking away from this invention idea and technical substances shall be covered within claimed area of this invention patent.
Number | Date | Country | Kind |
---|---|---|---|
200710147558.2 | Aug 2007 | CN | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/CN08/72171 | 8/27/2008 | WO | 00 | 10/25/2010 |