The patent or application file contains at least one drawing executed in color. Copies of this patent or patent application publication with color drawing(s) will be provided by the office upon request and payment of the necessary fee.
They will be full comprehensibility of the present invention from the detailed description given herein below for illustration only, and thus are not limitative of the present invention, and wherein:
Referring to
The detailed method flow is illustrated with reference to
First, when each image frame in the image sequence 10 is input according to the time point, the part of the time series analysis 110 executes the processes described in
(1) The state initializing process 110 mainly includes two situations. The first executing situation of the initializing process is executed as that, when a first image frame is input, i.e., when time t=0, the initializing process directly generates a feature point group {Y0} of the first image frame (Step 200), and generates a three-dimensional state collection {X0} corresponding to the feature point group (Step 210) as the feature point information thereof. As shown by the triangular pattern in
{X0} is obtained by initializing three-dimensional states on each feature point y0 of the {Y0} through the Kalman Filter time series analysis model. The three-dimensional state x0 includes the horizontal position, vertical position and depth position of the feature point in the three-dimensional space. The depth position can be generated by first computing {Y0} and {Y1} through the three-dimensional vision manner, and then projecting through a camera model. {Y1} is obtained by finding corresponding point {Y0} in the first image frame.
Another situation of the initializing process is executed as that, when the image frame is input not for the first time (i.e., when the input image frame is not the first image frame), i.e., the image frame input at the time t, since the reliable feature point information 50 (i.e., the feature point information including the feature point group and the three-dimensional state collection) remained after the operation of the former image frame is updated into the feature point information of the former image frame, the next input image frame takes the updated feature point information as the initialization result (Step 260). This situation is shown as the triangular pattern in
(2) After the state initializing process is finished, the feature point information of the image frame being processed is transferred into the system modeling 112 for predicting the state, and the predicting of the feature point information ({Yt+1}, {xt+1}) of the initialized feature point information {Yt} and the three-dimensional state {xt} at the next time point is performed through the established analysis model (Step 220). As described above, the analysis model used in the embodiment of the present invention is the Kalman Filter time series analysis model, wherein the descriptions on Yt and Xt are represented by the following expressions:
X
t+1
=F
t
X
t
+U
t
+Q
t;
Y
t
=H
t
X
t
+R
t;
Ft simulates the linear variation process of the state Xt along with the time, Ut is a known translation amount at the state of Xt, Ht simulates the relationship between Xt and Yt, Qt and Rt simulate the interference of noise, wherein Qt also can be represented as Qt˜N(0, qt), Rt also can be represented as Rt˜N (0, rt). Therefore, the prediction value of each Xt+1 is represented as
wherein
As shown in
(3) After the state is predicted, the three-dimensional state collection {Xt+1} in the prediction result must be properly corrected (Step 230). The above can be mainly achieved by the correcting model existing in the Kalman Filter time series analysis model, and the correcting model is represented by the following expression.
X
t+1
˜N({circumflex over (X)}t+1, Pt+1);
wherein
In the Kalman Filter time series analysis model, the error E and gain K are
respectively defined as Et+1=(Yt+1−Ht+1{circumflex over (X)}t+1) and
As shown in
The above correcting model can be properly adjusted according to the employed analysis model, and further can be properly adjusted according to the motion mode of the input image sequence 10 in the three-dimensional space. The related adjusting method differs according to the different analysis models, but the present invention holds the design flexibility of the adjustment of this part.
(4) The corrected feature point information ({Yt+1}, {Xt+1}) is transferred from the time series analysis 110 stage to the feature evolving 120 stage. In this stage, the reliable feature point information ({{tilde over (Y)}t+1}, {{tilde over (X)}t+1}) to be remained is screened by the evolving operation on the feature points (Step 240), which mainly includes two parts of screening given below in detail.
The first part is generating a new feature point, which includes the following steps.
(a) The new feature point is found out according to the method of corresponding the feature points between {Yt} and {Yt+1}, and added into {{tilde over (Y)}t+1}.
(b) A weight
is set for initializing the three-dimensional state collection of the collection {{tilde over (X)}t+1}, wherein the weight is determined by the state value of the neighboring feature points. The weight value is represented by the existing time of neighboring feature points in the whole image sequence 10, or represented by the distance from the neighboring feature points.
The definition of the neighboring feature points is represented by the following expression:
X′
t+1
={x
t+1
εX
t+1
|∥y
t+1
−{tilde over (y)}
t+1∥<η};
And, the expression of the weight
is:
Therefore, the three-dimensional state of each x,+, after being initialized can be further represented by the following expression:
After the process of generating the new feature point, the feature point information is shown by the hollow circular pattern connected by dashed line in
The second part is deleting the feature point, which includes the following steps.
(c) The feature point generating an error larger than the threshold during the feature matching process is deleted from the existing feature point collection {{tilde over (Y)}t+1}. This part may generate errors in the feature matching process when the feature point is found out by the feature matching method, thus the feature points with large errors must be deleted.
(c) The feature point generating an error larger than the threshold during the feature corresponding is deleted from the newly generated feature point collection {{tilde over (Y)}t+1}. This part is the same as the former process, which is used to delete the feature points generated by the feature matching error.
(c) and (d) mainly define Pt() as a rectangular region taking as the center at the time t, and thus at the time t+1, the feature matching error Et+1(yt, yt+1) of each feature point yt+1 in the existing feature point collection {Yt+1} is defined as the following expression:
E
t+1(yt, yt+1)=∥Pt(yt)−Pt+1(yt+1)∥;
and the feature corresponding error Et+1({tilde over (y)}t, {tilde over (y)}t+1) of each feature point {tilde over (y)}t+1, in the newly generated feature point collection {{tilde over (Y)}t+1} is defined as the following expression:
E
t+1({tilde over (y)}t+1, {tilde over (y)}t)=∥Pt+1({tilde over (y)}t+1)−Pt({tilde over (y)}t)∥;
and when Et+1(yt, yt+1) and Et+1({tilde over (y)}t+1, {tilde over (y)}t) are respectively larger than the preset threshold, the feature points yt+1 and {tilde over (y)}t+1 are deleted.
(e) The feature point with an error calculated by the system model analysis during the prediction of {Xt+1} larger than the threshold is to be deleted. This part is mainly directed to delete the feature point with large error when the three-dimensional state is transferred through the Kalman Filter time series analysis model.
The error is defined as the difference between the state {circumflex over (X)}t+1 (obtained by correcting each xt+1 in the collection {Xt+1} through the Kalman Filter time series analysis model) multiplying Ht+1 and each y1 in the feature point collection {Yt}, which can be represented as
After the feature point is deleted, the deleted feature point information is represented by connecting the part marked with “X” by dashed lines, and the survival feature point includes the feature point group {{tilde over (Y)}t+1} 330 after the deletion at the time t+1 and the three-dimensional state collection {{tilde over (X)}t+1} 430 after the deletion at the time t+.
After the Step 240, the generated feature point information ({{tilde over (Y)}t+1}, {{tilde over (X)}t+1}) is the so-called reliable feature point information, and the feature point groups in the feature point information all have a strong corresponding relationship. At this point, whether other images in the image sequence 10 still need to be processed or not is determined (Step 250). If yes, it is necessary to return to the Step 220 to execute in a recursion manner, however, before that, the step 260 of transferring the newly obtained reliable feature point information back to the stage of time series analysis 110 for being updated must be performed. That is, during the state updating process 113, {Yt+1}={Yt+1}+{{tilde over (Y)}t+1} and {Xt+1}={Xt+1}+{{tilde over (X)}t+1} are taken as the state ({Yt+1},{Xt+1}) of the reliable feature point information 50 when a next image frame in the image sequence 10 is processed.
Now the reliable feature point information 50 ({Yt+1},{Xt+1}) is the triangular pattern connected by dashed lines as in
On the contrary, if all the image frames in the image sequence 10 are processed, the finally remained reliable feature point information 50 (i.e., the so-called three-dimensional feature point information) is output to the state updating process 113 (Step 270), and the {Yt+1}={Yt+1}+{{tilde over (Y)}t+1} and {Xt+1}={Xt+1}+{{tilde over (X)}t+1} are taken as the final three-dimensional scene information of the whole image sequence 10.
As mentioned before, the input object is a group of space points having three-dimensional track or motion on the time axis. Since the space point itself has the content of the three-dimensional feature point information, when the above-mentioned state initializing process 111 executes the first situation of the initializing process, the process of generating three-dimensional state collection (i.e., the part of Step 210) executed when the first space point is input can be omitted.
The reconstruction result of-the three-dimensional scene will become more accurate by means of the operation of the present invention-as demonstrated in the comparison between
With the descriptions of invention, it will be obvious that the same may be varied in many ways. Such variations are not to be regarded as a departure from the principle and scope of the invention, and all such modifications as would be obvious to one skilled in the art are intended to be included within the scope of the following claims.
Number | Date | Country | Kind |
---|---|---|---|
095136372 | Sep 2006 | TW | national |