The present disclosure relates to field space data normalization processing methods applicable to GBP (Gaussian Belief Propagation) algorithm and, more particularly, to a field space data normalization processing method applicable to GBP algorithm for simultaneous localization and mapping (SLAM) related to machine learning.
SLAM aims to enable robots to achieve localization, mapping, and navigation in unknown surroundings. SLAM technology is currently widely applied to robots, unmanned aerial vehicles (UAVs), self-driving cars, AR and VR to achieve autonomous localization, mapping, navigation of machines with sensors. Existing mainstream SLAM technology application falls under two categories: laser SLAM for use in mapping and navigation using laser radar, and vision SLAM based on monocular/binocular camera vision mapping and navigation.
When it comes to developing vision SLAM, bundle adjustment is the last step of nearly all the algorithms which use feature points in 3D model reconstruction and the best solution for optimization of 3D models, camera poses, camera matrix, and deformation parameters.
The concepts of neural networks are incorporated into GBP (Gaussian Belief Propagation) algorithm to compute bundle adjustment in SLAM. Initial field space data value varies from data and processing technique to data and processing technique; thus, neural network models are extremely unstable during training, and computation indicates that abnormal values cause a training failure. Furthermore, when initial field space data is used to train neural networks, deviations of overlearned data lead to low generalization of models
Therefore, data preprocessing is an important step in machine learning and usually involves performing normalization on data. However, SLAM technology has a drawback: camera pose data exists in the space of mathematically defined SE(3)-Special Euclidean Group, and SE(3) is unable to perform normalization on data directly.
Referring to
The present disclosure provides a field space data normalization processing method applicable to GBP algorithm to overcome drawbacks of machine learning, namely, the initial field space data value is incorrect, and model generalization is low.
The present disclosure provides a field space data normalization processing method applicable to GBP algorithm and adapted to perform on a computer the steps of: performing vector matrix transformation by inputting a first vector data with a first computation module and transforming it into a first matrix expression
in a se(3) corresponding to SE(3), where
performing first matrix transformation by transforming, in a second computation module, a first matrix expression
into a second matrix expression
in SE(3); providing a means of normalization, comprising providing a data normalization algorithm
where X denotes unnormalized source data, μ denotes the mean of unnormalized source data X, Xmax and Xmin denote the maximum and minimum of unnormalized source data X, respectively, Xnom denotes normalized data; performing normalization and third matrix transformation by performing data normalization on the second matrix expression
with the data normalization algorithm and transforming it into a third matrix expression
where
S=Xmax−Xmin, and S denotes range; and performing normalization by performing data normalization on the second matrix expression
in a third computation module; performing second matrix transformation by transforming, in a fourth computation module, data-normalized second matrix expression
into a third matrix expression
in SE(3), where
S=Xmax−Xmin, μ is the mean of X; and transforming third matrix and outputting second vector data by transforming, in a fifth computation module, the third matrix expression
into a fourth matrix expression
in se(3) corresponding to SE(3), and then transforming it into a second vector data and outputting it, thereby finalizing normalization of the first vector data.
The implementation of the present disclosure at least achieves inventive benefits as follows:
The structure and the technical means adopted by the present invention to achieve the above and other objects can be best understood by referring to the following detailed description of the preferred embodiments and the accompanying drawings, wherein:
Referring to
Lie groups are continuous and smooth; they are groups and manifolds. Intuitively, a rigid body is able to move continuously within a space, and thus both SO(3) and SE (3) are Lie groups. SO(3) is a special orthogonal group, whereas SE(3) is a special Euclidean group. A three-dimensional rotational matrix constitutes a special orthogonal group SO(3), whereas a transformation matrix constitutes a special Euclidean group SE(3). Both SO(3) and SE(3) can work with well-defined multiplication rather than addition and thus are hardly useful in taking limits and taking derivatives.
A Lie algebra corresponds to a Lie group and lies in a vector space. They are commonly abbreviated to so(3) and se(3). In fact, a Lie algebra is a tangent space at an identity element of a Lie group. A tangent space itself is a vector space, and thus the tangent space can be defined with a good addition computation. Therefore, characteristics of a Lie group can be evaluated with a corresponding Lie algebra (i.e., tangent space). For the sake of illustration, se(3) corresponding to SE(3), which the SE(3) in this embodiment corresponds to, is abbreviated to “se(3) corresponding to SE(3).”
The step of performing vector matrix transformation S10 is explained below. With the first computation module A block, a first vector data, for example, (ρ,ω)=(ρ1, ρ2, ρ3, ω1, ω2, ω3)∈R6, is inputted, and then the first vector data is transformed into a first matrix expression
in a se(3) corresponding to SE(3), where
Therefore, R6 space includes therein: the leading three dimensions ρ1, ρ2, ρ3 being translational, and the trailing three dimensions ω1, ω2, ω3 being rotational. Finally, a six-dimensional vector is transformed into a four-dimensional matrix.
The step of performing first matrix transformation S20 involves performing exponential mapping from “se(3) corresponding to SE(3)” to “SE(3),” with details being explained below. In a second computation module B block, a first matrix expression
in se(3) corresponding to SE(3) in a first computation module A block is transformed with, for example, an exponential function exp(.), into a second matrix expression
in SE(3). In short, the first matrix transformation is carried out with, for example, an exponential function exp(.).
The step of providing a means of normalization S30 is explained below. A third computation module C block is provided with a data normalization algorithm, for example, min-max normalization algorithm:
, where X denotes unnormalized source data, and μ denotes the
mean of unnormalized source data X, allowing the data to be restricted to the range [−1, 1] and to have the mean equal to 0, where Xmax and Xmin denote the maximum and minimum of unnormalized source data X, respectively, and Xnom denotes normalized data.
The step of performing normalization and third matrix transformation S40 is explained below. In a fourth computation module D block, the second matrix expression
in SE(3) undergoes data normalization with a data normalization algorithm of the third computation module C block and transformation to obtain a third matrix expression
where
S=Xma−Xmin, and S denotes a range.
The step of transforming a third matrix and outputting a second vector data S50 is explained below. In the fifth computation module E block, the normalized third matrix expression
in SE(3) of the fourth computation module D block is transformed back to a fourth matrix expression
in se(3) corresponding to SE(3) to undergo the third matrix transformation with, for example, a logarithmic function ln(.).
The step of outputting second vector data is explained below. In the fifth computation module (E block), the fourth matrix expression
is transformed once again into a second vector data to be outputted, where the second vector data, for example, is expressed by ({tilde over (ρ)}, ω)=(, , , ω1, ω2, ω3), thereby finalizing the normalization of the first vector data.
To prove that the present disclosure demonstrates significant inventiveness, both the data normalized in this embodiment and the data normalized according to the prior art undergo train loss simulation analysis and comparison to yield the results shown in Table 1. As shown in Table 1, the neural network model training carried out with the normalized data in this embodiment can attain full completion, but the neural network model training carried out with conventional unnormalized data ends up with NaN at the 20th epoch or thereafter and thus prevents the model from being fully, effectively trained and validated.
The above description is only the preferred embodiments of the present invention and is not intended to limit the present invention in any form. Although the invention has been disclosed as above in the preferred embodiments, they are not intended to limit the invention. A person skilled in the relevant art will recognize that equivalent embodiment modified and varied as equivalent changes disclosed above can be used without parting from the scope of the technical solution of the present invention. All the simple modification, equivalent changes and modifications of the above embodiments according to the material contents of the invention shall be within the scope of the technical solution of the present invention.
Number | Date | Country | Kind |
---|---|---|---|
111116523 | Apr 2022 | TW | national |