The present application claims priority to and the benefit of Chinese Patent Application No. 202210853238.3, filed on Jul. 20, 2022. The aforementioned application of which is incorporated by reference herein in its entirety.
The present document relates to the field of image processing, and in particular, to a prediction method for a target object, a computer device, and a storage medium.
At present, a LiDAR is a scanning sensor adopting the non-contact laser ranging technology, which has a similar working principle to that of a general radar system. Specifically, a target is detected by emitting laser beams, and point cloud data are formed by collecting the reflected beams and may sense the spatial range of the target object after certain processing. Therefore, the LiDAR is widely applied to the field of intelligent driving.
In the prior art, the point cloud data are mainly converted into dense voxel data; voxel features of the dense voxels are extracted; and calculation is performed on a dense voxel feature map. The calculation method may take a large computation cost for data in a larger spatial range, which is not conducive to the quick sensing of a remote target object.
A plurality of embodiments of the present document provide a prediction method and apparatus for a target object, a computer device, and a storage medium, to provide a method for improving the prediction efficiency of the target object.
Some embodiments of the present document provide a method for predicting a target object, which includes: performing a voxelization processing on point cloud data to obtain a plurality of voxels, wherein the plurality of voxels correspond to a plurality of point data in the point cloud data, and at least a portion of the plurality of voxels forms a voxel set; extracting voxel features of the voxels in the voxel set to obtain a plurality of voxel features; mapping the plurality of voxel features to a plurality of point data included in the plurality of voxels, respectively, to obtain a plurality of point data features of the plurality of point data;
and predicting, according to the plurality of point data features, the target object.
Some embodiments of the present document provide a prediction apparatus for a target object, which includes: a point cloud data voxelization processing module, configured to perform a voxelization processing on point cloud data to obtain a plurality of voxels, wherein the plurality of voxels correspond to a plurality of point data in the point cloud data, and at least a portion of the plurality of voxels forms a voxel set; a voxel feature extraction module, configured to extract voxel features of the voxels in the voxel set to obtain a plurality of voxel features; a voxel feature mapping module, configured to map the plurality of voxel features to a plurality of point data included in the plurality of voxels, respectively, to obtain a plurality of point data features of the plurality of point data; and a target object prediction module, configured to predict, according to the plurality of point data features, the target object.
Some embodiments of the present document provide a computer device, which includes memory and a processor. The memory has a computer program stored thereon, and the processor, when executing the computer program, implements the method according to any one of the above embodiments.
Some embodiments of the present document provide a computer-readable storage medium having a computer program stored thereon. The computer program, when executed by a processor, implements the method according to any one of the above embodiments.
In a plurality of embodiments provided by the present document, point cloud data, after a voxelization processing, are converted into voxel data; a voxel feature of a voxel is extracted; the voxel feature is mapped to point cloud data included in a voxel corresponding to the voxel feature to obtain a plurality of point data features; and the target object is predicted according to the point data features, so that the prediction efficiency of the target object is improved.
object according to some embodiments of the present document.
The technical solutions in the embodiments of the present document are described below with reference to the drawings in the embodiments of the present document. It is obvious that the described embodiments are only part of the embodiments of the present document, rather than all of the embodiments. All other embodiments obtained by those of ordinary skill in the art without making any creative effort on the basis of the embodiments of the present document shall fall within the protection scope of the present document.
In the related art, converting sparse point clouds into dense voxels is a common method for point cloud processing, which enables a detector to easily utilize efficient convolution operation and overcome the difficulties caused by sparse features on object detection. The method is a mainstream paradigm for point cloud-based 3D object detection in current autonomous driving scenes. Because the calculation is to be performed on the dense feature map, the paradigm has extremely high calculation cost in large-range point clouds, which is not conducive to the efficient remote sensing in an autonomous driving system. However, if the dense feature map in the detectors is directly removed, the problem of missing the center feature of the object is caused. The detectors based on point representation of sparse 3D objects perform feature extraction and prediction directly on the point representation. This approach involves multiple field point searches of quadratic complexity and usually involves down-sampling the number of point clouds to a small fixed number. This compromises the efficiency and flexibility of the detectors, making it difficult to apply them directly to a practical system.
Therefore, it is necessary to provide a prediction method for a target object, which may perform processing on sparse point clouds, without converting sparse point clouds into dense voxels, and without performing feature extraction and prediction directly on the point representation. The method includes: performing a voxelization processing on point cloud data to obtain a plurality of voxels, wherein the plurality of voxels correspond to a plurality of point data in the point cloud data, and at least a portion of the plurality of voxels forms a voxel set; extracting voxel features of the voxels in the voxel set to obtain a plurality of voxel features; mapping the plurality of voxel features to a plurality of point data included in the plurality of voxels, respectively, to obtain a plurality of point data features of the plurality of point data; and predicting, according to the plurality of point data features, the target object, so as to solve the technical problem of low efficiency in predicting the target object.
An application scene example of a prediction system for a target object is provided in the present document. In the process of vehicle traveling, a vehicle-borne LiDAR arranged on a vehicle may acquire the point cloud data of the surrounding environment of the vehicle according to a certain time interval. Certainly, the point cloud data of the surrounding environment of the vehicle may also be acquired at a certain distance interval in the process of vehicle traveling. Please refer to
Please refer to
Refer to the steps from sparse voxel feature extraction to point cloud classification and center prediction in
Then, instance feature extraction and instance prediction may be performed through a sparse instance recognition model. Specifically, instance features may be extracted from the plurality of point data by using algorithms for dynamic pooling and dynamic broadcasting. The dynamic pooling method may include weighting and summing the point data features in the first prediction point data set G1 to obtain a first instance feature, then assigning the first instance feature to the point data features in the first prediction point data set, and obtaining, according to the difference between the first instance feature and the point data features, a first target feature. As such, the above processes of dynamic pooling and dynamic broadcasting are repeated continuously and alternately multiple times to obtain the first target feature, which is used as a first prediction point data set feature. For example, an instance feature of a first instance IN1 may be extracted from the plurality of point data D1 by using the above algorithms for dynamic pooling and dynamic broadcasting, and an instance prediction for the first instance IN1 may be performed. In addition, an instance feature of a second instance IN2 may be extracted from the plurality of point data D2 by using the above algorithms for dynamic pooling and dynamic broadcasting, and an instance prediction for the second instance IN2 may be performed.
The method for extracting the first prediction point data set feature may further include repeating the above processes of dynamic pooling and dynamic broadcasting multiple times on coordinate vectors (x, y, z) of the point data to obtain a second target feature, and fusing the first target feature and the second target feature to obtain a third target feature, which is used as the first prediction point data set feature. A candidate prediction box for the target object is framed on the basis of the first prediction point data set feature. Specifically,
After obtaining the prediction boxes for the target objects, the server may perform the above processes of dynamic pooling and dynamic broadcasting repeatedly and alternately multiple times again on the foreground point data (e.g., the foreground point data outside the candidate prediction boxes and excluded therefrom and the foreground point data within the candidate prediction boxes and included therein) included in the candidate prediction boxes (e.g., the candidate prediction box F1 and the candidate prediction box F2), so as to correct the candidate prediction boxes, and further obtain corrected instances (a first instance INF and a second instance IN2′). According to the corrected instances, more accurate candidate prediction boxes may be obtained, and the prediction boxes for the target objects (e.g., target object O1 and object O2) are further obtained to predict the target objects. Through the above embodiments, the interference of background noise may be eliminated, and thus the accuracy of target object prediction is improved. Finally, the server may send the prediction box to a control system of a vehicle in an intelligent driving process, so that the vehicle may timely judge whether it is necessary to take braking measures, honk a horn, and the like.
Please refer to
LiDAR, an airborne LiDAR, a spaceborne LiDAR, a station LiDAR, a mobile LiDAR, or the like. The server 120 may be an electronic device having a certain computation processing capability, which may have a network communication module, a processor, a memory, and the like. Certainly, the server may also refer to software running on the electronic device. The server 120 may also be a distributed server, which may be a system with multiple processors, memories, network communication modules, and the like operating in coordination. Alternatively, the server 120 may also be a server cluster formed by several servers. Alternatively, with the development of science and technology, the server 120 may also be a new technical means capable of implementing the corresponding functions of the embodiments of the present document. For example, it may be a new form of “server” based on quantum computing implementations.
Please refer to
Step S210: performing a voxelization processing on point cloud data to obtain a plurality of voxels, wherein the plurality of voxels correspond to a plurality of point data in the point cloud data, and at least a portion of the plurality of voxels forms a voxel set. The point cloud data may include the target object.
In some cases, in the process of recognizing a target object by using point cloud data, the point cloud data needs to be converted into gridded data with a certain size. Therefore, the point cloud data can be first converted into volume data that can represent the range of an object.
The point cloud data may be a set of vectors in a three-dimensional coordinate system. The method for acquiring the point cloud data may include scanning through a LiDAR. Scanned data obtained by scanning are recorded in the form of points, and each of the points contains a three-dimensional coordinate. Certainly, the point cloud data may also include information such as color, reflectivity, and intensity that represents attributes of the scanned object. Certainly, the point cloud data may also be a set of vectors in a two-dimensional coordinate system.
According to the method for the voxelization processing, the point cloud data
can be divided into grids according to a spatial range, so that the point cloud data can be stored in a memory in order, which is conducive to reducing random memory access and increasing data computation efficiency. After voxels are obtained by voxelization of the point cloud data, a spatial convolution operation can be efficiently performed on the voxels, which is conducive to the extraction of multi-scale and multi-level local features. In another aspect, after the voxelization processing, down-sampling processing can be effectively performed on the point cloud data, which is suitable for processing the point cloud data of large order of magnitude.
The voxels may be pixels in a three-dimensional space. Specifically, the size of the voxels may be set in advance to obtain, e.g., a cube of 5 cm×5 cm×5 cm, a cube of 30 cm×30 cm×30 cm, any cube with length of sides between 5 cm and 30 cm, or a cube of other sizes. That is, one voxel can include several data points in the original point cloud data.
The voxel set includes a plurality of voxels. The voxels in the voxel set include point cloud data in the spatial range. Specifically, for example, if there is one voxel with a coordinate ranging from (0,0,0) to (1,1,1), and one of the point cloud data has a coordinate vector of (0.5,0.5,0.5), the voxel may belong to the voxel set. If there is no point cloud data in the spatial range of (0,0,0)-(1,1,1), the voxel does not belong to the voxel set.
Step S220: extracting voxel features of the voxels in the voxel set to obtain a plurality of voxel features.
In some cases, the voxel features may be used for classification of the voxels or segmentation of the target object. Therefore, in order to predict the target object through the point cloud data, the voxel features may be extracted first.
The voxel features may be obtained by performing convolution and pooling operations on the point cloud data in each voxel grid. Specifically, for example, the method for extracting the voxel features may include first using a common point cloud feature extractor such as PointNet to perform feature extraction to obtain initial voxel features; then, inputting the initial voxel features into a common sparse voxel encoder (such as a sparse convolution-based network and a sparse attention network) and the like, to extract the voxel features.
Step S230: mapping the plurality of voxel features to a plurality of point data included in the plurality of voxels, respectively, to obtain a plurality of point data features of the plurality of point data.
In some cases, if the feature extraction is directly performed on the point cloud data, the extracted features of the point cloud data may be poor, which directly affects the results of classification of the point cloud data or segmentation of the point cloud data. Since the point cloud data included in the voxel set are similar to the voxel features, the voxel feature of a voxel obtained after the voxelization processing can be mapped to the point cloud data included in the voxel.
The point data may be initial point cloud data. The point data may also be smaller voxels further divided on the basis of the currently divided voxels. Since a smaller voxel represents a smaller spatial range, this smaller voxel can be considered one point datum. There is little difference between the central vector (center of mass) of a smaller voxel and the vector of the point cloud data included in the voxel.
The point data features are those obtained by mapping a voxel feature to point clouds included in a voxel corresponding to the voxel feature. Specifically, for example, if four point data are included in one voxel, the four point data have the same point data feature as the voxel feature of the voxel. Certainly, in order to distinguish the four point data, it is also possible to add relative coordinates of the four point data with respect to the center of the voxel to which the four point data belong to the point data features.
Certainly, the point data features may also be obtained through a certain interpolation algorithm, such as an inverse distance-weighted interpolation algorithm, a linear interpolation method, a bilinear interpolation method, and a Kriging interpolation method. Specifically, for example, the step of finding the point data features using the inverse distance-weighted interpolation method may be: determining a reference pixel (i.e., voxel) to be used for the point data, calculating the distance between the vector of the point data and the central vector of the voxel, and using the following formula:
{circumflex over (Z)}(S0)=Σi=1NλiZ(Si) (formula 1)
where {circumflex over (Z)}(S0) represents features of point data to be interpolated, Z(Si) represents an attribute value of a reference pixel, N represents the number of the selected reference pixels, and A represents a weight coefficient calculated according to the reciprocal of the distance.
Step S240: predicting, according to the plurality of point data features, the target object.
In some cases, if the feature extraction of the point cloud data and the prediction are performed directly by a detector of points, the point cloud data needs to be down-sampled to a small amount, which damages the efficiency and flexibility of the detector to a certain extent. Therefore, the voxel features can be used as point data features after being assigned to the point data features, so that the features of the point clouds are enriched to a certain extent, and thus the accuracy of the target object prediction can be improved to a certain extent.
The prediction method for the target object includes performing a voxelization processing on point cloud data to obtain a plurality of voxels, wherein the plurality of voxels correspond to a plurality of point data in the point cloud data, and at least a portion of the plurality of voxels forms a voxel set; extracting voxel features of the voxels in the voxel set to obtain a plurality of voxel features; mapping the plurality of voxel features to a plurality of point data included in the plurality of voxels, respectively, to obtain a plurality of point data features of the plurality of point data; and predicting, according to the plurality of point data features, the target object, so as to solve the technical problem of low efficiency in predicting the target object.
In some embodiments, the step of performing the voxelization processing on the point cloud data to obtain the plurality of voxels may include: dividing, according to a spatial coordinate range of the point cloud data, the point cloud data into the plurality of voxels by a specified resolution; and on the basis of the point cloud data, extracting initial voxel features of the plurality of voxels.
In some cases, the point cloud data may be in an unordered form with a large data volume, so there is a certain difficulty in directly performing feature extraction on the point cloud data. Therefore, in order to change the point cloud data into ordered data with a small data volume, the voxelization processing can be performed on the point cloud data, which is conducive to the extraction of multi-scale and multi-level local feature information to a certain extent.
The method for performing the voxelization processing on the point cloud data may include determining, according to the minimum value and the maximum value of all of the point cloud data in three coordinate directions of X, Y, and Z, the smallest cuboid containing all of the point cloud data; setting, according to the size of the smallest cuboid and the resolution requirement, the size of a voxel, and dividing, according to the size of the voxel, the smallest cuboid into a plurality of voxels; and determining a voxel where each point in the point cloud data is located, and only displaying voxels containing the point cloud data after traversing all points of the point cloud data.
The initial voxel features may be extracted by using a common point cloud feature extractor. Specifically, for example, PointNet may be used to extract initial point cloud features. The PointNet is a neural network architecture that can process the point cloud data in an end-to-end manner. The PointNet includes two alignment networks, namely T-Net for aligning input data (input transform) and T-Net for aligning features (feature transform). They are recorded as T-Net-1 and T-Net-2, respectively. An output of T-Net-1 is a transformation matrix of 3×3 used for transforming an input (xyz). An output of T-Net-2 is a matrix of 64×64 used for aligning a feature space, such that a transformation matrix A of the feature space is as close to an orthogonal matrix as possible. Certainly, in the embodiments of the present document, the method for extracting the initial voxel features is not limited. The method for extracting the initial voxel features may also be VoxelNet, FaFNet, PointNet++, VoteNet, PointTrackNet, and the like.
In some embodiments, the step of extracting the voxel features of the voxels in the voxel set to obtain the plurality of voxel features may include: inputting the initial voxel features into a preset sparse voxel encoder to extract the voxel features.
In some cases, in order to enrich the extracted features, a sparse voxel encoder may be used to extract the voxel features.
The method for inputting the initial voxel features into the preset sparse voxel encoder to extract the voxel features may be performed by using a common sparse voxel encoder. Specifically, for example, the voxels included in the voxel set are first represented in the form of a sparse matrix, and then the sparse matrix is input into a 3D convolution network, for example, a sparse convolution-based network, to extract the features.
In some embodiments, the step of mapping the plurality of voxel features to the plurality of point data included in the plurality of voxels, respectively, to obtain the plurality of point data features of the plurality of point data may include: mapping the voxel features of the voxels in the voxel set, to point data included in voxels corresponding to the voxel features, to obtain initial point data features; calculating geometric features between the point data and centers of the voxels corresponding to the point data; and on the basis of the initial point data features and the geometric features, generating the point data features of the point data.
In some cases, since one voxel in the voxel set may include a plurality of point cloud data, if the voxel features are directly mapped to the point cloud data included in the voxel, the resulting point data features may be repeated. Therefore, in order to distinguish point cloud data included in one voxel, the geometric dimension of the point data can be increased, so that the acquired point data features can be distinguished.
The initial point data features may be obtained by directly mapping the voxel features to the point cloud data. The geometric features may be relational features between the point cloud data and the voxel to which the point cloud data belong. Specifically, for example, a distance relationship between the point cloud data and voxel data may be obtained according to the difference between the coordinate vectors of the point cloud data and the coordinate vector of the center of the voxel to which the point cloud data belong. Certainly, the geometric features may also include angles between straight lines connecting the point cloud data and the center of the voxel and a horizontal direction or a vertical direction or a certain specified direction.
In some embodiments, the step of predicting, according to the plurality of point data features, the target object includes: predicting, on the basis of the plurality of point data features, a plurality of point data belonging to a foreground of the target object, to obtain a plurality of foreground point data; performing clustering processing on the plurality of foreground point data to obtain a first prediction point data set belonging to the target object, wherein the first prediction point data set includes a plurality of point data corresponding to the target object; extracting point data features of the plurality of point data in the first prediction point data set to obtain a first prediction point data set feature; and on the basis of the first prediction point data set feature, predicting the target object.
In some cases, in order to achieve the prediction of the target object, one or more of an edge-based method, a region segmentation-based method, an attribute-based method, and a model-based method are generally adopted directly for the prediction. However, the point cloud-based segmentation method is prone to poor features extracted by the prediction method for the target object, and thus cannot be directly applied to the feature extraction. Therefore, the voxel features can be mapped to the point clouds to obtain point data features, and then the prediction is performed on the basis of the point data features.
The foreground point data are point cloud data representing the target object. Specifically, for example, the point cloud data obtained by scanning using a vehicle-borne LiDAR include buildings, trees, the ground, pedestrians, obstacles, and the like. Because the vehicle may be affected by the obstacles, the pedestrians, and the like during traveling, it is necessary to honk a horn or take an avoidance measure. Certainly, in some cases, ground data can also be used as foreground point cloud data to recognize lane lines, thus alerting the vehicle to the need for immediate lane changes during running.
The method for the clustering processing may be a process of classifying foreground point clouds, in which the point cloud data representing the same target object are clustered. Specifically, for example, the point cloud data are divided into n categories in advance and n objects are randomly selected as initial clustering centers. Then, the distance between each of the objects and each of the clustering centers is calculated, and each of the objects is assigned to the clustering center closest to the object. The cluster centers and the respective objects assigned to them represent a cluster.
When one sample is assigned, the cluster center of a cluster is recalculated on the basis of the existing objects in the cluster. This process may be repeated until a certain termination condition is met. The termination condition may be that no (or a minimum number of) objects are reassigned to different clusters, no (or a minimum number of) cluster centers are changed again, or the sum of squared errors is locally minimal. Certainly, the method for the clustering processing may also be a hierarchy-based clustering method, a grid-based clustering method, a maxi-min distance clustering method, a graph theory clustering method, or the like.
The first prediction point data set is a set of point cloud data representing the same target object. Specifically, for example, the foreground point data are clustered to obtain six categories of pedestrian A, pedestrian B, vehicle A, vehicle B, obstacle A, and obstacle B, and the first prediction point data set may be one of the pedestrian A, the pedestrian B, the vehicle A, the vehicle B, the obstacle A, and the obstacle B. The first prediction point data set feature includes features of the same target object. Specifically, for example, if an obstacle includes a plurality of point cloud data features, the feature of the obstacle may be a feature of a new point obtained from the features of the point cloud data representing the obstacle according to a certain algorithm (such as averaging) or the like, which is used as the first prediction point data set feature.
In some embodiments, the step of predicting, on the basis of the plurality of point data features, the plurality of point data belonging to the foreground of the target object, to obtain the plurality of foreground point data may include: predicting a type of a target object corresponding to the plurality of point data by using the plurality of point data features; and on the basis of the type of the target object, determining the plurality of foreground point data.
In some cases, if the point cloud data are directly divided into foreground points and background points by using a binary classification method, the case of misclassification may be caused by the difference in the properties of the point clouds. Therefore, the point data may be divided into a plurality of categories, each of which represents a different target object.
The type of the target object may be an obstacle, a pedestrian, a tree, a lane, a building, or the like.
In some embodiments, the step of performing clustering processing on the plurality of foreground point data to obtain the first prediction point data set belonging to the target object may include: determining a plurality of center point data representing the target object (e.g., a plurality of center point data C1 representing a first target object and a plurality of center point data C2 representing a second target object, as shown in
In some cases, in the process of clustering the point clouds, the centers with all foreground points predicted can be used as nodes, and the distances between the points can be used as the weights of the edges to construct a graph structure. If the distance between two points is less than a certain threshold, the two points are considered to be connected. The connected component algorithm in graph theory is used to find all connected components in the graph. Each connected component found is considered an instance. Thus, each foreground point has its unique instance identifier, which indicates to which instance the point belongs.
Please refer to
In some cases, after the instance identifier of each data point is obtained, in order to obtain the first prediction point data set feature, the point data features may be fused according to a preset weight coefficient to obtain an instance feature, and then the first prediction point data set feature may be obtained according to the difference between the instance feature and the point data features.
The first instance feature is obtained by fusing the point data features of the point data in the first prediction point data set. Specifically, for example, if an instance includes 1000 pieces of point data, features of the 1000 pieces of point data may be averaged to obtain an average feature, and then the average feature is used as an instance feature. Certainly, the instance feature may also be obtained by fusing the features of the 1000 pieces of point data according to a certain weight coefficient.
The first target feature may be the difference between the instance feature and the point data features. Specifically, for example, the instance feature may be assigned to the features of the 1000 pieces of point data by broadcasting. Then, the difference between the feature of each of the 1000 pieces of point data and the instance feature is calculated respectively to obtain the first target feature. Certainly, the instance feature and the first target feature may be iterated multiple times, that is, the first target feature may be used as a new point data feature, and then the above processes of dynamic pooling and dynamic broadcasting are repeated on the basis of the new point data feature, so as to finally obtain a first target feature data. Specifically, for example, in
In some embodiments, on the basis of a second weight coefficient, weighted coordinates of the point data in the first prediction point data set are calculated; the difference between coordinates and the weighted coordinates of the point data in the first prediction point data set is determined to obtain a second target feature; and the first target feature and the second target feature are fused to obtain a third target feature, wherein the third target feature is used as the first prediction point data set feature.
In some cases, the first prediction point data set feature obtained by simply performing computation on the point data features may be not rich enough. Therefore, features of the coordinate dimension may also be added to the process of acquiring the first prediction point data set feature.
The second target feature may be an average of the coordinate vectors of the point data in the instance. Specifically, for example, if the coordinate vectors of 5 points are (1,1,1), (2,2,2), (3,3,3), (4,4,4), and (5,5,5), respectively, the second target feature may be (3,3,3).
The third target feature is a result obtained after the first target feature and the second target feature are fused. Specifically, for example, if the first target feature is a 128-dimensional vector and the second target feature is a 3-dimensional vector, the third target feature may be a 131-dimensional vector obtained by splicing the first target feature and the second target feature.
In some embodiments, the prediction method for the target object may further include: on the basis of the first prediction point data set feature, generating a prediction box indicating that an object is predicted to be the target object, wherein at least one of the plurality of foreground point data within the prediction box is used as a second prediction point data set; calculating a second instance feature of the point data in the second prediction point data set by using a third weight coefficient; on the basis of the difference between the point data features of the point data in the second prediction point data set and the second instance feature, obtaining a fourth target feature; and adjusting, according to the fourth target feature, the predicted target object.
In some cases, the results after one target object recognition may be not accurate enough. Therefore, a spatial range of the target object can be framed after the first target object recognition, and then a prediction box is corrected or adjusted by using the above processes of dynamic pooling and dynamic broadcasting again for the foreground point data in the spatial range. In some embodiments, the above correction or adjustment process may be performed multiple times.
The second prediction point data set is the foreground point data included in the prediction box. Specifically, for example, a point A1 is excluded from the category A for computation in the clustering process, and a point A2 is included in the category A. However, the resulting prediction box includes the point A1 and does not include the point A2. Therefore, the point A1 should be included in the second prediction point data and the point A2 should be excluded.
The second instance feature is obtained after the features of the foreground point data included in the prediction box are processed according to the weight coefficients. The fourth target feature is derived from the difference between the second instance feature and the point data feature in the second prediction point data set.
Please refer to
The point cloud data voxelization processing module 610 is configured to perform a voxelization processing on point cloud data to obtain a plurality of voxels, wherein the plurality of voxels correspond to a plurality of point data in the point cloud data, and at least a portion of the plurality of voxels forms a voxel set.
The voxel feature extraction module 620 is configured to extract voxel features of the voxels in the voxel set to obtain a plurality of voxel features.
The voxel feature mapping module 630 is configured to map the plurality of voxel features to a plurality of point data included in the plurality of voxels, respectively, to obtain a plurality of point data features of the plurality of point data.
The target object prediction module 640 is configured to predict, according to the plurality of point data features, the target object.
For specific limitations of the prediction apparatus for the target object, reference may be made to the limitations of the above prediction method for the target object, which are not repeated here. The modules in the prediction apparatus for the target object described above may be implemented wholly or partially by software, hardware, or a combination thereof. The above modules may be embedded in or independent of a processor in a computer device in the form of hardware, or stored in a memory in the computer device in the form of software, so that the processor can call and execute the operations corresponding to the above modules.
Please refer to
The embodiments of the present document further provide a computer-readable storage medium having a computer program stored thereon. The computer program, when executed by a computer, causes the computer to perform the prediction method for the target object in any one of the above embodiments.
The embodiments of the present document further provide a computer program product including an instruction, which, when executed by a computer, causes the computer to perform the prediction method for the target object in any one of the above embodiments.
It should be understood that specific examples are set forth herein only to assist those skilled in the art in better understanding the embodiments of the present document and are not intended to limit the scope of the present document.
It should be understood that, in the various embodiments of the present document, the sequence number of a process or step does not mean the order of execution, and the order of execution of each process or step should be determined by its function and inherent logic, and shall not limit the implementation process of the embodiments of the present document in any way.
It should be understood that the various embodiments described in the present document may be implemented individually or in combination, which is not limited by the embodiments of the present document.
Unless otherwise defined, all technical and scientific terms used in the embodiments of the present document have the same meaning as commonly understood by those of ordinary skill in the art to which the present document belongs. The terms used in the present document are for the purpose of describing specific embodiments only and are not intended to limit the scope of the present document. The term “and/or” used herein includes any and all combinations of one or more of the associated listed items. The singular forms “a/an”, “said”, and “the” used in the embodiments of the present document and the appended claims are intended to include the plural forms as well, unless otherwise clearly indicated in the context.
It is understood that the processor of the embodiments of the present document may be an integrated circuit chip having a signal processing capability. In the implementation process, the steps of the above method embodiments may be performed by integrated logic circuits of hardware in a processor or by instructions in the form of software. The above processor may be a general-purpose processor, a digital signal processor (DSP), an application specific integrated circuit (ASIC), a field programmable gate array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, or discrete hardware components. The methods, steps, and logic block diagrams disclosed in the embodiments of the present document can be implemented or performed. The general-purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the methods disclosed in conjunction with the embodiments of the present document can be directly embodied as performed by a hardware decoding processor or performed with a combination of hardware and software modules in a decoding processor. The software module can be located in a random access memory, a flash memory, a read-only memory, a programmable read-only memory, or an electrically rewritable programmable memory, a register, and other storage media well known in the art. The storage medium is located in the memory, and the processor reads the information in the memory and completes the steps of the above method in combination with its hardware.
It is understood that the memory in the embodiments of the present document may be a volatile memory or a nonvolatile memory, or may include both a volatile memory and a nonvolatile memory. The nonvolatile memory may be a read-only memory (ROM), a programmable read-only memory (PROM), an erasable programmable read-only memory (EPROM), an electrically erasable programmable read-only memory (EEPROM), or a flash memory. The volatile memory may be a random access memory (RAM). It should be noted that the memories of the systems and methods described herein are intended to include, but are not limited to these and any other suitable types of memories.
Those of ordinary skill in the art will recognize that the units and algorithmic steps of the various examples described in conjunction with the embodiments disclosed herein are capable of being implemented in electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are performed in hardware or software depends on the particular application and design constraints of the technical solutions. The skilled person may use different methods to implement the described functions for each particular application, but such implementations should not be considered outside the scope of the present document.
It will be clear to those skilled in the field that, for the convenience and brevity of present document, the specific working processes of the systems, apparatuses, and units described above can be referred to the corresponding processes in the preceding method embodiments and will not be repeated here.
In the several embodiments provided in the present document, it should be understood that the disclosed systems, apparatuses, and methods can be implemented in other ways. For example, the apparatus embodiments described above are merely illustrative, for example, the division of the units described, which is only a logical functional division, can be practically implemented in another way. For example, a plurality of units or components may be combined or integrated into another system, or some features may be ignored or not performed. In another aspect, the mutual coupling or direct coupling or communication connection shown or discussed may be achieved through some interfaces. An indirect coupling or communication connection between apparatuses or units may be in electrical, mechanical, or other forms.
The units illustrated as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place or may be distributed to a plurality of network units. Some or all of these units may be selected according to practical needs to achieve the purpose of the embodiment solution.
In addition, the functional units in the embodiments of the present document may be integrated into one processing unit, or each unit may be physically present separately, or two or more units may be integrated into one unit.
The function, if implemented in the form of a software functional module and sold or used as an independent product, may be stored in a computer-readable storage medium. Based on such understanding, the technical solutions of the present document essentially can be, or part of the technical solutions contributing to the prior art can be, or part of the technical solutions can be embodied in the form of a software product. The computer software product is stored in a storage medium and includes several instructions for enabling a computer device (which can be a personal computer, a server, a network device, or the like) to implement all or part of the steps of the methods described in the embodiments of the present document. The preceding storage medium includes a U-disk, a removable hard disk, a read-only memory (ROM), a random access memory (RAM), a magnetic disk, an optical disk, or other media capable of storing program codes.
What is described above is only specific embodiments of the present document, but the protection scope of the present document is not limited thereto. Those skilled in the art can easily think of changes or substitutions within the technical scope disclosed in the present document, and these changes and substitutions shall fall within the protection scope of the present document. Therefore, the protection scope of the present document shall be subject to the protection scope of the claims.
Number | Date | Country | Kind |
---|---|---|---|
202210853238.3 | Jul 2022 | CN | national |