WO2022166400A1 - 一种处理三维点云的方法、装置、设备以及存储介质 - Google Patents
一种处理三维点云的方法、装置、设备以及存储介质 Download PDFInfo
- Publication number
- WO2022166400A1 WO2022166400A1 PCT/CN2021/137305 CN2021137305W WO2022166400A1 WO 2022166400 A1 WO2022166400 A1 WO 2022166400A1 CN 2021137305 W CN2021137305 W CN 2021137305W WO 2022166400 A1 WO2022166400 A1 WO 2022166400A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- point
- points
- indistinguishable
- neural network
- point cloud
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 91
- 238000012545 processing Methods 0.000 title claims abstract description 88
- 238000003860 storage Methods 0.000 title claims abstract description 13
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 81
- 230000004927 fusion Effects 0.000 claims abstract description 33
- 238000012549 training Methods 0.000 claims description 51
- 238000012360 testing method Methods 0.000 claims description 13
- 238000004590 computer program Methods 0.000 claims description 12
- 238000012795 verification Methods 0.000 claims description 9
- 230000002776 aggregation Effects 0.000 claims description 8
- 238000004220 aggregation Methods 0.000 claims description 8
- 238000011176 pooling Methods 0.000 claims description 7
- 230000004931 aggregating effect Effects 0.000 claims description 3
- 238000000691 measurement method Methods 0.000 claims description 3
- 230000008569 process Effects 0.000 description 33
- 230000011218 segmentation Effects 0.000 description 26
- 238000010586 diagram Methods 0.000 description 13
- 230000006870 function Effects 0.000 description 10
- 238000011156 evaluation Methods 0.000 description 8
- 238000013528 artificial neural network Methods 0.000 description 6
- 230000008859 change Effects 0.000 description 6
- 238000000605 extraction Methods 0.000 description 5
- 238000001514 detection method Methods 0.000 description 4
- 238000004422 calculation algorithm Methods 0.000 description 3
- 239000003086 colorant Substances 0.000 description 3
- 238000013461 design Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 239000013598 vector Substances 0.000 description 3
- 230000003044 adaptive effect Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000002708 enhancing effect Effects 0.000 description 2
- 238000012854 evaluation process Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 230000000750 progressive effect Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000013501 data transformation Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000011426 transformation method Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Definitions
- the present application belongs to the field of computer technology, and in particular relates to a method for processing a 3D point cloud, a device for processing a 3D point cloud, a device for processing a 3D point cloud, and a storage medium.
- Point cloud (full name in English) is a collection of point data on the surface of the product obtained by measuring instruments in reverse engineering. In addition to geometric positions, point cloud data also has color information. Color information is usually obtained by acquiring a color image through a camera, and then assigning the color information (RGB) of the pixel at the corresponding position to the corresponding point in the point cloud. The acquisition of intensity information is the echo intensity collected by the receiving device of the laser scanner. This intensity information is related to the surface material, roughness, incident angle direction of the target, as well as the emitted energy and laser wavelength of the instrument.
- the 3D point cloud data is denormalized.
- the multi-view projection technology projects the denormalized 3D point cloud into a 2D image, and then converts the 2D point cloud into a 2D image.
- Image processing At present, the point cloud data processing needs to convert the point cloud data into other data formats, such as projecting a three-dimensional point cloud to a two-dimensional image as the input of the convolutional neural network; however, this process has the following shortcomings: (1) Due to occlusion, the projection process itself will cause some data to be missing. (2) The amount of calculation in the process of data transformation is relatively large. Therefore, it is necessary to directly construct a convolutional neural network to process 3D point cloud data.
- the embodiments of the present application provide a method for processing a 3D point cloud, a device for processing a 3D point cloud, a device for processing a 3D point cloud, and a storage medium, so as to solve the problem of existing methods that can directly process 3D point cloud data.
- the processed convolutional neural network cannot accurately extract the feature information of each point, which leads to the problem of inaccurate prediction results when predicting the category of these points.
- a first aspect of the embodiments of the present application provides a method for processing a three-dimensional point cloud, including:
- the point cloud data is input into the trained convolutional neural network for processing, and the target feature corresponding to each point is obtained.
- the convolutional neural network includes a geometric attention fusion module and a focusing module.
- the geometric attention fusion module For extracting the local enhancement feature of each said point, the focusing module is used for extracting the target feature of each said point based on the local enhancement feature of each said point;
- the prediction category corresponding to each point is determined.
- a second aspect of the embodiments of the present application provides an apparatus for processing a three-dimensional point cloud, including:
- the processing unit is used to input the point cloud data into the trained convolutional neural network for processing, and obtain the target feature corresponding to each point.
- the convolutional neural network includes a geometric attention fusion module and a focusing module.
- the geometric attention fusion module is used for extracting the local enhancement feature of each said point
- the said focusing module is used for extracting the target feature of each said point based on the local enhancement feature of each said point;
- the determining unit is used for determining the prediction category corresponding to each point based on the target feature corresponding to each point.
- a third aspect of the embodiments of the present application provides a device for processing a three-dimensional point cloud, including a memory, a processor, and a computer program stored in the memory and running on the processor, characterized in that: When the processor executes the computer program, the steps of the method for processing a three-dimensional point cloud according to the first aspect above are implemented.
- a fourth aspect of the embodiments of the present application provides a computer-readable storage medium, where a computer program is stored in the computer-readable storage medium, and when the computer program is executed by a processor, the three-dimensional processing according to the first aspect above is implemented.
- the steps of the point cloud method are described in detail below.
- a fifth aspect of the embodiments of the present application provides a computer program product that, when the computer program product runs on a device for processing a three-dimensional point cloud, causes a device for processing a three-dimensional point cloud to execute the above-mentioned first aspect. Steps of a method for processing a three-dimensional point cloud.
- a method for processing a three-dimensional point cloud, a device for processing a three-dimensional point cloud, a device for processing a three-dimensional point cloud, and a storage medium provided by the embodiments of the present application have the following beneficial effects:
- a device for processing a three-dimensional point cloud processes point cloud data through a trained convolutional neural network to obtain a target feature corresponding to each point, and determines each point based on the target feature corresponding to each point. the corresponding prediction category. Because when extracting the target feature corresponding to each point, the local enhancement feature of each point is extracted based on the geometric attention fusion module included in the convolutional neural network, and then based on the focusing module included in the convolutional neural network and each point The local enhancement features are extracted to obtain the target features of each point.
- the target feature of each point extracted based on this method contains important geometric information corresponding to each point, which makes the extracted target feature of each point more accurate and effective, and then predicts according to the target feature of each point. category, the resulting predictions are very accurate.
- Fig. 1 is a kind of schematic diagram of difficult division of complex point cloud scene provided by the present application
- FIG. 2 is a schematic flowchart of a method for processing a three-dimensional point cloud provided by an embodiment of the present application
- FIG. 3 is a schematic flowchart of a method for processing a three-dimensional point cloud provided by another embodiment of the present application.
- FIG. 4 is a schematic diagram of a geometric attention fusion module provided by the present application.
- FIG. 5 is a schematic flowchart of a method for processing a 3D point cloud provided by another embodiment of the present application.
- FIG. 6 is a schematic flowchart of a method for processing a three-dimensional point cloud provided by yet another embodiment of the present application.
- FIG. 7 is a schematic diagram of a focusing module provided by the present application.
- FIG. 8 is an evaluation process of the new evaluation criteria for indivisible regions provided by the present application.
- Fig. 9 is a kind of semantic segmentation network oriented to large-scale complex scene point cloud provided by this application.
- FIG. 11 provides an application scenario diagram for this application
- FIG. 12 is a schematic diagram of an apparatus for processing a three-dimensional point cloud provided by an embodiment of the present application.
- FIG. 13 is a schematic diagram of a device for processing a three-dimensional point cloud provided by another embodiment of the present application.
- a point cloud (full name in English) is a collection of point data on the appearance surface of a product obtained by a measuring instrument in reverse engineering.
- the point cloud data also has color information.
- Color information is usually obtained by acquiring a color image through a camera, and then assigning the color information (RGB) of the pixel at the corresponding position to the corresponding point in the point cloud.
- the acquisition of intensity information is the echo intensity collected by the receiving device of the laser scanner. This intensity information is related to the surface material, roughness, incident angle direction of the target, as well as the emitted energy and laser wavelength of the instrument.
- the 3D point cloud data is denormalized.
- the multi-view projection technology projects the denormalized 3D point cloud into a 2D image, and then converts the 2D point cloud into a 2D image.
- Image processing At present, the point cloud data processing needs to convert the point cloud data into other data formats, such as projecting a three-dimensional point cloud to a two-dimensional image as the input of the convolutional neural network; however, this process has the following shortcomings: (1) Due to occlusion, the projection process itself will cause some data to be missing. (2) The data conversion process requires a large amount of calculation, which will lead to a large amount of memory consumption, occupy a lot of computer resources, and easily lose spatial geometric information during the conversion process.
- one-dimensional convolutional neural networks can directly manipulate and process non-normalized point cloud data.
- the basic idea is to learn the spatial encoding of each point, and then aggregate all single point features into an overall representation. But this design cannot fully capture the relationship between points.
- An enhanced version of point cloud convolution can divide point clouds into overlapping local regions based on distance measurements in the underlying space, and use 2D convolution to extract local feature neighborhood structures that capture fine geometry. However, it only considers the local area of each point and cannot correlate similar local features on the point cloud.
- the existing convolutional neural networks that can directly process 3D point cloud data cannot accurately extract the feature information of each point, resulting in inaccurate prediction results when predicting the category of these points.
- the existing 3D scene point cloud processing methods are particularly poor in segmentation of difficult-to-segment regions, and the problems are mainly concentrated in the segmentation edges of objects, inside objects that are easily confused, and some discrete and confusing small regions.
- FIG. 1 is a schematic diagram of a difficult region of a complex point cloud scene provided by the present application.
- the first type is the complex boundary region, which belongs to boundary points (object boundary and prediction boundary).
- boundary points object boundary and prediction boundary.
- the second type is the obfuscated interior region, which contains interior points from different classes of objects with similar textures and geometries. For example, doors and walls have a similar appearance, are almost flat, and have similar colors. In this case, even for humans, it is difficult to accurately identify whether certain points belong to a door or a wall.
- the third type is isolated small areas, which are scattered and difficult to predict. Also, objects in the scene are not fully captured by the device due to occlusion. Therefore, for points in isolated small regions, it is also impossible to accurately predict the class to which they belong.
- the present application provides a method for processing a three-dimensional point cloud.
- a device for processing a three-dimensional point cloud processes the point cloud data through a trained convolutional neural network, and obtains The target feature corresponding to each point, and the prediction category corresponding to each point is determined based on the target feature corresponding to each point.
- the local enhancement feature of each point is extracted based on the geometric attention fusion module included in the convolutional neural network, and then based on the focusing module included in the convolutional neural network and each point
- the local enhancement features are extracted to obtain the target features of each point.
- the target feature of each point extracted based on this method contains important geometric information corresponding to each point, which makes the extracted target feature of each point more accurate and effective, and then predicts according to the target feature of each point. category, the resulting predictions are very accurate.
- the method for processing 3D point cloud provided by this application can be applied to various fields that need to analyze 3D point cloud, such as automatic driving (such as obstacle detection and automatic path planning of automatic driving equipment, etc.), robots (object detection of home service robots, etc.) , route recognition, etc.) and other human-computer interaction fields, this method can provide users with real-time and accurate behavior recognition and detection functions, improve accuracy and interest, and ensure the safety of human-computer interaction activities. This is only an exemplary description, and it is not limited.
- FIG. 2 is a schematic flowchart of a method for processing a three-dimensional point cloud provided by an embodiment of the present application.
- the execution subject of the method for processing a three-dimensional point cloud in this embodiment is a device for processing a three-dimensional point cloud, and the device includes but is not limited to a smartphone, a tablet computer, a computer, a Personal Digital Assistant (PDA), a notebook computer, a supercomputer A mobile personal computer (ultra-mobile personal computer, UMPC), a netbook, an independent server, a distributed server, a server cluster or a cloud server, etc., may also include a terminal such as a desktop computer.
- the method for processing a three-dimensional point cloud as shown in FIG. 2 may include S101 to S103, and the specific implementation principles of each step are as follows.
- S101 Acquire point cloud data including multiple points.
- Point cloud data including a plurality of points may be acquired by a device that processes a three-dimensional point cloud.
- the device for processing the three-dimensional point cloud includes a laser device, a stereo camera, or a time-of-flight camera, etc.
- the collection can be performed by a laser device, a stereo camera, or a time-of-flight camera.
- point cloud data can be collected for three-dimensional objects by using a data collection method based on automatic point cloud splicing. During the collection process, multiple stations can be used to scan and stitch the data of each station together to obtain point cloud data. , and achieve accurate registration of point clouds from different angles by iteratively optimizing the coordinate transformation parameters.
- S102 Input the point cloud data into a trained convolutional neural network for processing to obtain a target feature corresponding to each point.
- the convolutional neural network includes a geometric attention fusion module and a focusing module.
- the geometric attention The fusion module is used for extracting the local enhancement feature of each said point
- the focusing module is used for extracting the target feature of each said point based on the local enhancement feature of each said point.
- a pre-trained convolutional neural network is pre-stored in the device for processing a three-dimensional point cloud.
- the convolutional neural network is obtained by training the initial convolutional neural network based on the training set and the test set using a machine learning algorithm.
- the convolutional neural network includes a geometric attention fusion module and a focusing module.
- the geometric attention fusion module is used to extract the local enhancement feature of each point
- the focusing module is used to extract each point based on the local enhancement feature of each point. target features.
- the training set includes sample point cloud data of multiple sample points
- the test set includes sample features and sample categories corresponding to each sample point.
- the convolutional neural network can be pre-trained by a device that processes 3D point clouds, or the files corresponding to the convolutional neural network can be transplanted to a device that processes 3D point clouds after being pre-trained by other devices. That is, the execution subject for training the convolutional neural network and the execution subject for using the convolutional neural network may be the same or different. For example, when other equipment is used to train the initial convolutional neural network, after the other equipment finishes training the initial convolutional neural network, the network parameters of the initial convolutional neural network are fixed to obtain a file corresponding to the convolutional neural network. This file is then ported to a device that processes 3D point clouds.
- the device for processing 3D point cloud uses the geometric attention fusion module included in the convolutional neural network to extract the local enhancement feature of each point; and then based on the local enhancement feature of each point, Using the focusing module included in the convolutional neural network, the target features of each point are extracted.
- the predicted probability value corresponding to each category corresponding to each point is determined; based on the predicted probability value corresponding to each category, the predicted category corresponding to each point is determined.
- the local enhancement feature of each point is first extracted based on the geometric attention fusion module included in the convolutional neural network, and then based on the focus included in the convolutional neural network The module and the local enhancement feature of each point are extracted to obtain the target feature of each point.
- the target feature of each point extracted based on this method contains important geometric information corresponding to each point, which makes the extracted target feature of each point more accurate and effective, and then predicts according to the target feature of each point. category, the resulting predictions are very accurate.
- FIG. 3 is a schematic flowchart of a method for processing a three-dimensional point cloud provided by another embodiment of the present application, which mainly involves a possible implementation of extracting local enhancement features of each point based on the geometric attention fusion module. Way.
- the method includes:
- the K-nearest neighbor query algorithm For each point in the point cloud data, use the K-nearest neighbor query algorithm to obtain the point's neighbors in Euclidean space, and determine the eigenvalue map structure based on the point's neighbors in the Euclidean space;
- the eigenvalue matrix is obtained by decomposing the three-dimensional structure tensor; the neighbor points of the point in the eigenvalue space are determined based on the eigenvalue matrix.
- a tuple of eigenvalues is calculated based on the original coordinates of each point, expressed as and as the input feature for that point
- the local feature corresponding to the point can be obtained by fusing the adjacent points of the point in the Euclidean space and the adjacent points of the point in the eigenvalue space by the following formula (1), the formula (1) is as follows:
- S203 Aggregate local features corresponding to the points to obtain local enhanced features corresponding to the points.
- the local features corresponding to each point calculated in S202 are aggregated, and the aggregation is That is, the local enhancement feature corresponding to the point is obtained.
- local features corresponding to the point may be aggregated based on an attention pooling manner to obtain local enhanced features corresponding to the point.
- the local feature corresponding to the point can be aggregated by the following formula (2) to obtain the local enhancement feature corresponding to the point.
- the formula (2) is as follows:
- FIG. 4 is a schematic diagram of a geometric attention fusion module provided by this application.
- the geometric attention fusion module can also be called a geometry-based attention fusion module.
- the point coordinates, point features and feature roots of each point are input, and the feature roots are performed based on the feature roots.
- Space K nearest neighbors based on the point coordinates and point features of each point, perform the Euclidean space K nearest neighbors, and obtain the point's nearest neighbors in the Euclidean space and the point's nearest neighbors in the eigenvalue space.
- the neighbors of the point in the Euclidean space and the neighbors of the point in the eigenvalue space are fused to obtain the local features corresponding to the point.
- the local feature corresponding to the point is processed by the dot product and summation of the multi-layer perceptron, and the local enhancement feature corresponding to the point is obtained. That is, in the geometry-based attention fusion module, the inputs are point-wise coordinates, point features, and feature roots. In this module, we aggregate features in eigenvalue space and Euclidean space, and then use attention pooling to generate output features for each point.
- FIG. 5 is a schematic flowchart of a method for processing a three-dimensional point cloud provided by another embodiment of the present application, which mainly involves a possibility of extracting the target feature of each point based on the local enhancement feature of each point. implementation.
- the method includes:
- S301 Perform a local difference on each point based on the local enhancement feature of each point to obtain a local difference corresponding to each point.
- S302 Determine the indistinguishable point among the multiple points according to the local difference corresponding to each point.
- the acquired points include indistinguishable points, which are points in which the prediction category is difficult to determine among the points, that is, points in the indistinguishable area in the indistinguishable area diagram of the complex point cloud scene shown in FIG. 1 . Namely, complex boundary regions, obfuscated interior regions, and points in isolated small regions.
- the focusing module also referred to as an indistinguishable area focusing (IAF) module, can adaptively select indistinguishable points and enhance the features of each point.
- IAF indistinguishable area focusing
- the IAF module is a new indistinguishable region model based on hierarchical semantic features, which can adaptively select indistinguishable points.
- the IAF model first obtains fine-grained features and high-level semantic features of indistinguishable points, and then enhances the features through non-local operations between the indistinguishable points and the corresponding overall point set.
- Local difference refers to the difference between each point and its neighbors.
- the local difference reflects the difference of each point to a certain extent, which depends on the low-level geometric features, latent space and high-level semantic features. So we use local difference as the criterion for mining indistinguishable points. For each point p i , we get the K-nearest neighbors in Euclidean space, then we get the following local differences for each point in each layer, then we accumulate these local differences together, we adjust the accumulated results according to the local differences Sort in descending order, and then select a portion of points with large local differences as indistinguishable points. Corresponding to the points in the three regions we mentioned earlier, these indistinguishable points change dynamically as the network is updated iteratively.
- the indistinguishable points are distributed in regions where the original attributes (coordinates and colors) change rapidly. As the training process progresses, the indistinguishable points are located in the indistinguishable regions mentioned in the introduction.
- this application utilizes a non-local mechanism through the following equations The features of all points are updated, thereby implicitly enhancing the features of indistinguishable points. In addition to this, we also compute the predicted output of the current layer.
- the local difference is performed on each point based on the local enhancement feature of each point, and the local difference corresponding to each point is obtained, which can be realized by the following formula (3), and the formula (3) is as follows:
- FIG. 6 is a schematic flowchart of a method for processing a three-dimensional point cloud provided by yet another embodiment of the present application, which mainly involves using a multilayer perceptron to extract a target feature corresponding to each indistinguishable point.
- the method includes:
- S401 Acquire a predicted label corresponding to each indistinguishable point, and acquire an intermediate feature corresponding to each indistinguishable point.
- the predicted label corresponding to each indistinguishable point can be obtained by the following formula (5).
- the following formula (6) can be used to aggregate the predicted labels and intermediate features corresponding to the indistinguishable point to obtain an aggregation result corresponding to the indistinguishable point.
- this paper uses a non-local mechanism to update the features of all points through the following formula (7), thereby implicitly enhancing the features of indistinguishable points.
- FIG. 7 is a schematic diagram of a focusing module provided by the present application.
- the focusing module may also be referred to as an indivisible area focusing processing module. Upsampling, multi-layer perceptron learning and other processing are performed on the input features, the corresponding features of the coding layer and the predicted values of the previous layer, and finally the indistinguishable points and the target features corresponding to the indistinguishable points are extracted, and the current layer is also calculated. prediction output.
- IAF-Net Indistinguishable Area Focusing Network
- the present application also provides a new evaluation criterion for indivisible regions. Based on the preset measurement method, it is possible to evaluate whether the prediction category corresponding to each indistinguishable point is accurate; when it is detected that the number of indistinguishable points whose predicted category is accurate does not meet the preset threshold, continue to train the convolutional neural network.
- the K nearest neighbors in the Euclidean space are predicted as ⁇ Z i ,1 ⁇ j ⁇ K ⁇ , and then we count the points m i that satisfy the condition Z i ⁇ Z i,gt , then put Use 0, ⁇ 1 , ⁇ 2 , 1 to divide into three parts S1, S2, S3, and finally use As a new evaluation criterion, it corresponds to the segmentation performance on three inseparable regions.
- FIG. 8 is an evaluation process of the new evaluation criteria for indivisible regions provided by this application.
- determining the prediction category corresponding to each point based on the target feature corresponding to each point includes: determining each indistinguishable point based on the target feature corresponding to each indistinguishable point. The predicted probability value corresponding to each category. Based on the predicted probability values corresponding to each category, the predicted category corresponding to each indistinguishable point is determined.
- FIG. 9 is a semantic segmentation network for large-scale complex scene point clouds provided by the present application.
- the semantic segmentation network includes a feature extraction unit and a segmentation unit.
- the network takes N points as input, and uses the geometric attention module and the indivisible region focusing processing module mentioned in the first aspect and the second aspect to extract the features of the point cloud.
- the network connects each layer and computes the predicted class probability for each point in the point cloud.
- the predicted probability value corresponding to each category corresponding to each indistinguishable point is determined.
- the predicted category corresponding to each indistinguishable point is determined. For example, if an indistinguishable point corresponds to the category of table with a predicted probability value of 0.6, and corresponds to the category of book with a predicted probability value of 0.9, then the predicted category corresponding to the indistinguishable point is book. This is only an exemplary description, and the comparison is not limited.
- indistinguishable points include points located on complex boundaries, points with similar local textures but different categories, and points in isolated small hard regions, which greatly affect the performance of 3D semantic segmentation.
- IAF-Net Indistinguishable Region Focusing Network
- IPBM indistinguishable point-based metric
- the embodiment of the present application directly processes the point cloud data by setting the point cloud convolutional neural network for sharing local geometric information, and does not need to convert the point cloud data into other complex data formats, which is beneficial to reduce memory occupation and computer resource consumption, and can Extract rich feature data faster. And the method of geometric change attention is more conducive to exploring the overall structural geometric features of point cloud edge contours, thereby improving the accuracy of classification and segmentation tasks.
- a method for processing a three-dimensional point cloud provided by the present application further includes: obtaining a training set and A test set, the training set includes sample point cloud data of multiple sample points, and the test set includes the sample features and sample categories corresponding to each sample point; the initial convolutional neural network is trained through the training set, and the training set is obtained.
- Convolutional neural network verify the convolutional neural network in training based on the sample set; when the verification result does not meet the preset conditions, adjust the network parameters of the convolutional neural network in the training, and continue based on the training set Train the convolutional neural network in the training; when the verification result meets the preset conditions, stop training the convolutional neural network in the training, and use the trained convolutional neural network as the trained convolutional neural network .
- the sample point cloud data of multiple sample points can be collected by the device, or it can be collected by other devices and transmitted to the device.
- the points in the point cloud data can be rotated and/or the points in the point source data can be rotated.
- the point coordinates are perturbed within a predetermined range around the point to enhance the point cloud data; and/or randomly delete points in the point cloud data.
- a random probability is randomly generated according to a preset maximum random probability, and points in the point cloud data are deleted according to the generated random probability.
- the convolutional neural network When the parameters of the convolutional neural network are input, it can be further carried out: manually classify and filter the collected 3D point cloud data according to the category, and complete the preliminary data preparation work.
- the convolution kernels of the trained convolutional neural network can be obtained using the first part of the point cloud data in the classified category to obtain the trained convolutional neural network; the second part of the point cloud data in the classified category can be used as validation data to evaluate The convolutional neural network. For example, according to the data sorting process, 90% of the data of each category of the 3D point cloud is selected as the training data for network training, and the remaining 10% of the data is reserved as the experimental verification data for later evaluation of the model recognition accuracy and generalization ability. Evaluate.
- FIG. 10 is an adaptive change process of indistinguishable points in the training process provided by the present application.
- the indistinguishable points are distributed in regions where the original attributes (coordinates and colors) change rapidly.
- the indistinguishable points are located in the indistinguishable regions mentioned in the introduction.
- the present application may further process the features of the point cloud data after extraction: after performing several deconvolution modules on the geometric feature information, the maximum K pooling operation may be used to extract the geometric features of the point cloud. , for subsequent classification, segmentation or registration.
- the feature obtained by the multi-layer convolution module is an NxM dimension vector
- N is the number of points
- M is the dimension of each point feature
- the maximum k pooling operation refers to taking the largest K value among the i-th dimension features of N points.
- the global feature vector of the KxM dimensional point cloud is obtained.
- the output features of the convolutional modules at each layer can be combined for a max pooling operation and finally passed through a fully connected layer.
- the cross-entropy function can be used as the loss function
- the back-propagation algorithm can be used to train and optimize the model.
- the global features and object category information of the point cloud are used as the local features of the point cloud, and a higher-dimensional local cloud feature is formed after the point cloud, and the previously extracted point cloud features are formed.
- segmentation prediction is performed by the predicted probabilities of the object segmentation parts obtained by the multilayer perceptron and the normalized exponential function.
- This application designs a convolutional neural network structure for 3D point cloud classification and segmentation, adjusts the network parameters of the neural network, including but not limited to (learning rate, batch size), and adopts different learning strategies to promote convolution
- the neural network converges to the best optimization direction of the network model; finally, the trained network model is used to test the verification data to realize the classification and segmentation of the point cloud.
- the geometric information disentanglement convolution designed by the present invention is a module in the neural network, which can directly extract features with large and small geometric changes from the signals distributed on the point cloud, so it can be used in combination with other modules in the neural network. . network.
- the number of input and output channels and the combination of output channels can be changed to achieve the best results in different tasks.
- Different neural network structures can be designed by using the geometric feature information sharing module.
- the present application can be applied to scene segmentation tasks and three-dimensional scene reconstruction tasks in the field of unmanned driving and robot vision.
- FIG. 11 provides an application scenario diagram for this application.
- FIG. 11 mainly shows the application of the present invention to the scene segmentation task of unmanned vehicles and robot vision.
- the class and location of objects can be obtained, which is the basis for other tasks in this field.
- a method for processing a three-dimensional point cloud provided in this application can be used for scene segmentation tasks of an unmanned intelligent robot.
- the depth camera is used to collect point cloud data of the actual scene, and then use the trained neural network to extract the local features of the point cloud, and then segment the scene.
- the segmentation results i.e. the different objects in the scene
- the characteristics of the input can be changed according to different tasks, for example, the distance between the point and its neighbors, the color information of the point, the combination of feature vectors, and the local shape context information of the point.
- Features are substituted or combined.
- the inseparable area focusing module in the network is a portable point cloud feature learning module, which can be used as a feature extractor for other point cloud-related tasks, such as 3D point cloud completion, 3D point cloud detection. and other tasks.
- FIG. 12 is a schematic diagram of an apparatus for processing a three-dimensional point cloud provided by an embodiment of the present application.
- Each unit included in the apparatus is used to execute each step in the embodiment corresponding to FIG. 2 , FIG. 3 , FIG. 5 , and FIG. 6 .
- FIG. 12 please refer to the relevant descriptions in the corresponding embodiments of FIG. 2 , FIG. 3 , FIG. 5 , and FIG. 6 .
- Figure 11 including:
- an acquiring unit 510 configured to acquire point cloud data including a plurality of points
- the processing unit 520 is configured to input the point cloud data into the trained convolutional neural network for processing, and obtain the target feature corresponding to each point.
- the convolutional neural network includes a geometric attention fusion module and a focusing module, so The geometric attention fusion module is used for extracting the local enhancement feature of each described point, and the focusing module is used for extracting the target feature of each described point based on the local enhancement feature of each described point;
- the determining unit 530 is configured to determine the prediction category corresponding to each point based on the target feature corresponding to each point.
- processing unit 520 is specifically configured to:
- the neighbors of the point in Euclidean space are obtained based on the geometric attention fusion module, and the neighbors of the point in the eigenvalue space are determined based on the neighbors of the point in the Euclidean space point;
- the local features corresponding to the points are aggregated to obtain local enhanced features corresponding to the points.
- processing unit 520 is further configured to:
- the local features corresponding to the points are aggregated based on the attention pooling method to obtain the local enhanced features corresponding to the points.
- the plurality of points include indistinguishable points, and the indistinguishable points are points in which the prediction category is not easily determined among the plurality of points, and the processing unit 520 is further configured to:
- a multi-layer perceptron is used to extract the target feature corresponding to each indistinguishable point.
- processing unit 520 is further configured to:
- a multi-layer perceptron is used to extract the target feature corresponding to each indistinguishable point.
- the determining unit 530 is specifically configured to:
- the predicted category corresponding to each indistinguishable point is determined.
- the device further includes:
- a sample acquisition unit configured to acquire a training set and a test set
- the training set includes sample point cloud data of a plurality of sample points
- the test set includes sample features and sample categories corresponding to each sample point
- the first training unit is used to train the initial convolutional neural network through the training set to obtain the convolutional neural network in training;
- a verification unit for verifying the convolutional neural network in the training based on the sample set
- an adjustment unit configured to adjust the network parameters of the convolutional neural network in the training when the verification result does not meet the preset conditions, and continue to train the convolutional neural network in the training based on the training set;
- the second training unit is configured to stop training the convolutional neural network in training when the verification result meets the preset condition, and use the trained convolutional neural network as the trained convolutional neural network.
- the device further includes:
- the evaluation unit is used to evaluate whether the prediction category corresponding to each indistinguishable point is accurate based on the preset measurement method
- the third training unit is configured to continue training the convolutional neural network when it is detected that the number of indistinguishable points with accurate predicted categories does not meet the preset threshold.
- FIG. 13 is a schematic diagram of a device for processing a three-dimensional point cloud provided by another embodiment of the present application.
- the apparatus 6 for processing a three-dimensional point cloud in this embodiment includes: a processor 60 , a memory 61 , and computer instructions 62 stored in the memory 61 and executable on the processor 60 .
- the processor 60 executes the computer instructions 62
- the steps in each of the foregoing method embodiments for processing a three-dimensional point cloud are implemented, for example, S101 to S103 shown in FIG. 2 .
- the processor 60 executes the computer instructions 62
- the functions of the units in the above embodiments for example, the functions of the units 510 to 530 shown in FIG. 12 are implemented.
- the computer instructions 62 may be divided into one or more units, and the one or more units are stored in the memory 61 and executed by the processor 60 to complete the present application.
- the one or more units may be a series of computer instruction segments capable of performing specific functions, and the instruction segments are used to describe the execution process of the computer instructions 62 in the apparatus 6 for processing three-dimensional point clouds.
- the computer instructions 62 can be divided into an acquisition unit, a processing unit, and a determination unit, and the specific functions of each unit are as described above.
- the device for processing a three-dimensional point cloud may include, but is not limited to, a processor 60 and a memory 61 .
- FIG. 6 is only an example of the device 6 for processing three-dimensional point clouds, and does not constitute a limitation on the device for processing three-dimensional point clouds, and may include more or less components than those shown, or combine certain Some components, or different components, for example, the device for processing a 3D point cloud may also include an input and output terminal, a network access terminal, a bus, and the like.
- the so-called processor 60 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuit (Application Specific Integrated Circuit, ASIC), Off-the-shelf programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
- a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
- the memory 61 may be an internal storage unit of the device for processing three-dimensional point clouds, such as a hard disk or memory of the device for processing three-dimensional point clouds.
- the memory 61 can also be an external storage terminal of the device for processing three-dimensional point clouds, such as a plug-in hard disk, a smart memory card (Smart Media Card, SMC), a secure digital (Secure Digital, SD) card, flash memory card (Flash Card), etc.
- the memory 61 may also include both an internal storage unit of the device for processing three-dimensional point clouds and an external storage terminal.
- the memory 61 is used to store the computer instructions and other programs and data required by the terminal.
- the memory 61 can also be used to temporarily store data that has been output or will be output.
- Embodiments of the present application also provide a computer storage medium, which may be non-volatile or volatile, and stores a computer program that implements the above products when the computer program is executed by a processor. Steps in the embodiment of the method for constructing the knowledge graph.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Computational Linguistics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Evolutionary Biology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Probability & Statistics with Applications (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (11)
- 一种处理三维点云的方法,其特征在于,包括:获取包括多个点的点云数据;将所述点云数据输入到已训练的卷积神经网络中处理,得到每个点对应的目标特征,所述卷积神经网络包括几何注意力融合模块和聚焦模块,所述几何注意力融合模块用于提取每个所述点的局部增强特征,所述聚焦模块用于基于每个所述点的局部增强特征,提取每个所述点的目标特征;基于每个点对应的目标特征,确定每个点对应的预测类别。
- 如权利要求1所述的方法,其特征在于,所述提取每个所述点的局部增强特征,包括:针对点云数据中的每个点,基于所述几何注意力融合模块获取所述点在欧式空间的近邻点,且基于所述点在欧式空间的近邻点确定所述点在特征值空间的近邻点;融合所述点在欧式空间的近邻点以及所述点在特征值空间的近邻点,得到所述点对应的局部特征;聚合所述点对应的局部特征,得到所述点对应的局部增强特征。
- 如权利要求2所述的方法,其特征在于,所述聚合所述点对应的局部特征,得到所述点对应的局部增强特征,包括:基于注意力池化方式聚合所述点对应的局部特征,得到所述点对应的局部增强特征。
- 如权利要求1所述的方法,其特征在于,所述多个点包括不可分辨点,所述不可分辨点为所述多个点中不易确定预测类别的点,所述基于每个所述点的局部增强特征,提取每个所述点的目标特征,包括:基于每个所述点的局部增强特征对每个点进行局部差分,得到每个点对应的局部差异;根据每个点对应的局部差异,在所述多个点中确定所述不可分辨点;采用多层感知器提取每个不可分辨点对应的目标特征。
- 如权利要求4所述的方法,其特征在于,所述采用多层感知器提取每个不可分辨点对应的目标特征,包括:获取每个不可分辨点对应的预测标签,以及获取每个不可分辨点对应的中间特征;针对每个不可分辨点,聚集所述不可分辨点对应的预测标签和中间特征,得到所述不可分辨点对应的聚集结果;基于每个不可分辨点对应的聚集结果,采用多层感知器提取每个不可分辨点对应的目标特征。
- 如权利要求4所述的方法,其特征在于,所述基于每个点对应的目标特 征,确定每个点对应的预测类别,包括:基于每个不可分辨点对应的目标特征,确定每个不可分辨点对应的各个类别所对应的预测概率值;基于各个类别所对应的预测概率值,确定每个不可分辨点对应的预测类别。
- 如权利要求1至6任一项所述的方法,其特征在于,所述将所述点云数据输入到已训练的卷积神经网络中处理,得到每个点对应的目标特征之前,所述方法还包括:获取训练集和测试集,所述训练集包括多个样本点的样本点云数据,所述测试集包括每个样本点对应的样本特征以及样本类别;通过所述训练集对初始卷积神经网络进行训练,得到训练中的卷积神经网络;基于所述样本集对所述训练中的卷积神经网络进行验证;当验证结果不满足预设条件时,调整所述训练中的卷积神经网络的网络参数,并继续基于所述训练集对所述训练中的卷积神经网络进行训练;当验证结果满足预设条件时,停止训练所述训练中的卷积神经网络,并将训练后的卷积神经网络作为所述已训练的卷积神经网络。
- 如权利要求4所述的方法,其特征在于,所述方法还包括:基于预设度量方法,评价每个不可分辨点对应的预测类别是否准确;当检测到预测类别准确的不可分辨点的数量不满足预设阈值时,继续训练所述卷积神经网络。
- 一种处理三维点云的装置,其特征在于,包括:获取单元,用于获取包括多个点的点云数据;处理单元,用于将所述点云数据输入到已训练的卷积神经网络中处理,得到每个点对应的目标特征,所述卷积神经网络包括几何注意力融合模块和聚焦模块,所述几何注意力融合模块用于提取每个所述点的局部增强特征,所述聚焦模块用于基于每个所述点的局部增强特征,提取每个所述点的目标特征;确定单元,用于基于每个点对应的目标特征,确定每个点对应的预测类别。
- 一种处理三维点云的设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至8任一项所述的方法。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至8任一项所述的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110163660.1 | 2021-02-05 | ||
CN202110163660.1A CN112966696B (zh) | 2021-02-05 | 2021-02-05 | 一种处理三维点云的方法、装置、设备以及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022166400A1 true WO2022166400A1 (zh) | 2022-08-11 |
Family
ID=76274715
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/137305 WO2022166400A1 (zh) | 2021-02-05 | 2021-12-12 | 一种处理三维点云的方法、装置、设备以及存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN112966696B (zh) |
WO (1) | WO2022166400A1 (zh) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115312119A (zh) * | 2022-10-09 | 2022-11-08 | 之江实验室 | 基于蛋白质三维结构图像鉴定蛋白质结构域的方法及系统 |
CN115311534A (zh) * | 2022-08-26 | 2022-11-08 | 中国铁道科学研究院集团有限公司 | 基于激光雷达的铁路周界入侵辨识方法、装置及存储介质 |
CN115457496A (zh) * | 2022-09-09 | 2022-12-09 | 北京百度网讯科技有限公司 | 自动驾驶的挡墙检测方法、装置及车辆 |
CN115546302A (zh) * | 2022-10-20 | 2022-12-30 | 上海人工智能创新中心 | 一种局部几何建模的点云数据解算方法 |
CN115661812A (zh) * | 2022-11-14 | 2023-01-31 | 苏州挚途科技有限公司 | 目标检测方法、目标检测装置与电子设备 |
CN115984331A (zh) * | 2023-02-13 | 2023-04-18 | 中国人民解放军国防科技大学 | 基于多尺度点云特征融合的运动小目标事件流分割方法 |
CN116137059A (zh) * | 2023-04-17 | 2023-05-19 | 宁波大学科学技术学院 | 基于多层次特征提取网络模型的三维点云质量评价方法 |
CN116413740A (zh) * | 2023-06-09 | 2023-07-11 | 广汽埃安新能源汽车股份有限公司 | 一种激光雷达点云地面检测方法及装置 |
CN116524197A (zh) * | 2023-06-30 | 2023-08-01 | 厦门微亚智能科技有限公司 | 一种结合边缘点和深度网络的点云分割方法、装置及设备 |
CN116520289A (zh) * | 2023-07-04 | 2023-08-01 | 东莞市新通电子设备有限公司 | 五金件加工过程智能控制方法及相关装置 |
CN117152363A (zh) * | 2023-10-30 | 2023-12-01 | 浪潮电子信息产业股份有限公司 | 基于预训练语言模型的三维内容生成方法、装置及设备 |
WO2024060395A1 (zh) * | 2022-09-19 | 2024-03-28 | 南京邮电大学 | 一种基于深度学习的高精度点云补全方法及装置 |
CN117994778A (zh) * | 2024-04-07 | 2024-05-07 | 吉林大学 | 一种基于点云数据的冲压件自动计数和分类方法 |
CN118037703A (zh) * | 2024-03-16 | 2024-05-14 | 中新聚能建设工程有限公司 | 一种基于多源信息的隧道掌子面分区方法及相关设备 |
CN118172398A (zh) * | 2024-05-14 | 2024-06-11 | 山东省计算中心(国家超级计算济南中心) | 基于双层聚焦-注意力特征交互的点云配准方法及系统 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112966696B (zh) * | 2021-02-05 | 2023-10-27 | 中国科学院深圳先进技术研究院 | 一种处理三维点云的方法、装置、设备以及存储介质 |
CN113327279B (zh) * | 2021-08-04 | 2021-09-28 | 腾讯科技(深圳)有限公司 | 一种点云数据处理方法、装置、计算机设备及存储介质 |
CN113486988B (zh) * | 2021-08-04 | 2022-02-15 | 广东工业大学 | 一种基于适应性自注意力变换网络的点云补全装置和方法 |
CN114998414B (zh) * | 2022-06-02 | 2024-08-09 | 华侨大学 | 一种基于点云数据的零件三维尺寸测量方法、装置和介质 |
CN117368876B (zh) * | 2023-10-18 | 2024-03-29 | 广州易而达科技股份有限公司 | 一种人体检测方法、装置、设备及存储介质 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111028327A (zh) * | 2019-12-10 | 2020-04-17 | 深圳先进技术研究院 | 一种三维点云的处理方法、装置及设备 |
EP3767521A1 (en) * | 2019-07-15 | 2021-01-20 | Promaton Holding B.V. | Object detection and instance segmentation of 3d point clouds based on deep learning |
CN112966696A (zh) * | 2021-02-05 | 2021-06-15 | 中国科学院深圳先进技术研究院 | 一种处理三维点云的方法、装置、设备以及存储介质 |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10650278B1 (en) * | 2017-07-21 | 2020-05-12 | Apple Inc. | Semantic labeling of point clouds using images |
CN109655019B (zh) * | 2018-10-29 | 2021-02-02 | 北方工业大学 | 一种基于深度学习和三维重建的货物体积测量方法 |
CN111414953B (zh) * | 2020-03-17 | 2023-04-18 | 集美大学 | 点云分类方法和装置 |
CN112287939B (zh) * | 2020-10-29 | 2024-05-31 | 平安科技(深圳)有限公司 | 三维点云语义分割方法、装置、设备及介质 |
-
2021
- 2021-02-05 CN CN202110163660.1A patent/CN112966696B/zh active Active
- 2021-12-12 WO PCT/CN2021/137305 patent/WO2022166400A1/zh active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3767521A1 (en) * | 2019-07-15 | 2021-01-20 | Promaton Holding B.V. | Object detection and instance segmentation of 3d point clouds based on deep learning |
CN111028327A (zh) * | 2019-12-10 | 2020-04-17 | 深圳先进技术研究院 | 一种三维点云的处理方法、装置及设备 |
CN112966696A (zh) * | 2021-02-05 | 2021-06-15 | 中国科学院深圳先进技术研究院 | 一种处理三维点云的方法、装置、设备以及存储介质 |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115311534A (zh) * | 2022-08-26 | 2022-11-08 | 中国铁道科学研究院集团有限公司 | 基于激光雷达的铁路周界入侵辨识方法、装置及存储介质 |
CN115311534B (zh) * | 2022-08-26 | 2023-07-18 | 中国铁道科学研究院集团有限公司 | 基于激光雷达的铁路周界入侵辨识方法、装置及存储介质 |
CN115457496B (zh) * | 2022-09-09 | 2023-12-08 | 北京百度网讯科技有限公司 | 自动驾驶的挡墙检测方法、装置及车辆 |
CN115457496A (zh) * | 2022-09-09 | 2022-12-09 | 北京百度网讯科技有限公司 | 自动驾驶的挡墙检测方法、装置及车辆 |
WO2024060395A1 (zh) * | 2022-09-19 | 2024-03-28 | 南京邮电大学 | 一种基于深度学习的高精度点云补全方法及装置 |
CN115312119A (zh) * | 2022-10-09 | 2022-11-08 | 之江实验室 | 基于蛋白质三维结构图像鉴定蛋白质结构域的方法及系统 |
US11908140B1 (en) | 2022-10-09 | 2024-02-20 | Zhejiang Lab | Method and system for identifying protein domain based on protein three-dimensional structure image |
CN115546302A (zh) * | 2022-10-20 | 2022-12-30 | 上海人工智能创新中心 | 一种局部几何建模的点云数据解算方法 |
CN115546302B (zh) * | 2022-10-20 | 2024-06-14 | 上海人工智能创新中心 | 一种局部几何建模的点云数据解算方法 |
CN115661812A (zh) * | 2022-11-14 | 2023-01-31 | 苏州挚途科技有限公司 | 目标检测方法、目标检测装置与电子设备 |
CN115984331A (zh) * | 2023-02-13 | 2023-04-18 | 中国人民解放军国防科技大学 | 基于多尺度点云特征融合的运动小目标事件流分割方法 |
CN116137059A (zh) * | 2023-04-17 | 2023-05-19 | 宁波大学科学技术学院 | 基于多层次特征提取网络模型的三维点云质量评价方法 |
CN116137059B (zh) * | 2023-04-17 | 2024-04-26 | 宁波大学科学技术学院 | 基于多层次特征提取网络模型的三维点云质量评价方法 |
CN116413740B (zh) * | 2023-06-09 | 2023-09-05 | 广汽埃安新能源汽车股份有限公司 | 一种激光雷达点云地面检测方法及装置 |
CN116413740A (zh) * | 2023-06-09 | 2023-07-11 | 广汽埃安新能源汽车股份有限公司 | 一种激光雷达点云地面检测方法及装置 |
CN116524197B (zh) * | 2023-06-30 | 2023-09-29 | 厦门微亚智能科技股份有限公司 | 一种结合边缘点和深度网络的点云分割方法、装置及设备 |
CN116524197A (zh) * | 2023-06-30 | 2023-08-01 | 厦门微亚智能科技有限公司 | 一种结合边缘点和深度网络的点云分割方法、装置及设备 |
CN116520289B (zh) * | 2023-07-04 | 2023-09-01 | 东莞市新通电子设备有限公司 | 五金件加工过程智能控制方法及相关装置 |
CN116520289A (zh) * | 2023-07-04 | 2023-08-01 | 东莞市新通电子设备有限公司 | 五金件加工过程智能控制方法及相关装置 |
CN117152363A (zh) * | 2023-10-30 | 2023-12-01 | 浪潮电子信息产业股份有限公司 | 基于预训练语言模型的三维内容生成方法、装置及设备 |
CN117152363B (zh) * | 2023-10-30 | 2024-02-13 | 浪潮电子信息产业股份有限公司 | 基于预训练语言模型的三维内容生成方法、装置及设备 |
CN118037703A (zh) * | 2024-03-16 | 2024-05-14 | 中新聚能建设工程有限公司 | 一种基于多源信息的隧道掌子面分区方法及相关设备 |
CN117994778A (zh) * | 2024-04-07 | 2024-05-07 | 吉林大学 | 一种基于点云数据的冲压件自动计数和分类方法 |
CN118172398A (zh) * | 2024-05-14 | 2024-06-11 | 山东省计算中心(国家超级计算济南中心) | 基于双层聚焦-注意力特征交互的点云配准方法及系统 |
Also Published As
Publication number | Publication date |
---|---|
CN112966696B (zh) | 2023-10-27 |
CN112966696A (zh) | 2021-06-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022166400A1 (zh) | 一种处理三维点云的方法、装置、设备以及存储介质 | |
US10885352B2 (en) | Method, apparatus, and device for determining lane line on road | |
CN108898086B (zh) | 视频图像处理方法及装置、计算机可读介质和电子设备 | |
CN106204522B (zh) | 对单个图像的联合深度估计和语义标注 | |
CN108229419B (zh) | 用于聚类图像的方法和装置 | |
CN111028327B (zh) | 一种三维点云的处理方法、装置及设备 | |
EP3844669A1 (en) | Method and system for facilitating recognition of vehicle parts based on a neural network | |
CN110084299B (zh) | 基于多头融合注意力的目标检测方法和装置 | |
WO2022193335A1 (zh) | 点云数据处理方法、装置、计算机设备和存储介质 | |
US20130346346A1 (en) | Semi-supervised random decision forests for machine learning | |
CN110390706B (zh) | 一种物体检测的方法和装置 | |
CN113807399A (zh) | 一种神经网络训练方法、检测方法以及装置 | |
CN112336342A (zh) | 手部关键点检测方法、装置及终端设备 | |
CN115294332B (zh) | 一种图像处理方法、装置、设备和存储介质 | |
Chen et al. | Lrgnet: Learnable region growing for class-agnostic point cloud segmentation | |
Ward et al. | RGB-D image-based object detection: from traditional methods to deep learning techniques | |
Chew et al. | Large-scale 3D point-cloud semantic segmentation of urban and rural scenes using data volume decomposition coupled with pipeline parallelism | |
Panda et al. | Kernel density estimation and correntropy based background modeling and camera model parameter estimation for underwater video object detection | |
Liu et al. | Fast and robust ellipse detector based on edge following method | |
CN116432060A (zh) | 基于雷达的目标自适应聚类方法、装置、设备及存储介质 | |
CN113989574B (zh) | 图像解释方法、图像解释装置、电子设备和存储介质 | |
CN114283343A (zh) | 基于遥感卫星图像的地图更新方法、训练方法和设备 | |
CN111340145B (zh) | 点云数据分类方法、装置、分类设备 | |
CN111813882B (zh) | 一种机器人地图构建方法、设备及存储介质 | |
Zhang et al. | A small target detection algorithm based on improved YOLOv5 in aerial image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21924393 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21924393 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21924393 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19/01/2024) |