CN113298952B - Incomplete point cloud classification method based on data expansion and similarity measurement - Google Patents
Incomplete point cloud classification method based on data expansion and similarity measurement Download PDFInfo
- Publication number
- CN113298952B CN113298952B CN202110652559.2A CN202110652559A CN113298952B CN 113298952 B CN113298952 B CN 113298952B CN 202110652559 A CN202110652559 A CN 202110652559A CN 113298952 B CN113298952 B CN 113298952B
- Authority
- CN
- China
- Prior art keywords
- point cloud
- cloud
- point
- classification
- loss
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000005259 measurement Methods 0.000 title claims abstract description 23
- 238000000034 method Methods 0.000 title claims description 53
- 230000007547 defect Effects 0.000 claims abstract description 48
- 230000007246 mechanism Effects 0.000 claims abstract description 10
- 230000008569 process Effects 0.000 claims description 30
- 238000005070 sampling Methods 0.000 claims description 27
- 238000011524 similarity measure Methods 0.000 claims description 12
- 238000012549 training Methods 0.000 claims description 12
- 238000011176 pooling Methods 0.000 claims description 10
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000013461 design Methods 0.000 claims description 6
- 238000011156 evaluation Methods 0.000 claims description 6
- 238000000605 extraction Methods 0.000 claims description 5
- 238000012360 testing method Methods 0.000 claims description 5
- 238000002474 experimental method Methods 0.000 claims description 3
- 230000000717 retained effect Effects 0.000 claims description 3
- 238000010276 construction Methods 0.000 claims description 2
- 230000006870 function Effects 0.000 description 22
- 238000012800 visualization Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000013434 data augmentation Methods 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 238000000844 transformation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/20—Finite element generation, e.g. wire-frame surface description, tesselation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Software Systems (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Computing Systems (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Image Analysis (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
Description
技术领域technical field
本发明属于三维计算机视觉领域,具体涉及到一种基于数据扩容与相似性度量的面向残缺点云分类方法。The invention belongs to the field of three-dimensional computer vision, and specifically relates to a residual defect-oriented cloud classification method based on data expansion and similarity measurement.
背景技术Background technique
点云不同于二维图像规则的网格结构,而是一组无序的3D点,所以点云分类具有一定挑战性。近年来出现了诸多应用于三维点云的深度学习方法。这些方法对点云进行精准的分类通常依赖于数据的丰富性和可用性,相较于二维图像领域中常用的大体量数据集MS COCO和ImageNet数据集,三维点云数据集的规模通常要小很多,且样本的丰富度不足。常用的点云分类数据集结构完整表义清晰,而实际应用中的点云数据来源主要有两种方式:第一是利用激光雷达通过测量传感器和被测物体之间的距离的方式,扫描获取离散点云数据;第二是利用RGB-D相机通过测量相机与物体之间的距离的方式获取点云数据。无论以哪种方式采集数据,由于场景中物体间遮挡、光线反射、传感器分辨率和视角的限制,会造成几何和语义信息的丢失,导致点云出现不同程度的残缺。在识别分类中,区域性的缺失会导致点云丢失部分特征,影响形状分类的精度,例如在VKITT数据集中,某些被拦腰截断的房屋容易被网络错误判定为卡车。Point clouds are different from the regular grid structure of 2D images, but a set of disordered 3D points, so point cloud classification is challenging. In recent years, there have been many deep learning methods applied to 3D point clouds. The accurate classification of point clouds by these methods usually depends on the richness and availability of data. Compared with MS COCO and ImageNet datasets, which are commonly used in the field of 2D images, the scale of 3D point cloud datasets is usually smaller. Many, and the richness of the sample is insufficient. The commonly used point cloud classification datasets have complete structures and clear definitions, and there are two main ways to source point cloud data in practical applications: the first is to use lidar to measure the distance between the sensor and the object to be measured. Discrete point cloud data; the second is to use an RGB-D camera to obtain point cloud data by measuring the distance between the camera and the object. No matter which method is used to collect data, due to the limitation of occlusion between objects in the scene, light reflection, sensor resolution and viewing angle, geometric and semantic information will be lost, resulting in different degrees of incomplete point cloud. In the identification and classification, the lack of regionality will cause the point cloud to lose some features and affect the accuracy of shape classification. For example, in the VKITT dataset, some houses that are truncated by the middle are easily misidentified as trucks by the network.
发明内容SUMMARY OF THE INVENTION
本发明的目的在于如何将将完整的点云分类网络扩展到残缺点云,从而为面向残缺点云的分类问题提供优秀的解决方案的一种基于数据扩容与相似性度量的面向残缺点云的分类方法。The purpose of the present invention is how to extend the complete point cloud classification network to the residual defect cloud, so as to provide an excellent solution for the classification problem oriented to the residual defect cloud. Classification.
本发明的目的是这样实现的:The object of the present invention is achieved in this way:
一种基于数据扩容与相似性度量的残缺点云分类方法,包括以下步骤:A cloud classification method based on data expansion and similarity measurement, comprising the following steps:
S1:对原始点云进行预处理;S1: Preprocess the original point cloud;
该步骤首先对未规范化的数据进行规范化,使得数据范围始终保持在[-1,1]中,以便于后续任务;This step first normalizes the unnormalized data, so that the data range is always kept in [-1, 1] for the convenience of subsequent tasks;
S2:提出了一种面向三维点云的数据扩容方式——随机区域擦除;S2: Propose a data expansion method for 3D point cloud - random area erasure;
该步骤能够兼顾学习完整点云及残缺点云的语义信息,针对完整点云随机去除部分区域的点云,剥离掉局部信息;这一过程模拟了实际采集点云过程中出现的遮挡、缺失等问题,且增加了数据的复杂性和丰富度,增强了网络的鲁棒性与泛化性能;其具体步骤如下所示:This step can take into account the semantic information of the complete point cloud and the residual defect cloud. For the complete point cloud, the point cloud in some areas is randomly removed, and the local information is stripped off; this process simulates the occlusion and lack of actual point cloud collection. It increases the complexity and richness of the data, and enhances the robustness and generalization performance of the network; the specific steps are as follows:
S21:在训练过程中,原始点云数据输入网络后复制本身获得与之相同的“复本”作为随机区域擦除模块的输入数据;S21: During the training process, after the original point cloud data is input into the network, the copy itself obtains the same "replica" as the input data of the random area erasing module;
S22:随机区域擦除的方法会在“复本”点集中随机选择一个点作为擦除的区域中心;S22: The method of random area erasing will randomly select a point in the "replica" point set as the erased area center;
S23:随机区域擦除模块设置了一个擦除的范围,并随机选定范围内的任意一大小作为擦除半径以形成一个可供擦除的球域;S23: The random area erasing module sets an erasing range, and randomly selects any size within the range as the erasing radius to form a sphere for erasing;
S24:将擦除球域内的点云数据清除,与此同时模块需要判断经过擦除的点云所保留的点数是否满足后续分类网络的要求下限,分类网络的输入点数为1024,故当擦除后剩余点数不满1024时,网络则放弃对该点云数据实行随机区域擦除;S24: Clear the point cloud data in the erasing sphere. At the same time, the module needs to determine whether the number of points retained by the erased point cloud meets the lower limit of the subsequent classification network. The input points of the classification network are 1024, so when erasing When the number of remaining points is less than 1024, the network will abandon the random area erasure of the point cloud data;
S3:设计一个点相似性度量模块,引入注意力机制,完成辅助损失函数的设计;S3: Design a point similarity measurement module, introduce an attention mechanism, and complete the design of the auxiliary loss function;
该步骤中的相似性度量模块以S2输出的完整点云和残缺点云共同作为输入,设定评价指标判断二者之间的相似性;对完整点云应用注意力机制获得不同点之间的权重值,该权重值作为参数参与构建相似性度量损失;相似性损失函数以辅助损失的形式引入总体损失函数,在网络模型的训练过程中引入并在网络的测试环节去除;其具体步骤如下:The similarity measurement module in this step takes the complete point cloud and the residual defect cloud output by S2 as the input, and sets the evaluation index to judge the similarity between the two; applies the attention mechanism to the complete point cloud to obtain the difference between the different points. The weight value, which is used as a parameter to participate in the construction of the similarity measurement loss; the similarity loss function is introduced into the overall loss function in the form of auxiliary loss, which is introduced in the training process of the network model and removed in the testing process of the network; the specific steps are as follows:
S31:将数据量不同的完整点云与残缺点云共同降采样,降采样的方法为最远点采样,S31: Downsampling the complete point cloud and the residual defect cloud with different amounts of data together, the downsampling method is the farthest point sampling,
经过最远点采样的完整点云与残缺点云各保留1024个采样点;The complete point cloud and the residual defect cloud sampled by the farthest point each retain 1024 sampling points;
S32:对下采样后的完整点云应用注意力机制,以此获得各个点之间的权重,获取注意力的公式为:S32: Apply the attention mechanism to the downsampled complete point cloud to obtain the weight between each point. The formula for obtaining the attention is:
A=g(Qi,W)A=g(Q i ,W)
其中Qi={qj|1≤j≤i}代表了完整点云,W是共享多层感知机的可学习权重;where Q i ={q j |1≤j≤i} represents the complete point cloud, and W is the learnable weight of the shared multilayer perceptron;
S33:利用点云补全中的评价指标评估残缺点云与完整点云的关系,该部分主要包括如下步骤:S33: Use the evaluation indicators in point cloud completion to evaluate the relationship between the residual defect cloud and the complete point cloud. This part mainly includes the following steps:
S331:采用点云补全中的倒角距离(CD)来评估完整点云Q与残缺点云P之间的关系,CD表示一个点云中每一个点到另一个点云中与之最近的邻居点之间的平均距离,由此构成CD损失,其具体公式为:S331: Use the chamfering distance (CD) in point cloud completion to evaluate the relationship between the complete point cloud Q and the residual defect cloud P, CD represents each point in one point cloud to the nearest point in the other point cloud. The average distance between neighbor points, which constitutes the CD loss, and its specific formula is:
S332:采用点云补全中的地球移动距离(EMD)来评估完整点云Q与残缺点云P之间的关系,EMD距离表示的是一个点云“移动”成另一个点云的最优规划路径下的最小消耗,由此构成EMD损失,其具体公式为:S332: Use the Earth Movement Distance (EMD) in point cloud completion to evaluate the relationship between the complete point cloud Q and the residual defect cloud P. The EMD distance represents the optimal point cloud "moving" into another point cloud. The minimum consumption under the planned path constitutes the EMD loss, and its specific formula is:
其中φ是一个可以使对应点之间的平均距离最小的双射,且和的大小必须相同;where φ is a bijection that minimizes the average distance between corresponding points, and must have the same magnitude;
S333:将CD损失与EMD损失相结合,CD计算效率高时间复杂度低,EMD对单个点云的个性特征及局部细节有更好的区分度,将二者结合起来共同构建相似性度量损失,该损失函数公式为:S333: Combining CD loss and EMD loss, CD calculation efficiency is high and time complexity is low, EMD has a better discrimination degree for the individual characteristics and local details of a single point cloud, and the two are combined to jointly construct the similarity measure loss, The loss function formula is:
S34:将完整点云经过注意力模块获得的权重作为参数,引入完整点云与残缺点云之间距离dij的计算中,dij是相似性度量损失的核心部分,其公式如下:S34: Take the weight obtained by the complete point cloud through the attention module as a parameter, and introduce it into the calculation of the distance d ij between the complete point cloud and the residual defect cloud, d ij is the core part of the similarity measurement loss, and its formula is as follows:
dij=A·||Q-P||d ij =A·||QP||
S4:对完整点云和残缺点云进行分类;S4: Classify complete point cloud and residual defect cloud;
该步骤利用S2得到的扩容后的数据作为分类网络的输入,分类网络部分延续了PointNet++的encoder-decoder分类结构;多层次的特征提取结构捕获了局部及全局特征,分类损失与相似性度量损失以特定比例结合构成网络的整体损失函数;其具体步骤如下所示:In this step, the expanded data obtained by S2 is used as the input of the classification network. The classification network part continues the encoder-decoder classification structure of PointNet++; the multi-level feature extraction structure captures local and global features, and the classification loss and similarity measurement loss are A specific ratio is combined to form the overall loss function of the network; its specific steps are as follows:
S41:将扩容后的完整点云与残缺点云共同输入分类网络实现残缺点云分类;具体包括步骤如下:S41: Input the expanded complete point cloud and the residual defect cloud together into the classification network to achieve residual defect cloud classification; the specific steps are as follows:
S411:利用PointNet++的编码部分中的SA层实现特征提取,SA层具体包括Sampling,Grouping,PointNet三个模块,具体步骤如下:S411: Use the SA layer in the coding part of PointNet++ to achieve feature extraction. The SA layer specifically includes three modules: Sampling, Grouping, and PointNet. The specific steps are as follows:
S4111:SA层中的Sampling模块首先对输入模块的点云进行下采样,采样方式是最远点采样,采样点数为1024个点;S4111: The Sampling module in the SA layer first downsamples the point cloud of the input module, the sampling method is the farthest point sampling, and the number of sampling points is 1024 points;
S4112:SA层中的Grouping模块以所选取点为质心,通过邻域算法选取固定区域内合适数目的点构成局部点集,每个邻域内选取的点数为32个;S4112: The Grouping module in the SA layer takes the selected point as the centroid, selects an appropriate number of points in the fixed area through the neighborhood algorithm to form a local point set, and the number of points selected in each neighborhood is 32;
S4113:SA层的PointNet部分则使用最大池化(max-pooling)作为对称函数实现点云的排列不变性,使用MLP模块提取局部特征,该过程的公式如下:S4113: The PointNet part of the SA layer uses max-pooling as a symmetric function to achieve the invariance of the arrangement of point clouds, and uses the MLP module to extract local features. The formula for this process is as follows:
f({x1,...,xn})≈g(h(x1),...,h(xn))f({x 1 ,...,x n })≈g(h(x 1 ),...,h(x n ))
在该实现过程中,h代表MLP,g代表max-pooling函数,max-pooling函数表示特征中最大值且满足置换不变性,该公式为:In this implementation process, h represents MLP, g represents the max-pooling function, and the max-pooling function represents the maximum value in the feature and satisfies the permutation invariance. The formula is:
max(x1,x2,x3,...)≡max(xα1,xα2,xα3,...),xi∈RD max(x 1 ,x 2 ,x 3 ,...)≡max(x α1 ,x α2 ,x α3 ,...), xi ∈R D
S412:重复两次S411的过程,逐步获得不同层次的局部特征及全局特征;S412: Repeat the process of S411 twice to gradually obtain local features and global features at different levels;
S413:解码部分将编码部分所得降采样得到的全局特征作为输入引入全连接网络,并通过Softmax完成分类;S413: The decoding part introduces the global feature obtained by the downsampling obtained from the encoding part as an input into the fully connected network, and completes the classification through Softmax;
S42:将分类网络得到的损失函数与S3得到的相似性度量损失相结合组成IPC-Net的损失函数,分类网络部分的损失函数LCLS为NLLLoss;相似性度量损失由S333所得,并以辅助函数的形式应用于模型的训练过程,在测试环节去除;超参数alpha用于均衡两种损失,alpha值经由实验确定为0.05,IPC-Net的损失函数为:S42: Combine the loss function obtained by the classification network and the similarity measure loss obtained by S3 to form the loss function of IPC-Net. The loss function L CLS of the classification network part is NLLLoss; The form of is applied to the training process of the model and removed in the testing process; the hyperparameter alpha is used to balance the two losses, the alpha value is determined to be 0.05 through experiments, and the loss function of IPC-Net is:
LCON=0.95LCLS+0.05LSM。L CON = 0.95L CLS + 0.05L SM .
与现有技术相比,本发明的有益效果是:Compared with the prior art, the beneficial effects of the present invention are:
1、本发明提出了一个残缺点云的分类网络IPC-Net。它可以将完整的点云分类网络扩展到不完整的点云,从而确保不同完整性的点云分类的鲁棒性。1. The present invention proposes a residual defect cloud classification network IPC-Net. It can extend the complete point cloud classification network to incomplete point clouds, thus ensuring the robustness of point cloud classification with different completeness.
2、本发明设计了一种针对3D点云的数据扩充方法,该方法在语义层面上注重数据的扩充,具有实际意义。我们提出了一种点相似度测量模块,其中引入并完善了注意力机制来设计辅助损失函数。2. The present invention designs a data expansion method for 3D point cloud, which pays attention to data expansion at the semantic level and has practical significance. We propose a point similarity measurement module in which an attention mechanism is introduced and refined to design an auxiliary loss function.
3、本发明不依赖于特定数据集,在不同输入数据集上均具有普适性,在ModelNet40数据集上达到优秀的分类效果。3. The present invention does not depend on a specific data set, is universal on different input data sets, and achieves an excellent classification effect on the ModelNet40 data set.
附图说明Description of drawings
图1是本发明提出的基于数据扩容与相似性度量的残缺点云分类网络IPC-Net的流程图;Fig. 1 is the flow chart of the residual defect cloud classification network IPC-Net based on data expansion and similarity measure proposed by the present invention;
图2是IPC-Net网络的各模块结构图;Fig. 2 is each module structure diagram of IPC-Net network;
图3是随即区域擦除的流程图;Fig. 3 is the flow chart of random area erasing;
图4是ModelNet40数据集随机擦除的可视化结果;Figure 4 is the visualization result of random erasure of the ModelNet40 dataset;
图5是用于判断IPC-Net泛化性能的自制数据集可视化图。Figure 5 is a visualization of the homemade dataset used to judge the generalization performance of IPC-Net.
具体实施方式Detailed ways
下面结合附图对本发明具体实施方式做进一步说明。The specific embodiments of the present invention will be further described below with reference to the accompanying drawings.
本申请中提到的点云,是在同一空间参考系下表达目标空间分布和目标表面特性的海量点集合,在获取物体表面每个采样点的空间坐标后,得到的是点的集合。The point cloud mentioned in this application is a massive set of points that express the spatial distribution of the target and the characteristics of the target surface under the same spatial reference system. After obtaining the spatial coordinates of each sampling point on the surface of the object, a set of points is obtained.
本申请中提到的点云分类,是为每个点分配一个语义标记。点云的分类是将点云分类到不同的点云集,同一个点云集具有相似或相同的属性。The point cloud classification mentioned in this application is to assign a semantic label to each point. The classification of point clouds is to classify point clouds into different point cloud sets, and the same point cloud set has similar or the same properties.
本申请中提到的点云补全,是一种在三维物体的检测识别领域应用广泛的用于修补有所缺失的点云的方法,从缺失点云出发估计完整点云,从而获得更高质量的点云。面向残缺点云的分类网络是对点云分类网络的进一步推广,残缺点云的常用分析方法主要是点云补全网络。点云补全是实现残缺点云分类的先验任务,获取完整点云数据后再进行形状分类。但是,在点云分类任务中,点云补全操作是冗余的。因为网络可以通过学习残缺点云的语义信息来实现分类,而不需要生成具有不明显特征的局部细节。此外,在分类网络之前添加点云完成网络也将增加模型的复杂性。与此同时在深度学习分析3D点云的研究中,训练数据起着举足轻重的作用。但是,实际的训练数据通常存在体积小,复杂度低,丰富性差的问题。因此,数据扩容是一种常用的训练方法。数据扩容是正则化的一种显示形式,旨在仅通过人工操作使用数据来扩展训练集的数量和丰富性。由于在此过程中进行了操作转换,数据扩充使网络可以学习的语义信息最大化,这也成为防止卷积神经网络模型过度拟合的关键部分。The point cloud completion mentioned in this application is a method widely used in the field of three-dimensional object detection and recognition for repairing missing point clouds. The complete point cloud is estimated from the missing point cloud, so as to obtain higher quality point cloud. The classification network for the residual defect cloud is a further promotion of the point cloud classification network. The common analysis method of the residual defect cloud is mainly the point cloud completion network. Point cloud completion is a priori task to achieve residual and defect cloud classification. After obtaining complete point cloud data, shape classification is performed. However, in point cloud classification tasks, point cloud completion operations are redundant. Because the network can achieve classification by learning the semantic information of residual defect clouds without generating local details with indistinct features. In addition, adding the point cloud completion network before the classification network will also increase the complexity of the model. At the same time, in the research of deep learning analysis of 3D point cloud, training data plays a pivotal role. However, the actual training data usually has the problems of small size, low complexity and poor richness. Therefore, data scaling is a commonly used training method. Data augmentation is an explicit form of regularization that aims to expand the number and richness of the training set using data only through human manipulation. Due to the operational transformations performed in the process, data augmentation maximizes the semantic information that the network can learn, which also becomes a key part of preventing overfitting of convolutional neural network models.
本发明内容主要包括数据扩容模块和点相似性度量模块。数据扩容部分主要通过随机区域擦除来实现。随机区域擦除过程在实际应用中模拟了点云的遮挡,从而提高了网络的泛化性能。点相似性度量模块将注意力机制与相似性度量结合在一起,以指导不完整的点特征向原始点特征学习。同时,本发明利用注意力获得原始点云不同点之间的权重,使网络更加关注具有高信息量的学习特征。本发明还使用PointNet++的分层分类结构作为参考,演示了IPC-Net,并演示了其在几何任务上的有效性。The content of the present invention mainly includes a data expansion module and a point similarity measurement module. The data expansion part is mainly realized by erasing random areas. The random region erasing process simulates the occlusion of point clouds in practical applications, thereby improving the generalization performance of the network. The point similarity measure module combines an attention mechanism with a similarity measure to guide the learning of incomplete point features towards the original point features. At the same time, the present invention uses attention to obtain the weights between different points of the original point cloud, so that the network pays more attention to the learning features with high information content. The present invention also uses the hierarchical classification structure of PointNet++ as a reference to demonstrate IPC-Net and demonstrate its effectiveness on geometric tasks.
下面结合说明书附图对本发明创造进一步的详细说明。The present invention will be further described in detail below with reference to the accompanying drawings.
图1为本发明提出的基于数据扩容与相似性度量的残缺点云分类网络IPC-Net的流程图,图2为IPC-Net网络各模块的具体结构图。具体包括如下步骤:FIG. 1 is a flowchart of the residual defect cloud classification network IPC-Net based on data expansion and similarity measurement proposed by the present invention, and FIG. 2 is a specific structural diagram of each module of the IPC-Net network. Specifically include the following steps:
S1,预处理输入数据;S1, preprocess input data;
S2,对原始点云进行数据扩容;S2, perform data expansion on the original point cloud;
S3,利用相似性度量模块评估完整点云与残缺点云之间差异,并输出相似性度量损失;S3, use the similarity measurement module to evaluate the difference between the complete point cloud and the residual defect cloud, and output the similarity measurement loss;
S4,利用分类网络实现残缺点云分类。S4, use the classification network to achieve cloud classification of residual defects.
下面结合附图所示,更加具体的描述前述基于数据扩容和点云相似性度量的残缺点云分类网络。The following describes the residual defect cloud classification network based on data expansion and point cloud similarity measurement in more detail with reference to the accompanying drawings.
步骤S1,对原始点云进行归一化,将不同大小的点云数据归一化到[-1,1]范围内,原始点云的可视化结果如图5所示。In step S1, the original point cloud is normalized, and the point cloud data of different sizes are normalized to the range of [-1, 1]. The visualization result of the original point cloud is shown in Figure 5.
步骤S2,中对结构完整的原始点云实行数据扩容,数据扩容过程主要是通过随机区域擦除实现的,随即区域擦除流程如图3所示,其具体步骤如下所示:In step S2, data expansion is performed on the original point cloud with complete structure. The data expansion process is mainly realized by erasing random regions. The process of erasing regions is shown in Figure 3. The specific steps are as follows:
S2.1,在训练过程中,原始点云数据输入网络后复制本身获得与之相同的“复本”作为随机区域擦除模块的输入数据。S2.1, in the training process, after the original point cloud data is input into the network, the copy itself obtains the same "replica" as the input data of the random area erasing module.
S2.2,随机区域擦除的方法会在“复本”点集中随机选择一个点作为擦除的区域中心。S2.2, the method of random area erasing will randomly select a point in the "replica" point set as the center of the area to be erased.
S2.3,随机区域擦除模块设置了一个擦除的范围,并随机选定范围内的任意一大小作为擦除半径以形成一个可供擦除的球域。S2.3, the random area erasing module sets an erasing range, and randomly selects any size within the range as the erasing radius to form a sphere for erasing.
S2.4,将擦除球域内的点云数据清除,与此同时模块需要判断经过擦除的点云所保留的点数是否满足后续分类网络的要求下限,分类网络的输入点数为1024,故当擦除后剩余点数不满1024时,网络则放弃对该点云数据实行随机区域擦除。S2.4, clear the point cloud data in the erased sphere. At the same time, the module needs to judge whether the number of points retained by the erased point cloud meets the lower limit of the requirements of the subsequent classification network. The number of input points of the classification network is 1024, so when When the number of remaining points after erasing is less than 1024, the network will give up the random area erasing of the point cloud data.
本发明的数据扩容模块保证了最大程度的随机原则,首先是擦除区域中心点的位置是随机的,其次是擦除区域的半径大小也是随机的,最后由于可能存在点高度集中于擦除区域的现象,所以对于一个点云数据,能否执行擦除操作也是随机的,图4可视化的显示了随机擦除的完整流程。The data expansion module of the present invention ensures the principle of randomness to the greatest extent. First, the position of the center point of the erasing area is random, secondly, the radius of the erasing area is also random. Finally, there may be points that are highly concentrated in the erasing area. Therefore, for a point cloud data, whether the erasing operation can be performed is also random. Figure 4 visually shows the complete process of random erasing.
步骤S3,将经过最远点采样的输入点云引入相似性度量模块,利用评估指标将完整点云与残缺点云联系起来,通过评估二者之间的差异性获得点云相似性度量损失,该过程具体步骤如下:Step S3, introduce the input point cloud sampled by the farthest point into the similarity measurement module, use the evaluation index to connect the complete point cloud and the residual defect cloud, and obtain the point cloud similarity measurement loss by evaluating the difference between the two, The specific steps of this process are as follows:
S3.1,将数据量不同的完整点云与残缺点云共同降采样,降采样的方法为最远点采样,经过最远点采样的完整点云与残缺点云各保留1024个采样点。S3.1, down-sampling the complete point cloud and residual defect cloud with different amounts of data together. The down-sampling method is the farthest point sampling, and the complete point cloud and residual defect cloud sampled by the farthest point each retain 1024 sampling points.
S3.2,对下采样后的完整点云应用注意力机制,以此获得各个点之间的权重,获取注意力的公式为:S3.2, the attention mechanism is applied to the downsampled complete point cloud to obtain the weight between each point. The formula for obtaining the attention is:
A=g(Qi,W)A=g(Q i ,W)
其中Qi={qj|1≤j≤i}代表了完整点云,W是共享多层感知机的可学习权重。where Q i ={q j |1≤j≤i} represents the complete point cloud, and W is the learnable weight of the shared multilayer perceptron.
S3.3,利用点云补全中的评价指标评估残缺点云与完整点云的关系,该部分主要包括如下步骤:S3.3, use the evaluation indicators in point cloud completion to evaluate the relationship between the residual defect cloud and the complete point cloud. This part mainly includes the following steps:
S3.3.1,采用点云补全中的倒角距离(CD)来评估完整点云Q与残缺点云P之间的关系,CD表示一个点云中每一个点到另一个点云中与之最近的邻居点之间的平均距离,由此构成CD损失,其具体公式为:S3.3.1, use the chamfering distance (CD) in point cloud completion to evaluate the relationship between the complete point cloud Q and the residual defect cloud P, CD represents each point in one point cloud to another point cloud with it The average distance between the nearest neighbor points, which constitutes the CD loss, and its specific formula is:
S3.3.2,采用点云补全中的地球移动距离(EMD)来评估完整点云Q与残缺点云P之间的关系,EMD距离表示的是一个点云“移动”成另一个点云的最优规划路径下的最小消耗,由此构成EMD损失,其具体公式为:S3.3.2, the Earth Movement Distance (EMD) in point cloud completion is used to evaluate the relationship between the complete point cloud Q and the residual defect cloud P. The EMD distance represents the "movement" of one point cloud into another point cloud. The minimum consumption under the optimal planning path constitutes the EMD loss, and its specific formula is:
其中φ是一个可以使对应点之间的平均距离最小的双射,且和的大小必须相同。where φ is a bijection that minimizes the average distance between corresponding points, and must have the same magnitude.
S3.3.3,将CD损失与EMD损失相结合,CD计算效率高时间复杂度低,EMD对单个点云的个性特征及局部细节有更好的区分度,将二者结合起来共同构建相似性度量损失,该损失函数公式为:S3.3.3, Combining CD loss and EMD loss, CD calculation efficiency is high and time complexity is low, EMD has a better degree of discrimination for the individual characteristics and local details of a single point cloud, and the two are combined to jointly construct a similarity measure loss, the loss function formula is:
S3.4,将完整点云经过注意力模块获得的权重作为参数,引入完整点云与残缺点云之间距离dij的计算中,dij是相似性度量损失的核心部分,其公式如下:S3.4, take the weight obtained by the complete point cloud through the attention module as a parameter, and introduce it into the calculation of the distance d ij between the complete point cloud and the residual defect cloud, d ij is the core part of the similarity measure loss, and its formula is as follows:
dij=A·||Q-P||d ij =A·||QP||
S4.1,将扩容后的完整点云与残缺点云共同输入分类网络实现残缺点云分类。其具体步骤如下:S4.1, input the expanded complete point cloud and residual defect cloud into the classification network to realize residual defect cloud classification. The specific steps are as follows:
S4.1.1,利用PointNet++的编码部分中的SA层实现特征提取,SA层具体包括Sampling,Grouping,PointNet三个模块,具体步骤如下:S4.1.1, use the SA layer in the coding part of PointNet++ to achieve feature extraction. The SA layer specifically includes three modules: Sampling, Grouping, and PointNet. The specific steps are as follows:
S4.1.1.1,SA层中的Sampling模块首先对输入模块的点云进行下采样,采样方式是最远点采样,采样点数为1024个点。S4.1.1.1, the Sampling module in the SA layer first downsamples the point cloud of the input module, the sampling method is the farthest point sampling, and the number of sampling points is 1024 points.
S4.1.1.2,SA层中的Grouping模块以所选取点为质心,通过邻域算法选取固定区域内合适数目的点构成局部点集,每个邻域内选取的点数为32个。S4.1.1.2, the Grouping module in the SA layer takes the selected point as the centroid, and selects an appropriate number of points in the fixed area through the neighborhood algorithm to form a local point set, and the number of points selected in each neighborhood is 32.
S4.1.1.3,SA层的PointNet部分则使用最大池化(max-pooling)作为对称函数实现点云的排列不变性,使用MLP模块提取局部特征,该过程的公式如下:In S4.1.1.3, the PointNet part of the SA layer uses max-pooling as a symmetric function to achieve the invariance of the arrangement of point clouds, and uses the MLP module to extract local features. The formula for this process is as follows:
f({x1,...,xn})≈g(h(x1),...,h(xn))f({x 1 ,...,x n })≈g(h(x 1 ),...,h(x n ))
在该过程中,h代表MLP,g代表max-pooling函数,max-pooling函数表示特征中最大值且满足置换不变性,该公式为:In this process, h represents the MLP, g represents the max-pooling function, and the max-pooling function represents the maximum value in the feature and satisfies the permutation invariance. The formula is:
max(x1,x2,x3,...)≡max(xα1,xα2,xα3,...),xi∈RD max(x 1 ,x 2 ,x 3 ,...)≡max(x α1 ,x α2 ,x α3 ,...), xi ∈R D
S4.1.2,重复两次S411的过程,逐步获得不同层次的局部特征及全局特征。S4.1.2, repeat the process of S411 twice to gradually obtain local features and global features at different levels.
S4.1.3,解码部分将编码部分所得降采样得到的全局特征作为输入引入全连接网络,并通过Softmax完成分类。S4.1.3, the decoding part takes the global features obtained by down-sampling obtained from the encoding part as input into the fully connected network, and completes the classification through Softmax.
S4.2,将分类网络得到的损失函数与S3得到的相似性度量损失相结合组成IPC-Net的损失函数,分类网络部分的损失函数LCLS为NLLLoss。相似性度量损失由S333所得,并以辅助函数的形式应用于模型的训练过程,在测试环节去除。超参数alpha用于均衡两种损失,本发明中的alpha值经由实验确定为0.05,IPC-Net的损失函数为:S4.2, the loss function obtained by the classification network and the similarity measure loss obtained in S3 are combined to form the loss function of IPC-Net, and the loss function L CLS of the classification network part is NLLLoss. The similarity measure loss is obtained by S333, and is applied to the training process of the model in the form of an auxiliary function, and is removed in the testing process. The hyperparameter alpha is used to balance the two losses. The alpha value in the present invention is determined to be 0.05 through experiments, and the loss function of IPC-Net is:
LCON=0.95LCLS+0.05LSM L CON = 0.95L CLS +0.05L SM
结合以上所描述的残缺点云分类方法,在输入点云为完整点云或为缺损程度不同的残缺点云时,利用本发明提出的ICP-Net网络,均能实现较好的分类效果,且网络具备较强的鲁棒性。Combining the above-described classification method of residual defect cloud, when the input point cloud is a complete point cloud or a residual defect cloud with different degrees of defect, the ICP-Net network proposed by the present invention can achieve a better classification effect, and The network has strong robustness.
虽然本发明已以较佳实例揭露如上,然其并非用以限定本发明。本发明所属技术领域中具有通常知识者,在不脱离本发明的精神和范围内,当可作各种的更动与润饰。因此,本发明的保护范围当视权利要求书所界定为准。Although the present invention has been disclosed above with preferred embodiments, it is not intended to limit the present invention. Those skilled in the art to which the present invention pertains can make various changes and modifications without departing from the spirit and scope of the present invention. Therefore, the protection scope of the present invention should be defined by the claims.
Claims (1)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110652559.2A CN113298952B (en) | 2021-06-11 | 2021-06-11 | Incomplete point cloud classification method based on data expansion and similarity measurement |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110652559.2A CN113298952B (en) | 2021-06-11 | 2021-06-11 | Incomplete point cloud classification method based on data expansion and similarity measurement |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113298952A CN113298952A (en) | 2021-08-24 |
CN113298952B true CN113298952B (en) | 2022-07-15 |
Family
ID=77328010
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110652559.2A Active CN113298952B (en) | 2021-06-11 | 2021-06-11 | Incomplete point cloud classification method based on data expansion and similarity measurement |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113298952B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112132958A (en) * | 2020-09-23 | 2020-12-25 | 哈尔滨工程大学 | A 3D reconstruction method of underwater environment based on binocular vision |
CN112614071A (en) * | 2020-12-29 | 2021-04-06 | 清华大学 | Self-attention-based diverse point cloud completion method and device |
CN112785526A (en) * | 2021-01-28 | 2021-05-11 | 南京大学 | Three-dimensional point cloud repairing method for graphic processing |
CN112927359A (en) * | 2021-03-22 | 2021-06-08 | 南京大学 | Three-dimensional point cloud completion method based on deep learning and voxels |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10176589B2 (en) * | 2017-01-31 | 2019-01-08 | Mitsubishi Electric Research Labroatories, Inc. | Method and system for completing point clouds using planar segments |
CN112488210B (en) * | 2020-12-02 | 2025-01-10 | 北京工业大学 | A method for automatic classification of 3D point clouds based on graph convolutional neural networks |
-
2021
- 2021-06-11 CN CN202110652559.2A patent/CN113298952B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112132958A (en) * | 2020-09-23 | 2020-12-25 | 哈尔滨工程大学 | A 3D reconstruction method of underwater environment based on binocular vision |
CN112614071A (en) * | 2020-12-29 | 2021-04-06 | 清华大学 | Self-attention-based diverse point cloud completion method and device |
CN112785526A (en) * | 2021-01-28 | 2021-05-11 | 南京大学 | Three-dimensional point cloud repairing method for graphic processing |
CN112927359A (en) * | 2021-03-22 | 2021-06-08 | 南京大学 | Three-dimensional point cloud completion method based on deep learning and voxels |
Non-Patent Citations (4)
Title |
---|
MinkLoc3D: Point Cloud Based Large-Scale Place Recognition;Jacek Komorowski;《https://arxiv.org/abs/2011.04530》;20201109;第1-10页 * |
基于时空注意力机制的视频行人再识别方法研究;李明哲;《中国优秀博硕士学位论文全文数据库(硕士) 信息科技辑》;20210515(第05期);第I138-811页 * |
基于生成对抗网络的点云形状保结构补全;缪永伟 等;《中国科学:信息科学》;20200420;第50卷(第05期);第675–691页 * |
基于神经网络的点云残缺数据修补方法;刘俊 等;《机械工程师》;20070210(第02期);第32-34页 * |
Also Published As
Publication number | Publication date |
---|---|
CN113298952A (en) | 2021-08-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110111345B (en) | A 3D Point Cloud Segmentation Method Based on Attention Network | |
CN110276316B (en) | A human keypoint detection method based on deep learning | |
CN113449736B (en) | Photogrammetry point cloud semantic segmentation method based on deep learning | |
CN113379646B (en) | An Algorithm for Dense Point Cloud Completion Using Generative Adversarial Networks | |
CN112347987A (en) | A 3D Object Detection Method Based on Multimodal Data Fusion | |
CN112949647B (en) | Three-dimensional scene description method and device, electronic equipment and storage medium | |
CN111127538B (en) | A 3D reconstruction method of multi-view images based on convolutional cyclic encoding-decoding structure | |
CN110132263A (en) | A star map recognition method based on representation learning | |
CN114842343B (en) | ViT-based aerial image recognition method | |
CN111753698A (en) | A multimodal three-dimensional point cloud segmentation system and method | |
CN113837166B (en) | Automatic pointer instrument reading method based on deep learning | |
CN116704137B (en) | Reverse modeling method for point cloud deep learning of offshore oil drilling platform | |
CN113628329B (en) | Zero-sample sketch three-dimensional point cloud retrieval method | |
CN116958420A (en) | A high-precision modeling method for the three-dimensional face of a digital human teacher | |
CN114638751A (en) | End-to-end point cloud completion method based on generation of countermeasure network | |
Shi et al. | An improved lightweight deep neural network with knowledge distillation for local feature extraction and visual localization using images and LiDAR point clouds | |
CN116563466A (en) | Deep learning-based three-dimensional Shan Mudian cloud completion method | |
Gao et al. | HDRNet: High‐Dimensional Regression Network for Point Cloud Registration | |
CN113281779B (en) | A method, device, equipment and medium for rapid detection of 3D objects | |
CN113343765B (en) | Scene retrieval method and system based on point cloud rigid registration | |
CN115019147A (en) | A Transformer-based Grasp Detection Model Applicable to Object Stacking and Chaotic Scenes | |
CN113298952B (en) | Incomplete point cloud classification method based on data expansion and similarity measurement | |
CN118397616A (en) | 3D target detection method based on density perception completion and sparse fusion | |
CN118037785A (en) | A method of image registration based on point hints | |
CN117593368A (en) | 6D pose estimation method based on iterative attention fusion network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |