CN117036370A - Plant organ point cloud segmentation method based on attention mechanism and graph convolution - Google Patents
Plant organ point cloud segmentation method based on attention mechanism and graph convolution Download PDFInfo
- Publication number
- CN117036370A CN117036370A CN202310704110.5A CN202310704110A CN117036370A CN 117036370 A CN117036370 A CN 117036370A CN 202310704110 A CN202310704110 A CN 202310704110A CN 117036370 A CN117036370 A CN 117036370A
- Authority
- CN
- China
- Prior art keywords
- point
- feature
- point cloud
- layer
- features
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000011218 segmentation Effects 0.000 title claims abstract description 53
- 238000000034 method Methods 0.000 title claims abstract description 33
- 230000007246 mechanism Effects 0.000 title claims abstract description 30
- 210000000056 organ Anatomy 0.000 title claims abstract description 16
- 238000012549 training Methods 0.000 claims abstract description 16
- 230000008878 coupling Effects 0.000 claims abstract description 14
- 238000010168 coupling process Methods 0.000 claims abstract description 14
- 238000005859 coupling reaction Methods 0.000 claims abstract description 14
- 238000013528 artificial neural network Methods 0.000 claims abstract description 6
- 238000010586 diagram Methods 0.000 claims abstract description 6
- 238000000605 extraction Methods 0.000 claims abstract description 6
- 230000002776 aggregation Effects 0.000 claims description 23
- 238000004220 aggregation Methods 0.000 claims description 23
- 239000011159 matrix material Substances 0.000 claims description 17
- 230000006870 function Effects 0.000 claims description 15
- 238000011176 pooling Methods 0.000 claims description 15
- 238000004422 calculation algorithm Methods 0.000 claims description 13
- 230000008569 process Effects 0.000 claims description 9
- 238000004364 calculation method Methods 0.000 claims description 8
- 239000000284 extract Substances 0.000 claims description 8
- 238000012545 processing Methods 0.000 claims description 6
- 239000013598 vector Substances 0.000 claims description 5
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 3
- 230000004913 activation Effects 0.000 claims description 3
- 238000011426 transformation method Methods 0.000 claims description 3
- 238000013507 mapping Methods 0.000 claims description 2
- 239000010410 layer Substances 0.000 claims 19
- 239000013047 polymeric layer Substances 0.000 claims 1
- 238000006116 polymerization reaction Methods 0.000 claims 1
- 238000005070 sampling Methods 0.000 claims 1
- 241000196324 Embryophyta Species 0.000 abstract description 37
- 238000011160 research Methods 0.000 abstract description 7
- 238000004458 analytical method Methods 0.000 abstract description 5
- 240000006394 Sorghum bicolor Species 0.000 abstract description 4
- 235000011684 Sorghum saccharatum Nutrition 0.000 abstract description 4
- 235000007688 Lycopersicon esculentum Nutrition 0.000 abstract description 3
- 244000061176 Nicotiana tabacum Species 0.000 abstract description 3
- 235000002637 Nicotiana tabacum Nutrition 0.000 abstract description 3
- 240000003768 Solanum lycopersicum Species 0.000 abstract description 3
- 235000021307 Triticum Nutrition 0.000 abstract description 2
- 244000098338 Triticum aestivum Species 0.000 abstract description 2
- 240000008042 Zea mays Species 0.000 abstract description 2
- 235000005824 Zea mays ssp. parviglumis Nutrition 0.000 abstract description 2
- 235000002017 Zea mays subsp mays Nutrition 0.000 abstract description 2
- 235000005822 corn Nutrition 0.000 abstract description 2
- 208000003643 Callosities Diseases 0.000 abstract 1
- 206010020649 Hyperkeratosis Diseases 0.000 abstract 1
- 230000000694 effects Effects 0.000 description 5
- 238000013527 convolutional neural network Methods 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 4
- 230000000877 morphologic effect Effects 0.000 description 3
- 238000012360 testing method Methods 0.000 description 3
- 241000209510 Liliopsida Species 0.000 description 2
- 238000002679 ablation Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 241000894007 species Species 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000010521 absorption reaction Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 235000013399 edible fruits Nutrition 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 230000029553 photosynthesis Effects 0.000 description 1
- 238000010672 photosynthesis Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Medical Informatics (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域Technical field
本发明属于三维点云实例分割技术领域,具体涉及一种基于注意力机制和空间图卷积的双分支并行植物器官点云分割方法。The invention belongs to the technical field of three-dimensional point cloud instance segmentation, and specifically relates to a dual-branch parallel plant organ point cloud segmentation method based on attention mechanism and spatial graph convolution.
背景技术Background technique
随着激光雷达设备的普及和各种消费级深度传感器的出现,点云数据越来越广泛的应用于各个领域,比如机器人、自动驾驶和城市规划等。在表型分析研究中,三维点云作为现实世界的低分辨率表示,已经成为研究植物结构和形态最直接有效的数据形式。许多研究已经采用植物三维结构来进行器官分割、监测长势和评价品种等。三维坐标系中的点作为构成点云的最基本单元,类似于二维图片中的像素点,但可以容纳更多高维度的语义信息。在表型研究中,植物器官的形态结构是直观且重要的性状,可以反映出植物对外界条件的适应能力和生长情况,例如光合作用效率和水分吸收效率等。植物器官点云分割是指将植物按照不同器官(如茎、叶、果实等)进行语义划分的过程,是后续对点云数据进行深入理解的基础,并对理解植物的功能结构具有重要意义,目前是一个颇具挑战性的研究方向。With the popularization of lidar equipment and the emergence of various consumer-grade depth sensors, point cloud data is increasingly used in various fields, such as robotics, autonomous driving, and urban planning. In phenotypic analysis research, three-dimensional point cloud, as a low-resolution representation of the real world, has become the most direct and effective data form for studying plant structure and morphology. Many studies have used plant three-dimensional structures to segment organs, monitor growth, and evaluate varieties. Points in a three-dimensional coordinate system, as the most basic units that constitute point clouds, are similar to pixels in two-dimensional pictures, but can accommodate more high-dimensional semantic information. In phenotypic research, the morphological structure of plant organs is an intuitive and important trait, which can reflect the plant's ability to adapt to external conditions and growth, such as photosynthesis efficiency and water absorption efficiency. Plant organ point cloud segmentation refers to the process of semantically dividing plants according to different organs (such as stems, leaves, fruits, etc.). It is the basis for subsequent in-depth understanding of point cloud data and is of great significance for understanding the functional structure of plants. It is currently a challenging research direction.
传统的植物点云分割算法需要人工预先进行特征描述,这种分割过程复杂繁琐,随着大数据时代的来临,传统处理方法难以满足快速准确分析的需求。因此,自动化分割方法的需求日益增大。而随着计算机图形处理器性能的飞速增长,作为人工智能的主导技术,深度学习已经成功的用于解决各类二维视觉问题。但由于点云在空间汇总的无序性和复杂性,深度学习方法在点云上的应用还面临诸多挑战。在视觉分割任务中有良好表现的卷积神经网络(CNN)通过共享核卷积来提取特征,提高模型效率,CNN固有的平移不变性使其对局部特征的把控更为精准。然而,CNN自身的感受野通常比较小,对于全局特征的捕获能力相对较弱,并且无法直接作用于原始点云数据中。另一种应用于点云数据中的神经网络结构是图卷积网络(GCN),它将点云中每个独立的点视为图数据结构中的顶点,可以通过直接在点云数据中进行类卷积操作来提取局部特征。在自然语言处理领域有着杰出表现的Transformer同样可以很好地捕获到全局特征,其核心思想注意力(Attention)机制也非常适合处理点云数据。这些深度学习方法在许多公共点云数据集上取得了令人满意的分割结果,展现了深度学习方法对于点云数据分割的有效性。The traditional plant point cloud segmentation algorithm requires manual feature description in advance. This segmentation process is complex and cumbersome. With the advent of the big data era, traditional processing methods cannot meet the needs of fast and accurate analysis. Therefore, there is an increasing need for automated segmentation methods. With the rapid growth of computer graphics processor performance, deep learning, as the leading technology of artificial intelligence, has been successfully used to solve various two-dimensional vision problems. However, due to the disorder and complexity of point clouds in spatial aggregation, the application of deep learning methods on point clouds still faces many challenges. Convolutional neural networks (CNN), which have good performance in visual segmentation tasks, extract features through shared kernel convolution and improve model efficiency. The inherent translation invariance of CNN makes it more accurate to control local features. However, the receptive field of CNN itself is usually relatively small, the ability to capture global features is relatively weak, and it cannot directly act on the original point cloud data. Another neural network structure applied to point cloud data is the graph convolutional network (GCN), which treats each independent point in the point cloud as a vertex in the graph data structure, which can be directly implemented in the point cloud data. Convolution-like operations to extract local features. Transformer, which has outstanding performance in the field of natural language processing, can also capture global features well, and its core idea attention mechanism is also very suitable for processing point cloud data. These deep learning methods have achieved satisfactory segmentation results on many public point cloud data sets, demonstrating the effectiveness of deep learning methods for point cloud data segmentation.
然而,相对来说,植物点云在结构上的复杂性导致在器官分割任务中需要识别的语义信息量也就更大。在获取点云时,叶片之间的遮挡问题常常会导致部分点云丢失,出现孔洞、稀疏的问题。此外,植物器官之间的相似度很高,不同叶片实例往往具有相同的颜色、形态结构和纹理等特征,这种高度重复的特征对于神经网络的学习并不友好。最后,不同品种植物具有的几何形态特征不同,即使同一品种的植物在不同生长环境下的表型特征也不尽相同,甚至会产生很大差异,这对于网络的泛化能力要求很高。综上,目前植物点云的分割精度尚不能满足要求。However, relatively speaking, the structural complexity of plant point clouds results in a greater amount of semantic information that needs to be recognized in the organ segmentation task. When obtaining point clouds, the occlusion problem between leaves often causes part of the point cloud to be lost, resulting in holes and sparse problems. In addition, the similarity between plant organs is very high, and different leaf instances often have the same characteristics such as color, morphological structure, and texture. This highly repetitive feature is not friendly to the learning of neural networks. Finally, different species of plants have different geometric morphological characteristics. Even the phenotypic characteristics of plants of the same species in different growth environments are different, and may even be very different, which requires high generalization capabilities of the network. In summary, the current segmentation accuracy of plant point clouds cannot meet the requirements.
发明内容Contents of the invention
本发明的目的是为了解决复杂植物点云中精准的器官分割问题,提供一种基于注意力机制和空间图卷积的双分支并行植物器官点云分割方法,该方法为复杂结构的植物三维点云提供了可靠且高效的器官分割方法,以番茄、玉米、烟草、高粱、小麦五种植物点云数据作为研究对象,基于注意力机制和空间图卷积全新设计一种双分支并行神经网络架构TRGCN,能够同时捕获点云的局部特征和全局特征,用于训练高鲁棒性的实例分割模型,可提高植物点云的分割精度,为快速、高效、准确的植物表型分析提供数据支持。The purpose of this invention is to solve the problem of accurate organ segmentation in complex plant point clouds and provide a two-branch parallel plant organ point cloud segmentation method based on attention mechanism and spatial graph convolution. This method is a three-dimensional plant point cloud with complex structure. The cloud provides a reliable and efficient organ segmentation method. Taking five plant point cloud data of tomato, corn, tobacco, sorghum, and wheat as the research object, a dual-branch parallel neural network architecture is newly designed based on the attention mechanism and spatial graph convolution. TRGCN can simultaneously capture local features and global features of point clouds and is used to train highly robust instance segmentation models. It can improve the segmentation accuracy of plant point clouds and provide data support for fast, efficient, and accurate plant phenotypic analysis.
为实现上述目的,本发明采取的技术方案如下:In order to achieve the above objects, the technical solutions adopted by the present invention are as follows:
一种基于注意力机制和图卷积的植物器官点云分割方法,所述方法为:A plant organ point cloud segmentation method based on attention mechanism and graph convolution. The method is:
步骤一:特征编码器以原始点云作为输入,采用多层感知机将特征映射到高维空间,并采用点云注意力机制初步提取特征,然后将初始特征数据输入TRGCN块,该模块能够多个级联叠加,以不断深化对高维特征的理解;所述TRGCN块中的特征聚合层提取邻域特征的同时下采样点云,然后进入双分支并行网络部分,分别是由空间图卷积构成的局部特征捕获分支和由点注意力机制构成的全局特征学习分支,最终将特征数据输入T-G特征耦合层,得到目标数量的点云以及对应的高维抽象特征,编码器部分通过叠加TRGCN块,从原始植物点云中抽取高维度的特征信息,用于分割任务;Step 1: The feature encoder takes the original point cloud as input, uses a multi-layer perceptron to map the features to a high-dimensional space, and uses the point cloud attention mechanism to initially extract features, and then inputs the initial feature data into the TRGCN block. This module can cascade superposition to continuously deepen the understanding of high-dimensional features; the feature aggregation layer in the TRGCN block extracts neighborhood features while downsampling point clouds, and then enters the dual-branch parallel network part, which is composed of spatial graph convolution The local feature capture branch composed of the local feature capture branch and the global feature learning branch composed of the point attention mechanism finally input the feature data into the T-G feature coupling layer to obtain the target number of point clouds and corresponding high-dimensional abstract features. The encoder part superimposes the TRGCN block , extract high-dimensional feature information from the original plant point cloud for segmentation tasks;
步骤二:特征解码器部分同样堆叠三个级联的TRGCN块,并分别接收编码器中三个TRGCN块的输出,但用插值层替换特征聚合层,插值层将高维度点集的特征恢复到低维度点集中,但仍然输出K近邻算法的分组结果供TRGCN的两个分支计算所需;对于分割结果预测,解码器在TRGCN块后设置独立的插值层,并采用单个点注意力层来保证信息完整性,网络最后采用多层感知机输出点云的分割结果;Step 2: The feature decoder part also stacks three cascaded TRGCN blocks and receives the output of the three TRGCN blocks in the encoder respectively, but replaces the feature aggregation layer with an interpolation layer. The interpolation layer restores the features of the high-dimensional point set to Low-dimensional points are concentrated, but the grouping results of the K nearest neighbor algorithm are still output for calculation by the two branches of TRGCN; for segmentation result prediction, the decoder sets an independent interpolation layer after the TRGCN block and uses a single point attention layer to ensure Information integrity, the network finally uses a multi-layer perceptron to output the segmentation results of the point cloud;
步骤三:网络训练:本研究所有实验均在独立服务器上进行,该服务器配备了12核心20线程的CPU、64GB内存和一个Nvidia GeForce RTX 3090Ti GPU;使用独立服务器进行神经网络训练,训练阶段,所有的植物点云分割模型采用相同的超参数,所述超参数具体为:训练批次大小设置为32,初始学习率设置为0.001,使用Adam方法对网络进行优化,共训练100个周期,每20个周期将学习率减半,权重衰减设置为0.0001,动量设置为0.9,K近邻算法的K值设置为12,点注意力层的特征维度设置为256。Step 3: Network training: All experiments in this study were conducted on an independent server equipped with a 12-core 20-thread CPU, 64GB memory, and an Nvidia GeForce RTX 3090Ti GPU; an independent server was used for neural network training. During the training phase, all The plant point cloud segmentation model uses the same hyperparameters. The hyperparameters are specifically: the training batch size is set to 32, the initial learning rate is set to 0.001, the Adam method is used to optimize the network, and a total of 100 cycles are trained, every 20 The learning rate is halved every cycle, the weight decay is set to 0.0001, the momentum is set to 0.9, the K value of the K nearest neighbor algorithm is set to 12, and the feature dimension of the point attention layer is set to 256.
进一步地,所述步骤一具体为:Further, the step one is specifically:
(1)特征聚合层(1) Feature aggregation layer
特征聚合具体过程为:输入具有feature维度的x个点,先经过随机最远点采样,然后采用K近邻算法将点云分组,输入到多层感知机中将邻居点特征聚合到中心点,最后采用最大池化操作计算得到具有feature’维度特征的y个点;The specific process of feature aggregation is: input x points with feature dimension, first randomly sample the farthest point, then use the K nearest neighbor algorithm to group the point cloud, input it into the multi-layer perceptron to aggregate the neighbor point features to the center point, and finally The maximum pooling operation is used to calculate y points with feature' dimension characteristics;
特征聚合层采用K近邻算法对输入点集进行采样分组;特征聚合层将计算的K近邻矩阵输出,与后续并行分支共享;The feature aggregation layer uses the K nearest neighbor algorithm to sample and group the input point set; the feature aggregation layer outputs the calculated K nearest neighbor matrix and shares it with subsequent parallel branches;
(2)局部特征捕获分支(2) Local feature capture branch
该分支基于动态空间图卷积构建,用于提取输入的植物点云中的局部特征;首先,基于点集V和近邻信息E构建特征图G=(V,E),采用边缘卷积在输入的特征空间上进行特征提取;提取某点xi特征的公式为:This branch is built based on dynamic spatial graph convolution and is used to extract local features in the input plant point cloud; first, a feature map G = (V, E) is constructed based on the point set V and neighbor information E, and edge convolution is used to extract the local features in the input plant point cloud. Feature extraction is performed on the feature space of ; the formula for extracting features of a certain point x i is:
fi=?h(xi,yi)f i =? h(x i ,y i )
其中,xj为点xi的邻居点之一,?和h分别代表某种聚合函数和某种关系操作;即采用一种关系操作将候选点周围的邻居点特征聚合,即得到候选点的特征信息,该关系操作定义为边缘卷积;Among them, x j is one of the neighbor points of point xi ,? and h respectively represent a certain aggregation function and a certain relational operation; that is, a relational operation is used to aggregate the features of neighbor points around the candidate point, that is, the characteristic information of the candidate point is obtained. The relational operation is defined as edge convolution;
采用最大池化作为聚合函数,具体过程为:Maximum pooling is used as the aggregation function. The specific process is:
convi=Max(MLP(h(xi,xi-xj)))conv i =Max(MLP(h( xi , xi -x j )))
关系操作h定义为点xi,xi和其邻居点xj的特征差值和点xi输出值之间的线性组合;The relational operation h is defined as the linear combination between the feature differences of point x i , x i and its neighbor point x j and the output value of point x i ;
(3)全局特征学习分支(3) Global feature learning branch
采用局部邻域中的向量注意力机制来提取特征,计算公式为:The vector attention mechanism in the local neighborhood is used to extract features. The calculation formula is:
其中,xj是点xi的K个邻居点之一,X为每个单株植物点云中的独立点集合,ρ是正则化函数,γ是映射函数,β是某种关系运算,在本研究中定义为邻域点与关注点的差值,φ、α是点级别的特征变换方法,分别得到自注意力机制中的Q、K、V(分别为Query、Key、Value,是注意力机制中的专有名词,对应中文是查询、键、值)值,δ是位置编码函数,根据以上注意力机制,提出点注意力层,改进后的计算公式为:Among them, x j is one of the K neighbor points of point x i , In this study, it is defined as the difference between the neighborhood point and the attention point, φ, α is a point-level feature transformation method, which is obtained from Q, K, and V in the attention mechanism (Query, Key, and Value respectively, which are proper nouns in the attention mechanism, corresponding to query, key, and value in Chinese) value, δ is the position encoding function. Based on the above attention mechanism, a point attention layer is proposed. The improved calculation formula is:
(4)T-G特征耦合层(4)T-G characteristic coupling layer
经过以上处理,得到两个维度以及形状完全相同的特征矩阵:具有显著局部特征的矩阵G和具有完整全局特征的矩阵T;将G和T拼接后输入特征耦合层得到目标特征矩阵:After the above processing, two feature matrices with identical dimensions and shapes are obtained: a matrix G with significant local features and a matrix T with complete global features; G and T are spliced and input into the feature coupling layer to obtain the target feature matrix:
TG=Linear(ReLU(Linear(T,G)))TG=Linear(ReLU(Linear(T,G)))
采用两个线性层和一个ReLU激活层设计了T-G特征耦合层,使网络能够学习两部分矩阵各自更重要的信息,并结合成为目标特征矩阵。The T-G feature coupling layer is designed using two linear layers and a ReLU activation layer, allowing the network to learn the more important information of the two matrices and combine them into the target feature matrix.
本发明相对于现有技术的有益效果为:本发明设计了一个全新的基于点注意力机制和空间图卷积的双分支并行实例分割网络TRGCN,直接输入三维点云,双分支分别关注局部特征提取和全局特征提取,并通过T-G特征耦合层融合两种特征。结果表明,TRGCN在不同的植物点云上均获得了优秀的性能,准确率高于其他主流点云分割网络,具有很好的泛化能力,可以为快速、高效、准确的植物表型分析提供良好的数据支撑。The beneficial effects of the present invention compared to the existing technology are: the present invention designs a brand new dual-branch parallel instance segmentation network TRGCN based on point attention mechanism and spatial graph convolution, which directly inputs the three-dimensional point cloud, and the dual branches focus on local features respectively. extraction and global feature extraction, and fuse the two features through the T-G feature coupling layer. The results show that TRGCN has achieved excellent performance on different plant point clouds, with an accuracy higher than other mainstream point cloud segmentation networks. It has good generalization ability and can provide fast, efficient and accurate plant phenotype analysis. Good data support.
附图说明Description of the drawings
图1为本发明TRGCN的网络架构图;Figure 1 is a network architecture diagram of the TRGCN of the present invention;
图2为本发明TRGCN块的结构图;Figure 2 is a structural diagram of the TRGCN block of the present invention;
图3为本发明TRGCN块全局特征学习层的架构图;Figure 3 is an architectural diagram of the global feature learning layer of the TRGCN block of the present invention;
图4为本发明对五种植物点云的分割结果图。Figure 4 is a diagram showing the segmentation results of five plant point clouds according to the present invention.
具体实施方式Detailed ways
为使本发明的目的、技术方案及优点更加清楚明白,以下结合具体实施方式,对本发明进行进一步的详细说明。应当理解的是,此处所描述的具体实施方式仅用以解释本发明,并不限定本发明的保护范围。In order to make the purpose, technical solutions and advantages of the present invention more clear, the present invention will be further described in detail below in conjunction with specific embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention and do not limit the scope of the present invention.
实施例1:Example 1:
本研究基于点云自注意力机制和空间图卷积,创新性的提出一种采用编码器-解码器架构设计的双分支并行网络Transformer Graph Convolution Network(TRGCN)(图1)。Based on the point cloud self-attention mechanism and spatial graph convolution, this research innovatively proposes a dual-branch parallel network Transformer Graph Convolution Network (TRGCN) designed using an encoder-decoder architecture (Figure 1).
特征编码器以原始点云作为输入,采用多层感知机将特征映射到高维空间(默认为32维),并采用点云注意力机制初步提取特征。然后将初始特征数据输入TRGCN模块(图2),该模块可多个级联叠加,以不断深化对高维特征的理解。具体来说,TRGCN块中的特征聚合层提取邻域特征的同时下采样点云,然后进入双分支并行网络部分,分别是由空间图卷积构成的局部特征捕获分支和由点注意力机制构成的全局特征学习分支。最终,将特征数据输入特别设计的T-G特征耦合层,得到目标数量的点云以及对应的高维抽象特征。编码器部分通过叠加TRGCN块,从原始植物点云中抽取高维度的特征信息,用于分割任务。The feature encoder takes the original point cloud as input, uses a multi-layer perceptron to map features to a high-dimensional space (default is 32 dimensions), and uses a point cloud attention mechanism to initially extract features. The initial feature data is then input into the TRGCN module (Figure 2), which can be superimposed in multiple cascades to continuously deepen the understanding of high-dimensional features. Specifically, the feature aggregation layer in the TRGCN block extracts neighborhood features while downsampling the point cloud, and then enters the dual-branch parallel network part, which is the local feature capture branch composed of spatial graph convolution and the point attention mechanism. The global feature learning branch. Finally, the feature data is input into the specially designed T-G feature coupling layer to obtain the target number of point clouds and corresponding high-dimensional abstract features. The encoder part extracts high-dimensional feature information from the original plant point cloud by superimposing TRGCN blocks for segmentation tasks.
(1)特征聚合层(1) Feature aggregation layer
TRGCN块中特征聚合层的作用是在多个模块堆叠过程中减少输入的点集基数同时抽象出跟高维度的特征向量。例如从原始输入到第一个TRGCN块,点数量从N减少为N/4,点云特征维度从F提升到2F;The function of the feature aggregation layer in the TRGCN block is to reduce the input point set base and abstract high-dimensional feature vectors during the stacking process of multiple modules. For example, from the original input to the first TRGCN block, the number of points is reduced from N to N/4, and the point cloud feature dimension is increased from F to 2F;
特征聚合具体过程如下:输入具有feature维度的x个点,先经过随机最远点采样,然后采用K近邻算法将点云分组,输入到多层感知机中将邻居点特征聚合到中心点,最后采用最大池化操作计算得到具有feature’维度特征的y个点(默认y=x/4,feature’=2*feature)。The specific process of feature aggregation is as follows: input x points with feature dimension, first randomly sample the farthest point, then use the K nearest neighbor algorithm to group the point cloud, input it into the multi-layer perceptron to aggregate the neighbor point features to the center point, and finally The maximum pooling operation is used to calculate y points with feature' dimension characteristics (default y=x/4, feature'=2*feature).
特征聚合层采用K近邻算法对输入点集进行采样分组。并且,为了节省训练时的显存空间,此层会将计算的K近邻矩阵输出,与后续并行分支共享。The feature aggregation layer uses the K nearest neighbor algorithm to sample and group the input point set. Moreover, in order to save video memory space during training, this layer will output the calculated K nearest neighbor matrix and share it with subsequent parallel branches.
(2)局部特征捕获分支(2) Local feature capture branch
该分支基于动态空间图卷积构建,用于提取输入的植物点云中的局部特征。首先,基于点集V和近邻信息E构建特征图G=(V,E),采用边缘卷积在输入的特征空间上进行特征提取。提取某点xi特征的公式如下:This branch is built based on dynamic spatial graph convolution and is used to extract local features in the input plant point cloud. First, a feature map G = (V, E) is constructed based on the point set V and neighbor information E, and edge convolution is used to extract features on the input feature space. The formula for extracting features of a certain point x i is as follows:
fi=?h(xi,yi)f i =? h(x i ,y i )
其中,xj代表点xi的邻居点之一,?和h分别代表某种聚合函数和某种关系操作。即采用一种关系操作将候选点周围的邻居点特征聚合,即可得到候选点的特征信息,该关系操作定义为边缘卷积。为了增强对点云中局部特征的理解,本研究采用最大池化作为聚合函数,具体过程如下:Among them, x j represents one of the neighbor points of point xi ,? and h represent some kind of aggregate function and some kind of relational operation respectively. That is, a relational operation is used to aggregate the features of neighbor points around the candidate point to obtain the feature information of the candidate point. This relational operation is defined as edge convolution. In order to enhance the understanding of local features in point clouds, this study uses max pooling as the aggregation function. The specific process is as follows:
convi=Max(MLP(h(xi,xi-xj)))conv i =Max(MLP(h( xi , xi -x j )))
关系操作h定义为点xi xi和其邻居点xj的特征差值和点xi输出值之间的线性组合。这种选择不仅保留了局部点集中彼此影响的特征,也部分考虑到整体的全局特征。The relational operation h is defined as the linear combination between the characteristic difference between point x i x i and its neighbor point x j and the output value of point x i . This choice not only retains the characteristics of local point sets that influence each other, but also partially takes into account the overall global characteristics.
(3)全局特征学习分支(3) Global feature learning branch
如图3所示,此分支基于点云注意力机制构建,非常适合处理点云数据,本质上,点云数据可以看作是嵌入在注意力空间的词向量,本研究采用局部邻域中的向量注意力机制来提取特征,计算公式如下:As shown in Figure 3, this branch is built based on the point cloud attention mechanism and is very suitable for processing point cloud data. In essence, point cloud data can be regarded as word vectors embedded in the attention space. This study uses Vector attention mechanism is used to extract features. The calculation formula is as follows:
其中,xj是点xi的K个邻居点之一,φ、α是点级别的特征变换方法,分别得到自注意力机制中的Q、K、V值,δ是位置编码函数,ρ是正则化函数,γ是映射函数,β是某种关系运算,在本研究中定义为邻域点与关注点的差值。根据以上注意力机制,本研究提出了点注意力层,改进后的计算公式如下:Among them, x j is one of the K neighbor points of point x i , φ, α is a point-level feature transformation method, which is obtained from the Q, K, and V values in the attention mechanism respectively. δ is the position encoding function, ρ is the regularization function, γ is the mapping function, and β is some kind of relational operation. In this paper In the study, it is defined as the difference between the neighborhood point and the attention point. Based on the above attention mechanism, this study proposes a point attention layer. The improved calculation formula is as follows:
与普通注意力机制不同的是,在α函数中也加入了位置编码,以此强化对特征的理解。在点注意力层的基础上,TRGCN编码器在全局特征学习分支构建了一个残差结构。在点注意力层前后各添加一个线性层,并将最终输出与输入进行残差连接,以此促进信息交换,加速网络收敛,并为训练深层网络提供可能性。Different from the ordinary attention mechanism, position coding is also added to the α function to enhance the understanding of features. Based on the point attention layer, the TRGCN encoder constructs a residual structure in the global feature learning branch. Add a linear layer before and after the point attention layer, and conduct residual connection between the final output and the input to promote information exchange, accelerate network convergence, and provide the possibility to train deep networks.
(4)T-G特征耦合层(4)T-G characteristic coupling layer
经过以上处理,可以得到两个维度以及形状完全相同的特征矩阵:具有显著局部特征的矩阵G和具有完整全局特征的矩阵T。将G和T拼接后输入特征耦合层得到目标特征矩阵:After the above processing, two feature matrices with exactly the same dimensions and shapes can be obtained: a matrix G with significant local features and a matrix T with complete global features. After splicing G and T, input into the feature coupling layer to obtain the target feature matrix:
TG=Linear(ReLU(Linear(T,G)))TG=Linear(ReLU(Linear(T,G)))
采用两个线性层和一个ReLU激活层设计了T-G特征耦合层,使网络可以学习两部分矩阵各自更重要的信息,并结合成为目标特征矩阵。The T-G feature coupling layer is designed using two linear layers and a ReLU activation layer, so that the network can learn the more important information of the two matrices and combine them into the target feature matrix.
综上所述,TRGCN网络特征编码器部分可以通过改变TRGCN块的堆叠数量,设计适应不同视觉任务的模型。较少的TRGCN块可以用于轻量级分类网络,而更多的级联TRGCN块可以用于更细粒度的任务,例如点云分割和目标识别。In summary, the feature encoder part of the TRGCN network can design models adapted to different visual tasks by changing the stacking number of TRGCN blocks. Fewer TRGCN blocks can be used for lightweight classification networks, while more cascaded TRGCN blocks can be used for finer-grained tasks such as point cloud segmentation and object recognition.
特征解码器部分同样堆叠三个级联的TRGCN块,并分别接收编码器中三个TRGCN块的输出,但用插值层替换特征聚合层。与特征聚合层相反的是解码器中的插值层将高维度点集的特征恢复到低维度点集中,但仍然输出K近邻算法的分组结果供TRGCN的两个分支计算所需。对于实例分割结果预测,解码器在TRGCN块后设置独立的插值层,并采用单个点注意力层来保证信息完整性,网络最后采用多层感知机输出点云的分割结果。The feature decoder part also stacks three cascaded TRGCN blocks and receives the outputs of the three TRGCN blocks in the encoder respectively, but replaces the feature aggregation layer with an interpolation layer. Contrary to the feature aggregation layer, the interpolation layer in the decoder restores the features of the high-dimensional point set to the low-dimensional point set, but still outputs the grouping results of the K nearest neighbor algorithm for the calculation of the two branches of TRGCN. For instance segmentation result prediction, the decoder sets an independent interpolation layer after the TRGCN block and uses a single point attention layer to ensure information integrity. The network finally uses a multi-layer perceptron to output the segmentation result of the point cloud.
网络训练。本研究所有实验均在独立服务器上进行,该服务器配备了12核心20线程的CPU、64GB内存和一个Nvidia GeForce RTX 3090Ti GPU。在训练阶段,5种植物点云分割模型采用相同的超参数,具体为:训练批次大小设置为32,初始学习率设置为0.001,使用Adam方法对网络进行优化,共训练100个周期,每20个周期将学习率减半,权重衰减设置为0.0001,动量设置为0.9,K近邻算法的K值设置为12,点注意力层的特征维度设置为256。Network training. All experiments in this study were conducted on a dedicated server equipped with a 12-core 20-thread CPU, 64GB memory, and an Nvidia GeForce RTX 3090Ti GPU. In the training phase, the five plant point cloud segmentation models use the same hyperparameters, specifically: the training batch size is set to 32, the initial learning rate is set to 0.001, the Adam method is used to optimize the network, and a total of 100 cycles are trained. The learning rate is halved for 20 epochs, the weight decay is set to 0.0001, the momentum is set to 0.9, the K value of the K nearest neighbor algorithm is set to 12, and the feature dimension of the point attention layer is set to 256.
在5种植物点云数据上进行器官实例分割测试,取得了最高86.38%的平均交并比和88.58%的平均准确率。为了验证TRGCN的分割能力,另外选择了三个主流的点云分割网络与TRGCN进行对比,在5种分割任务中,TRGCN有9项指标领先其他三种方法,在绝大部分分割任务中取得了最优的精度,尤其在高粱叶片上的精度提高更为明显,这表明TRGCN更擅长处理单子叶植物点云。由于双子叶作物的冠层结构较为拥挤,容易造成遮挡问题,烟草和番茄点云的分割效果并不如单子叶作物,但分割效果仍然好于其他三种分割网络。具体的测试结果如表1所示,图4是五种植物点云的分割效果图。The organ instance segmentation test was conducted on five kinds of plant point cloud data, and the highest average intersection and union ratio of 86.38% and an average accuracy of 88.58% were achieved. In order to verify the segmentation ability of TRGCN, three mainstream point cloud segmentation networks were selected for comparison with TRGCN. Among the 5 segmentation tasks, TRGCN has 9 indicators leading the other three methods, and has achieved the best results in most segmentation tasks. The optimal accuracy, especially the improvement in accuracy on sorghum leaves, is more obvious, which shows that TRGCN is better at processing monocot point clouds. Since the canopy structure of dicotyledonous crops is relatively crowded and can easily cause occlusion problems, the segmentation effect of tobacco and tomato point clouds is not as good as that of monocot crops, but the segmentation effect is still better than the other three segmentation networks. The specific test results are shown in Table 1. Figure 4 is a segmentation rendering of five plant point clouds.
本发明还采用高粱点云作为研究对象,对TRGCN池化层和级联TRGCN块的堆叠数量进行探讨,结果表明采用最大池化的网络分割性能最佳,准确率比平均池化和求和池化高约2%。在级联的TRGCN块数量为3的时候网络在训练时间和分割效果上达到最优,牺牲一定的时间得到了更高的分割精度。具体的测试结果如表2和3所示。This invention also uses sorghum point cloud as the research object to discuss the stacking number of TRGCN pooling layers and cascaded TRGCN blocks. The results show that the network segmentation performance using maximum pooling is the best, and the accuracy is better than average pooling and summation pooling. The chemical is about 2% higher. When the number of cascaded TRGCN blocks is 3, the network achieves optimal training time and segmentation effect, sacrificing a certain amount of time to obtain higher segmentation accuracy. The specific test results are shown in Tables 2 and 3.
表1为本发明TRGCN网络与其他主流网络的分割精度对比表Table 1 is a comparison table of segmentation accuracy between the TRGCN network of the present invention and other mainstream networks.
表2本发明消融实验1,不同池化层的分割效果表Table 2 Ablation experiment 1 of the present invention, segmentation effect table of different pooling layers
表3本发明消融实验2,不同TRGCN块堆叠数量的分割效果表。Table 3: Ablation experiment 2 of the present invention, segmentation effect table with different stacking numbers of TRGCN blocks.
Claims (2)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310704110.5A CN117036370A (en) | 2023-06-14 | 2023-06-14 | Plant organ point cloud segmentation method based on attention mechanism and graph convolution |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310704110.5A CN117036370A (en) | 2023-06-14 | 2023-06-14 | Plant organ point cloud segmentation method based on attention mechanism and graph convolution |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117036370A true CN117036370A (en) | 2023-11-10 |
Family
ID=88625100
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310704110.5A Pending CN117036370A (en) | 2023-06-14 | 2023-06-14 | Plant organ point cloud segmentation method based on attention mechanism and graph convolution |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117036370A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117455929A (en) * | 2023-12-26 | 2024-01-26 | 福建理工大学 | Teeth segmentation method and terminal based on dual-stream self-attention graph convolution network |
CN117726822A (en) * | 2024-02-18 | 2024-03-19 | 安徽大学 | Three-dimensional medical image classification segmentation system and method based on double-branch feature fusion |
CN117745148A (en) * | 2024-02-10 | 2024-03-22 | 安徽省农业科学院烟草研究所 | Multi-source data-based rice stubble flue-cured tobacco planting quality evaluation method and system |
CN118632027A (en) * | 2024-08-08 | 2024-09-10 | 华侨大学 | A point cloud compression method based on graph convolutional network |
-
2023
- 2023-06-14 CN CN202310704110.5A patent/CN117036370A/en active Pending
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117455929A (en) * | 2023-12-26 | 2024-01-26 | 福建理工大学 | Teeth segmentation method and terminal based on dual-stream self-attention graph convolution network |
CN117455929B (en) * | 2023-12-26 | 2024-03-15 | 福建理工大学 | Teeth segmentation method and terminal based on dual-stream self-attention graph convolution network |
CN117745148A (en) * | 2024-02-10 | 2024-03-22 | 安徽省农业科学院烟草研究所 | Multi-source data-based rice stubble flue-cured tobacco planting quality evaluation method and system |
CN117745148B (en) * | 2024-02-10 | 2024-05-10 | 安徽省农业科学院烟草研究所 | Multi-source data-based rice stubble flue-cured tobacco planting quality evaluation method and system |
CN117726822A (en) * | 2024-02-18 | 2024-03-19 | 安徽大学 | Three-dimensional medical image classification segmentation system and method based on double-branch feature fusion |
CN117726822B (en) * | 2024-02-18 | 2024-05-03 | 安徽大学 | Three-dimensional medical image classification and segmentation system and method based on dual-branch feature fusion |
CN118632027A (en) * | 2024-08-08 | 2024-09-10 | 华侨大学 | A point cloud compression method based on graph convolutional network |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN117036370A (en) | Plant organ point cloud segmentation method based on attention mechanism and graph convolution | |
Shen et al. | ICAFusion: Iterative cross-attention guided feature fusion for multispectral object detection | |
CN109948691B (en) | Image description generation method and device based on depth residual error network and attention | |
CN108960212A (en) | Based on the detection of human joint points end to end and classification method | |
CN110458085B (en) | Video behavior identification method based on attention-enhanced three-dimensional space-time representation learning | |
CN112651329B (en) | Low-resolution ship classification method for generating countermeasure network through double-flow feature learning | |
CN115908772A (en) | Target detection method and system based on Transformer and fusion attention mechanism | |
CN117218351A (en) | Three-dimensional point cloud semantic segmentation method based on local and global context awareness | |
Zhang et al. | Recognition of bird nests on power transmission lines in aerial images based on improved YOLOv4 | |
Cheng et al. | A survey on image semantic segmentation using deep learning techniques | |
CN116630387A (en) | Monocular Image Depth Estimation Method Based on Attention Mechanism | |
Wang et al. | Hierarchical kernel interaction network for remote sensing object counting | |
CN114898157B (en) | Global learning device and method for hyperspectral image classification | |
Zhao et al. | PCUNet: A context-aware deep network for coarse-to-fine point cloud completion | |
Song et al. | Lightweight multi-level feature difference fusion network for RGB-DT salient object detection | |
CN115330703A (en) | Cloud and cloud shadow detection method in remote sensing images based on context information fusion | |
Zhao et al. | Recurrent adaptive graph reasoning network with region and boundary interaction for salient object detection in optical remote sensing images | |
Jiang et al. | An improved action recognition network with temporal extraction and feature enhancement | |
CN117994623A (en) | Image feature vector acquisition method | |
Wang et al. | Image target recognition based on improved convolutional neural network | |
CN114494284B (en) | Scene analysis model and method based on explicit supervision area relation | |
Shah et al. | A review of deep learning models for computer vision | |
Lin et al. | Patchmixing masked autoencoders for 3d point cloud self-supervised learning | |
CN117422914A (en) | Target perception method for multi-task learning | |
Gong et al. | Multi-feature fusion network for facial expression recognition in the wild |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |