CN115731137A - Outdoor large scene point cloud segmentation method based on A-EdgeConv - Google Patents

Outdoor large scene point cloud segmentation method based on A-EdgeConv Download PDF

Info

Publication number
CN115731137A
CN115731137A CN202211513938.4A CN202211513938A CN115731137A CN 115731137 A CN115731137 A CN 115731137A CN 202211513938 A CN202211513938 A CN 202211513938A CN 115731137 A CN115731137 A CN 115731137A
Authority
CN
China
Prior art keywords
point
super
points
characteristic
point cloud
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202211513938.4A
Other languages
Chinese (zh)
Other versions
CN115731137B (en
Inventor
张良
廉飞宇
丁航
时文博
靳于康
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hubei University
Original Assignee
Hubei University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hubei University filed Critical Hubei University
Priority to CN202211513938.4A priority Critical patent/CN115731137B/en
Publication of CN115731137A publication Critical patent/CN115731137A/en
Application granted granted Critical
Publication of CN115731137B publication Critical patent/CN115731137B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Image Analysis (AREA)

Abstract

The invention belongs to the technical field of point cloud image processing, and particularly relates to an outdoor large-scene point cloud segmentation method based on A-EdgeConv, which combines local geometric information and a graph segmentation algorithm to achieve the acquisition of a super-point, adopts a local adjacency graph and an attention mechanism to more accurately extract a super-point characteristic, and achieves the characteristic aggregation of the super-point and the super-edge through a GRU (generalized regression) cyclic neural network.

Description

Outdoor large scene point cloud segmentation method based on A-EdgeConv
Technical Field
The invention belongs to the technical field of point cloud image processing, and particularly relates to an outdoor large scene point cloud segmentation method based on A-EdgeConv.
Background
In recent years, laser scanning technology is rapidly developed, becomes an important way for rapidly acquiring three-dimensional information, and has wide application in the fields of automatic driving, smart city construction and the like. The efficient and accurate semantic segmentation of the point cloud is a precondition of the application, and the accurate extraction of the ground object target in an outdoor large scene with huge data volume and uneven scene distribution is always a challenging problem.
The information of the valid features is a direct factor in obtaining the correct semantic label. In a large outdoor scene, the requirement on detail characteristic information is more urgent due to the complex relation between the ground class and the space, and how to quickly acquire the effective information of the object becomes a research hotspot. Aiming at the extraction of object feature information, most researches adopt a PointNet model to extract point cloud features, low-dimensional information is projected to a high-dimensional space by using a shared multilayer perceptron, and the point cloud features are extracted by combining a maximum pooling function. The information extraction mode does not consider the point cloud space relation in the neighborhood, weakens the feature expression of different ground objects in a complex environment, and loses a large amount of effective information.
Disclosure of Invention
In view of the above problems, the present invention provides an a-EdgeConv-based outdoor large scene point cloud segmentation method, which overcomes or at least partially solves the above problems, and can efficiently and accurately segment the outdoor large scene point cloud.
Specifically, the invention provides an outdoor large scene point cloud segmentation method based on A-EdgeConv, which comprises the following steps:
step 1, describing local information of a point cloud by using local geometric features;
step 2, carrying out geometric homogeneity division on the point cloud in a graph cutting mode to obtain a super point;
step 3, constructing a local adjacency graph based on the over points obtained in the step 2, describing local characteristics, and highlighting key information through an attention mechanism algorithm;
step 4, constructing a super-point diagram by utilizing the adjacency relation of the super-points to acquire super-edge characteristics;
and 5, combining the GRU gating cycle unit and the GNN graph neural network model to aggregate the characteristics of the super point and the super edge, and realizing accurate semantic segmentation of the point cloud.
Optionally, the step 1 includes:
step 1.1, preprocessing outdoor large scene point cloud and removing noise points;
step 1.2, selecting an optimal neighborhood by minimizing the characteristic entropy;
step 1.3, combining each optimal neighborhood of the point cloud, obtaining a characteristic value by calculating a covariance matrix of adjacent points, and calculating linear L according to the covariance characteristic value λ Flatness P λ Degree of scattering S λ All-round difference O λ Change of curvature C λ Five characteristic indexes.
Optionally, in step 1.2, the method for calculating the minimum characteristic entropy is as shown in formula (1):
E λ =-λ 1 ln(λ 1 )-λ 2 ln(λ 2 )-λ 3 ln(λ 3 ) (1)
wherein λ is 1 、λ 2 、λ 3 Respectively representing the covariance eigenvalues of the neighborhood of the target point.
Optionally, in step 1.3, L is linear λ Flatness P λ Degree of scattering S λ All-round difference O λ Change of curvature C λ The calculation methods of the five characteristic indexes are shown in formulas (2) to (6):
Figure BDA0003970103640000021
Figure BDA0003970103640000022
Figure BDA0003970103640000023
Figure BDA0003970103640000024
Figure BDA0003970103640000025
optionally, the step 2 includes:
step 2.1, constructing a minimum energy function according to the point cloud local information described in the step 1;
and 2.2, solving the problem of minimizing energy by adopting a one-cut algorithm, adaptively adjusting the size of the divided segments according to the local geometric complexity, and realizing the acquisition of the over point through graph cutting.
Optionally, in step 2.1, the energy minimization function is as shown in equation (7):
E(G)=∑ p∈P D p (G p )+λ∑ (p,q)∈E V p,q (G p ,G q ) (7)
wherein D is p (G p ) Is a data cost term, which is the sum of all points and their penalties for assigning tag numbers, D p (G p ) Indicates a label G p Penalty in assigning to node p; sigma (p,q)∈E V p,q (G p ,G q ) For the smoothing cost term, is the sum of penalties for all neighboring labels to be different, V p,q (G p ,G q ) The penalty of different label numbers between adjacent nodes p and q is represented; lambda is a weight coefficient, and the numerical value reflects the proportion of a data cost item and a smooth cost item in graph cutting; suppose that the feature of point p is denoted f p Which is associated with a label L p Data cost D of p (G p ) The calculation is shown in equation (8):
D p (G p )=(g p -f p ) 2 (8)
wherein, g p Indicates a tag number G p Point cloud characteristics corresponding to the segmented object;
the smoothing cost is used to penalize the inconsistency of label numbers between adjacent nodes p and q, and generally speaking, if the distance between the nodes p and q is smaller, the smoothing cost V between the nodes p and q is smaller p,q (G p ,G q ) The larger the probability that the node p and q label numbers are the same; smoothing cost V between adjacent nodes p,q (G p ,G q ) The calculation formula is shown in formulas (9) to (10):
V p,q (G p ,G q )=exp(-ds) (9)
Figure BDA0003970103640000031
wherein f is p 、f q Respectively representing the geometric features of the adjacent point clouds p and q.
Optionally, the step 3 includes:
step 3.1, constructing a local adjacency graph in the super-point by taking the super-point obtained in the step 2 as a basis;
step 3.2, randomly selecting a clustering initial point x from the three-dimensional point cloud data; finding a neighborhood feature point set M of an initial point x, assuming that there are n neighboring points, defined as: x = X 1 ,X 2 ,…X n Performing convolution operation on the obtained data;
step 3.3, for n neighborhood points X in the neighborhood range of the point cloud data X in the super point ij1 ,X ij2 ,…,X ijn The neighborhood point feature vector is expressed as:
Figure BDA0003970103640000041
a unique attention score is learned for the feature vector of each neighborhood point by a sharing function g.
Step 3.4, when the neighborhood relationship also exists between the points q and p in the point set M in space, grouping q and p into one class; weight coefficient
Figure BDA0003970103640000042
Expressing the importance of each feature vector, and performing weighted summation on each feature vector to obtain the final rich feature vector representing the over-point information
Figure BDA0003970103640000043
Step 3.5, performing the operation on all the over points in the point set to obtain all over point characteristics;
and 3.6, overlapping a plurality of layers of A-EdgeConv modules to realize layer-by-layer extraction and summarization of local information, and realizing accurate description of the over point by overlapping different characteristic spatial information.
Optionally, in step 3.3, the definition of the sharing function g is shown in formula (11):
Figure BDA0003970103640000044
wherein, W is the weight learnable by the multi-layer perceptron,
Figure BDA0003970103640000045
as feature vectors
Figure BDA0003970103640000046
And obtaining the weight coefficient through the sharing function.
Optionally, in the step 3.4, the feature vector
Figure BDA0003970103640000047
The calculation is shown in equation (12):
Figure BDA0003970103640000048
optionally, the step 4 includes:
step 4.1, constructing a super-point map based on the super-points obtained in step 2, and assuming that adjacent super-points S 'and T' exist, if one or more edges formed by three-dimensional points in the two super-points S 'and T' belong to E, the offset set between the super-points is as shown in formula (13):
δ(S′,T′)={(p i -p j )∣(i,j)∈E∩(S′×T′)} (13)
in the formula, p i 、p j Respectively representing point clouds on two sides of an edge formed by adjacent super points;
step 4.2, the super-edge characteristics are obtained according to the relation of the super-point graph, and the characteristic value decomposition is carried out on the three-dimensional point set in the adjacent super-points to obtain the characteristic value lambda 1 ≥λ 2 ≥λ 3 ≥0;
And 4.3, obtaining the one-dimensional and/or three-dimensional super-edge characteristics through different combinations of the characteristic values.
Optionally, the step 5 includes:
step 5.1, embedding the over-point characteristic of the step 3 and the over-edge characteristic of the step 4 into a circulation unit as initial states;
step 5.2, inputting the over-point characteristic and the over-edge characteristic as initial values h of the hidden state 0 And input x 0 The method determines whether the candidate state at the current time needs to depend on the network state at the last time and how much the candidate state needs to depend on by resetting the gate Wr, and determines h in the state for controlling the output at the current time by the updating gate Wz t How many historical states h to keep t-1 And how many candidate states at the current time are retained
Figure BDA0003970103640000051
Continuously updating the input hidden state h at the next moment t
And 5.3, matching the aggregated features with the point labels, and optimizing by minimizing the residual error between the synthesis and the real labels in the training process to realize accurate segmentation of the point cloud.
The invention has the beneficial effects that:
1. according to the method, the point cloud is segmented in an object-oriented mode, so that the dependence on an optimal neighborhood is reduced to a certain extent, the influence of noise in the point cloud on feature calculation is reduced, the calculation of geometric features is facilitated, and the uncertainty of information extraction is reduced; secondly, context information between some new characteristic attributes derived by the object and the contained point clouds is beneficial to improving the classification precision, and good effect is shown in a complex outdoor scene;
2. the A-EdgeConv (Attention-EdgeConv) -based outdoor large scene point cloud segmentation method combines the local neighborhood characteristics of the hyper-points to realize the extraction of context information between point clouds, provides more local details and provides more information for the outdoor scene point cloud segmentation. Meanwhile, the attention mechanism module can focus on key points more effectively, the problem that key feature information is lost due to the fact that a maximum/average pool is used for hard integration of adjacent features is avoided, the problems that the complexity of model training is increased due to the fact that the key feature information is lost, the feature information is difficult to extract fully and the like are solved, and the segmentation precision is improved.
The above and other objects, advantages and features of the present invention will become more apparent to those skilled in the art from the following detailed description of specific embodiments thereof, taken in conjunction with the accompanying drawings.
Drawings
Some specific embodiments of the invention will be described in detail hereinafter, by way of illustration and not limitation, with reference to the accompanying drawings. The same reference numbers in the drawings identify the same or similar elements or components. Those skilled in the art will appreciate that the drawings are not necessarily drawn to scale. In the drawings:
FIG. 1 is a flow chart of the method for segmenting the outdoor large scene point cloud based on A-EdgeConv according to the present invention;
fig. 2 is a schematic diagram of step 3 of the a-EdgeConv-based outdoor large scene point cloud segmentation method according to the embodiment of the present invention;
FIG. 3 is a schematic diagram of the A-EdgeConv module of the present invention;
FIG. 4 is a diagram of the result of semantic segmentation of the outdoor large scene according to the present invention.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings. The embodiments described below with reference to the accompanying drawings are illustrative and intended to explain the present invention and should not be construed as limiting the present invention.
The following describes an outdoor large scene point cloud segmentation method based on the a-EdgeConv according to the embodiment of the invention with reference to fig. 1 to 4.
As shown in fig. 1, an embodiment of the present invention provides an outdoor large scene point cloud segmentation method based on a-EdgeConv, including the following steps:
step 1, describing local information of a point cloud by using local geometric features;
step 1.1, preprocessing outdoor large scene point cloud and removing noise points;
step 1.2, selecting an optimal neighborhood by minimizing the characteristic entropy;
step 1.3, combining each optimal neighborhood of the point cloud, obtaining an eigenvalue by calculating a covariance matrix of adjacent points, and calculating linear L according to the covariance eigenvalue λ Flatness P λ Degree of scattering S λ All-round difference O λ Change of curvature C λ Five characteristic indexes;
step 2, carrying out geometric homogeneity division on the point cloud in a graph cutting mode to obtain a super point;
step 2.1, constructing a minimum energy function according to the point cloud local information described in the step 1;
step 2.2, solving the problem of minimizing energy by adopting a one-cut algorithm, adaptively adjusting the size of a divided segment according to the local geometric complexity, and realizing the acquisition of the over point by graph cutting;
step 3, as shown in fig. 2, constructing a local adjacency graph based on the over points obtained in the step 2, describing local characteristics, and highlighting key information through an attention mechanism algorithm;
step 3.1, constructing a local adjacency graph inside the super-point by taking the super-point obtained in the step 2 as a basis;
step 3.2, randomly selecting a clustering initial point x from the three-dimensional point cloud data; finding a neighborhood feature point set M of an initial point x, assuming that there are n neighboring points, defined as: x = X 1 ,X 2 ,…X n Performing convolution operation on the obtained data;
step 3.3, for over-point interiorN neighborhood points X in neighborhood range of partial point cloud data X ij1 ,X ij2 ,…,X ijn The neighborhood point feature vector is expressed as:
Figure BDA0003970103640000071
a unique attention score is learned for the feature vector of each neighborhood point by a sharing function g.
Step 3.4, when the point q and the point p in the point set M also have a neighborhood relationship in space, the q and the p are gathered into one type; weight coefficient
Figure BDA0003970103640000072
Expressing the importance of each feature vector, and performing weighted summation on each feature vector to obtain the final rich feature vector representing the over-point information
Figure BDA0003970103640000073
Step 3.5, performing the operation on all the over points in the point set to obtain all over point characteristics;
step 3.6, overlapping a plurality of layers of A-EdgeConv modules to realize layer-by-layer extraction and summarization of local information, and realizing accurate description of the over point by overlapping different characteristic space information;
step 4, constructing a super-point diagram by utilizing the adjacency relation of the super-points to acquire super-edge characteristics;
step 4.1, constructing a hyper-point map based on the hyper-points obtained in step 2, assuming that there are adjacent hyper-points S 'and T', and if there is one or more edges formed by three-dimensional points in two hyper-points S 'and T' belonging to E, the offset set δ (S ', T') between the hyper-points is as shown in formula (13):
δ(S′,T′)={(p i -p j )∣(i,j)∈E∩(S′×T′)} (13)
in the formula, p i 、p j Respectively representing point clouds on two sides of an edge formed by adjacent super points;
step 4.2, the super-edge characteristics are obtained according to the relation of the super-point graph, and the characteristic value decomposition is carried out on the three-dimensional point set in the adjacent super-points to obtain the characteristic valueλ 1 ≥λ 2 ≥λ 3 ≥0;
4.3, obtaining one-dimensional and/or three-dimensional super-edge characteristics through different combinations of characteristic values;
step 5, combining a GRU gating cycle unit and a GNN graph neural network model to aggregate the characteristics of the super point and the super edge to realize accurate semantic segmentation of the point cloud;
step 5.1, embedding the over-point characteristic in the step 3 and the over-edge characteristic in the step 4 into a circulation unit as initial states;
step 5.2, inputting the over-point characteristic and the over-edge characteristic as initial values h of the hidden state 0 And input x 0 The reset gate Wr determines whether the candidate state at the current moment needs to depend on the network state at the last moment and how much the candidate state needs to depend on, and the update gate Wz determines h in the state for controlling the output at the current moment t How many historical states h to keep t-1 And how many candidate states at the current time are retained
Figure BDA0003970103640000081
Continuously updating the input hidden state h at the next moment t As shown in fig. 3;
and 5.3, matching the aggregated features with the point labels, and optimizing through minimizing the residual error between the synthesis and the real labels in the training process to realize accurate segmentation of the point cloud, wherein the result of the semantic segmentation of the outdoor large scene is shown in FIG. 4.
Another embodiment of the present invention provides an outdoor large scene point cloud segmentation method based on a-EdgeConv, including the following steps:
step 1, describing local information of point cloud by using local geometric features;
step 1.1, preprocessing outdoor large scene point cloud and removing noise points;
step 1.2, selecting the optimal neighborhood through the minimized characteristic entropy, wherein the calculation method of the minimized characteristic entropy is shown as a formula (1):
E λ =-λ 1 ln(λ 1 )-λ 2 ln(λ 2 )-λ 3 ln(λ 3 ) (1)
wherein λ is 1 、λ 2 、λ 3 Respectively representing the covariance eigenvalues of the neighborhood of the target point.
Step 1.3, combining each optimal neighborhood of the point cloud, obtaining an eigenvalue by calculating a covariance matrix of adjacent points, and calculating linear L according to the covariance eigenvalue λ Flatness P λ Degree of scattering S λ All formula (O) λ Change of curvature C λ Five characteristic indexes, linearity L λ Flatness P λ Degree of scattering S λ All formula (O) λ Change of curvature C λ The calculation methods of the five characteristic indexes are shown in formulas (2) to (6):
Figure BDA0003970103640000091
Figure BDA0003970103640000092
Figure BDA0003970103640000093
Figure BDA0003970103640000094
Figure BDA0003970103640000095
step 2, carrying out geometric homogeneity division on the point cloud in a graph cutting mode to obtain a super point;
step 2.1, constructing a minimum energy function according to the point cloud local information described in the step 1;
step 2.2, solving the problem of minimizing energy by adopting a one-cut algorithm, adaptively adjusting the size of a division segment according to the local geometric complexity, and realizing the acquisition of the over point by graph cutting;
step 3, as shown in fig. 2, constructing a local adjacency graph based on the over points obtained in the step 2, describing local features, and highlighting key information through an attention mechanism algorithm;
step 3.1, constructing a local adjacency graph inside the super-point by taking the super-point obtained in the step 2 as a basis;
step 3.2, randomly selecting a clustering initial point x from the three-dimensional point cloud data; finding a neighborhood feature point set M of an initial point x, assuming that there are n neighboring points, defined as: x = X 1 ,X 2 ,…X n Performing convolution operation on the obtained data;
step 3.3, for n neighborhood points X in the neighborhood range of the point cloud data X in the super point ij1 ,X ij2 ,…,X ijn The neighborhood point feature vector is expressed as:
Figure BDA0003970103640000096
learning a unique attention score for the feature vector of each neighborhood point by a sharing function g, which is defined as formula (11):
Figure BDA0003970103640000097
wherein W is the weight learnable by the multi-layer perceptron,
Figure BDA0003970103640000098
as feature vectors
Figure BDA0003970103640000099
And obtaining the weight coefficient through the sharing function.
Step 3.4, when the point q and the point p in the point set M also have a neighborhood relationship in space, the q and the p are gathered into one type; weight coefficient
Figure BDA0003970103640000101
Expressing the importance of each feature vector, and performing weighted summation on each feature vector to obtainFinal rich feature vector characterizing the hyper-point information
Figure BDA0003970103640000102
Feature vector
Figure BDA0003970103640000103
The calculation is shown in equation (12):
Figure BDA0003970103640000104
step 3.5, performing the operation on all the over points in the point set to obtain all over point characteristics;
step 3.6, overlapping a plurality of layers of A-EdgeConv modules to realize layer-by-layer extraction and summarization of local information, and realizing accurate description of the over point by overlapping different characteristic space information;
step 4, constructing a super-point diagram by utilizing the adjacency relation of the super-points to acquire super-edge characteristics;
step 4.1, constructing a hyper-point map based on the hyper-points obtained in step 2, assuming that there are adjacent hyper-points S 'and T', and if there is one or more edges formed by three-dimensional points in two hyper-points S 'and T' belonging to E, the offset set δ (S ', T') between the hyper-points is as shown in formula (13):
δ(S′,T′)={(p i -p j )∣(i,j)∈E∩(S′×T′)} (13)
in the formula, p i 、p j Respectively representing point clouds on two sides of an edge formed by adjacent super points;
step 4.2, the super-edge characteristics are obtained according to the relation of the super-point graph, and the characteristic value decomposition is carried out on the three-dimensional point set in the adjacent super-points to obtain the characteristic value lambda 1 ≥λ 2 ≥λ 3 ≥0;
4.3, obtaining one-dimensional and/or three-dimensional super-edge characteristics through different combinations of characteristic values;
step 5, combining a GRU gating cycle unit and a GNN graph neural network model to aggregate the characteristics of the super point and the super edge, and realizing accurate semantic segmentation of the point cloud;
step 5.1, embedding the over-point characteristic of the step 3 and the over-edge characteristic of the step 4 into a circulation unit as initial states;
step 5.2, inputting the over-point characteristic and the over-edge characteristic as initial values h of the hidden state 0 And input x 0 The reset gate Wr determines whether the candidate state at the current moment needs to depend on the network state at the last moment and how much the candidate state needs to depend on, and the update gate Wz determines h in the state for controlling the output at the current moment t How many historical states h to keep t-1 And how many candidate states at the current time are retained
Figure BDA0003970103640000105
Continuously updating the input hidden state h at the next moment t As shown in fig. 3;
and 5.3, matching the aggregated features with the point labels, and optimizing through minimizing the residual error between the synthesis and the real labels in the training process to realize accurate segmentation of the point cloud, wherein the result of the semantic segmentation of the outdoor large scene is shown in FIG. 4.
Another embodiment of the present invention provides an outdoor large scene point cloud segmentation method based on a-EdgeConv, including the following steps:
step 1, describing local information of a point cloud by using local geometric features;
step 1.1, preprocessing outdoor large scene point cloud and removing noise points;
step 1.2, selecting the optimal neighborhood through the minimized characteristic entropy, wherein the calculation method of the minimized characteristic entropy is shown as a formula (1):
E λ =-λ 1 ln(λ 1 )-λ 2 ln(λ 2 )-λ 3 ln(λ 3 ) (1)
wherein λ is 1 、λ 2 、λ 3 Respectively representing covariance eigenvalues of the neighborhood of the target point.
Step 1.3, combining each optimal neighborhood of the point cloud, obtaining an eigenvalue by calculating a covariance matrix of adjacent points, and calculating linear L according to the covariance eigenvalue λ Flatness P λ Degree of scattering S λ All formula (O) λ Change in curvature C λ Five characteristic indexes, linearity L λ Flatness P λ Degree of scattering S λ All-round difference O λ Change of curvature C λ The calculation methods of the five characteristic indexes are shown in formulas (2) to (6):
Figure BDA0003970103640000111
Figure BDA0003970103640000112
Figure BDA0003970103640000113
Figure BDA0003970103640000114
Figure BDA0003970103640000115
step 2, carrying out geometric homogeneity division on the point cloud in a graph cutting mode to obtain a super point;
step 2.1, constructing a minimum energy function according to the point cloud local information described in the step 1; the minimum energy function is shown in equation (7):
E(G)=∑ p∈P D p (G p )+λ∑ (p,q)∈E V p,q (G p ,G q ) (7)
wherein D is p (G p ) Is a data cost term which is the sum of all points and their penalties for assigning tag numbers, D p (G p ) Indicates the label G p Penalty when assigning to node p; sigma (p,q)∈E V p,q (G p ,G q ) To be smoothedCost term is the sum of penalties for all neighboring tags being different, V p,q (G p ,G q ) Represents the penalty of different label numbers between adjacent nodes p and q; lambda is a weight coefficient, and the numerical value reflects the proportion of the data cost item and the smooth cost item in graph cut; let the characteristic of the point p be denoted f p Which is associated with a label L p Data cost D of p (G p ) The calculation is shown in equation (8):
D p (G p )=(g p -f p ) 2 (8)
wherein, g p Indicates a label number G p Point cloud characteristics corresponding to the segmented object;
the smoothing cost is used to penalize the inconsistency of label numbers between adjacent nodes p and q, and generally speaking, if the distance between the nodes p and q is smaller, the smoothing cost V between the nodes p and q is smaller p,q (G p ,G q ) The larger the probability that the node p and q label numbers are the same; smoothing cost V between adjacent nodes p,q (G p ,G q ) The calculation formula is shown in formulas (9) to (10):
V p,q (G p ,G q )=exp(-ds) (9)
Figure BDA0003970103640000121
wherein f is p 、f q Respectively representing the geometric features of the adjacent point clouds p and q.
Step 2.2, solving the problem of minimizing energy by adopting a one-cut algorithm, adaptively adjusting the size of a division segment according to the local geometric complexity, and realizing the acquisition of the over point by graph cutting;
step 3, as shown in fig. 2, constructing a local adjacency graph based on the over points obtained in the step 2, describing local features, and highlighting key information through an attention mechanism algorithm;
step 3.1, constructing a local adjacency graph inside the super-point by taking the super-point obtained in the step 2 as a basis;
step 3.2, random three-dimensionalSelecting a clustering initial point x from the point cloud data; finding a neighborhood feature point set M of an initial point x, assuming that there are n neighboring points, defined as: x = X 1 ,X 2 ,…X n Performing convolution operation on the data;
step 3.3, for n neighborhood points X in the neighborhood range of the point cloud data X in the super point ij1 ,X ij2 ,…,X ijn The neighborhood point feature vector is expressed as:
Figure BDA0003970103640000131
learning a unique attention score for the feature vector of each neighborhood point by a sharing function g, which is defined as formula (11):
Figure BDA0003970103640000132
wherein, W is the weight learnable by the multi-layer perceptron,
Figure BDA0003970103640000133
as feature vectors
Figure BDA0003970103640000134
And obtaining the weight coefficient through the sharing function.
Step 3.4, when the neighborhood relationship also exists between the points q and p in the point set M in space, grouping q and p into one class; weight coefficient
Figure BDA0003970103640000135
Expressing the importance of each feature vector, and performing weighted summation on each feature vector to obtain the final rich feature vector representing the over-point information
Figure BDA0003970103640000136
Feature vector
Figure BDA0003970103640000137
The calculation is shown in equation (12):
Figure BDA0003970103640000138
step 3.5, performing the operation on all the over points in the point set to obtain all over point characteristics;
step 3.6, overlapping a plurality of layers of A-EdgeConv modules to realize layer-by-layer extraction and summarization of local information, and realizing accurate description of the over point by overlapping different characteristic space information;
step 4, constructing a super-point diagram by utilizing the adjacency relation of the super-points to acquire super-edge characteristics;
step 4.1, constructing a hyper-point map based on the hyper-points obtained in step 2, assuming that there are adjacent hyper-points S 'and T', and if there is one or more edges formed by three-dimensional points in two hyper-points S 'and T' belonging to E, the offset set δ (S ', T') between the hyper-points is as shown in formula (13):
δ(S′,T′)={(p i -p j )∣(i,j)∈E∩(S′×T′)} (13)
in the formula, p i 、p j Respectively representing point clouds on two sides of an edge formed by adjacent super points;
step 4.2, the super-edge characteristics are obtained according to the relation of the super-point graph, and the characteristic value decomposition is carried out on the three-dimensional point set in the adjacent super-points to obtain the characteristic value lambda 1 ≥λ 2 ≥λ 3 ≥0;
And 4.3, obtaining one-dimensional and/or three-dimensional super-edge characteristics through different combinations of the characteristic values, wherein the super-edge characteristics are shown in a table 1:
TABLE 1 superceding characteristics
Figure BDA0003970103640000141
Step 5, combining a GRU gating cycle unit and a GNN graph neural network model to aggregate the characteristics of the super point and the super edge, and realizing accurate semantic segmentation of the point cloud;
step 5.1, embedding the over-point characteristic of the step 3 and the over-edge characteristic of the step 4 into a circulation unit as initial states;
step 5.2, inputting the over-point characteristic and the over-edge characteristic as the initial value h of the hidden state 0 And input x 0 The reset gate Wr determines whether the candidate state at the current moment needs to depend on the network state at the last moment and how much the candidate state needs to depend on, and the update gate Wz determines h in the state for controlling the output at the current moment t How many historical states h to keep t-1 And how many candidate states at the current time are retained
Figure BDA0003970103640000142
Continuously updating the input hidden state h at the next moment t As shown in fig. 3;
and 5.3, matching the aggregated features with the point labels, and optimizing through minimizing the residual error between the synthesis and the real labels in the training process to realize accurate segmentation of the point cloud, wherein the result of the semantic segmentation of the outdoor large scene is shown in FIG. 4.
According to the method, local geometric information and a graph cut algorithm are combined to achieve the acquisition of the super point, the local adjacent graph and an attention mechanism are combined to extract the super point characteristics more accurately, the characteristic aggregation of the super point and the super edge is achieved through a GRU (generalized regression analysis unit) cyclic neural network, the point cloud segmentation speed in a large scene is greatly improved, the characteristic expression is improved due to the combination of the local adjacent graph and the attention mechanism, the effective information and the point cloud local structure are concerned, the segmentation precision is improved, and the optimization complexity is reduced.
In the description of the present invention, furthermore, the terms "first", "second" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implying any number of technical features indicated. Thus, a feature defined as "first" or "second" may explicitly or implicitly include at least one such feature. In the description of the present invention, "a plurality" means at least two, e.g., two, three, etc., unless specifically limited otherwise.
In the present invention, unless otherwise expressly stated or limited, the terms "mounted," "connected," "secured," and the like are to be construed broadly and can, for example, be fixedly connected, detachably connected, or integrally formed; may be mechanically coupled, may be electrically coupled or may be in communication with each other; they may be directly connected or indirectly connected through intervening media, or they may be connected internally or in any other suitable relationship, unless expressly stated otherwise. The specific meanings of the above terms in the present invention can be understood according to specific situations by those of ordinary skill in the art.
In the present disclosure, the terms "one embodiment," "some embodiments," "an example," "a specific example," or "some examples" and the like mean that a specific feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present disclosure. In this specification, the schematic representations of the terms used above are not necessarily intended to refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Moreover, various embodiments or examples and features of various embodiments or examples described in this specification can be combined and combined by one skilled in the art without being mutually inconsistent.
Thus, it should be appreciated by those skilled in the art that while a number of exemplary embodiments of the invention have been illustrated and described in detail herein, many other variations or modifications consistent with the principles of the invention may be directly determined or derived from the disclosure of the present invention without departing from the spirit and scope of the invention. Accordingly, the scope of the invention should be understood and interpreted to cover all such other variations or modifications.

Claims (10)

1. An outdoor large scene point cloud segmentation method based on A-EdgeConv is characterized by comprising the following steps:
step 1, describing local information of a point cloud by using local geometric features;
step 2, carrying out geometric homogeneity division on the point cloud in a graph cutting mode to obtain a super point;
step 3, constructing a local adjacency graph based on the super points obtained in the step 2, describing local characteristics, and highlighting key information through an attention mechanism algorithm;
step 4, constructing a super-point diagram by utilizing the adjacency relation of the super-points to acquire super-edge characteristics;
and 5, combining the GRU gating cycle unit and the GNN graph neural network model to aggregate the characteristics of the super point and the super edge, and realizing accurate semantic segmentation of the point cloud.
2. The segmentation method according to claim 1, wherein the step 1 comprises:
step 1.1, preprocessing outdoor large scene point cloud and removing noise points;
step 1.2, selecting an optimal neighborhood by minimizing the characteristic entropy;
step 1.3, combining each optimal neighborhood of the point cloud, obtaining an eigenvalue by calculating a covariance matrix of adjacent points, and calculating linear L according to the covariance eigenvalue λ Flatness P λ Degree of scattering S λ All formula (O) λ Change of curvature C λ Five characteristic indexes.
3. Segmentation method according to claim 2, characterized in that in step 1.2, the characteristic entropy E is minimized λ The calculation method is shown in formula (1):
E λ =-λ 1 ln(λ 1 )-λ 2 ln(λ 2 )-λ 3 ln(λ 3 ) (1)
wherein λ is 1 、λ 2 、λ 3 Respectively representing the covariance eigenvalues of the neighborhood of the target point.
4. Segmentation method according to claim 2, characterized in that in step 1.3, the L is linear λ Flatness P λ Degree of scattering S λ All-round difference O λ Change of curvature C λ The calculation methods of the five characteristic indexes are shown in formulas (2) to (6):
Figure FDA0003970103630000011
Figure FDA0003970103630000012
Figure FDA0003970103630000021
Figure FDA0003970103630000022
Figure FDA0003970103630000023
5. the segmentation method according to claim 1, wherein the step 2 comprises:
step 2.1, constructing a minimized energy function according to the point cloud local information described in the step 1;
and 2.2, solving the problem of minimizing energy by adopting a one-cut algorithm, adaptively adjusting the size of the divided segments according to the local geometric complexity, and realizing the acquisition of the over point through graph cutting.
6. Segmentation method according to claim 5, characterized in that in step 2.1, the minimization of the energy function E (G) is constructed as shown in equation (7):
E(G)=∑ p∈P D p (G p )+λ∑ (p,q)∈E V p,q (G p ,G q ) (7)
wherein D is p (G p ) Is a data cost term, which is the sum of all points and their penalties for assigning tag numbers, D p (G p ) Indicates a label G p Penalty in assigning to node p; sigma (p,q)∈E V p,q (G p ,G q ) For the smoothing cost term, is the sum of penalties for all neighboring labels to be different, V p,q (G p ,G q ) Represents the penalty of different label numbers between adjacent nodes p and q; lambda is a weight coefficient, and the numerical value reflects the proportion of a data cost item and a smooth cost item in graph cutting; suppose that the feature of point p is denoted f p Which is associated with a label L p Data cost D of p (G p ) The calculation is shown in equation (8):
D p (G p )=(g p -f p ) 2 (8)
wherein, g p Indicates a label number G p Point cloud characteristics corresponding to the segmented object;
the smoothing cost is used to punish the inconsistency of label number between the adjacent nodes p and q, generally speaking, if the distance between the nodes p and q is smaller, the smoothing cost V between the nodes p and q is p,q (G p ,G q ) The greater the probability that the node p and q tag numbers are the same; smoothing cost V between adjacent nodes p,q (G p ,G q ) The calculation formula is shown in formulas (9) to (10):
V p,q (G p ,G q )=exp(-ds) (9)
Figure FDA0003970103630000024
wherein f is p 、f q Respectively representing the geometrical characteristics of the adjacent point clouds p and q.
7. The segmentation method according to claim 1, wherein the step 3 comprises:
step 3.1, constructing a local adjacency graph inside the super-point by taking the super-point obtained in the step 2 as a basis;
step 3.2, randomly selecting a clustering initial point x from the three-dimensional point cloud data; finding a neighborhood feature point set M of an initial point x, assuming that there are n neighboring points, defined as: x = X 1 ,X 2 ,…X n Performing convolution operation on the obtained data;
step 3.3, for n neighborhood points X in the neighborhood range of the point cloud data X in the super point ij1 ,X ij2 ,…,X ijn The neighborhood point feature vector is expressed as:
Figure FDA0003970103630000031
learning a unique attention score for the feature vector of each neighborhood point through a sharing function g;
step 3.4, when the point q and the point p in the point set M also have a neighborhood relationship in space, the q and the p are gathered into one type; weight coefficient
Figure FDA0003970103630000032
Expressing the importance of each feature vector, and performing weighted summation on each feature vector to obtain the final rich feature vector representing the over-point information
Figure FDA0003970103630000033
Step 3.5, performing the operation on all the over points in the point set to obtain all over point characteristics;
and 3.6, overlapping a plurality of layers of A-EdgeConv modules to realize layer-by-layer extraction and summarization of local information, and accurately describing the super point by overlapping different characteristic space information.
8. The segmentation method according to claim 7, wherein in step 3.3, the definition of the sharing function g is shown in formula (11):
Figure FDA0003970103630000034
wherein, W is the weight learnable by the multi-layer perceptron,
Figure FDA0003970103630000035
as feature vectors
Figure FDA0003970103630000036
A weight coefficient obtained by a sharing function;
in said step 3.4, the feature vectors
Figure FDA0003970103630000037
The calculation is shown in equation (12):
Figure FDA0003970103630000038
9. the segmentation method according to claim 1, wherein the step 4 comprises:
step 4.1, constructing a hyper-point map based on the hyper-points obtained in step 2, assuming that there are adjacent hyper-points S 'and T', and if there is one or more edges formed by three-dimensional points in two hyper-points S 'and T' belonging to E, the offset set δ (S ', T') between the hyper-points is as shown in formula (13):
δ(S′,T′)={(p i -p j )∣(i,j)∈E∩(S′×T′)} (13)
in the formula, p i 、p j Respectively representing point clouds on two sides of an edge formed by adjacent super points;
step 4.2, the super-edge characteristics are obtained according to the relation of the super-point graph, and the characteristic value decomposition is carried out on the three-dimensional point set in the adjacent super-points to obtain the characteristic value lambda 1 ≥λ 2 ≥λ 3 ≥0;
And 4.3, obtaining the one-dimensional and/or three-dimensional super-edge characteristics through different combinations of the characteristic values.
10. The segmentation method according to claim 1, wherein the step 5 comprises:
step 5.1, embedding the over-point characteristic of the step 3 and the over-edge characteristic of the step 4 into a circulation unit as initial states;
step 5.2, inputting the over-point characteristic and the over-edge characteristic to doIs an initial value h of a hidden state 0 And input x 0 The reset gate Wr determines whether the candidate state at the current moment needs to depend on the network state at the last moment and how much the candidate state needs to depend on, and the update gate Wz determines h in the state for controlling the output at the current moment t How many historical states h to keep t-1 And how many candidate states at the current time are retained
Figure FDA0003970103630000041
Continuously updating the input hidden state h at the next moment t
And 5.3, matching the aggregated features with the point labels, and optimizing by minimizing the residual error between the synthesis and the real labels in the training process to realize accurate segmentation of the point cloud.
CN202211513938.4A 2022-11-29 2022-11-29 Outdoor large scene point cloud segmentation method based on A-EdgeConv Active CN115731137B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211513938.4A CN115731137B (en) 2022-11-29 2022-11-29 Outdoor large scene point cloud segmentation method based on A-EdgeConv

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211513938.4A CN115731137B (en) 2022-11-29 2022-11-29 Outdoor large scene point cloud segmentation method based on A-EdgeConv

Publications (2)

Publication Number Publication Date
CN115731137A true CN115731137A (en) 2023-03-03
CN115731137B CN115731137B (en) 2024-07-26

Family

ID=85299105

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211513938.4A Active CN115731137B (en) 2022-11-29 2022-11-29 Outdoor large scene point cloud segmentation method based on A-EdgeConv

Country Status (1)

Country Link
CN (1) CN115731137B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117392386A (en) * 2023-10-13 2024-01-12 浙江省测绘科学技术研究院 Classification training method and device for superside mask generation network based on instance segmentation

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108319957A (en) * 2018-02-09 2018-07-24 深圳市唯特视科技有限公司 A kind of large-scale point cloud semantic segmentation method based on overtrick figure
CN110097556A (en) * 2019-04-29 2019-08-06 东南大学 Large-scale point cloud semantic segmentation algorithm based on PointNet
CN113674286A (en) * 2021-08-31 2021-11-19 浙江工商大学 Dental model point cloud segmentation method based on cross-image attention machine mechanism and cost function learning
US20220076432A1 (en) * 2020-05-06 2022-03-10 Luminar, Llc Neural network for object detection and tracking
US20220222824A1 (en) * 2020-09-15 2022-07-14 Sri International Fully automated multimodal system architecture for semantic segmentation of large-scale 3d outdoor point cloud data
CN114758129A (en) * 2022-04-12 2022-07-15 西安理工大学 RandLA-Net outdoor scene semantic segmentation method based on local feature enhancement

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108319957A (en) * 2018-02-09 2018-07-24 深圳市唯特视科技有限公司 A kind of large-scale point cloud semantic segmentation method based on overtrick figure
CN110097556A (en) * 2019-04-29 2019-08-06 东南大学 Large-scale point cloud semantic segmentation algorithm based on PointNet
US20220076432A1 (en) * 2020-05-06 2022-03-10 Luminar, Llc Neural network for object detection and tracking
US20220222824A1 (en) * 2020-09-15 2022-07-14 Sri International Fully automated multimodal system architecture for semantic segmentation of large-scale 3d outdoor point cloud data
CN113674286A (en) * 2021-08-31 2021-11-19 浙江工商大学 Dental model point cloud segmentation method based on cross-image attention machine mechanism and cost function learning
CN114758129A (en) * 2022-04-12 2022-07-15 西安理工大学 RandLA-Net outdoor scene semantic segmentation method based on local feature enhancement

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YEN-PO LIN等: "Attention EdgeConv For 3D Point Cloud Classification", PROCEEDINGS, APSIPA ANNUAL SUMMIT AND CONFERENCE 2021, 17 December 2021 (2021-12-17), pages 1 - 5 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117392386A (en) * 2023-10-13 2024-01-12 浙江省测绘科学技术研究院 Classification training method and device for superside mask generation network based on instance segmentation
CN117392386B (en) * 2023-10-13 2024-05-10 浙江省测绘科学技术研究院 Classification training method and device for superside mask generation network based on instance segmentation

Also Published As

Publication number Publication date
CN115731137B (en) 2024-07-26

Similar Documents

Publication Publication Date Title
He et al. Dyco3d: Robust instance segmentation of 3d point clouds through dynamic convolution
CN108132968B (en) Weak supervision learning method for associated semantic elements in web texts and images
CN107066559B (en) Three-dimensional model retrieval method based on deep learning
Lin et al. Active and incremental learning for semantic ALS point cloud segmentation
Oliva et al. Scene-centered description from spatial envelope properties
CN105809672B (en) A kind of image multiple target collaboration dividing method constrained based on super-pixel and structuring
Ren et al. 3d-a-nets: 3d deep dense descriptor for volumetric shapes with adversarial networks
Xia et al. Exploiting deep features for remote sensing image retrieval: A systematic investigation
CN108897791B (en) Image retrieval method based on depth convolution characteristics and semantic similarity measurement
Chen et al. Learning attentive and hierarchical representations for 3d shape recognition
Amer et al. Monocular extraction of 2.1 d sketch using constrained convex optimization
Yang et al. Local label descriptor for example based semantic image labeling
CN112364747B (en) Target detection method under limited sample
Zhao et al. DAR-Net: Dynamic aggregation network for semantic scene segmentation
CN115731137A (en) Outdoor large scene point cloud segmentation method based on A-EdgeConv
Wen et al. Learnable skeleton-aware 3d point cloud sampling
CN112860936B (en) Visual pedestrian re-identification method based on sparse graph similarity migration
Mukilan et al. Human and object detection using hybrid deep convolutional neural network
Ghosal et al. A comparative study among clustering techniques for leaf segmentation in rosette plants
Fabijanska Graph Convolutional Networks for Semi-Supervised Image Segmentation
Shyr et al. Supervised hierarchical Pitman-Yor process for natural scene segmentation
Guo Research on sports video retrieval algorithm based on semantic feature extraction
Li et al. A non-rigid 3D model retrieval method based on scale-invariant heat kernel signature features
CN111488923B (en) Enhanced anchor point image semi-supervised classification method
Özkan et al. Comparing feature extraction techniques for urban land‐use classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant