CN112149725A - Spectral domain graph convolution 3D point cloud classification method based on Fourier transform - Google Patents

Spectral domain graph convolution 3D point cloud classification method based on Fourier transform Download PDF

Info

Publication number
CN112149725A
CN112149725A CN202010991678.6A CN202010991678A CN112149725A CN 112149725 A CN112149725 A CN 112149725A CN 202010991678 A CN202010991678 A CN 202010991678A CN 112149725 A CN112149725 A CN 112149725A
Authority
CN
China
Prior art keywords
graph
convolution
point cloud
fourier transform
local
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010991678.6A
Other languages
Chinese (zh)
Other versions
CN112149725B (en
Inventor
陈苏婷
陈怀新
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN202010991678.6A priority Critical patent/CN112149725B/en
Publication of CN112149725A publication Critical patent/CN112149725A/en
Application granted granted Critical
Publication of CN112149725B publication Critical patent/CN112149725B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • G06F18/24143Distances to neighbourhood prototypes, e.g. restricted Coulomb energy networks [RCEN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/10Image enhancement or restoration using non-spatial domain filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a spectral domain graph convolution 3D point cloud classification method based on Fourier transform, which comprises the following steps: performing geometric sampling processing on the input original point cloud by using a G-PointNet network model: dividing points with a neighborhood included angle value larger than that of a point into a geometric feature region G and the rest points into other regions T by setting an angle threshold V, and sampling to obtain point clouds of the regions; introducing an expansion rate E based on a Dynamic KNN local graph construction method, and selectively establishing a local geometric graph every E adjacent point clouds. And performing spectrum domain graph convolution by using a spectrum domain graph convolution method based on Fourier transform to obtain a plurality of pooled graph local features, obtaining global features through G-PointNet, classifying, and obtaining a classification result. The invention effectively solves the problem of uneven distribution of the density degree of the point cloud, retains the space geometric information, can efficiently distinguish the edge points of the point cloud and simultaneously separate the noise points, and improves the classification precision.

Description

Spectral domain graph convolution 3D point cloud classification method based on Fourier transform
Technical Field
The invention relates to a spectral domain graph convolution 3D point cloud classification method based on Fourier transform, and belongs to the technical field of remote sensing image processing.
Background
With the image processing technique of hundreds of buzzes, classification methods based on two-dimensional images are diversified, and a few achievements are achieved. However, the deep learning processing method based on three-dimensional data has a far lower effect than the two-dimensional image classification effect. Three-dimensional data is typically represented in depth images, voxels, meshes, and point clouds. Compared with the method for acquiring three-dimensional data through an RGB-D camera or a mainstream sensor, the three-dimensional point cloud acquired by the laser radar can provide more reliable depth and outline information of the three-dimensional object, and is gradually applied to three-dimensional object classification in recent years.
In the previous work, computer vision researchers mostly use the success of CNN in image processing for point cloud classification, extract two-dimensional features by using a three-dimensional object, or obtain multiple two-dimensional view images from different "views" of the object, project the three-dimensional object into multiple views, extract corresponding view features, and then fuse these features to perform accurate object identification. However, the methods derive the shape of the 3D object from the 2D image, abandon the inherent spatial structure of the 3D point cloud, lose a large amount of spatial structure information, and consume excessive memory. In view of the deficiencies of 2D multi-view, researchers have attempted to voxelate 3D point clouds. However, voxel construction does not build complete edge information, which makes it difficult for such methods to capture high fine granularity.
Therefore, several problems are common in the current point cloud classification task. Firstly, image convolution can determine the size of a convolution kernel to define a local area in an image, unlike a grid structure with a regular arrangement in the image, point clouds are a set of dispersed points in a three-dimensional space, the point clouds are continuously distributed in the space, and the arrangement sequence of the point clouds does not change the space distribution, so that the point clouds cannot be directly convolved by using a traditional deep neural network. Secondly, the non-uniform density of the point cloud distribution in the space also brings great challenges to the classification problem.
Disclosure of Invention
In order to overcome the defects of the prior art and effectively solve the problem that the traditional 3D point cloud classification method is affected by the spatial relationship and the uneven distribution of the point cloud, the invention provides a spectral domain graph convolution 3D point cloud classification method based on Fourier transform, under the premise of not changing the spatial information of the point cloud, a new expression form-graph is provided, the graph structure effectively solves the problem of the adjacent relationship between points in most point cloud deep learning models, the spatial geometrical information is reserved, and the graph is very suitable for arranging irregular non-European data; deep learning lacks a lot of research work in a spectral domain, the model combines a spectral domain graph convolution method in a 3D point cloud frame for the first time, the spectral domain convolution has a solid mathematical theoretical basis, and graph convolution more emphasizes the adjacency relation between key points; the G-PointNet of the invention makes great improvement in the aspects of feature point acquisition and local area division, provides geometric sampling pretreatment and designs a Dynamic K neighbor graph construction method, namely Dynamic KNN, and effectively solves the problem of uneven distribution of point cloud density.
The invention specifically adopts the following technical scheme to solve the technical problems:
a spectral domain graph convolution 3D point cloud classification method based on Fourier transform comprises the following steps:
performing geometric sampling processing on the input original point cloud by using a G-PointNet network model: by setting an angle threshold V, dividing points with neighborhood included angle values larger than the angle threshold V into a geometric characteristic region G and dividing the rest points into other regions T, and uniformly sampling point clouds in the two regions respectively to obtain point clouds after geometric sampling of the regions;
constructing an undirected graph from the point clouds after geometric sampling of each region, introducing an expansion rate E based on a Dynamic KNN local graph construction method, and selectively establishing a local geometric graph at intervals of E neighboring point clouds to obtain a plurality of local geometric graphs;
and performing spectrum domain graph convolution on each local geometric graph by using a spectrum domain graph convolution method based on Fourier transform to obtain a plurality of pooled graph local features, obtaining global features through a G-PointNet network model, classifying, and obtaining a classification result.
Further, as a preferred technical solution of the present invention, the method performs spectrum domain graph convolution on each local geometric graph by using a spectrum domain graph convolution method based on fourier transform, specifically:
inputting a local geometric graph G ═ V, (V,) wherein V represents a corresponding node set and an edge set respectively, and [ mu ], V ∈ V represents a node in the graph, and [ mu ], V ∈ V represents an edge in the graph;
the laplacian matrix L ═ D-a defining the local geometry, where a denotes the adjacency matrix of the graph with the elements a insidei,j=Aj,iA matrix for representing the connection of nodes in the graph; d represents the degree matrix of the graph, within which the element Di,i=∑jAi,jThe degree of a node represents the number of edges connected to the node;
normalization yields a Laplace matrix, denoted as
Figure BDA0002689387150000021
Wherein InIs a unit matrix and is characterized by a set of laplacian eigenvectors U ═ U (U)1,u2,...un);
Taking the decomposed Laplace matrix eigenvector as a group of basis, taking the local geometric figure as an input x and carrying out Fourier transformation
Figure BDA0002689387150000022
T is the transposition of the matrix; obtaining a convolution kernel hθAnd (Λ) in a Fourier transform diagonal matrix form to obtain Fourier transform convolution in a spectral domain, and then performing inverse transformation on the Fourier transform convolution to finally obtain the convolution output of the spectral domain graph.
Further, as a preferred technical solution of the present invention, the spectral domain plot convolution output formula obtained by the method is:
y=σ(Uhθ(Λ)UTx)
wherein y is the output of the spectral domain graph convolution; x is the local geometry of the input and σ () is the activation function relu.
By adopting the technical scheme, the invention can produce the following technical effects:
the invention discloses a spectral domain graph convolution 3D point cloud classification method based on Fourier transform, which combines a deep network model of PointNet and spectral domain graph convolution operation based on Fourier transform to obtain a G-PointNet network model of the method. G-PointNet preserves the spatial transformation network T-Net in PointNet, and the semantic labels of point clouds must be invariant if they undergo some specific set transformations, such as rigid transformations. The representation of the learned set of points is therefore also invariant to these transformations. The solution is to fit all input sets into a canonical space before feature extraction. The three-dimensional space is aligned through sampling and interpolation, and a network layer is specially designed to be implemented on a GPU.
And in the preprocessing stage, geometric sampling processing is carried out on the original point cloud. Inspired by cavity convolution, the invention provides a graph construction method, namely a Dynamic KNN (D-KNN), a plurality of local geometric graph structures are constructed to be used as graph convolution input, the graph convolution input is mapped to a spectrum domain through Fourier transform and is subjected to convolution operation, the graph convolution operation is returned to a space domain through inverse Fourier transform, and finally global features of a plurality of pooled graph local features are obtained through PointNet and are classified. The process can be divided into three parts: geometric sampling, Dynamic KNN local map construction and spectrum domain map convolution based on Fourier transformation. Respectively has the following advantages:
1. geometric sampling: the advantages of geometric sampling are very significant. In the area with more obvious point cloud geometric characteristics, the number distribution of the sampled point clouds is more, the edge characteristics are very obvious, the calculation efficiency is high, and the anti-noise capability of the sampling result is stronger.
Dynamic KNN local map construction: the Dynamic KNN is a composition method for dynamically selecting a K near neighborhood, which is inspired from hole convolution, and on a two-dimensional image task, the hole convolution effectively increases the receptive field by introducing a hyper-parameter called ' expansion rate ' (dilation rate) ' on the premise of not changing the output size of an image. The expansion rate E is introduced into the KNN algorithm by the Dynamic KNN, and the expansion rate can be selected according to the density degree of the point cloud.
3. Fourier transform-based spectral domain map convolution: the spectral domain graph convolution network based on Fourier transform is used as a point cloud feature extraction network, and the advantage of the spectral domain graph convolution on a point cloud classification task is also quite obvious. In an end-to-end deep learning task relative to space-domain convolution, a mathematical theory cannot be used for explaining, and Fourier transformation has a solid theoretical basis for explaining the feasibility. Secondly, points and noise points with large edge changes in the point cloud are generally regarded as high-frequency signals, Fourier transform energy conversion distinguishes high-frequency signals from low-frequency signals, and for a classification task, the Fourier transform energy conversion effectively distinguishes the edge points of the point cloud and simultaneously separates the noise points, which is important for the classification task.
Drawings
FIG. 1 is a schematic diagram of the method of the present invention.
FIG. 2 is a schematic diagram of geometric sampling processing in the method of the present invention.
FIG. 3 is a schematic diagram of a partial view of a Dynamic KNN according to the present invention.
Detailed Description
The following describes embodiments of the present invention with reference to the drawings.
As shown in fig. 1, the invention designs a spectral domain graph convolution 3D point cloud classification method based on fourier transform, and the process can be divided into three parts: geometric sampling, Dynamic KNN local graph construction and spectrum domain graph convolution based on Fourier transform specifically comprise the following steps:
step 1, carrying out geometric sampling processing on input original point cloud by utilizing a G-PointNet network model. The method comprises the following specific steps:
firstly, obtaining the G-PointNet network model of the invention under the enlightenment of a PointNet depth network model and a spectral domain graph convolution operation based on Fourier transform. The G-PointNet network model preserves the spatial transformation network T-Net in PointNet. The G-PointNet network model directly uses as input a point cloud, which is a collection of consecutive points in three-dimensional space, represented as three-dimensional space coordinates (x, y, z), sometimes with additional features such as color, laser reflection intensity, etc. Unless specified, G-PointNet uses only three-dimensional coordinates (x, y, z) as point features.
The characteristics of the point cloud, which means that direct convolution with depth models in image processing cannot be used, have 3 characteristics:
1. the point set is disordered. The image can be changed when any two pixels in the image exchange positions, and unlike the two-dimensional image, the point cloud has no specific arrangement sequence, namely the arrangement sequence of any point in the space is changed, and the shape change of the point cloud cannot be influenced.
2. The interrelationship between points. Each point has three-dimensional coordinates (x, y, z), which means that each point has spatial information representing its shape, each point is not independent, and a neighboring set of local points may represent meaningful spatial information. Therefore, the model needs to be able to capture local structures from nearby points, as well as the combined interactions between local structures.
3. The permutation is unchanged. The point cloud is three-dimensional data, which represents any rotation or translation operation, and should not affect the final point cloud classification effect.
The G-PointNet network model of the method performs geometric sampling processing on the original point cloud in a preprocessing stage, reserves a spatial transformation network T-Net in the PointNet, and the T-Net can adjust the point cloud to a position suitable for point cloud classification.
Fig. 2 is a schematic diagram of geometric sampling processing of the method of the present invention. The method selects geometric sampling to process point cloud data. A traditional Point cloud classification model, such as PointNet, adopts a Farthest Point Sampling algorithm to randomly select an initial Point, selects a Point which is Farthest away from the initial Point to add a starting Point, and then continues iteration until the required number is iterated. Every time when the Farthest Point Sampling samples one Point, the distance between the sets needs to be calculated, the algorithm time complexity is high, and the edge characteristics of the obtained Sampling points are not obvious. By contrast, the advantages of geometric sampling are quite significant. The geometric sampling processing process of the invention is as follows:
assuming the number C of input point clouds, the number S of target sampling and the uniform sampling rate U, setting an angle threshold V, dividing points with neighborhood included angle values larger than the angle threshold V into geometric characteristic regions G, and dividing the rest points into other regions T; the point cloud is divided into two parts, namely a geometric characteristic region G and other regions T, and the point clouds in the two regions G, T are respectively and uniformly sampled to obtain point clouds after geometric sampling of each region.
Geometric sampling can acquire more point cloud numbers at places with larger point cloud curvature, however, calculating the point cloud curvature is time-consuming and greatly increases the work difficulty, so that a simple method is used to approximately achieve the curvature effect: and calculating the normal angle value from the characteristic point to the neighborhood point in the local point cloud picture structure, wherein the bigger the normal angle value is, the bigger the curvature value is. In fig. 2, c1, c2 and c3 are 3 point cloud points respectively, alpha represents a normal line included angle value, the curvature effect of c1 is replaced by the normal line included angle values of c2 and c3, and the two are in positive correlation.
Step 2, constructing an undirected graph from the point clouds after geometric sampling of each region, introducing an expansion rate E based on a Dynamic KNN local graph construction method, and selectively establishing a local geometric graph every E neighboring point clouds to obtain a plurality of local geometric graphs, wherein the method specifically comprises the following steps:
fig. 3 is a schematic diagram of a local map construction method dynamicknn. The receptive field of a point cloud refers to a set of point cloud nodes including a central node and its neighbors, however, the non-uniformity of the point cloud distribution results in some nodes possibly having only one neighbor, while other nodes may have as many as thousands of neighbors.
According to the local map construction method, the expansion rate E is introduced into a KNN algorithm by the Dynamic KNN, and the expansion rate can be selected according to the density degree of the point cloud. The Dynamic KNN sets two threshold points M, N, where M < N. And M and N are the number of target point clouds, the number of the point clouds after geometric sampling is X, and the expansion rate is E.
Figure BDA0002689387150000051
The method comprises the steps of giving an undirected graph G (v) to represent a graph structure of a point cloud, wherein v is { 1., n }, an edge belongs to v x v, and the idea of Dynamic KNN is that a local geometric graph is selectively established every E adjacent point clouds according to the sparsity degree of the point clouds, and a plurality of local geometric graphs are obtained after multiple selections are carried out. When the point cloud is sparse, the traditional KNN nearest neighbor idea is adopted, where E is 1. When the point cloud is dense, E points close to each other realize the connection of the points, a local geometric graph structure is established as graph convolution input and is sent to a spectral domain for graph convolution. The Dynamic KNN effectively solves the problem of overhigh node coincidence degree in dense point cloud, and simultaneously reduces the calculation complexity.
And 3, performing spectrum domain graph convolution on each local geometric graph by using a spectrum domain graph convolution method based on Fourier transform to obtain a plurality of pooled graph local features, obtaining global features through a PointNet network model, classifying, and obtaining a classification result. The method comprises the following steps of respectively taking a plurality of constructed local geometric graphs as graph convolution input x by a local graph construction method Dynamic KNN, mapping the graph convolution input x to a spectral domain by Fourier transform for convolution operation, returning the graph convolution operation to a spatial domain by inverse Fourier transform, and finally obtaining global features of a plurality of pooled graph local features by a PointNet network model for classification, wherein the specific process is as follows:
firstly, inputting a local geometric graph G ═ V ═ to represent an undirected graph, V respectively represents a corresponding node set and an edge set, mu, V ∈ V represents nodes in the graph, and (mu, V ∈ V represents edges in the graph;
next, a laplacian matrix L ═ D-a of the local geometry G is defined, where a denotes the adjacency matrix (adjacency matrix) of the graph, in which the element a isi,j=Aj,iI and j respectively represent the row number and the column number of elements positioned in the matrix, and are used for representing the matrix of the connection condition of the nodes in the graph; for an undirected graph with N nodes, the adjacency matrix is an N × N real-symmetric matrix. D represents a degree matrix (degree matrix) of the graph, in which the element Di,i=∑jAi,jAnd the degree of a node indicates the number of edges connected to the node. L represents the laplacian matrix of the diagram, which may be binary or weighted.
Then, normalization yields a Laplace matrix, which can be expressed as
Figure BDA0002689387150000061
Wherein InIs a unit matrix, and since the Laplace matrix is a symmetric matrix, the characteristic decomposition into a group of Laplace matrices can be performedThe placian eigenvector U ═ (U ═ U)1,u2,...un) (ii) a The characteristic vector of the Laplace characteristic vector and the characteristic vector associated with the larger characteristic value represents a signal carrying fast change and is regarded as a high-frequency signal; the eigenvectors associated with the smaller eigenvalues carry slower varying signals and are considered low frequency signals. In the point cloud classification task, the edge information of the object can be found as signals for distinguishing high frequency from low frequency.
Then, taking the decomposed Laplace matrix eigenvector as a group of basis, Fourier transform is performed using the local geometry as input x
Figure BDA0002689387150000062
T is the transposition of the matrix; inverse Fourier transform
Figure BDA0002689387150000063
The traditional Fourier transform and convolution are transferred to graph convolution, and the core work is to change the characteristic function of the Laplace operator into the characteristic vector of the Laplace matrix corresponding to the local geometric graph G. I.e. obtaining the convolution kernel hθAnd (Λ) in a Fourier transform diagonal matrix form to obtain Fourier transform convolution in a spectral domain, and then performing inverse transformation on the Fourier transform convolution to finally obtain spectral domain graph convolution output, wherein the derivation process is as follows:
considering the input as f, the Fourier transform of f
Figure BDA0002689387150000071
② convolution kernel hθA fourier transform diagonal matrix form of (Λ):
Figure BDA0002689387150000072
where theta is the kernel, lambda is the feature vector,
Figure BDA0002689387150000073
ulis a feature vector.
And thirdly, obtaining Fourier convolution in a spectral domain:
Figure BDA0002689387150000074
inverse transform of Fourier transform product
Figure BDA0002689387150000075
Fifthly, the final convolution formula of the spectrum domain diagram is as follows:
y=σ(Uhθ(Λ)UTx)
wherein y is the output of the spectral domain graph convolution; x is the local geometry of the input and σ () is the activation function relu.
And finally, obtaining a plurality of pooled local features of the map according to the output y of the spectrum domain map convolution, and obtaining global features through a PointNet network model for classification to obtain a final classification result.
Therefore, the method can be greatly improved in the aspects of feature point acquisition and local area division, geometric sampling pretreatment is provided, a Dynamic local graph construction method Dynamic KNN is designed, and the problem of uneven distribution of the density degree of the point cloud is effectively solved. On the premise of not changing the spatial information of the point cloud, a new expression form-graph is provided, the graph structure effectively solves the problem of the adjacent relation between points in most point cloud deep learning models, the spatial geometrical information is kept, and the receptive field is effectively increased on the premise of not changing the output size of the image by introducing a hyper-parameter called expansion rate. The spectrum domain graph convolution network based on Fourier transform is used as a point cloud feature extraction network, the Fourier transform energy conversion is used for efficiently distinguishing edge points of point clouds aiming at a classification task, and meanwhile, noise points are separated, so that the classification precision is improved, and the advantage of spectrum domain graph convolution on the point cloud classification task is very obvious.
The embodiments of the present invention have been described in detail with reference to the drawings, but the present invention is not limited to the above embodiments, and various changes can be made within the knowledge of those skilled in the art without departing from the gist of the present invention.

Claims (3)

1. The spectral domain graph convolution 3D point cloud classification method based on Fourier transform is characterized by comprising the following steps of:
performing geometric sampling processing on the input original point cloud by using a G-PointNet network model: by setting an angle threshold V, dividing points with neighborhood included angle values larger than the angle threshold V into a geometric characteristic region G and dividing the rest points into other regions T, and uniformly sampling point clouds in the two regions respectively to obtain point clouds after geometric sampling of the regions;
constructing an undirected graph from the point clouds after geometric sampling of each region, introducing an expansion rate E based on a Dynamic KNN local graph construction method, and selectively establishing a local geometric graph at intervals of E neighboring point clouds to obtain a plurality of local geometric graphs;
and performing spectrum domain graph convolution on each local geometric graph by using a spectrum domain graph convolution method based on Fourier transform to obtain a plurality of pooled graph local features, obtaining global features through a G-PointNet network model, classifying, and obtaining a classification result.
2. The fourier transform-based spectral domain graph convolution 3D point cloud classification method according to claim 1, wherein the method performs spectral domain graph convolution on each local geometric graph by using a fourier transform-based spectral domain graph convolution method, specifically:
inputting a local geometric graph G ═ V, (V,) wherein V represents a corresponding node set and an edge set respectively, and [ mu ], V ∈ V represents a node in the graph, and [ mu ], V ∈ V represents an edge in the graph;
the laplacian matrix L ═ D-a defining the local geometry, where a denotes the adjacency matrix of the graph with the elements a insidei,j=Aj,iA matrix for representing the connection of nodes in the graph; d represents the degree matrix of the graph, within which the element Di,i=∑jAi,jThe degree of a node represents the number of edges connected to the node;
normalization yields a Laplace matrix, denoted as
Figure FDA0002689387140000011
Wherein InIs a unit matrix and is characterized by a set of laplacian eigenvectors U ═ U (U)1,u2,...un);
Taking the decomposed Laplace matrix eigenvector as a group of basis, taking the local geometric figure as an input x and carrying out Fourier transformation
Figure FDA0002689387140000012
T is the transposition of the matrix; obtaining a convolution kernel hθAnd (Λ) in a Fourier transform diagonal matrix form to obtain Fourier transform convolution in a spectral domain, and then performing inverse transformation on the Fourier transform convolution to finally obtain the convolution output of the spectral domain graph.
3. The method for classifying a 3D point cloud based on convolution of a Fourier transform spectrum domain map according to claim 2, wherein the convolution output formula of the spectrum domain map obtained by the method is as follows:
y=σ(Uhθ(Λ)UTx)
wherein y is the output of the spectral domain graph convolution; x is the local geometry of the input and σ () is the activation function relu.
CN202010991678.6A 2020-09-18 2020-09-18 Fourier transform-based spectrum domain map convolution 3D point cloud classification method Active CN112149725B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010991678.6A CN112149725B (en) 2020-09-18 2020-09-18 Fourier transform-based spectrum domain map convolution 3D point cloud classification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010991678.6A CN112149725B (en) 2020-09-18 2020-09-18 Fourier transform-based spectrum domain map convolution 3D point cloud classification method

Publications (2)

Publication Number Publication Date
CN112149725A true CN112149725A (en) 2020-12-29
CN112149725B CN112149725B (en) 2023-08-22

Family

ID=73892671

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010991678.6A Active CN112149725B (en) 2020-09-18 2020-09-18 Fourier transform-based spectrum domain map convolution 3D point cloud classification method

Country Status (1)

Country Link
CN (1) CN112149725B (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112967296A (en) * 2021-03-10 2021-06-15 重庆理工大学 Point cloud dynamic region graph convolution method, classification method and segmentation method
CN113157864A (en) * 2021-04-25 2021-07-23 平安科技(深圳)有限公司 Key information extraction method and device, electronic equipment and medium
CN113763260A (en) * 2021-02-20 2021-12-07 京东鲲鹏(江苏)科技有限公司 Noise filtering method, device, equipment and storage medium based on water droplet noise
CN113936176A (en) * 2021-10-15 2022-01-14 哈尔滨理工大学 Three-dimensional point cloud classification method based on graph volume and shape descriptor
US11295170B1 (en) 2021-08-17 2022-04-05 FPT USA Corp. Group-equivariant convolutional neural networks for 3D point clouds
CN114359902A (en) * 2021-12-03 2022-04-15 武汉大学 Three-dimensional point cloud semantic segmentation method based on multi-scale feature fusion
CN114565774A (en) * 2022-02-21 2022-05-31 辽宁师范大学 3D (three-dimensional) image volume integral classification method based on local geometry and global structure joint learning
CN115099287A (en) * 2022-08-24 2022-09-23 山东大学 Space variable gene identification and analysis system based on graph Fourier transform
CN116977572A (en) * 2023-09-15 2023-10-31 南京信息工程大学 Building elevation structure extraction method for multi-scale dynamic graph convolution
WO2023222923A1 (en) * 2022-05-20 2023-11-23 Cobra Simulation Ltd Method of content generation from sparse point datasets
CN113936176B (en) * 2021-10-15 2024-06-07 哈尔滨理工大学 Three-dimensional point cloud classification method based on graph convolution and shape descriptors

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102292187A (en) * 2008-11-21 2011-12-21 普雷茨特两合公司 Method and device for monitoring a laser machining operation to be performed on a workpiece and laser machining head having such a device
US20150371431A1 (en) * 2013-01-29 2015-12-24 Andrew Robert Korb Methods for analyzing and compressing multiple images
CN106897707A (en) * 2017-03-02 2017-06-27 苏州中科天启遥感科技有限公司 Characteristic image time series synthetic method and device based in multi-source points
CN110348299A (en) * 2019-06-04 2019-10-18 上海交通大学 The recognition methods of three-dimension object
CN111027559A (en) * 2019-10-31 2020-04-17 湖南大学 Point cloud semantic segmentation method based on expansion point convolution space pyramid pooling
CN111160171A (en) * 2019-12-19 2020-05-15 哈尔滨工程大学 Radiation source signal identification method combining two-domain multi-features

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102292187A (en) * 2008-11-21 2011-12-21 普雷茨特两合公司 Method and device for monitoring a laser machining operation to be performed on a workpiece and laser machining head having such a device
US20150371431A1 (en) * 2013-01-29 2015-12-24 Andrew Robert Korb Methods for analyzing and compressing multiple images
CN106897707A (en) * 2017-03-02 2017-06-27 苏州中科天启遥感科技有限公司 Characteristic image time series synthetic method and device based in multi-source points
CN110348299A (en) * 2019-06-04 2019-10-18 上海交通大学 The recognition methods of three-dimension object
CN111027559A (en) * 2019-10-31 2020-04-17 湖南大学 Point cloud semantic segmentation method based on expansion point convolution space pyramid pooling
CN111160171A (en) * 2019-12-19 2020-05-15 哈尔滨工程大学 Radiation source signal identification method combining two-domain multi-features

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
AOKI Y等: "PointNetLK: Robust & Efficient Point Cloud Registration Using PointNet", 《PROCEEDINGS OF THE IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
ZHENZHONG KUANG等: "Discovering the Latent Similarities of the KNN Graph by Metric Transformation", 《ACM》 *
王旭娇: "基于深度学习的场景三维点云分类方法研究", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 *
陈万军;张二虎;: "基于深度信息的人体动作识别研究综述", 《西安理工大学学报》 *
陈苏婷等: "基于图卷积神经网络的三维点云分割算法Graph⁃PointNet", 《现代电子技术》 *
黄星奕等: "霉变出芽花生的近红外光谱无损检测研究", 《中国农业科技导报》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113763260B (en) * 2021-02-20 2024-04-16 京东鲲鹏(江苏)科技有限公司 Water drop noise-based noise filtering method, device, equipment and storage medium
CN113763260A (en) * 2021-02-20 2021-12-07 京东鲲鹏(江苏)科技有限公司 Noise filtering method, device, equipment and storage medium based on water droplet noise
CN112967296A (en) * 2021-03-10 2021-06-15 重庆理工大学 Point cloud dynamic region graph convolution method, classification method and segmentation method
CN113157864A (en) * 2021-04-25 2021-07-23 平安科技(深圳)有限公司 Key information extraction method and device, electronic equipment and medium
US11295170B1 (en) 2021-08-17 2022-04-05 FPT USA Corp. Group-equivariant convolutional neural networks for 3D point clouds
CN113936176A (en) * 2021-10-15 2022-01-14 哈尔滨理工大学 Three-dimensional point cloud classification method based on graph volume and shape descriptor
CN113936176B (en) * 2021-10-15 2024-06-07 哈尔滨理工大学 Three-dimensional point cloud classification method based on graph convolution and shape descriptors
CN114359902A (en) * 2021-12-03 2022-04-15 武汉大学 Three-dimensional point cloud semantic segmentation method based on multi-scale feature fusion
CN114359902B (en) * 2021-12-03 2024-04-26 武汉大学 Three-dimensional point cloud semantic segmentation method based on multi-scale feature fusion
CN114565774A (en) * 2022-02-21 2022-05-31 辽宁师范大学 3D (three-dimensional) image volume integral classification method based on local geometry and global structure joint learning
CN114565774B (en) * 2022-02-21 2024-04-05 辽宁师范大学 3D (three-dimensional) graph volume integration class method based on local geometry and global structure joint learning
WO2023222923A1 (en) * 2022-05-20 2023-11-23 Cobra Simulation Ltd Method of content generation from sparse point datasets
CN115099287A (en) * 2022-08-24 2022-09-23 山东大学 Space variable gene identification and analysis system based on graph Fourier transform
CN116977572B (en) * 2023-09-15 2023-12-08 南京信息工程大学 Building elevation structure extraction method for multi-scale dynamic graph convolution
CN116977572A (en) * 2023-09-15 2023-10-31 南京信息工程大学 Building elevation structure extraction method for multi-scale dynamic graph convolution

Also Published As

Publication number Publication date
CN112149725B (en) 2023-08-22

Similar Documents

Publication Publication Date Title
CN112149725B (en) Fourier transform-based spectrum domain map convolution 3D point cloud classification method
Chen et al. LSANet: Feature learning on point sets by local spatial aware layer
CN111028327A (en) Three-dimensional point cloud processing method, device and equipment
CN110675379A (en) U-shaped brain tumor segmentation network fusing cavity convolution
CN112634149A (en) Point cloud denoising method based on graph convolution network
CN115222625A (en) Laser radar point cloud denoising method based on multi-scale noise
CN111815640B (en) Memristor-based RBF neural network medical image segmentation algorithm
CN115690188A (en) Human body three-dimensional measurement method based on point cloud model optimization
CN116721121A (en) Plant phenotype color image feature extraction method
CN116310095A (en) Multi-view three-dimensional reconstruction method based on deep learning
CN111340133A (en) Image classification processing method based on deep convolutional neural network
CN108921853B (en) Image segmentation method based on super-pixel and immune sparse spectral clustering
Zhou et al. Image copy-move forgery passive detection based on improved PCNN and self-selected sub-images
CN112085066B (en) Voxelized three-dimensional point cloud scene classification method based on graph convolution neural network
Wang et al. Multispectral point cloud superpoint segmentation
CN113657522A (en) Multi-view three-dimensional model clustering method
Feng et al. Infrared and visible image fusion based on the total variational model and adaptive wolf pack algorithm
Cheng et al. Visual information quantification for object recognition and retrieval
Tan et al. Accurate detection of built-up areas from high-resolution remote sensing imagery using a fully convolutional network
CN104751449A (en) Particle swarm optimization based SAR image segmentation method
CN112884884A (en) Candidate region generation method and system
CN112633376A (en) Point cloud data ground feature classification method and system based on deep learning and storage medium
Li et al. Graph Convolution Network with Double Filter for Point Cloud Segmentation
Zhang et al. A method for identifying and repairing holes on the surface of unorganized point cloud
Ren et al. Steel Surface Defect Detection Using Improved Deep Learning Algorithm: ECA-SimSPPF-SIoU-Yolov5

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant