CN110322453A - 3D point cloud semantic segmentation method based on position attention and auxiliary network - Google Patents
3D point cloud semantic segmentation method based on position attention and auxiliary network Download PDFInfo
- Publication number
- CN110322453A CN110322453A CN201910604264.0A CN201910604264A CN110322453A CN 110322453 A CN110322453 A CN 110322453A CN 201910604264 A CN201910604264 A CN 201910604264A CN 110322453 A CN110322453 A CN 110322453A
- Authority
- CN
- China
- Prior art keywords
- network
- feature
- point cloud
- segmentation
- indicate
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
Abstract
The 3D point cloud semantic segmentation method based on position attention and auxiliary network that the invention proposes a kind of, mainly solves the problems, such as that prior art segmentation precision is lower, implementation is: obtaining training set T and test set V;3D point cloud semantic segmentation network is constructed, and sets the loss function of the network, which includes that successively cascade feature down-sampling network, position notice that power module, feature up-sample network and auxiliary network;The training that P wheel has supervision is carried out to the segmentation network using training set T: according to loss function in the training process of every wheel, network parameter is adjusted, after the completion of P takes turns training, using the highest network model of segmentation precision as trained network model;Test set V is input in trained network model and carries out semantic segmentation, obtains the segmentation result of each point.The present invention improves 3D point cloud semantic segmentation precision, can be used for automatic Pilot, robot, 3D scene rebuilding, quality testing, 3D drawing and smart city construction.
Description
Technical field
The invention belongs to technical field of data processing, in particular to a kind of 3D point cloud semantic segmentation method can be used for automatic
Driving, robot, 3D scene rebuilding, quality testing, 3D drawing and smart city construction.
Background technique
In recent years, with laser radar, the 3D such as RGBD camera sensor is answered in robot, the extensive of unmanned field
With deep learning has become one of research hotspot in the application of 3D point cloud data.3D point cloud data refer to: in a three-dimensional seat
The set of one group of vector in mark system, usually with x, the form of y, z three-dimensional coordinate indicate these vectors, are generally used to represent one
The external surface shape of a object.In addition, may also contain RGB color, intensity, ash outside the geological information represented in addition to (x, y, z)
The information such as angle value, depth or recycle time.Point cloud data is usually obtained by 3D scanning device, such as laser radar, RGBD
Camera etc..The mode of these sensors automation measures the information largely put in body surface, then with certain data text
Part exports point cloud data.Point cloud data has randomness, unstructured feature and may have in the 3 d space different
Consistency.This applies deep learning to face huge challenge in the research of 3D point cloud data.
3D point cloud semantic segmentation, which refers to, distributes a classification to each of point cloud data of input point.In grinding for early stage
Study carefully in work, 3D point cloud data are generally converted into the characteristics of image of manual voxel grid feature either multi-angle of view, then send
Enter deep learning network and carry out feature extraction, not only data volume is big but also calculates complexity for the method for such converting characteristic, if drop
Low resolution, then segmentation precision can decline.Therefore, point cloud data is directly handled using the method for deep learning seems especially heavy
It wants.
2017, entitled " the PointNet:Deep Learning on that Charles R Qi et al. is delivered on CVPR
The paper of Point Sets for 3D Classification and Segmentation " discloses a kind of directly processing 3D
The deep learning frame of point cloud data, this method solve the problems, such as some cloud randomnesses using the symmetric function of max-pooling, from
And extract the global characteristics of each point.But this method only considered global characteristics, have ignored the local feature of each point.Cause
This, PointNet soon, the team of Charles R Qi has delivered entitled " PointNet++:Deep in NIPS
The paper of Hierarchical Feature Learning on Point Sets in a Metric Space ", PointNet
++ it is the layering version of PointNet, every layer all there are three the stages: sampling, grouping and feature extraction.Some comparisons are chosen first
Central point of the important point as each regional area, then chooses k according to Euclidean distance around these central points
Neighbor Points.Feature is extracted using PointNet network using k Neighbor Points as a partial points cloud again, later to further feature
It is returned, to obtain 3D point cloud data semantic segmentation result, this method is promoted compared with PointNet precision.
Above-mentioned both methods is compared with the traditional method, and due to directly handling 3D point cloud data, calculates simple, effectively solution
The characteristics of point cloud randomness and segmentation precision is improved, still, PointNet++ is not due to accounting for each central point
Relationship namely contextual information between feature, so character representation is relatively weak, in addition, PointNet++ has deferred to coding-
Decoded general framework does not account for the more information of bottom, and therefore, segmentation precision is not high, still has improved space.
Summary of the invention
It is a kind of based on position attention and auxiliary net it is an object of the invention in view of the above shortcomings of the prior art, propose
The 3D point cloud data semantic dividing method of network, with the auxiliary net of the position attention of associated context feature and reconstruction bottom-up information
Network improves segmentation precision.
To achieve the above object, technical solution of the present invention includes the following steps:
(1) from the training file and test file of the official website ScanNet downloading 3D point cloud data, and classification statistics is carried out to it
And dicing treatment, obtain training set T and test set V;
(2) 3D point cloud semantic segmentation network is constructed comprising successively cascade feature down-sampling network, position attention mould
Block, feature up-sample network and auxiliary network;
(3) polytypic cross entropy loss function, the loss function as 3D point cloud semantic segmentation network are used;
(4) training set T is used, the training that P wheel has supervision, P >=500 are carried out to 3D point cloud data semantic segmentation network;
(4a), according to the loss function of semantic segmentation network, adjusts network parameter, obtains network during the training of every wheel
Model;
(4b) is every P1Wheel, assesses the segmentation precision of current network model using the sample of test set, if currently
The segmentation precision of network model is higher than previously stored network model, then is saved, P1≥2;
After the completion of (4c) P wheel training, using the highest network model of segmentation precision as trained network model;
(5) test set V is input in trained network model and carries out semantic segmentation, obtain the segmentation knot of each point
Fruit.
Compared with prior art, the present invention having the advantage that
The present invention pays attention to power module due to constructing 3D point cloud semantic segmentation network, and by position therein, calculates it
Correlation between feature representated by each mass center of input data is the local centroid feature of 3D point cloud semantic segmentation network
Increase contextual information;Simultaneously because by auxiliary network therein, to the low-level image feature of 3D point cloud semantic segmentation network into
Row passback, has rebuild the bottom-up information of 3D point cloud semantic segmentation network, has effectively increased the segmentation precision of 3D point cloud semantic segmentation.
Detailed description of the invention
Fig. 1 is implementation flow chart of the invention;
Fig. 2 is the 3D point cloud semantic segmentation network overall structure figure constructed in the present invention;
Fig. 3 is attention function structure chart in position in the present invention.
Specific embodiment
In the following with reference to the drawings and specific embodiments, present invention is further described in detail.
Referring to Fig.1, the realization step of this example includes as follows.
Step 1, training set T and test set V is obtained.
1.1) from the training file and test file of the official website ScanNet downloading 3D point cloud data, wherein trained file includes
f0A cloud scene includes f in test file1A cloud scene, f in the present embodiment0=1201, f1=312;
1.2) using all f in statistics with histogram training file0The number of each classification of the point cloud data of a scene, and count
Calculate the weight w of each classificationk:
Wherein, GkIndicate the number of kth class point cloud data, M indicates the number of all point cloud datas, and L indicates segmentation classification
It counts, L >=2, L=21 in the present embodiment;
1.3) to each scene in training file, point centered on a point is randomly selected, coordinate is (x, y, z),
(x-0.75, x+0.75) is taken around it, (y-0.75, y+0.75), the point in (z-0.75, z+0.75) range forms a number
According to block;
1.4) sampling number N is set0, points in data block and sampling number N that (1.3) are obtained0It is compared, sentences
Breaking, whether it is reasonable:
If the points in the data block are greater than sampling number N0, then it is reasonable to be judged to the data block, and in the data block with
Machine samples N0It is a, a sample data is formed, otherwise, abandons the data block, thus obtains training set T, in the present embodiment, N0
=8192;
1.5) for f all in test file1Each of a scene scene is 1.5 × 1.5 × 3 using size
Cube window carries out sliding window stripping and slicing, to each data block, stochastical sampling N0It is a, a sample data is formed, is tested
Collect V.
Step 2,3D point cloud semantic segmentation network is constructed.
Referring to Fig. 2, the 3D point cloud semantic segmentation network of the building of this step includes successively cascade feature down-sampling network,
Position notices that power module, feature up-sample network and auxiliary network.
2.1) feature down-sampling network is set:
This feature down-sampling network includes n cascade PointSA modules, and each PointSA module includes successively cascading
The sampling of point cloud mass center and packet layer, point cloud feature extraction layer, wherein n >=2, the parameter is set as n=4 in the present embodiment;
For the mass center sampling of m-th of PointSA module and packet layer, m=1,2 ..., n use iteration most first
Far point sampling method is from input point cluster samplingA point as center of mass point, secondly, withCentered on the center of mass point of a sampling, make
With spherical search algorithm, in its certain radius rmIn the range of search forIt is a, form a grouping.The 1st in the present embodiment
PointSA module, settingr1=0.1;2nd PointSA module, setting r2=0.2;3rd PointSA module, settingr3=0.4;4th PointSA module,
Settingr4=0.8;
For the point cloud feature extraction layer of m-th of PointSA module, including 3 successively cascade 2D convolutional layers, for mentioning
The feature of mass center sampling and packet layer output data is taken, and pond is carried out to the provincial characteristics extracted using maximum pondization strategy
Change.In the present embodiment the convolution kernel size of 3 2D convolutional layers of the point cloud feature extraction layer of the 1st PointSA module be 1 ×
1, step-length is 1, and output channel number is 32,32,64 respectively;3 2D of the point cloud feature extraction layer of the 2nd PointSA module
The convolution kernel size of convolutional layer is 1 × 1, and step-length is 1, and output channel number is 64,64,128 respectively;3rd PointSA mould
The convolution kernel size of 3 2D convolutional layers of the point cloud feature extraction layer of block is 1 × 1, and step-length is 1, output channel number difference
It is 128,128,256;The convolution kernel size of 3 2D convolutional layers of the point cloud feature extraction layer of the 4th PointSA module is 1
× 1, step-length is 1, and output channel number is 256,256,512 respectively;
2.2) setting position pays attention to power module, for calculating between feature representated by each mass center of its input data F
Correlation, obtain the strengthened feature E of position attention:
Referring to Fig. 3, the module working principle is as follows:
2.2.1) input data F passes through the first 1D convolutional layer Q respectively and obtains the feature Q of i-th of mass centeri, i=1,2 ...,
The mass center quantity of N, N expression F;Using the 2nd 1D convolutional layer U, the feature U of j-th of mass center is obtainedj, j=1,2 ..., N, then
The feature V of j-th of mass center is obtained by the 3rd 1D convolutional layer Vj;Wherein, the convolution kernel size of these three 1D convolutional layers Q, U, V is equal
It is 1, step-length is 1, and the output feature port number of the first 1D convolutional layer Q and the 2nd 1D convolutional layer U are that input data F feature is logical
Road numberThe output feature port number of 3rd 1D convolutional layer V is identical as the feature port number of input data F;
2.2.2 the attention influence value t between feature representated by each mass center) is calculatedij:Use tijConstitute matrix A:
2.2.3) calculating position attention feature
2.2.4 the strengthened feature E of attention) is exported:
E=[E1;E2;...;Ei;...;EN],
Wherein, Ei=α Ji+FiIndicate the feature of i-th of mass center in E, α indicates position attention feature JiWeight, FiTable
Show the feature of i-th of mass center of input;
2.3) setting feature up-samples network:
This feature up-sampling network include successively cascade a PointFP module, 1D convolutional layer, Dropout layers and be used for
The 1D convolutional layer of classification, each PointFP module include successively cascade feature interpolated layer and feature extraction layer, wherein a >=2,
The parameter is set as a=4 in the present embodiment;
The structure of a PointFP module, feature interpolated layer and feature extraction layer is different, in which:
For the 1st PointFP module, feature interpolated layer carries out interpolation to the output feature of position attention module,
And cascade the output feature of feature and the 3rd PointSA module after interpolation, the output for obtaining feature interpolated layer is special
Sign;Its feature extraction layer includes 2 successively cascade 2D convolutional layers, for further extracting the output feature, 2 2D convolution
It is 1 × 1 that the convolution kernel of layer, which is convolution kernel size, and step-length is 1, and output channel number is 256,256 respectively;
For the 2nd PointFP module, feature interpolated layer carries out the output feature of the 1st PointFP module slotting
Value, and the output feature of feature and the 2nd PointSA module after interpolation is cascaded, obtain the output of feature interpolated layer
Feature;Its feature extraction layer, comprising 2 successively cascade 2D convolutional layers, for further extracting the output feature, 2 2D volumes
The convolution kernel of lamination is that convolution kernel size is 1 × 1, and step-length is 1, and output channel number is 256,256 respectively;
For the 3rd PointFP module, feature interpolated layer carries out the output feature of the 2nd PointFP module slotting
Value, and the output feature of feature and the 1st PointSA module after interpolation is cascaded, obtain the output of feature interpolated layer
Feature;Its feature extraction layer, comprising 2 successively cascade 2D convolutional layers, for further extracting the output feature, 2 2D volumes
The convolution kernel of lamination is that convolution kernel size is 1 × 1, and step-length is 1, and output channel number is 256,128 respectively;
For the 4th PointFP module, feature interpolated layer carries out the output feature of the 3rd PointFP module slotting
It is worth, the feature after obtaining interpolation, output feature of the feature after the interpolation as its feature interpolated layer;Its feature extraction layer, packet
Containing 3 successively cascade 2D convolutional layers, for further extracting the output feature, the convolution kernel of 3 2D convolutional layers is convolution
Core size is 1 × 1, and step-length is 1, and output channel number is 128,128,128 respectively.
The 1D convolutional layer, convolution kernel size are 1, and step-length 1, output feature port number is set as 128;
It is Dropout layers described, retain probability and is set as 0.5;
The 1D convolutional layer for classification, convolution kernel size are 1, and step-length 1, output feature port number is set as point
The classification number L cut.
2.4) setting auxiliary network:
The auxiliary network includes successively cascade b PointAux module and the 1D convolutional layer for classification, each
PointAux module includes 1D convolutional layer and feature interpolated layer, wherein b >=1, b=2 in the present embodiment;
For the 1st PointAux module, 1D convolutional layer is used to extract the spy of the 2nd PointFP module output data
Sign, convolution kernel size are 1, step-length 1, and output feature channel is the classification number L of segmentation;Its feature interpolated layer is used to 1D convolution
The feature that layer extracts carries out interpolation;
For the 2nd PointAux module, 1D convolutional layer is used to extract the spy of the 1st PointAux module output data
Sign, convolution kernel size are 1, step-length 1, and output feature channel is the classification number L of segmentation;Its feature interpolated layer is used to 1D convolution
The feature that layer extracts carries out interpolation;
For the 1D convolutional layer of classification, classify for the output feature to the 2nd PointAux module, convolution kernel
Size is 1, step-length 1, and output feature port number is set as the classification number L of segmentation.
Step 3, the loss function of 3D point cloud semantic segmentation network is set.
This example is by polytypic cross entropy loss function, as the loss function of 3D point cloud semantic segmentation network, table
Show that formula is as follows:
Wherein, C represents the sample points of training, and L represents classification sum, wkFor the weight of kth class, waFor auxiliary network
The weight of loss, wa∈ [0,1], w in the present embodimenta=0.5;
pi,kThe true probability that i-th of sample point belongs to kth class is represented, if i-th of sample point belongs to kth class, probability value
It is 1, otherwise, probability value 0;
WithIt respectively indicates feature up-sampling network and i-th of sample point of neural network forecast is assisted to belong to kth class
Probability, calculation formula are as follows:
Wherein,It respectively indicates feature up-sampling network and assists k-th of i-th of sample point of network output and lead to
Road characteristic value, calculation formula are as follows:
Wherein, xiIndicate the input feature vector of i-th of sample point, f1Indicate that feature up-samples network, θ1Indicate feature up-sampling
The parameter of network, f2Indicate auxiliary network, θ2Indicate the parameter of auxiliary network.
Step 4, using training set T, the training that P wheel has supervision, P >=500 are carried out to 3D point cloud semantic segmentation network.
P=1000 is taken in the present embodiment, training step is as follows:
4.1) during q takes turns training, l is setqThe learning rate of training process is taken turns for q, and θ is setqIt takes turns and trains for q
The parameter of the network model of process uses formula according to the loss function that step 3 is setAdjust θq, obtain
To the network model parameter θ for q+1 wheel training processq+1, thus obtain q and take turns the network model after training process;
4.2) every P1Wheel, test set is input in current network model, all point cloud datas in test set are obtained
Predict classification, P1>=2, in the present embodiment, P1=5;
4.3) statistical test concentrates the prediction classification of all point cloud datas number identical with its true classification, calculates segmentation
Precision:Wherein, R indicates the prediction classification of all point cloud datas number identical with its true classification in test set,
H indicates the number of all point cloud datas in test set;
4.4) the segmentation precision acc of the segmentation precision acc of current network model and previously stored network model is carried out
Compare, if the segmentation precision acc of current network model is higher than the segmentation precision acc of previously stored network model, shows to work as
Preceding network model is more preferable, and saves to it, otherwise, without saving.
4.5) after the completion of the training of P wheel, using the highest network model of segmentation precision as trained network model;
Step 5, test set V is input in the trained network model that step 4.5) obtains and carries out semantic segmentation, obtained
To the segmentation result of each point.
Below in conjunction with emulation experiment, technical effect of the invention is explained:
1. simulated conditions
Emulation experiment of the invention carries out in following environment.
Hardware platform: Intel (R) Xeon (R) CPU E5-2650v4@2.20GHz, 64GB running memory,
Ubuntu16.04 operating system, GeForce GTX TITAN X;
Software platform: Tensorflow deep learning frame, Python3.5, data set used by testing is a cloud number
According to collection ScanNet.
ScanNet is the indoor scene point cloud data collection for scanning and rebuilding by RGB-D camera.Include in total
1513 scenes use 1201 scenes as training set, and for 312 scenes as test set, the classification number for including has 21 classes.
2. emulation experiment:
Training set and test set are obtained according to the present invention, 3D point cloud semantic segmentation network are constructed, using training set to 3D point
Cloud semantic segmentation network carries out Training, is then predicted using trained network model the point in test set,
3D point cloud segmentation network is calculated to the segmentation precision of test set V according to the method for step 4.3.
Compare the precision of the invention for doing semantic segmentation to point cloud data with existing PointNet++ method, and uses segmentation
Evaluation index of the precision as the evaluation present invention and prior art quality, the results are shown in Table 1:
1 ScanNet Segmentation of Data Set accuracy comparison table of table
Evaluation index | The prior art | The present invention |
Segmentation precision | 0.836 | 0.852 |
From table 1 it follows that segmentation precision of the present invention on ScanNet data set has been more than the prior art
Pointnet++ improves 1.6%, shows that the present invention is better than PointNet++ to the semantic segmentation effect of 3D point cloud.
Claims (9)
1. a kind of 3D point cloud semantic segmentation method based on position attention and auxiliary network, which is characterized in that include the following:
(1) from the training file and test file of the official website ScanNet downloading 3D point cloud data, and classification statistics is carried out to it and is cut
Block processing, obtains training set T and test set V;
(2) 3D point cloud semantic segmentation network is constructed comprising successively cascade feature down-sampling network, position pay attention to power module,
Feature up-samples network and auxiliary network;
(3) polytypic cross entropy loss function, the loss function as 3D point cloud semantic segmentation network are used;
(4) training set T is used, the training that P wheel has supervision is carried out to 3D point cloud data semantic segmentation network, P >=500:
(4a), according to the loss function of semantic segmentation network, adjusts network parameter, obtains network mould during the training of every wheel
Type;
(4b) is every P1Wheel, assesses the segmentation precision of current network model using the sample of test set, if current network mould
The segmentation precision of type is higher than previously stored network model, then is saved, P1≥2;
After the completion of (4c) P wheel training, using the highest network model of segmentation precision as trained network model;
(5) test set V is input in trained network model and carries out semantic segmentation, obtain the segmentation result of each point.
2. the method according to claim 1, wherein being carried out at classification statistics and stripping and slicing in (1) to point cloud data
Reason, is accomplished by
(1a) uses all f in statistics with histogram training file0The number of each classification of the point cloud data of a scene, and calculate each
The weight w of a classificationk:
Wherein, GkIndicate the number of kth class point cloud data, M indicates the number of all point cloud datas, and L indicates segmentation classification number, f0
>=1000, L >=2;
(1b) randomly selects point centered on a point, coordinate is (x, y, z), in its week to each scene in training file
It enclosing and takes (x-0.75, x+0.75), (y-0.75, y+0.75), the point in (z-0.75, z+0.75) range forms a data block,
And by the data block points and sampling number N0It is compared, judges whether it is reasonable:
If the points in the data block are greater than sampling number N0, then it is reasonable to be judged to the data block, and the stochastical sampling in the data block
N0It is a, a sample data is formed, otherwise, the data block is abandoned, thus obtains training set T, wherein N0≥4096;
(1c) is for f all in test file1Each of a scene scene, the cube for the use of size being 1.5 × 1.5 × 3
Window carries out sliding window stripping and slicing, to each data block, stochastical sampling N0It is a, a sample data is formed, test set V, f are obtained1≥
300。
3. the method according to claim 1, wherein feature down-sampling network described in (2), including n grade
The PointSA module of connection, each PointSA module include successively cascade cloud mass center sampling and packet layer, point Yun Tezheng
Extract layer, wherein n >=2.
4. the method according to claim 1, wherein position described in (2) pays attention to power module, including 3 solely
Vertical 1D convolutional layer Q, U, V, the feature of the input F for extracting the module, and calculate between feature representated by each mass center
Attention influence value tijWith the strengthened feature E of attention:
E=[E1;E2;...;Ei;...;EN]
Wherein, UiIndicate that position pays attention to feature of the input data F of power module by 1D convolutional layer U i-th of the mass center extracted, Qj T
Indicate that position notices that the transposition of the feature of j-th mass center of the input data F of power module by 1D convolutional layer Q extraction, N indicate F
Mass center quantity, EiIndicate the feature of i-th of mass center in E, calculation formula are as follows:
Wherein, VjIndicate the feature of j-th mass center of the F by 1D convolutional layer V extraction,It indicates by position attention
The feature of i-th of mass center afterwards, α indicate the weight of position attention feature, FiIndicate the feature of i-th of mass center of input.
5. the method according to claim 1, wherein feature described in (2) up-samples network, including successively grade
A PointFP module of connection, 1D convolutional layer, Dropout layers and for classification 1D convolutional layer, each PointFP module includes
Successively cascade feature interpolated layer and feature extraction layer, wherein a >=2.
6. the method according to claim 1, wherein auxiliary network described in (2), including successively cascade b
A PointAux module and 1D convolutional layer for classification, each PointAux module include 1D convolutional layer and feature interpolated layer,
Wherein, b >=1.
7. the method according to claim 1, wherein 3D point cloud semantic segmentation network described in step (3)
Loss function, calculation formula are as follows:
Wherein, C represents the sample points of training, and L represents classification sum, wkFor the weight of kth class, waFor the loss for assisting network
Weight, wa∈[0,1];pi,kThe true probability that i-th of sample point belongs to kth class is represented, if i-th of sample point belongs to kth
Class, then probability value is 1, otherwise, probability value 0;WithIt respectively indicates feature up-sampling network and assists neural network forecast
I-th of sample point belongs to the probability of kth class,WithCalculation formula it is as follows:
Wherein,It respectively indicates feature up-sampling network and assists k-th of channel of i-th of sample point of network output special
Value indicative, calculation formula are as follows:
Wherein, xiIndicate the input feature vector of i-th of sample point, f1Indicate that feature up-samples network, θ1Indicate that feature up-samples network
Parameter, f2Indicate auxiliary network, θ2Indicate the parameter of auxiliary network.
8. the method according to claim 1, wherein being adjusted in (4a) according to the loss function of semantic segmentation network
Whole network parameter is carried out by following formula:
Wherein, lqIndicate the learning rate of q wheel training process, θqIndicate the 3D point cloud semantic segmentation network of q wheel training process
Parameter, θq+1It indicates to θqParameter after adjustment, for q+1 wheel training process.
9. the method according to claim 1, wherein every P in (4b)1Wheel, to the segmentation essence of current network model
Degree is assessed, and is accomplished by
(4b1) is every P1Wheel, test set is input in current network model, the prediction of all point cloud datas in test set is obtained
Classification;
(4b2) statistical test concentrates the prediction classification of all point cloud datas number identical with its true classification, calculates segmentation essence
Degree:Wherein, R indicates the prediction classification of all point cloud datas number identical with its true classification, H in test set
Indicate the number of all point cloud datas in test set;
(4b3) compares the segmentation precision acc of current network model and the segmentation precision acc of previously stored network model
Compared with if segmentation precision acc of the segmentation precision acc of current network model higher than previously stored network model, shows currently
Network model is more preferable, and saves to it, otherwise, without saving.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910604264.0A CN110322453B (en) | 2019-07-05 | 2019-07-05 | 3D point cloud semantic segmentation method based on position attention and auxiliary network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910604264.0A CN110322453B (en) | 2019-07-05 | 2019-07-05 | 3D point cloud semantic segmentation method based on position attention and auxiliary network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110322453A true CN110322453A (en) | 2019-10-11 |
CN110322453B CN110322453B (en) | 2023-04-18 |
Family
ID=68122807
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910604264.0A Active CN110322453B (en) | 2019-07-05 | 2019-07-05 | 3D point cloud semantic segmentation method based on position attention and auxiliary network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110322453B (en) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110827398A (en) * | 2019-11-04 | 2020-02-21 | 北京建筑大学 | Indoor three-dimensional point cloud automatic semantic segmentation algorithm based on deep neural network |
CN111127493A (en) * | 2019-11-12 | 2020-05-08 | 中国矿业大学 | Remote sensing image semantic segmentation method based on attention multi-scale feature fusion |
CN111192270A (en) * | 2020-01-03 | 2020-05-22 | 中山大学 | Point cloud semantic segmentation method based on point global context reasoning |
CN111223120A (en) * | 2019-12-10 | 2020-06-02 | 南京理工大学 | Point cloud semantic segmentation method |
CN111428619A (en) * | 2020-03-20 | 2020-07-17 | 电子科技大学 | Three-dimensional point cloud head attitude estimation system and method based on ordered regression and soft labels |
CN111583263A (en) * | 2020-04-30 | 2020-08-25 | 北京工业大学 | Point cloud segmentation method based on joint dynamic graph convolution |
CN112560865A (en) * | 2020-12-23 | 2021-03-26 | 清华大学 | Semantic segmentation method for point cloud under outdoor large scene |
CN112633330A (en) * | 2020-12-06 | 2021-04-09 | 西安电子科技大学 | Point cloud segmentation method, system, medium, computer device, terminal and application |
CN112927248A (en) * | 2021-03-23 | 2021-06-08 | 重庆邮电大学 | Point cloud segmentation method based on local feature enhancement and conditional random field |
CN113205509A (en) * | 2021-05-24 | 2021-08-03 | 山东省人工智能研究院 | Blood vessel plaque CT image segmentation method based on position convolution attention network |
CN113470048A (en) * | 2021-07-06 | 2021-10-01 | 北京深睿博联科技有限责任公司 | Scene segmentation method, device, equipment and computer readable storage medium |
CN115619963A (en) * | 2022-11-14 | 2023-01-17 | 吉奥时空信息技术股份有限公司 | City building entity modeling method based on content perception |
WO2023072175A1 (en) * | 2021-10-30 | 2023-05-04 | 华为技术有限公司 | Point cloud data processing method, neural network training method, and related device |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102034267A (en) * | 2010-11-30 | 2011-04-27 | 中国科学院自动化研究所 | Three-dimensional reconstruction method of target based on attention |
CN102036073A (en) * | 2010-12-21 | 2011-04-27 | 西安交通大学 | Method for encoding and decoding JPEG2000 image based on vision potential attention target area |
CN103871050A (en) * | 2014-02-19 | 2014-06-18 | 小米科技有限责任公司 | Image partition method, device and terminal |
US20190026956A1 (en) * | 2012-02-24 | 2019-01-24 | Matterport, Inc. | Employing three-dimensional (3d) data predicted from two-dimensional (2d) images using neural networks for 3d modeling applications and other applications |
US20190108639A1 (en) * | 2017-10-09 | 2019-04-11 | The Board Of Trustees Of The Leland Stanford Junior University | Systems and Methods for Semantic Segmentation of 3D Point Clouds |
US20190147245A1 (en) * | 2017-11-14 | 2019-05-16 | Nuro, Inc. | Three-dimensional object detection for autonomous robotic systems using image proposals |
CN109871532A (en) * | 2019-01-04 | 2019-06-11 | 平安科技(深圳)有限公司 | Text subject extracting method, device and storage medium |
-
2019
- 2019-07-05 CN CN201910604264.0A patent/CN110322453B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102034267A (en) * | 2010-11-30 | 2011-04-27 | 中国科学院自动化研究所 | Three-dimensional reconstruction method of target based on attention |
CN102036073A (en) * | 2010-12-21 | 2011-04-27 | 西安交通大学 | Method for encoding and decoding JPEG2000 image based on vision potential attention target area |
US20190026956A1 (en) * | 2012-02-24 | 2019-01-24 | Matterport, Inc. | Employing three-dimensional (3d) data predicted from two-dimensional (2d) images using neural networks for 3d modeling applications and other applications |
CN103871050A (en) * | 2014-02-19 | 2014-06-18 | 小米科技有限责任公司 | Image partition method, device and terminal |
US20190108639A1 (en) * | 2017-10-09 | 2019-04-11 | The Board Of Trustees Of The Leland Stanford Junior University | Systems and Methods for Semantic Segmentation of 3D Point Clouds |
US20190147245A1 (en) * | 2017-11-14 | 2019-05-16 | Nuro, Inc. | Three-dimensional object detection for autonomous robotic systems using image proposals |
CN109871532A (en) * | 2019-01-04 | 2019-06-11 | 平安科技(深圳)有限公司 | Text subject extracting method, device and storage medium |
Non-Patent Citations (3)
Title |
---|
JUN FU ET AL.: "Dual Attention Network for Scene Segmentation", 《ARXIV.ORG》 * |
LEI WANG ET AL.: "Graph Attention Convolution for Point Cloud Semantic Segmentation", 《2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 * |
陈晓范等: "结合注意力机制的人脸超分辨率重建", 《西安电子科技大学学报》 * |
Cited By (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110827398A (en) * | 2019-11-04 | 2020-02-21 | 北京建筑大学 | Indoor three-dimensional point cloud automatic semantic segmentation algorithm based on deep neural network |
CN110827398B (en) * | 2019-11-04 | 2023-12-26 | 北京建筑大学 | Automatic semantic segmentation method for indoor three-dimensional point cloud based on deep neural network |
CN111127493A (en) * | 2019-11-12 | 2020-05-08 | 中国矿业大学 | Remote sensing image semantic segmentation method based on attention multi-scale feature fusion |
CN111223120A (en) * | 2019-12-10 | 2020-06-02 | 南京理工大学 | Point cloud semantic segmentation method |
CN111192270A (en) * | 2020-01-03 | 2020-05-22 | 中山大学 | Point cloud semantic segmentation method based on point global context reasoning |
CN111428619B (en) * | 2020-03-20 | 2022-08-05 | 电子科技大学 | Three-dimensional point cloud head attitude estimation system and method based on ordered regression and soft labels |
CN111428619A (en) * | 2020-03-20 | 2020-07-17 | 电子科技大学 | Three-dimensional point cloud head attitude estimation system and method based on ordered regression and soft labels |
CN111583263B (en) * | 2020-04-30 | 2022-09-23 | 北京工业大学 | Point cloud segmentation method based on joint dynamic graph convolution |
CN111583263A (en) * | 2020-04-30 | 2020-08-25 | 北京工业大学 | Point cloud segmentation method based on joint dynamic graph convolution |
CN112633330A (en) * | 2020-12-06 | 2021-04-09 | 西安电子科技大学 | Point cloud segmentation method, system, medium, computer device, terminal and application |
CN112633330B (en) * | 2020-12-06 | 2024-02-02 | 西安电子科技大学 | Point cloud segmentation method, system, medium, computer equipment, terminal and application |
CN112560865A (en) * | 2020-12-23 | 2021-03-26 | 清华大学 | Semantic segmentation method for point cloud under outdoor large scene |
CN112560865B (en) * | 2020-12-23 | 2022-08-12 | 清华大学 | Semantic segmentation method for point cloud under outdoor large scene |
CN112927248A (en) * | 2021-03-23 | 2021-06-08 | 重庆邮电大学 | Point cloud segmentation method based on local feature enhancement and conditional random field |
CN113205509A (en) * | 2021-05-24 | 2021-08-03 | 山东省人工智能研究院 | Blood vessel plaque CT image segmentation method based on position convolution attention network |
CN113205509B (en) * | 2021-05-24 | 2021-11-09 | 山东省人工智能研究院 | Blood vessel plaque CT image segmentation method based on position convolution attention network |
CN113470048A (en) * | 2021-07-06 | 2021-10-01 | 北京深睿博联科技有限责任公司 | Scene segmentation method, device, equipment and computer readable storage medium |
WO2023072175A1 (en) * | 2021-10-30 | 2023-05-04 | 华为技术有限公司 | Point cloud data processing method, neural network training method, and related device |
CN115619963A (en) * | 2022-11-14 | 2023-01-17 | 吉奥时空信息技术股份有限公司 | City building entity modeling method based on content perception |
Also Published As
Publication number | Publication date |
---|---|
CN110322453B (en) | 2023-04-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110322453A (en) | 3D point cloud semantic segmentation method based on position attention and auxiliary network | |
CN110245709B (en) | 3D point cloud data semantic segmentation method based on deep learning and self-attention | |
CN111832655B (en) | Multi-scale three-dimensional target detection method based on characteristic pyramid network | |
CN109344821A (en) | Small target detecting method based on Fusion Features and deep learning | |
CN108805070A (en) | A kind of deep learning pedestrian detection method based on built-in terminal | |
CN107609525A (en) | Remote Sensing Target detection method based on Pruning strategy structure convolutional neural networks | |
CN110175613A (en) | Street view image semantic segmentation method based on Analysis On Multi-scale Features and codec models | |
CN111160108A (en) | Anchor-free face detection method and system | |
CN105956560A (en) | Vehicle model identification method based on pooling multi-scale depth convolution characteristics | |
CN112487862B (en) | Garage pedestrian detection method based on improved EfficientDet model | |
CN105243139A (en) | Deep learning based three-dimensional model retrieval method and retrieval device thereof | |
CN113160062B (en) | Infrared image target detection method, device, equipment and storage medium | |
CN111860587B (en) | Detection method for small targets of pictures | |
CN104809469A (en) | Indoor scene image classification method facing service robot | |
CN110334584B (en) | Gesture recognition method based on regional full convolution network | |
CN108446616A (en) | Method for extracting roads based on full convolutional neural networks integrated study | |
CN112750125B (en) | Glass insulator piece positioning method based on end-to-end key point detection | |
CN115909052A (en) | Hyperspectral remote sensing image classification method based on hybrid convolutional neural network | |
CN112950780B (en) | Intelligent network map generation method and system based on remote sensing image | |
CN109800756A (en) | A kind of text detection recognition methods for the intensive text of Chinese historical document | |
CN111339924A (en) | Polarized SAR image classification method based on superpixel and full convolution network | |
CN107832753B (en) | Face feature extraction method based on four-value weight and multiple classification | |
CN109859222A (en) | Edge extracting method and system based on cascade neural network | |
CN109671055A (en) | Pulmonary nodule detection method and device | |
CN113032613B (en) | Three-dimensional model retrieval method based on interactive attention convolution neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |