CN111738306A - Multi-view three-dimensional model retrieval method based on block convolution neural network - Google Patents

Multi-view three-dimensional model retrieval method based on block convolution neural network Download PDF

Info

Publication number
CN111738306A
CN111738306A CN202010487922.5A CN202010487922A CN111738306A CN 111738306 A CN111738306 A CN 111738306A CN 202010487922 A CN202010487922 A CN 202010487922A CN 111738306 A CN111738306 A CN 111738306A
Authority
CN
China
Prior art keywords
view
dimensional
model
block
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010487922.5A
Other languages
Chinese (zh)
Other versions
CN111738306B (en
Inventor
高赞
邵煜翔
程志勇
陈达
舒明雷
聂礼强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shandong Computer Science Center National Super Computing Center in Jinan
Shandong Institute of Artificial Intelligence
Original Assignee
Shandong Computer Science Center National Super Computing Center in Jinan
Shandong Institute of Artificial Intelligence
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shandong Computer Science Center National Super Computing Center in Jinan, Shandong Institute of Artificial Intelligence filed Critical Shandong Computer Science Center National Super Computing Center in Jinan
Priority to CN202010487922.5A priority Critical patent/CN111738306B/en
Publication of CN111738306A publication Critical patent/CN111738306A/en
Application granted granted Critical
Publication of CN111738306B publication Critical patent/CN111738306B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/53Querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24147Distances to closest patterns, e.g. nearest neighbour classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Abstract

A multi-view three-dimensional model retrieval method based on a block convolutional neural network is used for mining intrinsic relations among views in the process of extracting view features by using block convolutional layers aiming at multi-view images. Different weights are distributed to each view according to cosine similarity between each view feature and the features after the maximal view pooling, and the distinguishing performance between the view features is utilized to obtain more distinguishing model features. When the loss function is generated, not only the model characteristic but also the view characteristic are considered, so that the network can be better constrained to learn. The multi-view three-dimensional model retrieval method based on the block convolutional neural network achieves excellent performance in the relevant three-dimensional model retrieval data set.

Description

Multi-view three-dimensional model retrieval method based on block convolution neural network
Technical Field
The invention relates to the field of three-dimensional vision, in particular to a multi-view three-dimensional model retrieval method based on a block convolution neural network.
Background
Three-dimensional vision is receiving increasing attention from researchers as three-dimensional representation technology and computer hardware capabilities develop. Three-dimensional vision is a more realistic description of the real world, containing spatial structure information of three-dimensional objects and characteristics of solid geometry and contour curves, compared to traditional two-dimensional images. Three-dimensional model retrieval is a research hotspot in the field of three-dimensional vision. The related research method can be divided into two stages, namely (1) a three-dimensional model retrieval method based on the traditional method and (2) a three-dimensional model retrieval method based on deep learning.
The three-dimensional model retrieval method based on the conventional method is to generate feature descriptors from information of the three-dimensional model, such as geometric moments, surface distribution, volume descriptors and surface geometry, and then to perform a measure of similarity between features using euclidean distances. However, in practical applications, the traditional method for rendering and feature extraction of the three-dimensional model is not only computationally intensive but also not ideal. These difficulties have limited the development of the related conventional methods. The three-dimensional model retrieval method based on deep learning comprises a model-based retrieval algorithm and a multi-view-based retrieval algorithm. The model-based search algorithm is to directly extract features on a three-dimensional model by using a deep neural network, and can be divided into a voxel-based method, a grid-based method and a point cloud-based method according to different representations of the model. Since deep learning has made a significant development in the related task of two-dimensional images, multi-view based three-dimensional model algorithms have been proposed. Images of a three-dimensional model at various angles are taken by a camera set at a plurality of angles, and the three-dimensional model is represented by two-dimensional images at different angles. And processing the strong performance of the two-dimensional image by means of a deep neural network such as AlexNet, VGG, GoogleNet, ResNet and the like, extracting the characteristics of each view, and fusing all the view characteristics to obtain the final model characteristics. In retrieving matches, Euclidean distances are used to make a measure of similarity between features.
In a multi-view-based three-dimensional model retrieval method, each three-dimensional model is collectively represented by multiple views. The existing method extracts the features of each view respectively, and then mines the relationship among the features of the views. The relationships between the views are not taken into account in the process of view feature extraction. How to mine the relationship between the views in the extraction process of the view features becomes a key for improving the performance of the retrieval algorithm.
Disclosure of Invention
The invention provides a multi-view three-dimensional model retrieval method based on a block convolutional neural network to overcome the defects of the above technology, which can realize the relation mining among multiple views according to the block characteristics and block coordinates of each view, mine the discriminative information among the views by distributing different weights of the views through a self-adaptive view weight layer, and extract more discriminative model characteristics by using a discriminative loss function constraint network in training.
The technical scheme adopted by the invention for overcoming the technical problems is as follows:
a multi-view three-dimensional model retrieval method based on a block convolutional neural network comprises the following steps:
a) rendering the three-dimensional model to obtain N two-dimensional views of the three-dimensional model;
b) inputting each two-dimensional view into a convolutional neural network, wherein each two-dimensional view obtains M feature blocks, and the feature dimension of each feature block is P dimension;
c) searching k adjacent blocks of each feature block in a P-dimensional feature space according to Euclidean distances, constructing an adjacent graph by using the central block and the k adjacent blocks, wherein each adjacent graph comprises k +1 vertexes and k undirected edges, and the k undirected edges are respectively connected with the central block and the 12 adjacent blocks, and adopting (P)i,pj-pi) Edge features representing undirected edges, where piAs a central block, pjFor the jth adjacent block, the three-dimensional coordinates (x, y, z) are used to represent the block in the xth row and the yth column of the zth two-dimensional view, and the edge feature between the two blocks is represented as (p)i,pj-pi,ci,cj-ci),ciThree-dimensional coordinates of the central block, cjIs the three-dimensional coordinate of the jth adjacent block, the dimension of the edge feature is E dimension, and the three-dimensional coordinate is obtained by a formula
Figure BDA0002518062540000021
Performing convolution operation on the block to obtain new block characteristics pi', where h is a nonlinear function, and N (i) is a neighborhood sample obtained by the ith sample through a KNN algorithm;
d) new block characteristics piUsing average pooling layer fusion block characteristics to obtain view characteristics of each two-dimensional view, using one-dimensional convolution to extract the relation between adjacent views, obtaining pooled characteristics g of the two-dimensional views through maximum view pooling, and obtaining pooled characteristics g of the two-dimensional views through a formula
Figure BDA0002518062540000022
Computing the pooled feature g and the jth view feature fjCosine similarity s betweenj(fjG), where D is the characteristic dimension and k is the kth dimension, by means of the formula
Figure BDA0002518062540000031
Calculating the weight α of the cosine similarity of the jth two-dimensional viewjIn the formula si(fiG) features g after pooling and the ith view feature fiCosine similarity between them, by the formula fj′=αj×fjCalculating weighted view features fj', by the formula
Figure BDA0002518062540000032
Calculating weighted model characteristics g';
e) according to the formula LDis=β×Lmodel+γ×LviewsCalculating a loss function LDisIn the formula LmodelGenerating for model features gValue of loss function of LviewsAs a view feature fj' value of the resulting loss function, β is denoted LmodelWeight of the super parameter, gamma is LviewsThe size of the weight is super-parameter,
Figure BDA0002518062540000033
in the formula
Figure BDA0002518062540000034
A loss function value generated for a view feature of a jth two-dimensional view;
f) for a given three-dimensional model, the three-dimensional model that most closely resembles the model is retrieved from the target dataset.
Further, the three-dimensional model in the step a) is a three-dimensional model generated by a computer, and the computer renders the three-dimensional model to obtain a plurality of two-dimensional views.
Further, the three-dimensional model in the step a) is a real-world object, and X cameras with different angles are arranged around the object to acquire two-dimensional views of the object at different angles.
Further, one camera is disposed every 30 degrees in the circumferential direction around the object, and 12 two-dimensional views are obtained by disposing 12 cameras.
Further, each two-dimensional view in step b) is represented by 7 × 7 block features after passing through the convolutional neural network, and the feature dimension of each block feature is 512.
Further, k is set to 12 and E is set to 1030 in step c).
Further, using 7 × 7 average pooling layer fusion block features in step d), a one-dimensional convolution with a convolution kernel size of 3 is used to extract the relationship between adjacent views.
Further, in step e), β is 0.5 and γ is 0.5.
Further, the loss value weight of each two-dimensional view is set to α in step e)j′,
Figure BDA0002518062540000041
By the formula
Figure BDA0002518062540000042
Computing view characteristics fj' loss function value generated.
Further, in step f), model characteristics of the retrieved model and all models in the query set are respectively extracted, and a formula is utilized
Figure BDA0002518062540000043
Calculating to obtain Euclidean distance D (q, p) between model feature p of the retrieved model and model features q of all models in the query set, wherein p in the formulaiFor the feature value of the i-th dimension in the model feature p of the model to be retrieved, qiFor querying the characteristic value of ith dimension in model characteristic q of all models in the set, n is the characteristic dimension, and by the formula Redist (q)i,gj)=1-prob(gj|label=qi) × D (p, q) calculating the rearranged distance Redist (q)i,gj) In which prob (g)j|label=qi) Is the probability of belonging to the same class as the retrieved model.
The invention has the beneficial effects that: by using block convolution layers for multi-view images, the intrinsic relations between views are mined in the extraction process of view features. Different weights are distributed to each view according to cosine similarity between each view feature and the features after the maximal view pooling, and the distinguishing performance between the view features is utilized to obtain more distinguishing model features. When the loss function is generated, not only the model characteristic but also the view characteristic are considered, so that the network can be better constrained to learn. The multi-view three-dimensional model retrieval method based on the block convolutional neural network achieves excellent performance in the relevant three-dimensional model retrieval data set.
Drawings
FIG. 1 is a flow chart of the present invention.
Detailed Description
The invention is further described below with reference to fig. 1.
A multi-view three-dimensional model retrieval method based on a block convolutional neural network comprises the following steps:
a) and rendering the three-dimensional model to obtain N two-dimensional views of the three-dimensional model. The three-dimensional model may be a computer-generated three-dimensional model, the computer rendering the three-dimensional model resulting in a plurality of two-dimensional views. The three-dimensional model can also be a real world object, and X cameras with different angles are arranged around the object to acquire two-dimensional views of the object at different angles. Preferably, one camera is arranged at intervals of 30 degrees in the circumferential direction around the object, and 12 two-dimensional views are obtained by arranging 12 cameras.
b) In the previous multi-view method, the obtained multiple views are directly and simultaneously input into a convolutional neural network with shared weight to obtain the view characteristics of each view. In the method, after the block features are obtained, view features are not directly obtained by using average pooling, relationship mining needs to be carried out on all the block features, each two-dimensional view is input into a convolutional neural network, M feature blocks are obtained from each two-dimensional view, and the feature dimension of each feature block is P dimension. Each three-dimensional model is now represented collectively by 7 x 7 block features of 12 views, totaling 588 block features.
c) Aiming at the problem that the relation among multiple views cannot be mined in the process of extracting view features by the current method, a new neural network module is provided, and the neural network module is called block convolution. Relationship mining between multiple views can be implemented according to block features and block coordinates of each view. Firstly, searching k adjacent blocks of each feature block in a P-dimensional feature space according to Euclidean distance, wherein k can be set to be 12, constructing an adjacent map by using a central block and the k adjacent blocks, wherein each adjacent map comprises k +1 vertexes and k undirected edges, and the k undirected edges are respectively connected with the central block and the 12 adjacent blocks, and adopting (P)i,pj-pi) Edge features representing undirected edges, where piAs a central block, pjFor the jth neighbor block, considering that each block feature has its specific location in the view, the block is represented in the xth row and yth column of the z-th two-dimensional view using three-dimensional coordinates (x, y, z), and the edge feature between two blocks is represented by (p) by adding the coordinate information of the blocki,pj-pi,ci,cj-ci),ciThree-dimensional coordinates of the central block, cjFor the three-dimensional seating of the jth neighbor blockDimension E of edge feature is set to 1030 through formula
Figure BDA0002518062540000051
Performing convolution operation on the block to obtain new block characteristics pi' where h is a non-linear function, and is implemented by convolution operation of 1 × 1, reducing the edge feature of 1030 dimension to 515 dimension, n (i) a neighborhood sample obtained by KNN algorithm for the ith sample, new block feature contains information between the current block and the neighboring block in the feature space.
d) New block characteristics piThe present invention contemplates adaptive view weight layers assigning different weights to each view, unlike the previous fusion method using max-view pooling, the invention obtains pooled feature g of two-dimensional views by max-view pooling, by formulating a formula
Figure BDA0002518062540000061
Computing the pooled feature g and the jth view feature fjCosine similarity s betweenj(fjG), where D is the characteristic dimension and k is the kth dimension, by means of the formula
Figure BDA0002518062540000062
Calculating the weight α of the cosine similarity of the jth two-dimensional viewjIn the formula si(fiG) features g after pooling and the ith view feature fiCosine similarity between them, by the formula fj′=αj×fjCalculating weighted view features fj', by the formula
Figure BDA0002518062540000063
Calculating weighted model characteristics g';
e) previous methods all require the network to accurately classify the model features only, thereby producing loss function values. In this process, the view characteristics are not taken into account. Although each view feature cannot represent a model individually, it contains information about the model at a particular angle. Two classifiers are set in the network, which require accurate classification of model features and view features, respectively. The loss function is thus composed of two parts, one part of the loss function being calculated from the model features and the other part of the loss function being calculated from the view features. The obtained discriminative loss function can better constrain the network to train and extract more discriminative model features. According to the formula LDis=β×Lmodel+γ×LviewsCalculating a loss function LDisIn the formula LmodelValue of the loss function generated for model feature g', LviewsAs a view feature fj' value of the resulting loss function, β is denoted LmodelWeight of the super parameter, gamma is LviewsThe weight of the super parameter is calculated in the process of calculating LviewsIn the process of (1), firstly, an averaging mode is used to take the average value of the loss function value of each view,
Figure BDA0002518062540000064
in the formula
Figure BDA0002518062540000065
The value of the loss function generated for the view feature of the jth two-dimensional view is preferably β ═ 0.5 and γ ═ 0.5.
f) For a given three-dimensional model, the three-dimensional model that most closely resembles the model is retrieved from the target dataset.
By using block convolution layers for multi-view images, the intrinsic relations between views are mined in the extraction process of view features. Different weights are distributed to each view according to cosine similarity between each view feature and the features after the maximal view pooling, and the distinguishing performance between the view features is utilized to obtain more distinguishing model features. When the loss function is generated, not only the model characteristic but also the view characteristic are considered, so that the network can be better constrained to learn. The multi-view three-dimensional model retrieval method based on the block convolutional neural network achieves excellent performance in the relevant three-dimensional model retrieval data set.
Each view is assigned a different weight, with smaller weights indicating that the view features are less similar to the maximally pooled features, indicating that the view is more difficult to resolve. If the network can focus more on these views that are difficult to distinguish, more discriminative information can be obtained, which is very beneficial to the retrieval of the model. Weighting the loss value of each view to
Figure 1
Indicating that views with low view weights are assigned higher penalty weights and views with high view weights are assigned smaller penalty weights, thereby allowing the indistinguishable views to have a greater effect on feature extraction step e) to set the penalty weight of each two-dimensional view to αj′,
Figure BDA0002518062540000072
By the formula
Figure BDA0002518062540000073
Computing view characteristics fj' loss function value generated. Given any one three-dimensional model, the three-dimensional model most similar to the model is retrieved in the target dataset. The process of retrieval is as follows: and respectively extracting model features of the retrieved model and all models in the query set, calculating Euclidean distances pairwise, sequencing the models in the query set according to the Euclidean distances, and returning a query result according to the sequence. Thus, in step f) the models of the retrieved model and all models in the query set are extracted separatelyCharacteristic of, using a formula
Figure BDA0002518062540000074
Calculating to obtain Euclidean distance D (q, p) between model feature p of the retrieved model and model features q of all models in the query set, wherein p in the formulaiFor the feature value of the i-th dimension in the model feature p of the model to be retrieved, qiThe feature value of the ith dimension in the model feature q of all models in the query set is defined, and n is the feature dimension. In the retrieval task, the initially obtained sequencing result can be rearranged so as to improve the effect. In the present invention, the classification information of the model features is reordered from the initial ranking results by means of a trained network. First, the network classifies the retrieved model and determines that it belongs to a certain class. Then, all models in the query dataset are classified, resulting in a probability that they belong to each class respectively. By the formula Redist (q)i,gj)=1-prob(gj|label=qi) × D (p, q) calculating the rearranged distance Redist (q)i,gj) In which prob (g)j|label=qi) Is the probability of belonging to the same class as the retrieved model. This means that the higher the probability of belonging to the same class, the smaller the weight of their euclidean distance will be compared to the model with lower probability, so that the euclidean distance between the two becomes smaller. By means of the classification capability of the network, the effect of network retrieval is improved.

Claims (10)

1. A multi-view three-dimensional model retrieval method based on a block convolution neural network is characterized by comprising the following steps:
a) rendering the three-dimensional model to obtain N two-dimensional views of the three-dimensional model;
b) inputting each two-dimensional view into a convolutional neural network, wherein each two-dimensional view obtains M feature blocks, and the feature dimension of each feature block is P dimension;
c) searching k adjacent blocks of each feature block in a P-dimensional feature space according to Euclidean distances, constructing an adjacent graph by using the central block and the k adjacent blocks, wherein each adjacent graph comprises k +1 vertexes and k undirected edges, and the k undirected edges are respectively connected with the central block and the k undirected edges12 neighboring blocks, using (p)i,pj-pi) Edge features representing undirected edges, where piAs a central block, pjFor the jth adjacent block, the three-dimensional coordinates (x, y, z) are used to represent the block in the xth row and the yth column of the zth two-dimensional view, and the edge feature between the two blocks is represented as (p)i,pj-pi,ci,cj-ci),ciThree-dimensional coordinates of the central block, cjIs the three-dimensional coordinate of the jth adjacent block, the dimension of the edge feature is E dimension, and the three-dimensional coordinate is obtained by a formula
Figure FDA0002518062530000011
Performing convolution operation on the block to obtain new block characteristics pi', where h is a nonlinear function, and N (i) is a neighborhood sample obtained by the ith sample through a KNN algorithm;
d) new block characteristics piUsing average pooling layer fusion block characteristics to obtain view characteristics of each two-dimensional view, using one-dimensional convolution to extract the relation between adjacent views, obtaining pooled characteristics g of the two-dimensional views through maximum view pooling, and obtaining pooled characteristics g of the two-dimensional views through a formula
Figure FDA0002518062530000012
Computing the pooled feature g and the jth view feature fjCosine similarity s betweenj(fjG), where D is the characteristic dimension and k is the kth dimension, by means of the formula
Figure FDA0002518062530000013
Calculating the weight α of the cosine similarity of the jth two-dimensional viewjIn the formula si(fiG) features g after pooling and the ith view feature fiCosine similarity between them, by the formula fj′=αj×fjCalculating weighted view features fj', by the formula
Figure FDA0002518062530000021
Calculating weighted model characteristics g';
e) according to the formula LDis=β×Lmodel+γ×LviewsCalculating a loss function LDisIn the formula LmodelValue of the loss function generated for model feature g', LviewsAs a view feature fj' value of the resulting loss function, β is denoted LmodelWeight of the super parameter, gamma is LviewsThe size of the weight is super-parameter,
Figure FDA0002518062530000022
in the formula
Figure FDA0002518062530000023
A loss function value generated for a view feature of a jth two-dimensional view;
f) for a given three-dimensional model, the three-dimensional model that most closely resembles the model is retrieved from the target dataset.
2. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 1, wherein: the three-dimensional model in the step a) is a three-dimensional model generated by a computer, and the computer renders the three-dimensional model to obtain a plurality of two-dimensional views.
3. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 1, wherein: the three-dimensional model in the step a) is a real-world object, and X cameras with different angles are arranged around the object to acquire two-dimensional views of the object at different angles.
4. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 3, wherein: one camera is arranged every 30 degrees in the circumferential direction around the object, and 12 two-dimensional views are obtained by arranging 12 cameras.
5. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 1, wherein: each two-dimensional view in the step b) is represented by 7 multiplied by 7 block features after passing through the convolutional neural network, and the feature dimension of each block feature is 512.
6. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 1, wherein: in step c) k is set to 12 and E is set to 1030.
7. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 1, wherein: using 7 × 7 average pooling layer fusion block features in step d), extracting the relationship between adjacent views using a one-dimensional convolution with a convolution kernel size of 3.
8. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 1, wherein: in step e), β is 0.5 and γ is 0.5.
9. The method for retrieving the multi-view three-dimensional model based on the block convolutional neural network as claimed in claim 1, wherein the weight of the loss value of each two-dimensional view is set to α in step e)j′,
Figure FDA0002518062530000031
By the formula
Figure FDA0002518062530000032
Computing view characteristics fj' loss function value generated.
10. The block convolutional neural network-based multi-view three-dimensional model retrieval method of claim 1, wherein: in step f), the model characteristics of the retrieved model and all models in the query set are respectively extracted, and a formula is utilized
Figure FDA0002518062530000033
Is calculated to obtain the quiltSearching Euclidean distance D (q, p) between model characteristic p of model and model characteristic q of all models in query set, wherein p isiFor the feature value of the i-th dimension in the model feature p of the model to be retrieved, qiFor querying the characteristic value of ith dimension in model characteristic q of all models in the set, n is the characteristic dimension, and by the formula Redist (q)i,gj)=1-prob(gj|label=qi) × D (p, q) calculating the rearranged distance Redist (q)i,gj) In which prob (g)j|label=qi) Is the probability of belonging to the same class as the retrieved model.
CN202010487922.5A 2020-06-01 2020-06-01 Multi-view three-dimensional model retrieval method based on block convolution neural network Active CN111738306B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010487922.5A CN111738306B (en) 2020-06-01 2020-06-01 Multi-view three-dimensional model retrieval method based on block convolution neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010487922.5A CN111738306B (en) 2020-06-01 2020-06-01 Multi-view three-dimensional model retrieval method based on block convolution neural network

Publications (2)

Publication Number Publication Date
CN111738306A true CN111738306A (en) 2020-10-02
CN111738306B CN111738306B (en) 2022-05-13

Family

ID=72646627

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010487922.5A Active CN111738306B (en) 2020-06-01 2020-06-01 Multi-view three-dimensional model retrieval method based on block convolution neural network

Country Status (1)

Country Link
CN (1) CN111738306B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110292219A1 (en) * 2010-05-25 2011-12-01 Nelson Liang An Chang Apparatus and methods for imaging system calibration
WO2012071688A1 (en) * 2010-12-03 2012-06-07 中国科学院自动化研究所 Method for analyzing 3d model shape based on perceptual information
CN106910252A (en) * 2017-01-20 2017-06-30 东北石油大学 A kind of online mask method of threedimensional model based on semantic space projective transformation and system
WO2018065073A1 (en) * 2016-10-07 2018-04-12 Toyota Motor Europe Electronic device, system and method for recognizing and locating an object
US20190050981A1 (en) * 2017-08-09 2019-02-14 Shenzhen Keya Medical Technology Corporation System and method for automatically detecting a target object from a 3d image

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110292219A1 (en) * 2010-05-25 2011-12-01 Nelson Liang An Chang Apparatus and methods for imaging system calibration
WO2012071688A1 (en) * 2010-12-03 2012-06-07 中国科学院自动化研究所 Method for analyzing 3d model shape based on perceptual information
WO2018065073A1 (en) * 2016-10-07 2018-04-12 Toyota Motor Europe Electronic device, system and method for recognizing and locating an object
CN106910252A (en) * 2017-01-20 2017-06-30 东北石油大学 A kind of online mask method of threedimensional model based on semantic space projective transformation and system
US20190050981A1 (en) * 2017-08-09 2019-02-14 Shenzhen Keya Medical Technology Corporation System and method for automatically detecting a target object from a 3d image

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
YU T, ET AL: "Multi-view Harmonized Bilinear Network for 3D Object Recognition", 《IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
徐磊: "基于多视图的三维模型检索系统", 《中国优秀博硕士学位论文全文数据库(硕士)信息科技辑》 *

Also Published As

Publication number Publication date
CN111738306B (en) 2022-05-13

Similar Documents

Publication Publication Date Title
Lin et al. Convolution in the cloud: Learning deformable kernels in 3d graph convolution networks for point cloud analysis
Zhu et al. Deep learning in remote sensing: A comprehensive review and list of resources
Gao et al. Exploring deep learning for view-based 3D model retrieval
CN110852182B (en) Depth video human body behavior recognition method based on three-dimensional space time sequence modeling
CN108875076B (en) Rapid trademark image retrieval method based on Attention mechanism and convolutional neural network
CN110929080B (en) Optical remote sensing image retrieval method based on attention and generation countermeasure network
CN109101981B (en) Loop detection method based on global image stripe code in streetscape scene
CN111625667A (en) Three-dimensional model cross-domain retrieval method and system based on complex background image
CN109034035A (en) Pedestrian's recognition methods again based on conspicuousness detection and Fusion Features
CN110543581A (en) Multi-view three-dimensional model retrieval method based on non-local graph convolution network
CN106844620B (en) View-based feature matching three-dimensional model retrieval method
Gao et al. Group-pair convolutional neural networks for multi-view based 3d object retrieval
CN109920050B (en) Single-view three-dimensional flame reconstruction method based on deep learning and thin plate spline
Liu et al. Upright orientation of 3D shapes with convolutional networks
CN112597324A (en) Image hash index construction method, system and equipment based on correlation filtering
CN105574545B (en) The semantic cutting method of street environment image various visual angles and device
CN113032613B (en) Three-dimensional model retrieval method based on interactive attention convolution neural network
Su et al. 3d-assisted image feature synthesis for novel views of an object
CN111597367B (en) Three-dimensional model retrieval method based on view and hash algorithm
CN106951501B (en) Three-dimensional model retrieval method based on multi-graph matching
Wu et al. Deep texture exemplar extraction based on trimmed T-CNN
Zhou et al. Retrieval and localization with observation constraints
CN106951873A (en) A kind of Remote Sensing Target recognition methods
CN111738306B (en) Multi-view three-dimensional model retrieval method based on block convolution neural network
Ren et al. A multi-scale UAV image matching method applied to large-scale landslide reconstruction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant