CN112199533B - Unsupervised deep hash model training method and image retrieval method based on node characterization - Google Patents
Unsupervised deep hash model training method and image retrieval method based on node characterization Download PDFInfo
- Publication number
- CN112199533B CN112199533B CN202011100159.2A CN202011100159A CN112199533B CN 112199533 B CN112199533 B CN 112199533B CN 202011100159 A CN202011100159 A CN 202011100159A CN 112199533 B CN112199533 B CN 112199533B
- Authority
- CN
- China
- Prior art keywords
- network
- gcn
- node
- layer
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 61
- 238000012512 characterization method Methods 0.000 title claims abstract description 48
- 238000012549 training Methods 0.000 title claims abstract description 38
- 239000013598 vector Substances 0.000 claims description 21
- 239000011159 matrix material Substances 0.000 claims description 17
- 230000004913 activation Effects 0.000 claims description 12
- 230000008569 process Effects 0.000 claims description 8
- 238000004364 calculation method Methods 0.000 claims description 4
- 230000003213 activating effect Effects 0.000 claims description 3
- 238000011176 pooling Methods 0.000 claims description 3
- 230000006870 function Effects 0.000 abstract description 30
- 230000000694 effects Effects 0.000 description 6
- 238000012360 testing method Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000005096 rolling process Methods 0.000 description 2
- 101100498818 Arabidopsis thaliana DDR4 gene Proteins 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/53—Querying
- G06F16/532—Query formulation, e.g. graphical querying
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Library & Information Science (AREA)
- Mathematical Physics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an unsupervised depth hash model training and image retrieval method based on node characterization. According to the method, each image is regarded as a node, similarity of the images is measured by utilizing a similarity distance between image initialization features, a self-encoder based on a graph convolution network is designed, node characterization information of each image is generated in an unsupervised mode according to the similarity between the images, then learning of a hash function is guided on a lightweight network by utilizing the characterization information, semantic hash codes of the images are generated, and the unsupervised hash image retrieval performance is improved. The invention effectively learns the similarity between images, and the image retrieval performance is better than that of the current unsupervised hash method.
Description
Technical Field
The invention belongs to the field of image retrieval, and in particular relates to an unsupervised depth hash model training method and an image retrieval method based on node characterization.
Background
The hash code has the advantages of light weight storage and efficient exclusive OR operation, so that the image hash becomes an important technical means in the field of image retrieval.
The image retrieval method based on the image hash comprises a supervised hash retrieval method and an unsupervised hash retrieval method, wherein the supervised hash method is added with label information of the image, such as class labels of the image, in the learning process of a hash function, so that semantic similarity between data can be maintained better, and a high-quality hash code is generated. The unsupervised hash method does not need label information of the data, and the original geometric structure similarity of the data is kept to train the hash function by directly mining the internal connection of the data.
Although the supervised hash search method achieves good search effect, in an actual scene, the supervised hash search method cannot realize the search of the label image, so that the unsupervised hash search method is more common. However, the unsupervised hash search method does not sufficiently learn the similarity between images, thereby limiting the quality of the hash code, and thus resulting in poor image search results.
Disclosure of Invention
Aiming at the defects or improvement demands of the prior art, the invention provides an unsupervised depth hash model training method and an image retrieval method based on node characterization, which aim to solve the technical problems that the quality of hash codes is limited and the image retrieval effect is poor due to insufficient similarity between images in the conventional unsupervised hash retrieval method.
To achieve the above object, according to one aspect of the present invention, there is provided a training method of an unsupervised depth hash model based on node characterization, the unsupervised depth hash model including a res net-101 network, a GCN-based self-encoder AE, and a res net-18 network connected in sequence, the training method comprising the steps of:
(1) Acquiring a training set formed by N images, inputting the training set into a ResNet-101 network to extract the feature vector of each image from a pooling layer of the network, wherein the feature vectors of all the images form a vector set FV:
FV={v 1 ,v 2 ,…,v N },
wherein N is a positive integer, v q Represents the q-th feature vector in the vector set FV and has q ε [1, N];
(2) Selecting feature vectors of M images from the vector set FV obtained in the step (1) to form M nodes, and constructing a graph structure G= (X, E) according to the M nodes, wherein X represents the set formed by the M nodes and has X= [ X ] 1 ,x 2 ,…,x M ],Representing any two nodes x of the M nodes i And x j Edges joined, where i and j are positive integers and i and j are E [1, M];
(3) Acquiring a relation matrix A of M nodes from the graph structure G constructed in the step (2), performing iterative training on the AE based on GCN by using the acquired relation matrix A and the M nodes until the AE converges, and acquiring each node x at the moment k G-dimensional characterization information g (X) k ) Wherein k is [1, M],g∈{16,32,48,64,96,128};
(4) Using the node x obtained in step (3) k G-dimensional characterization information g (X) k ) Iterative training of the ResNet-18 network as tag information until the ResNet-18 network converges and node x is obtained at this time k Corresponding g-dimensional semantic hash code b k 。
Preferably, the coding network comprises a K-layer GCN for updating node characteristics, and the output of the coding network is node x k G-dimensional characterization information g (X) k ) Where K is any natural number, preferably 5.
Preferably, the step (3) specifically comprises:
first, a potential representation X of all M nodes in a layer 1 GCN in a coding network is obtained l+1 Wherein l is [1, K ]]:
Wherein X is l Representing potential characterizations of all M nodes in a layer I GCN in a coding network, X 1 =X,W l Representing the weight of the layer I GCN in the coding network, f l (-) represents a nonlinear activation function of the layer one GCN in the coding network,representing a normalized version of the relationship matrix a;
next, an output of the layer 1 GCN in the decoding network is generated in the reconstruction space
Wherein the method comprises the steps ofRepresenting the decoding result of the layer I GCN in the decoding network,/I>Indicating the weight of the layer I GCN in the decoding network,/->Representing a nonlinear activation function of a layer I GCN in the decoding network;
next, based on the output of the layer 1 GCN in the decoding networkAnd the set X calculates a loss function value L 1 The above process is repeated until the loss function value L 1 Obtaining trained AE until the data is minimum, and taking output of a K-th layer GCN in a coding network of the AE at the moment as a node x k G-dimensional characterization information g (X) k )。
Preferably, the normalized version of the relationship matrix AEqual to:
wherein the method comprises the steps ofI C Is an identity matrix>Is a diagonal matrix and satisfies->
Loss function value L 1 Equal to:
wherein,representing the use of a two-norm calculation.
Preferably, the step (4) specifically comprises:
first, the node x obtained in the step (3) is obtained k G-dimensional characterization information g (X) k ) The ResNet-18 network is input, and then at the last full connection layer of the ResNet-18 network, the mean square error function L is used fc So that the output result fc (x k ) Approximation node x k Characterization information g (x k );
The last active layer of the ResNet-18 network is then used to output the result fc (x k ) Activating, thereby obtaining an activation result fch (x k )=Tanh(fc(x k ) And) wherein
Then, a loss function value L of the ResNet-18 network is calculated according to the mean square error function and the binary cross entropy loss function 2 ;
The above process is then repeated until the loss function value L 2 Minimum, thereby obtaining a trained ResNet-18 network, when using a binary cross entropy loss function L fch To the activation result fch (x k ) Binarizing to generate each node x k G-dimensional semantic hash code of e X, i.e. g-dimensional language corresponding to each image in step (1)The hash code is defined.
Preferably, the mean square error function is
The binary cross entropy loss function is:
wherein coefficient b k =0.5×(sign(fch(x k ) +1)), sign (·) is a sign function,representation b k Is selected from the group consisting of the (i) th bit,representing fch (x k ) Is the ith bit of (c).
Preferably, the ResNet-18 network has a loss function value L 2 Equal to:
L 2 =μL fc +ωL fch ,
where μ and ω are hyper-parameters and satisfy μ+ω=1.
In general, the above technical solutions conceived by the present invention, compared with the prior art, enable the following beneficial effects to be obtained:
because the invention uses the step (2) and the step (3) in the model training stage, firstly, each image is regarded as a node and the nodes are organized into a graph structure, and further, the association relations among different nodes are learned by GCN and are fused into the characterization information of the nodes, the technical problem that the quality of hash codes is limited due to insufficient similarity between the learned images in the existing unsupervised hash retrieval method, and further, the image retrieval effect is poor can be solved.
(2) In the invention, step (4) is adopted in the model training stage, node characterization information is used as label information to guide the training of the hash function, and the ResNet-18 network is a lightweight network, so that the semantic similarity between the hash codes can be ensured to maintain the correlation between the nodes, and the hash codes with high quality can be generated efficiently.
According to another aspect of the present invention, there is provided an image retrieval method including the steps of:
(1) Obtaining images to be retrieved from a user, and obtaining each contrast image from a database;
(2) Generating g-dimensional semantic hash codes corresponding to the images to be searched and each contrast image respectively according to the unsupervised depth hash model obtained by training based on the training method of the unsupervised depth hash model of node characterization;
(3) And carrying out exclusive OR operation on the g-dimensional semantic hash codes corresponding to the images to be retrieved and the g-dimensional semantic hash codes corresponding to the contrast images so as to calculate the Hamming distance between the g-dimensional semantic hash codes and the g-dimensional semantic hash codes.
(4) And (3) selecting a contrast image corresponding to the Hamming distance within the specified Hamming radius from all the Hamming distances obtained in the step (3) as a final image retrieval result to output.
Preferably, the specified hamming radius range is {1,2,3}, preferably 2.
In general, the above technical solutions conceived by the present invention, compared with the prior art, enable the following beneficial effects to be obtained:
the invention combines GCN and AutoEncoder to generate potential characterization information of the image in an unsupervised mode, uses lightweight ResNet-18 network to fit the characterization information of the nodes and generates semantic hash codes of the image, thereby improving the unsupervised hash image retrieval performance and further solving the problem that the existing scheme cannot fully learn the similarity between the images.
Drawings
FIG. 1 is a schematic diagram of the overall framework of the training method of the unsupervised deep hash model based on node characterization of the present invention;
FIG. 2 is a flow chart of a method of training an unsupervised deep hash model based on node characterization of the present invention;
FIG. 3 is a flow chart of an image retrieval method provided by the invention;
FIG. 4 is a search result on a CIFAR-10 dataset using a 48-bit hash code length with a Hamming radius of 2 in accordance with the present invention.
Detailed Description
The present invention will be described in further detail with reference to the drawings and examples, in order to make the objects, technical solutions and advantages of the present invention more apparent. It should be understood that the specific embodiments described herein are for purposes of illustration only and are not intended to limit the scope of the invention. In addition, the technical features of the embodiments of the present invention described below may be combined with each other as long as they do not collide with each other.
The invention is realized on the basis of a graph rolling network (GCN) and a self-encoder (AutoEncoder), and the image retrieval performance of a model is improved by forming images into a graph structure, integrating the similarity between images into the characteristics of the images by using the GCN, and further generating the characterization information of the images by using the AutoEncoder in an unsupervised mode.
In the experiment of the invention, through testing on a plurality of image data sets, the advantages of the invention on the multi-label image data set are more obvious, the reason is that each image in the multi-label data set has more abundant semantic information, the GCN fully captures and learns the semantic similarity among the images, more accurate node characterization is generated, and a high-quality semantic hash code is obtained, so that the image retrieval performance of the model is further improved.
As shown in fig. 1 and 2, the present invention provides a training method of an unsupervised deep hash model based on node characterization, the unsupervised deep hash model including a res net-101 network, a self encoder (AutoEncoder, AE) based on a graph rolling network (Graph Convolutional Network, GCN), and a res net-18 network, which are sequentially connected, the training method comprising the steps of:
(1) Acquiring a training set formed by N images, inputting the training set into a ResNet-101 network to extract the feature vector of each image from a pooling layer of the network, wherein the feature vectors of all the images form a vector set FV:
FV={v 1 ,v 2 ,…,v N },
wherein N is a positive integer, v q Represents the q-th feature vector in the vector set FV and has q ε [1, N]。
(2) Selecting feature vectors of M images from the vector set FV obtained in the step (1) to form M nodes, and constructing a graph structure G= (X, E) according to the M nodes, wherein X represents the set formed by the M nodes and has X= [ X ] 1 ,x 2 ,…,x M ],Representing any two nodes x of the M nodes i And x j Edges joined, where i and j are positive integers and i and j are E [1, M];
(3) Acquiring a relation matrix A of M nodes from the graph structure G constructed in the step (2), performing iterative training on the AE based on GCN by using the acquired relation matrix A and the M nodes until the AE converges, and acquiring each node x at the moment k G-dimensional characterization information g (X) k ) Wherein k is [1, M],g∈{16,32,48,64,96,128};
AE is composed of coding network and decoding network, wherein the coding network comprises K-layer GCN for updating node characteristics, and the output of the coding network is node x k G-dimensional characterization information g (X) k ) Where K is any natural number, preferably 5.
The step (3) is specifically that firstly, potential representation X of all M nodes in the GCN of the layer 1 in the coding network is obtained l+1 Wherein l is [1, K ]]:
Wherein X is l Representing potential characterizations of all M nodes in a layer I GCN in a coding network, X 1 =X,W l Representing the weight of the layer I GCN in the coding network, f l (-) represents a nonlinear activation function of the layer one GCN in the coding network,representing a normalized version of the relationship matrix a, the calculation process is as follows:
wherein the method comprises the steps ofI C Is an identity matrix>Is a diagonal matrix and satisfies->
Next, in order for the input of the encoding network to approximate the output of the decoding network (which also uses K-layer GCN) to obtain high quality g-dimensional characterization information g (x k ) In the decoding network, the encoding network is then inverted to generate the output of the layer 1 GCN in the decoding network in the reconstruction space
Wherein the method comprises the steps ofRepresenting the decoding result of the layer I GCN in the decoding network,/I>Indicating the weight of the layer I GCN in the decoding network,/->Representing a nonlinear activation function of a layer I GCN in the decoding network;
next, based on the output of the layer 1 GCN in the decoding networkAnd the set X calculates a loss function value L 1 The above process is repeated until the loss function value L 1 Obtaining trained AE until the data is minimum, and taking output of a K-th layer GCN in a coding network of the AE at the moment as a node x k G-dimensional characterization information g (X) k );
Wherein,representing the use of a two-norm calculation.
The steps (2) and (3) have the advantages that each image is firstly regarded as a node, the nodes are organized into a graph structure, the GCN is adopted to learn the association relations among different nodes, and the relations are integrated into the characterization information of the nodes, so that the technical problem that the quality of hash codes is limited due to insufficient similarity between the images in the existing unsupervised hash retrieval method, and the image retrieval effect is poor is solved.
(4) Using the node x obtained in step (3) k G-dimensional characterization information g (X) k ) Iterative training of the ResNet-18 network as tag information until the ResNet-18 network converges and node x is obtained at this time k Corresponding g-dimensional semantic hash code b k 。
The step (4) is specifically that the node x obtained in the step (3) is first selected k G-dimensional characterization information g (X) k ) ResNet-18 network is input, and then a mean square error function L is used at the last Full Connection (FC) layer of the ResNet-18 network fc So that the output of the full connection layerResults fc (x) k ) Approximation node x k Characterization information g (x k ) The method comprises the steps of carrying out a first treatment on the surface of the The final active layer (Full connection hash, FCH) of the ResNet-18 network is then used to output the result fc (x k ) Activating, thereby obtaining an activation result fch (x k )=Tanh(fc(x k ) And) whereinThen, according to the mean square error function->And binary cross entropy loss functionTo calculate the loss function value L of ResNet-18 network 2 (wherein coefficient b k =0.5×(sign(fch(x k ) +1)) sign (·) is a sign function, ++>Representation b k Is selected from the group consisting of the (i) th bit,representing fch (x k ) I-th bit of (c) and repeatedly iterating the above process until the loss function value L 2 Minimum, thereby obtaining a trained ResNet-18 network, when using a binary cross entropy loss function L fch To the activation result fch (x k ) Binarizing to generate each node x k G-dimensional semantic hash code of e X, i.e. g-dimensional semantic hash code corresponding to each image in step (1), loss function value L of ResNet-18 network 2 Equal to:
L 2 =μL fc +ωL fch ,
where μ and ω are hyper-parameters and satisfy μ+ω=1.
The method has the advantages that node characterization information is used as label information to guide the training of the hash function, and the ResNet-18 network is a lightweight network, so that semantic similarity among hash codes can be ensured to maintain the correlation among nodes, and high-quality hash codes can be generated efficiently.
As shown in fig. 3, the present invention further provides an image retrieval method, which includes the following steps:
(1) Obtaining images to be retrieved from a user, and obtaining each contrast image from a database;
(2) Generating g-dimensional semantic hash codes corresponding to the images to be searched and each contrast image respectively according to the unsupervised depth hash model obtained by training based on the training method of the unsupervised depth hash model of node characterization;
(3) And carrying out exclusive OR operation on the g-dimensional semantic hash codes corresponding to the images to be retrieved and the g-dimensional semantic hash codes corresponding to the contrast images so as to calculate the Hamming distance between the g-dimensional semantic hash codes and the g-dimensional semantic hash codes.
(4) And (3) selecting a contrast image corresponding to the Hamming distance within the specified Hamming radius from all the Hamming distances obtained in the step (3) as a final image retrieval result to output.
Specifically, the specified hamming radius range is {1,2,3}, preferably 2.
In summary, the invention combines GCN and AutoEncoder to generate potential characterization information of images in an unsupervised mode, uses lightweight ResNet-18 network to fit characterization information of nodes and generate semantic hash codes of the images, thereby improving unsupervised hash image retrieval performance and further solving the problem that the existing scheme cannot fully learn the similarity between images.
Experimental results
The experimental environment of the invention: the CPU is 56 pieces of Inter Xeon (R) @2.4GHz, the GPU is 8 pieces of NVIDIA Tesla P40 GB, the memory is 256GB of DDR4, the hard disk capacity is 6TB, and the algorithm is realized by adopting Pytorch programming under a Linux operating system. The specific parameter settings are as follows: the batch size was 32 and the learning rate was 0.001.
To demonstrate the effectiveness of GCN in the present invention and the superiority of the search results of the present invention, the present invention conducted related tests on CIFAR-10 dataset and MS-COCO dataset. In the node characterization learning stage, different feature extraction networks VGG and ResNet-101 are used, and the retrieval performance mAP of the model under the two conditions of using GCN and not using GCN is recorded. "-" indicates that GCN was not used, "+" indicates that GCN was used, and the test results are given in Table 1.
TABLE 1 mAP values with/without GCN according to the invention
As can be seen from the results in Table 1, the present invention can achieve the highest mAP value by using a combination of ResNet-101 and GCN. Furthermore, the present invention has found that good results can be obtained using a combination of VGG and GCN. However, if GCN is not used, neither VGG nor ResNet-101 alone can achieve satisfactory results. These results indicate that GCN plays a critical role in improving the performance of hashing. The invention further observes that under the same conditions, the ResNet-101 is relatively better than VGG effect, and also shows that the network with stronger feature extraction capability is more helpful to improving hash performance.
FIG. 4 shows the result of the search on the CIFAR-10 dataset using a 48-bit hash code length with a Hamming radius of 2 according to the present invention. According to the invention, 4 images are randomly selected as search images: automobiles, birds, horses, and boats. From the returned search results, it can be seen that the search image still contains automobiles, birds, horses and boats. This reflects that the invention does learn semantic similarity between images and achieves good retrieval effect.
It will be readily appreciated by those skilled in the art that the foregoing description is merely a preferred embodiment of the invention and is not intended to limit the invention, but any modifications, equivalents, improvements or alternatives falling within the spirit and principles of the invention are intended to be included within the scope of the invention.
Claims (6)
1. A training method of an unsupervised deep hash model based on node characterization, the unsupervised deep hash model comprising a res net-101 network, a GCN-based self-encoder AE, and a res net-18 network connected in sequence, the training method comprising the steps of:
(1) Acquiring a training set formed by N images, inputting the training set into a ResNet-101 network to extract the feature vector of each image from a pooling layer of the network, wherein the feature vectors of all the images form a vector set FV:
FV={v 1 ,v 2 ,…,v N },
wherein N is a positive integer, v q Represents the q-th feature vector in the vector set FV and has q ε [1, N];
(2) Selecting feature vectors of M images from the vector set FV obtained in the step (1) to form M nodes, and constructing a graph structure G= (X, E) according to the M nodes, wherein X represents the set formed by the M nodes and has X= [ X ] 1 ,x 2 ,…,x M ],Representing any two nodes x of the M nodes i And x j Edges joined, where i and j are positive integers and i and j are E [1, M];
(3) Acquiring a relation matrix A of M nodes from the graph structure G constructed in the step (2), performing iterative training on the AE based on GCN by using the acquired relation matrix A and the M nodes until the AE converges, and acquiring each node x at the moment k G-dimensional characterization information g (X) k ) Wherein k is [1, M]G is {16,32,48,64,96,128}; the step (3) is specifically as follows:
first, a potential representation X of all M nodes in a layer 1 GCN in a coding network is obtained l+1 Wherein l is [1, K ]]:
Wherein X is l Representing potential characterizations of all M nodes in a layer I GCN in a coding network, X 1 =X,W l Representing the weights of the layer i GCN in the coding network,f l (-) represents a nonlinear activation function of the layer one GCN in the coding network,representing a normalized version of the relationship matrix a;
next, an output of the layer 1 GCN in the decoding network is generated in the reconstruction space
Wherein the method comprises the steps ofRepresenting the decoding result of the layer I GCN in the decoding network,/I>Representing the weights of the layer i GCN in the decoding network,representing a nonlinear activation function of a layer I GCN in the decoding network;
next, based on the output of the layer 1 GCN in the decoding networkAnd the set X calculates a loss function value L 1 The above process is repeated until the loss function value L 1 Obtaining trained AE until the data is minimum, and taking output of a K-th layer GCN in a coding network of the AE at the moment as a node x k G-dimensional characterization information g (X) k );
(4) Using the node x obtained in step (3) k G-dimensional characterization information g (X) k ) Iterative training of the ResNet-18 network as tag information until the ResNet-18 network converges and node x is obtained at this time k Corresponding g-dimensional semantic hash code b k The method comprises the steps of carrying out a first treatment on the surface of the Step (a)(4) The method comprises the following steps:
first, the node x obtained in the step (3) is obtained k G-dimensional characterization information g (X) k ) The ResNet-18 network is input, and then at the last full connection layer of the ResNet-18 network, the mean square error function L is used fc So that the output result fc (x k ) Approximation node x k Characterization information g (x k ) The method comprises the steps of carrying out a first treatment on the surface of the Mean square error function of
The binary cross entropy loss function is:
wherein coefficient b k =0.5×(sign(fch(x k ) +1)), sign (·) is a sign function,representation b k Is selected from the group consisting of the (i) th bit,representing fch (x k ) Is the ith bit of (2);
the last active layer of the ResNet-18 network is then used to output the result fc (x k ) Activating, thereby obtaining an activation result fch (x k )=Tanh(fc(x k ) And) wherein
Then, a loss function value L of the ResNet-18 network is calculated according to the mean square error function and the binary cross entropy loss function 2 ;
The above process is then repeated until the loss function value L 2 Minimum, thereby obtaining a trained ResNet-18 network, when using a binary cross entropy loss function L fch To the activation result fch (x k ) Binarizing to generate each node x k G-dimensional semantic hash codes of e X, i.e., g-dimensional semantic hash codes corresponding to the respective images in step (1).
2. Training method according to claim 1, characterized in that the coding network comprises a K-layer GCN for updating node characteristics, the output of the coding network being node x k G-dimensional characterization information g (X) k ) Wherein K is any natural number.
3. The training method of claim 2, wherein,
normalized version of the relationship matrix AEqual to:
wherein the method comprises the steps ofI C Is an identity matrix>Is a diagonal matrix and satisfies->
Loss function value L 1 Equal to:
wherein,representing the use of a two-norm calculation.
4. A training method as claimed in claim 3Characterized in that the loss function value L of the ResNet-18 network 2 Equal to:
L 2 =μL fc +ωL fch ,
where μ and ω are hyper-parameters and satisfy μ+ω=1.
5. An image retrieval method, characterized by comprising the steps of:
(1) Obtaining images to be retrieved from a user, and obtaining each contrast image from a database;
(2) The method for training the unsupervised depth hash model based on node characterization according to any one of claims 1 to 4, wherein the unsupervised depth hash model obtained by training is used for generating g-dimensional semantic hash codes corresponding to the image to be retrieved and each contrast image respectively;
(3) Performing exclusive OR operation on the g-dimensional semantic hash codes corresponding to the images to be retrieved and the g-dimensional semantic hash codes corresponding to the contrast images so as to calculate the Hamming distance between the g-dimensional semantic hash codes and the g-dimensional semantic hash codes;
(4) And (3) selecting a contrast image corresponding to the Hamming distance within the specified Hamming radius from all the Hamming distances obtained in the step (3) as a final image retrieval result to output.
6. The image retrieval method as recited in claim 5, wherein the specified hamming radius range is {1,2,3}.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011100159.2A CN112199533B (en) | 2020-10-15 | 2020-10-15 | Unsupervised deep hash model training method and image retrieval method based on node characterization |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011100159.2A CN112199533B (en) | 2020-10-15 | 2020-10-15 | Unsupervised deep hash model training method and image retrieval method based on node characterization |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112199533A CN112199533A (en) | 2021-01-08 |
CN112199533B true CN112199533B (en) | 2024-02-06 |
Family
ID=74009069
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011100159.2A Active CN112199533B (en) | 2020-10-15 | 2020-10-15 | Unsupervised deep hash model training method and image retrieval method based on node characterization |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112199533B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113537495A (en) * | 2021-08-05 | 2021-10-22 | 南方电网数字电网研究院有限公司 | Model training system, method and device based on federal learning and computer equipment |
CN115017366B (en) * | 2022-07-11 | 2024-04-02 | 中国科学技术大学 | Unsupervised video hash retrieval method based on multi-granularity contextualization and multi-structure preservation |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111078911A (en) * | 2019-12-13 | 2020-04-28 | 宁波大学 | Unsupervised hashing method based on self-encoder |
CN111753190A (en) * | 2020-05-29 | 2020-10-09 | 中山大学 | Meta learning-based unsupervised cross-modal Hash retrieval method |
-
2020
- 2020-10-15 CN CN202011100159.2A patent/CN112199533B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111078911A (en) * | 2019-12-13 | 2020-04-28 | 宁波大学 | Unsupervised hashing method based on self-encoder |
CN111753190A (en) * | 2020-05-29 | 2020-10-09 | 中山大学 | Meta learning-based unsupervised cross-modal Hash retrieval method |
Also Published As
Publication number | Publication date |
---|---|
CN112199533A (en) | 2021-01-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110837836B (en) | Semi-supervised semantic segmentation method based on maximized confidence | |
US8676725B1 (en) | Method and system for entropy-based semantic hashing | |
CN108920720A (en) | The large-scale image search method accelerated based on depth Hash and GPU | |
CN112199533B (en) | Unsupervised deep hash model training method and image retrieval method based on node characterization | |
CN112925962B (en) | Hash coding-based cross-modal data retrieval method, system, device and medium | |
CN112732864B (en) | Document retrieval method based on dense pseudo query vector representation | |
CN109446414B (en) | Software information site rapid label recommendation method based on neural network classification | |
CN106033426A (en) | A latent semantic min-Hash-based image retrieval method | |
CN111125469B (en) | User clustering method and device of social network and computer equipment | |
CN110941734A (en) | Depth unsupervised image retrieval method based on sparse graph structure | |
CN107180079B (en) | Image retrieval method based on convolutional neural network and tree and hash combined index | |
CN112861976B (en) | Sensitive image identification method based on twin graph convolution hash network | |
CN113312505A (en) | Cross-modal retrieval method and system based on discrete online hash learning | |
CN112417289A (en) | Information intelligent recommendation method based on deep clustering | |
CN110598022A (en) | Image retrieval system and method based on robust deep hash network | |
CN114328988A (en) | Multimedia data feature extraction method, multimedia data retrieval method and device | |
CN108805280B (en) | Image retrieval method and device | |
CN116049467A (en) | Non-supervision image retrieval method and system based on label visual joint perception | |
CN107451617B (en) | Graph transduction semi-supervised classification method | |
CN114996493A (en) | Electric power scene image data screening method based on data elimination and redundancy elimination | |
Zhang et al. | Semantic hierarchy preserving deep hashing for large-scale image retrieval | |
CN113298892A (en) | Image coding method and device, and storage medium | |
CN110659375A (en) | Hash model training method, similar object retrieval method and device | |
CN110674333A (en) | Large-scale image high-speed retrieval method based on multi-view enhanced depth hashing | |
CN116433922A (en) | Training method of graph representation learning system, graph data node classification method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |