CN110110120B - Image retrieval method and device based on deep learning - Google Patents

Image retrieval method and device based on deep learning Download PDF

Info

Publication number
CN110110120B
CN110110120B CN201810597022.9A CN201810597022A CN110110120B CN 110110120 B CN110110120 B CN 110110120B CN 201810597022 A CN201810597022 A CN 201810597022A CN 110110120 B CN110110120 B CN 110110120B
Authority
CN
China
Prior art keywords
neural network
image
binary hash
deep neural
images
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810597022.9A
Other languages
Chinese (zh)
Other versions
CN110110120A (en
Inventor
曾凡锋
胡胜达
王宝成
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Tianfenghaizhiyuan Communication Power Technology Co ltd
Original Assignee
North China University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by North China University of Technology filed Critical North China University of Technology
Priority to CN201810597022.9A priority Critical patent/CN110110120B/en
Publication of CN110110120A publication Critical patent/CN110110120A/en
Application granted granted Critical
Publication of CN110110120B publication Critical patent/CN110110120B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content

Abstract

The invention relates to an image retrieval method and device based on deep learning. The method comprises the following steps: building a deep neural network; inputting the images of the training set into a deep neural network for training; loading a trained deep neural network model, inputting all images of a training set into the deep neural network to obtain binary hash codes with semantic information, simultaneously obtaining the binary hash codes with visual information by adopting a traditional binary hash coding method, and establishing a local feature library; inputting an image to be retrieved into a deep neural network, acquiring a corresponding binary hash code with semantic information, acquiring the binary hash code with visual information of the image to be retrieved by adopting a traditional binary hash coding method, comparing the binary hash code with a local feature library, and calculating the similarity to obtain a retrieval result. The invention can keep the important information of the image as much as possible and can realize the rapid and accurate image retrieval of mass image data.

Description

Image retrieval method and device based on deep learning
Technical Field
The invention belongs to the application of deep learning in the field of image processing, and particularly relates to a method and a device for performing binary hash coding on an image by adopting a deep neural network and using the binary hash coding for retrieval.
Background
The image retrieval technology aims at image contents which are interested by a user, and presents related images to the user in a way that the similarity is from top to bottom according to a specific similarity measurement standard. The core problem is how to condense the information of the image, obtain the feature descriptor of the image and fully express the content information of the image.
The traditional image retrieval technology is used for extracting image features aiming at basic features such as texture, color, shape and the like of an image and calculating the similarity by adopting a corresponding image similarity measurement method. However, these basic image features cannot depict semantic content of an image, and in practical applications, image content concerned by a user is often semantic-level, such as concepts of scenes, objects, and the like in the image. In addition, with the development of the internet era, image retrieval tasks tend to be large-scale and diversified at the present stage, and an image retrieval algorithm is required to be capable of finding out an image which is interested by a user from massive image data in a shorter time, so that the image retrieval technology is more difficult.
In view of the above problems, in recent years, people have tried to extract features of an image by using a deep neural network, and achieve a certain result by using the strong image expression capability and the efficient calculation efficiency of binary hash coding of the deep neural network, especially the deep convolutional neural network. However, there are some limitations to the existing methods, especially the configuration of the pooling layer. In the structural design of the deep neural network, the pooling layer plays an important role. As the number of layers of the neural network continuously increases, on one hand, the pooling layer plays a role of down-sampling, so that the interference information of the image is reduced, and the important information of the image is obtained; on the other hand, the down sampling reduces the number of pixels of the image, reduces parameters of the neural network to a certain extent, and solves the problems of excessive network parameters, memory consumption and difficulty in convergence. In the existing design network, people often adopt a single pooling method, especially a maximum pooling method, to realize a pooling layer. Aiming at the defect, the invention provides a deep neural network structure combining multiple pooling methods to improve the image retrieval effect.
Disclosure of Invention
The invention provides a novel image retrieval method and device based on deep learning, aiming at the limitation of the existing image retrieval method based on deep learning.
Aiming at the defect that the conventional image retrieval method based on deep learning usually adopts a single pooling method in the design process of a network, the invention adopts two pooling methods of maximum pooling and average pooling for the output structure of each layer in the design process of a network structure, thereby greatly retaining the semantic information of the image and realizing better image retrieval effect. And further combining the hash code with semantic information acquired from the deep neural network with a plurality of traditional hash coding methods to acquire a final image retrieval result. The image retrieval method is based on deep learning, and the deep neural network is adopted for image retrieval, namely, the binary Hash codes of the images are obtained by constructing and training the corresponding deep neural network for the retrieval process of the images.
Specifically, the invention adopts the following technical scheme:
an image retrieval method based on deep learning, comprising the following steps:
1) constructing a deep neural network for binary Hash coding extraction of the image;
2) inputting the images of the training set into the deep neural network in batches for model training of the deep neural network, and storing the trained deep neural network model;
3) loading a deep neural network model, inputting all images of a training set into the deep neural network to obtain binary hash codes with semantic information, simultaneously obtaining the binary hash codes with visual information (texture, color and shape) by adopting a traditional binary hash coding method, and establishing a local feature library by utilizing the binary hash codes with the semantic information and the binary hash codes with the visual information;
4) inputting an image to be retrieved into a deep neural network model, acquiring a corresponding binary hash code with semantic information, acquiring the binary hash code with bottom-layer visual information of the image to be retrieved by adopting a traditional binary hash coding method, comparing the acquired binary hash code with a local feature library, and acquiring a retrieval result through similarity.
Further, the data set partitioning method of the deep neural network in step 1) comprises the following steps: if a large number of available training images do not exist in an actual application scene, an image data set close to the actual application scene is additionally taken as a training set, otherwise, the existing data set is divided into the training set and a testing set, each image has a respective label, the labels are set according to actual requirements, and the label of each image can be not unique.
Further, each pooling layer of the deep neural network in the step 1) has an average pooling layer and a maximum pooling layer, the number of network output layers of each pooling layer increases with the increase of the network depth, and the sizes of feature maps output by the same pooling layer are the same; all the output feature maps of the pooling layer are connected in series in a specific network layer, and are subjected to convolution operation again through a plurality of convolution layers, so that the feature fusion effect of all the feature maps is realized; and in the output part of the deep neural network, converting the output of the deep neural network into vectors by adopting a global average pooling layer, and finally reducing the output of the deep neural network to the specified binary hash coding length through a plurality of full connection layers.
Further, step 2) training the deep neural network by using a contrast loss function, wherein the contrast loss function is calculated by using the following steps:
a) given an arbitrary batch of images I ═ I0,I1,...IN-1Y ═ Y, its label0,y1...yN-1N is the size of a batch of images, and any image pair is taken as Ii、Ij,i,j∈[0,N]With loss value L of the contrast loss functionpair(Ii,Ij):
Figure BDA0001692013420000031
Wherein δ (I)i,Ij) Indicating whether the image pairs are similar, d (I)i,Ij) Representing the distance between the image pair, margin is the edge coefficient. Delta (I)i,Ij) And d (I)i,Ij) Are defined as follows:
Figure BDA0001692013420000032
wherein y isiAnd yjAre labels, y 'of the i, j images respectively'i,kAnd y'j,kThe output values of the ith and the jth images of the neural network at the kth node respectivelyM is the number of the output nodes of the neural network, and the length of the output nodes is the same as that of the extracted binary hash code in terms of value;
b) there is a total loss value L for all image pairs within a batch:
Figure BDA0001692013420000033
binding of Lpair(Ii,Ij) Definition, L is rewritten as:
Figure BDA0001692013420000034
further, step 3) combines the binary hash codes with semantic information obtained by the neural network and the binary hash codes with visual information obtained by adopting a plurality of traditional binary hash coding methods as the characteristics of the image.
Further, step 4) adopts the following similarity calculation method to measure the similarity between the images:
a) given image I to be retrievedtestInputting the data into a deep neural network to obtain binary Hash code H with semantic information0:
H0=sign(Y')
Wherein Y' is the output of the neural network, sign (#) is a sign taking function;
b) assuming that n traditional image hash code generation methods are adopted, the obtained binary hash codes are respectively H1,H2...HnThen, ItestTotal binary hash encoding Htest={H0,H1,...Hn}; then for each image I from the feature librarytrainIts binary hash encoding is H'train={H'0,H'1,...H'n};
c) Calculating similarity sim (I) by adopting one of the following two schemestest,Itrain):
The first method comprises the following steps:
Figure BDA0001692013420000041
and the second method comprises the following steps:
Figure BDA0001692013420000042
wherein λi∈(0,1]And lambda01+...λn=1,s(Hi,H'i) Comprises the following steps:
Figure BDA0001692013420000043
wherein h isk、h'kAre respectively HiAnd H'iValue of the k-th digit, ziThe number of bits for the ith hash code.
An image retrieval device based on deep learning, comprising:
the deep neural network building module is responsible for building a deep neural network and is used for extracting the binary Hash codes of the image;
the deep neural network training module is responsible for inputting the training set images into the deep neural network in batches for model training of the deep neural network and storing the trained deep neural network model;
the local feature library establishing module is responsible for loading a deep neural network model, inputting all images of a training set into the deep neural network, acquiring binary hash codes with semantic information, acquiring the binary hash codes with visual information by adopting a traditional binary hash coding method, and establishing a local feature library by utilizing the binary hash codes with the semantic information and the binary hash codes with the visual information;
and the retrieval module is responsible for inputting the image to be retrieved into the deep neural network, acquiring the corresponding binary hash code with semantic information, acquiring the binary hash code with visual information of the image to be retrieved by adopting a traditional binary hash coding method, comparing the acquired binary hash code with the local feature library, and acquiring a retrieval result by calculating the similarity.
The semantic information refers to the image features obtained from the neural network. The image features obtained from the neural network are generally considered to be high-level semantic features, namely, the image features are not only superficial visual expression of the image, but also expression of concepts such as an object, a scene and the like.
The pooling method of the invention can greatly retain the semantic information of the image because the invention adopts a plurality of pooling methods on the same layer of the network, the pooling methods have differences in information screening, and the information of the plurality of pooling methods is retained in the pooling process of each layer.
The rapid image retrieval method based on deep learning emphasizes improvement on the structure of a network, retains important information of images as much as possible by adopting various pooling methods, improves the accuracy of image retrieval, and simultaneously utilizes the inherent calculation overhead advantage of binary hash coding to realize rapid and accurate image retrieval in massive image data.
Drawings
FIG. 1 is a flow chart of the main steps of the method of the present invention.
Fig. 2 is a network structure diagram of the deep neural network of the present invention.
Detailed Description
The present invention will be described in detail below with reference to examples and the accompanying drawings.
FIG. 1 is a flow chart of the main steps of the method of the present invention.
Step 1: if a large number of available training images are not available in an actual application scene, an image data set close to the application scene is additionally taken as a training set, otherwise, the existing data set is divided into the training set and a testing set, each image has a respective label, the labels are set according to actual requirements, and the label of each image can be not unique.
Step 2: a deep neural network as shown in fig. 2 was constructed and designed. Each layer of the pooling layer of the network has an average pooling layer and a maximum pooling layer, the number of network output layers of each layer of the pooling layer is increased along with the increase of the network depth, and the feature graphs output by the same layer of the pooling layer have the same size. And (3) connecting all the output feature maps of the pooling layer in series (namely 'splicing' in fig. 2) at a specific network layer, and performing convolution operation again through a plurality of convolution layers to realize the feature fusion effect of all the feature maps. And in the output part of the deep neural network, converting the output of the deep neural network into vectors by adopting a global average pooling layer, and finally reducing the output of the deep neural network to the specified binary hash coding length through a plurality of full connection layers.
And step 3: and randomly dividing the images of the training set into a plurality of small batches to be input into the neural network, and performing deep neural network training by adopting a contrast loss function. Given an arbitrary batch of images I ═ I0,I1,...IN-1Y ═ Y, its label0,y1...yN-1N is the size of a batch of images. Get any intention pair Ii、Ij,i,j∈[0,N]With loss value L of the contrast loss functionpair(Ii,Ij):
Figure BDA0001692013420000051
Wherein, delta (I)i,Ij) Indicating whether the image pairs are similar, d (I)i,Ij) Representing the distance between the image pair, margin is the edge coefficient. Delta (I)i,Ij) And d (I)i,Ij) Are defined as follows:
Figure BDA0001692013420000052
wherein, yiAnd yjAre labels, y 'of the i, j images respectively'i,kAnd y'j,kRespectively the output values of the ith image and the jth image of the neural network at the kth node,and m is the number of the output nodes of the neural network, and the length of the output nodes is the same as that of the extracted binary hash codes in numerical value.
Therefore, there is a total loss value L for all image pairs within a batch:
Figure BDA0001692013420000061
binding of Lpair(Ii,Ij) By definition, L can be rewritten as:
Figure BDA0001692013420000062
and 4, step 4: loading a trained neural network model, inputting all training set images into a neural network to obtain corresponding binary Hash codes with semantic information, and simultaneously carrying out Hash coding on the images of the training set by adopting various traditional Hash coding methods such as common Hash, locality sensitive Hash and the like to obtain the binary Hash codes with visual information. And combining the two binary hash codes to establish a local image feature library.
And 5: inputting the images of the test set into a neural network, repeating the step 4) to obtain two types of binary hash codes, comparing the binary hash codes with the binary hash codes of all training images of the local feature library, and returning the results of the image retrieval to the user from top to bottom according to the similarity. The similarity calculation method is as follows:
given test image ItestInputting the binary hash code into a neural network to obtain a binary hash code H with semantic information0:
H0=sign(Y')
Wherein, Y' is the output of the neural network, sign (#) is a sign taking function.
Assuming that n traditional image hash code generation methods are adopted, the obtained binary hash codes with visual information are respectively H1,H2…HnThen, ItestTotal binary hash encoding Htest={H0,H1,...Hn}. Then for each image I from the feature librarytrainIts binary hash encoding is H'train={H'0,H'1,...H'nThe similarity sim (I) is shown as the following two kindstest,Itrain) And (3) calculating a scheme:
the first method comprises the following steps:
Figure BDA0001692013420000063
and the second method comprises the following steps:
Figure BDA0001692013420000064
wherein λi∈(0,1]And lambda01+...λn=1,s(Hi,H'i) Comprises the following steps:
Figure BDA0001692013420000071
wherein h isk、h'kAre respectively HiAnd H'iValue of the k-th digit, ziThe number of bits for the ith hash code.
In order to verify the effectiveness of the method, the invention carries out comparison tests on public data sets CIFAR-10 and NUS-WIDE, and compares the method with the existing method.
The CIFAR-10 dataset contains 60,000 images, which the authorities classify into 10 categories of 50,000 training images and 10,000 test images. The present invention uses a given 50,000 training images to train a deep neural network, randomly partitioning 10,000 test images into 1,000 query sets and 9,000 image sets.
The NUS-WIDE dataset contains 19,5834 images from Flickr, for a total of 81 classes. The invention uses the most common partitioning method on this data set to select the 21 most common categories, comprising 19,5834 images, each category containing at least 5,000 images. Then, 10,000 images are randomly selected from the images as a test set, and the rest images are used as a training set. Finally, 1,000 pieces of images are randomly selected from the test set as a query set, and the rest 9,000 pieces of images are used as the queried image set.
In order to simplify the experiment, the invention only adopts the proposed binary hash codes with semantic information to carry out experiment comparison, and thoroughly inspects the average accuracy of various methods under different hash code lengths. The results of the experiment are shown in table 1. From experimental results, the method provided by the invention has better image retrieval effect than the existing advanced method no matter on a CIFAR-10 data set or a NUS-WIDE data set.
TABLE 1 average accuracy of image retrieval at different Hash code lengths
Figure BDA0001692013420000072
In Table 1 above, the CNNH + method reference, "Supervised Hashing for Image Retrieval Learning"; DNNH methods reference "Simultaneous feed Learning and Hash Coding with Deep Neural Networks"; DLBHC methods reference "Deep Learning of Binary Hash Codes for Fast Image Retrieval"; DSH methods references "Deep Supervised Hashing for Fast Image Retrieval"; SUBIC method reference "SUBIC A super, structured bind code for image search".
The pooling method employed in the present invention at each level is not limited to average pooling and maximum pooling, but may be a variety of combinations of existing pooling methods.
The above-mentioned embodiments and the drawings are only for illustrating the technical principles of the present invention and are not to be construed as limiting the present invention. The technical solution of the present invention can be changed and modified equally by those skilled in the art, and the protection scope of the present invention should be subject to the limitation of the claims.

Claims (6)

1. An image retrieval method based on deep learning is characterized by comprising the following steps:
1) constructing a deep neural network for extracting binary Hash codes of the image;
2) inputting the images of the training set into the deep neural network in batches, training the model of the deep neural network, and storing the trained deep neural network model;
3) loading a deep neural network model, inputting all images of a training set into the deep neural network, acquiring binary hash codes with semantic information, acquiring the binary hash codes with visual information by adopting a traditional binary hash coding method, and establishing a local feature library by utilizing the binary hash codes with the semantic information and the binary hash codes with the visual information;
4) inputting an image to be retrieved into a deep neural network, acquiring a corresponding binary hash code with semantic information, acquiring the binary hash code with visual information of the image to be retrieved by adopting a traditional binary hash coding method, comparing the acquired binary hash code of the image to be retrieved with a local feature library, and acquiring a retrieval result by calculating similarity;
step 1) each layer of the pooling layer of the deep neural network is provided with an average pooling layer and a maximum pooling layer, the number of network output layers of each layer of pooling layer is increased along with the increase of the network depth, and the sizes of characteristic graphs output by the same layer of pooling layer are the same; all the output feature maps of the pooling layer are connected in series in a specific network layer, and are subjected to convolution operation again through a plurality of convolution layers, so that the feature fusion effect of all the feature maps is realized; in the output part of the deep neural network, converting the output of the deep neural network into vectors by adopting a global average pooling layer, and finally reducing the output of the deep neural network to the specified binary hash coding length through a plurality of full connection layers;
step 2) training the deep neural network by adopting a contrast loss function, wherein the contrast loss function is calculated by adopting the following steps:
a) given an arbitrary batch of images I ═ I0,I1,...IN-1Y ═ Y, its label0,y1...yN-1N is the size of a batch of images, and any image pair is taken as Ii、Ij,i,j∈[0,N]With loss value L of the contrast loss functionpair(Ii,Ij):
Figure FDA0002967715030000011
Wherein, delta (I)i,Ij) Indicating whether the image pairs are similar, d (I)i,Ij) Representing the distance between the image pairs, margin being an edge coefficient;
δ(Ii,Ij) And d (I)i,Ij) Are defined as follows:
Figure FDA0002967715030000012
wherein, yiAnd yjAre labels, y 'of the i, j images respectively'i,kAnd y'j,kThe output values of the ith and the j images of the neural network at the kth node and the number of the output nodes of the m neural network are the same as the length of the extracted binary hash codes in terms of numerical values;
b) there is a total loss value L for all image pairs within a batch:
Figure FDA0002967715030000021
binding of Lpair(Ii,Ij) Definition, L is rewritten as:
Figure FDA0002967715030000022
2. the method of claim 1, wherein the deep neural network data set partitioning method of step 1) is: if a large number of available training images do not exist in an actual application scene, an image data set close to the actual application scene is additionally taken as a training set, otherwise, the existing data set is divided into the training set and a testing set, each image has a respective label, the labels are set according to actual requirements, and the label of each image can be not unique.
3. The method of claim 1, wherein step 3) combines the binary hash code with semantic information obtained by the neural network with the binary hash code with visual information obtained by using a plurality of conventional binary hash coding methods as the features of the image.
4. The method of claim 1, wherein step 4) measures the similarity between images using the following similarity calculation method:
a) given image I to be retrievedtestInputting the data into a deep neural network to obtain binary Hash code H with semantic information0:
H0=sign(Y'),
Wherein, Y' is the output of the neural network, sign (#) is a sign taking function;
b) assuming that n traditional image hash code generation methods are adopted, the obtained binary hash codes are respectively H1,H2…HnThen, ItestTotal binary hash encoding Htest={H0,H1,...Hn}; then for each image I from the feature librarytrainIts binary hash encoding is H'train={H'0,H'1,...H'n};
c) Calculating the similarity sim (I) by adopting one of the following two schemestest,Itrain):
The first method comprises the following steps:
Figure FDA0002967715030000023
and the second method comprises the following steps:
Figure FDA0002967715030000031
λ0=0,
wherein λi∈(0,1]And lambda01+...λn=1,s(Hi,H'i) Comprises the following steps:
Figure FDA0002967715030000032
wherein h isk、h'kAre respectively HiAnd H'iValue of the k-th digit, ziThe number of bits for the ith hash code.
5. An image retrieval apparatus based on deep learning, characterized by comprising:
the deep neural network building module is responsible for building a deep neural network and is used for extracting binary Hash codes of the image;
the deep neural network training module is responsible for inputting the training set images into the deep neural network in batches, training the model of the deep neural network and storing the trained deep neural network model;
the local feature library establishing module is responsible for loading a deep neural network model, inputting all images of a training set into the deep neural network, acquiring binary hash codes with semantic information, acquiring the binary hash codes with visual information by adopting a traditional binary hash coding method, and establishing a local feature library by utilizing the binary hash codes with the semantic information and the binary hash codes with the visual information;
the retrieval module is responsible for inputting an image to be retrieved into a deep neural network, acquiring a corresponding binary hash code with semantic information, acquiring the binary hash code with visual information of the image to be retrieved by adopting a traditional binary hash coding method, comparing the acquired binary hash code of the image to be retrieved with a local feature library, and acquiring a retrieval result by calculating similarity;
each layer of the pooling layer of the deep neural network is provided with an average pooling layer and a maximum pooling layer, the number of network output layers of each layer of the pooling layer is increased along with the increase of the network depth, and the sizes of feature graphs output by the same layer of the pooling layer are the same; all the output feature maps of the pooling layer are connected in series in a specific network layer, and are subjected to convolution operation again through a plurality of convolution layers, so that the feature fusion effect of all the feature maps is realized; in the output part of the deep neural network, converting the output of the deep neural network into vectors by adopting a global average pooling layer, and finally reducing the output of the deep neural network to the specified binary hash coding length through a plurality of full connection layers;
the deep neural network training module trains the deep neural network by adopting a contrast loss function, and the contrast loss function is calculated by adopting the following steps:
a) given an arbitrary batch of images I ═ I0,I1,...IN-1Y ═ Y, its label0,y1...yN-1N is the size of a batch of images, and any image pair is taken as Ii、Ij,i,j∈[0,N]With loss value L of the contrast loss functionpair(Ii,Ij):
Figure FDA0002967715030000041
Wherein, delta (I)i,Ij) Indicating whether the image pairs are similar, d (I)i,Ij) Representing the distance between the image pairs, margin being an edge coefficient; delta (I)i,Ij) And d (I)i,Ij) Are defined as follows:
Figure FDA0002967715030000042
wherein,yiAnd yjAre labels, y 'of the i, j images respectively'i,kAnd y'j,kThe output values of the ith and the j images of the neural network at the kth node and the number of the output nodes of the m neural network are the same as the length of the extracted binary hash codes in terms of numerical values; b) there is a total loss value L for all image pairs within a batch:
Figure FDA0002967715030000043
binding of Lpair(Ii,Ij) Definition, L is rewritten as:
Figure FDA0002967715030000044
6. the apparatus of claim 5, wherein the retrieval module measures similarity between images using the following similarity calculation method:
a) given image I to be retrievedtestInputting the data into a deep neural network to obtain binary Hash code H with semantic information0:
H0=sign(Y'),
Wherein, Y' is the output of the neural network, sign (#) is a sign taking function;
b) assuming that n traditional image hash code generation methods are adopted, the obtained binary hash codes are respectively H1,H2…HnThen, ItestTotal binary hash encoding Htest={H0,H1,...Hn}; then for each image I from the feature librarytrainIts binary hash encoding is H'train={H'0,H'1,...H'n};
c) Calculating the similarity sim (I) by adopting one of the following two schemestest,Itrain):
The first method comprises the following steps:
Figure FDA0002967715030000045
and the second method comprises the following steps:
Figure FDA0002967715030000051
λ0=0,
wherein λi∈(0,1]And lambda01+...λn=1,s(Hi,H'i) Comprises the following steps:
Figure FDA0002967715030000052
wherein h isk、h'kAre respectively HiAnd H'iValue of the k-th digit, ziThe number of bits for the ith hash code.
CN201810597022.9A 2018-06-11 2018-06-11 Image retrieval method and device based on deep learning Active CN110110120B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810597022.9A CN110110120B (en) 2018-06-11 2018-06-11 Image retrieval method and device based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810597022.9A CN110110120B (en) 2018-06-11 2018-06-11 Image retrieval method and device based on deep learning

Publications (2)

Publication Number Publication Date
CN110110120A CN110110120A (en) 2019-08-09
CN110110120B true CN110110120B (en) 2021-05-25

Family

ID=67483331

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810597022.9A Active CN110110120B (en) 2018-06-11 2018-06-11 Image retrieval method and device based on deep learning

Country Status (1)

Country Link
CN (1) CN110110120B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114419712A (en) * 2020-05-14 2022-04-29 支付宝(杭州)信息技术有限公司 Feature extraction method for protecting personal data privacy, model training method and hardware
CN111914652B (en) * 2020-07-02 2022-07-29 河海大学 Face feature binary coding and recognition method based on two-layer self-organizing neural network
CN112381074B (en) * 2021-01-15 2021-05-25 北京每日优鲜电子商务有限公司 Image recognition method and device, electronic equipment and computer readable medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529605A (en) * 2016-11-28 2017-03-22 东华大学 Image identification method of convolutional neural network model based on immunity theory
CN106909905A (en) * 2017-03-02 2017-06-30 中科视拓(北京)科技有限公司 A kind of multi-modal face identification method based on deep learning
CN107510452A (en) * 2017-09-30 2017-12-26 扬美慧普(北京)科技有限公司 A kind of ECG detecting method based on multiple dimensioned deep learning neutral net

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7831531B1 (en) * 2006-06-22 2010-11-09 Google Inc. Approximate hashing functions for finding similar content
US20180068344A1 (en) * 2013-10-15 2018-03-08 Jennifer L. Sultzaberger Systems and methods for management of media campaigns
CN106354735A (en) * 2015-07-22 2017-01-25 杭州海康威视数字技术股份有限公司 Image target searching method and device
CN105260412A (en) * 2015-09-24 2016-01-20 东方网力科技股份有限公司 Image storage method and device, and image retrieval method and device
CN105512289B (en) * 2015-12-07 2018-08-14 郑州金惠计算机系统工程有限公司 Image search method based on deep learning and Hash
WO2017142629A1 (en) * 2016-02-18 2017-08-24 Google Inc. Image classification neural networks
CN106126581B (en) * 2016-06-20 2019-07-05 复旦大学 Cartographical sketching image search method based on deep learning
CN106407352B (en) * 2016-09-06 2019-09-10 广东顺德中山大学卡内基梅隆大学国际联合研究院 Traffic image search method based on deep learning
CN107092661A (en) * 2017-03-28 2017-08-25 桂林明辉信息科技有限公司 A kind of image search method based on depth convolutional neural networks
CN107402947B (en) * 2017-03-29 2020-12-08 北京猿力教育科技有限公司 Picture retrieval model establishing method and device and picture retrieval method and device
CN107330446B (en) * 2017-06-05 2020-08-04 浙江工业大学 Image classification-oriented deep convolutional neural network optimization method
CN107229757B (en) * 2017-06-30 2020-09-29 中国科学院计算技术研究所 Video retrieval method based on deep learning and Hash coding

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106529605A (en) * 2016-11-28 2017-03-22 东华大学 Image identification method of convolutional neural network model based on immunity theory
CN106909905A (en) * 2017-03-02 2017-06-30 中科视拓(北京)科技有限公司 A kind of multi-modal face identification method based on deep learning
CN107510452A (en) * 2017-09-30 2017-12-26 扬美慧普(北京)科技有限公司 A kind of ECG detecting method based on multiple dimensioned deep learning neutral net

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Detect face in the wild using CNN cascade with feature aggregation at multi-resolution;Jingjing Deng 等;《Image Processing 》;20180222;第4167-4171页 *
基于卷积神经网络的中文微博观点分类;廖祥文 等;《模式识别与人工智能》;20161215;第1075-1082页 *
结合全局与局部池化的深度哈希人脸识别算法;曾燕 等;《西安电子科技大学学报(自然科学版)》;20180320;第163-169页 *

Also Published As

Publication number Publication date
CN110110120A (en) 2019-08-09

Similar Documents

Publication Publication Date Title
CN108920720B (en) Large-scale image retrieval method based on depth hash and GPU acceleration
CN107330074B (en) Image retrieval method based on deep learning and Hash coding
CN106033426B (en) Image retrieval method based on latent semantic minimum hash
CN108984642B (en) Printed fabric image retrieval method based on Hash coding
CN107506793B (en) Garment identification method and system based on weakly labeled image
CN104036012B (en) Dictionary learning, vision bag of words feature extracting method and searching system
CN111753189A (en) Common characterization learning method for few-sample cross-modal Hash retrieval
CN101710334A (en) Large-scale image library retrieving method based on image Hash
CN110110120B (en) Image retrieval method and device based on deep learning
CN113377981B (en) Large-scale logistics commodity image retrieval method based on multitask deep hash learning
CN103186538A (en) Image classification method, image classification device, image retrieval method and image retrieval device
CN110619051A (en) Question and sentence classification method and device, electronic equipment and storage medium
CN113177141B (en) Multi-label video hash retrieval method and device based on semantic embedded soft similarity
CN110598061A (en) Multi-element graph fused heterogeneous information network embedding method
CN113127632B (en) Text summarization method and device based on heterogeneous graph, storage medium and terminal
CN108491430A (en) It is a kind of based on the unsupervised Hash search method clustered to characteristic direction
CN111723692B (en) Near-repetitive video detection method based on label features of convolutional neural network semantic classification
CN108805280B (en) Image retrieval method and device
CN113269224A (en) Scene image classification method, system and storage medium
CN113988147A (en) Multi-label classification method and device for remote sensing image scene based on graph network, and multi-label retrieval method and device
CN109933682A (en) A kind of image Hash search method and system based on semanteme in conjunction with content information
CN112163114A (en) Image retrieval method based on feature fusion
Al-Jubouri Content-based image retrieval: Survey
CN110674334B (en) Near-repetitive image retrieval method based on consistency region deep learning features
Zhang et al. Improved image retrieval algorithm of GoogLeNet neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20211021

Address after: 101399 No. 119, A2 CAIDA Second Street, Nancai Town, Shunyi District, Beijing

Patentee after: Beijing Ford Technology Development Co.,Ltd.

Address before: 100144 Beijing City, Shijingshan District Jin Yuan Zhuang Road No. 5, North China University of Technology

Patentee before: NORTH CHINA University OF TECHNOLOGY

TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20230118

Address after: Room 503, Building 1, Xifuhe Industry University Research Collaborative Innovation Center, No. 20, Xinyuan Road, Xinhe Street, Yannan High tech Zone, Yancheng City, Jiangsu Province, 224000

Patentee after: Jiangsu Tianfenghaizhiyuan Communication Power Technology Co.,Ltd.

Address before: 101399 No. 119, A2 CAIDA Second Street, Nancai Town, Shunyi District, Beijing

Patentee before: Beijing Ford Technology Development Co.,Ltd.

TR01 Transfer of patent right