CN108304573A - Target retrieval method based on convolutional neural networks and supervision core Hash - Google Patents
Target retrieval method based on convolutional neural networks and supervision core Hash Download PDFInfo
- Publication number
- CN108304573A CN108304573A CN201810157751.2A CN201810157751A CN108304573A CN 108304573 A CN108304573 A CN 108304573A CN 201810157751 A CN201810157751 A CN 201810157751A CN 108304573 A CN108304573 A CN 108304573A
- Authority
- CN
- China
- Prior art keywords
- image
- hash
- formula
- layer
- convolutional neural
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 82
- 238000013527 convolutional neural network Methods 0.000 title claims abstract description 58
- 238000012549 training Methods 0.000 claims abstract description 48
- 230000002708 enhancing effect Effects 0.000 claims abstract description 8
- 230000004304 visual acuity Effects 0.000 claims abstract description 5
- 239000011159 matrix material Substances 0.000 claims description 31
- 238000005070 sampling Methods 0.000 claims description 27
- 239000013598 vector Substances 0.000 claims description 19
- 238000013459 approach Methods 0.000 claims description 8
- 238000013507 mapping Methods 0.000 claims description 8
- 230000000644 propagated effect Effects 0.000 claims description 8
- 230000008569 process Effects 0.000 claims description 7
- 238000011478 gradient descent method Methods 0.000 claims description 6
- 210000002569 neuron Anatomy 0.000 claims description 5
- 238000010183 spectrum analysis Methods 0.000 claims description 5
- 238000010606 normalization Methods 0.000 claims description 4
- 230000006870 function Effects 0.000 description 52
- 238000010586 diagram Methods 0.000 description 11
- 238000000605 extraction Methods 0.000 description 10
- 201000011243 gastrointestinal stromal tumor Diseases 0.000 description 6
- 238000010276 construction Methods 0.000 description 5
- 238000013135 deep learning Methods 0.000 description 3
- 230000001537 neural effect Effects 0.000 description 3
- 238000001228 spectrum Methods 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 238000005457 optimization Methods 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000009467 reduction Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000013480 data collection Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 210000005036 nerve Anatomy 0.000 description 1
- 210000004218 nerve net Anatomy 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 238000013139 quantization Methods 0.000 description 1
- 230000002040 relaxant effect Effects 0.000 description 1
- 229920006395 saturated elastomer Polymers 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/51—Indexing; Data structures therefor; Storage structures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- Software Systems (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Biomedical Technology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
The present invention relates to search method field, it is based especially on convolutional neural networks and supervises the target retrieval method of core Hash.The search method includes:(1) it introduces convolutional neural networks to learn training image, the high-order for implicitly being learnt to obtain image data using its special networks structure is indicated, further feature is generated;(2) supervision core hash method of the enhancing to the resolving power of linearly inseparable data is introduced, simultaneously object function is proposed using the equivalence relation of Hash codes inner product and Hamming distance, and the affinity information of combined training image exercises supervision study to dimensional images feature, and generate Hash codes;(3) image index is constructed using trained hash function, realizes the retrieval to large-scale image data.The present invention greatly improves target retrieval efficiency by the target retrieval method based on convolutional neural networks and supervision core Hash, enhances the practicability under big data environment.
Description
Technical field
The present invention relates to search method field, it is based especially on convolutional neural networks and supervises the target retrieval side of core Hash
Method.
Background technology
With the arrival in big data epoch, internet video image resource rapidly increases, how to extensive video, image
Concern target in resource is fast and effeciently retrieved urgently to be resolved hurrily to meet user demand.Although local feature region goes out
It is existing, such as SIFT (Scale-Invariant Feature Transform), HOG (Histogram of Orientated
) etc. Gradients excellent performance is shown in image processing field, but generate these and describe the fixed coding step of the period of the day from 11 p.m. to 1 a.m
So that description lacks learning ability, its picture material ability to express is limited, it is difficult to various image data is adapted to, certain
Large-scale image target retrieval performance is reduced in degree.The inherent implication relation of great amount of images data in order to obtain generates more
With distinction and representative feature, deep learning (Deep Learning) is applied to image procossing by the scholars such as Hinton
In field, new approaches are provided to extract significantly more efficient characteristics of image.
Depth confidence network (Deep Belief Network, DBN) top layer uses three rank Boltzmann machine (Boltzmann
Machine, BM), improved DBN is used for objective feature extraction, this feature is rotationally-varying to target to have higher Shandong
Stick.In addition, researcher construct convolution depth confidence network (Convolutional Deep Belief Network,
CDBN), learn effective high-order character representation in the natural image that can be never marked using CDBN.And by convolutional Neural
SPP (Spatial are added in the convolutional layer of network (Convolutional Neural Network, CNN) and full articulamentum
Pyramid Pooling) layer, directly different size image is learnt and generates Analysis On Multi-scale Features.
But deep learning generate characteristics of image dimension it is higher, there are problems that dimension disaster, when image data scale compared with
When big, retrieval rate will be made drastically by carrying out retrieval according to traditional nearest _neighbor retrieval method (such as R-tree, KD-tree)
Decline, it is difficult to be suitable for large-scale data.Extensive high dimensional image is effectively retrieved to realize, researcher proposes
Approximate KNN search strategy (Approximate Nearest Neighbor, ANN).Wherein, salted hash Salted (Hashing) is
The main stream approach of approximate KNN search problem is solved, thought is by dimensional images Feature Mapping using hash function race to low
In dimension space, while making distance is closer in former space point is mapped to after lower dimensional space still to keep closer distance.LSH and its
The hash function of innovatory algorithm construction is all unrelated with data, and in recent years, how researchers are in conjunction with data characteristics structure
It makes hash function that is effective, compacting and proposes many algorithms.Hash method (Spetral Hashing, SH) is composed first to similar
The Laplacian Matrix characteristic value and feature vector of figure are analyzed, then by the condition of relaxing the restriction, will be to image feature vector
The dimensionality reduction problem that encoded question is converted to Laplce's characteristic pattern solves, this method rely on data itself generate index than with
Machine generates hash function method and reaches higher accuracy rate.But unsupervised method does not consider the semantic information of image,
And user is often more likely to the semantic information of retrieval result.For this purpose, researcher proposes semi-supervised hash method (Semi-
Supervised Hashing, SSH).Researchers also proposed some full supervision and breathe out on the basis of semi-supervised learning method
Uncommon method, such as SH (Semantic Hashing), BRE (Binary Reconstructive Embedding), MLH
(Minimal Loss Hashing) etc., full hash method of supervising can reach higher accuracy rate compared to non-supervisory method, but
The problems such as to be that there are optimization process complex, training effectiveness is low, the application that this severely limits it on large-scale dataset.
Invention content
The technical problem to be solved by the present invention is to:In order to solve existing unsupervised method, there is no the languages for considering image
The deficiency of adopted information, the target retrieval method based on convolutional neural networks and supervision core Hash that the present invention provides a kind of, passes through
Target retrieval method based on convolutional neural networks and supervision core Hash can utilize convolutional neural networks to Large Scale Graphs
As data characteristics progress autonomous learning, enhance the ability to express of characteristics of image.Secondly, by supervision core hash method to dimensional images
Further feature exercises supervision study, and high dimensional feature is mapped in low-dimensional Hamming space, generates the Hash codes compacted, greatly
Target retrieval efficiency is improved, the practicability under big data environment is enhanced.
The technical solution adopted by the present invention to solve the technical problems is:
A kind of target retrieval method based on convolutional neural networks and supervision core Hash, includes the following steps:
(1) convolutional neural networks are introduced to learn training image, is implicitly learnt using its special networks structure
High-order to image data indicates, generates further feature;
(2) supervision core hash method of the enhancing to the resolving power of linearly inseparable data is introduced, while using in Hash codes
Product and the equivalence relation of Hamming distance propose object function, and the affinity information of combined training image to dimensional images feature into
Row supervised learning, and generate Hash codes;
(3) image index is constructed using trained hash function, realizes the retrieval to large-scale image data.
Specifically, the input picture size of the convolutional neural networks is 227 × 227, is exported deep for 4096 × 1 image
Layer feature includes 5 convolutional layers, 3 sub- sample levels altogether;In the characteristic pattern of convolutional layer preceding layerWith the volume that can learn
Product core KijConvolution is carried out, the result of convolution generates the characteristic pattern of this layer through nonlinear function f ()Concrete form is such as
Under:
Formula one:
For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijIt can be with preceding layer
One or more features figure determines convolution relation, MjRepresent input feature vector set of graphs, common nonlinear function have f (x)=
Tanh (x) and f (x)=(1+e-x)-1, the characteristic pattern size h of convolutional layer generationlIt can be calculated by formula two:
Formula two:
hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel moving step length, ρl
To the columns of preceding layer characteristic pattern edge zero padding when expression convolution algorithm;Each layer convolution kernel size Z={ z1=11, z2=5, z3=
z4=z5=3 }, moving step length Λ={ λ1=4, λ2=λ3=λ4=λ5=1 }, characteristic pattern edge zero padding columns P={ ρ1=0, ρ2
=2, ρ3=ρ4=ρ5=1 };Sub-sampling layer carries out maximum value sampling, sampling area 3 using overlap sampling method to characteristic pattern
× 3, sampling step length is 2 pixels;
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
The propagated forward stage:Sample (X, a Y are chosen from training samplep), X is transmitted to from input layer through converting step by step
Output layer calculates corresponding reality output:
Formula three:Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n))
The back-propagating stage:The stage is the error propagation stage, calculates reality output OpY is exported with corresponding idealp
Error:
Formula four:
By error EpReversely successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized,
As overall error E≤ε, the training of the batch training sample is completed, after the completion of the training of all batches, by image input convolution god
Through in network, image data step by step by each network layer after, can be obtained the further feature of image in output end.
Specifically, measure image between apart from when, give the Hash codes dimension r, then need r coefficient vector
a1,…,arConstruct hash functionThe label information of training image can lead to
The semantic dependency and space length for crossing image obtain, and are defined as description label image collection χl={ x1,…,xlIn between element
Correlation check matrix
Formula five:
Make image xi,xjHamming distance Dh(xi,xj) meet:
Formula six:
Hash codes distance, the Hash codes code of image x are calculated using inner product of vectors operationr(x)=[h1(x),…,hr(x)]
∈{1,-1}1×r, then image xi,xjDistance calculate as shown in formula 11:
D(xi,xj)=coder(xi)·coder(xj)
=| k | hk(xi)=hk(xj),1≤k≤r}|-|{k|hk(xi)≠hk(xj),1≤k≤r}
=r-2 | and k | hk(xi)≠hk(xj),1≤k≤r}|
Formula seven:=r-2Dh(xi,xj)
D(xi,xj) ∈ [- r, r], to D (xi,xj) obtain after normalizationDefinition is so that similar
MatrixThe minimum object function with check matrix S distances:
Formula eight:
Matrix F robenius norms are sought in expression,For the Kazakhstan of label image collection χ l
Uncommon code matrix, is generalized to matrix form, according to formula (3) H by sgn ()lIt can be expressed as:
Formula nine:
Hl is substituted into public
Formula eight
Formula ten:
Formula 11:
Define matrixWherein R0=rS can then lead to
It crosses greedy algorithm minimum formula (11) and gradually estimates ak:
Formula 12:
Remove constant term, more succinct object function can be obtained:
Formula 13:
WithSgn (x) is replaced, then approximate objective functionAs shown in formula 14:
Formula 14:
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Formula 15:
⊙ indicates Hadamard inner product operations, to accelerateIt restrains and is breathed out with spectrum
Spectral analysis method in uncommon generates initial valueGradient search process is further accelerated, it is right after obtaining hash function H and Hash table H
The further feature of query image carries out Hash mapping and obtains coder(xq), calculate coder(xq) with Hash table H in Hash codes away from
From the closer image of layback is as retrieval result.
The beneficial effects of the invention are as follows:The target based on convolutional neural networks and supervision core Hash that the present invention provides a kind of
Search method can utilize convolutional Neural by the target retrieval method based on convolutional neural networks and supervision core Hash
Network carries out autonomous learning to large-scale image data feature, enhances the ability to express of characteristics of image.Secondly, by supervision core Hash
Method exercises supervision study to dimensional images further feature, and high dimensional feature is mapped in low-dimensional Hamming space, and generation is compacted
Hash codes, greatly improve target retrieval efficiency, enhance the practicability under big data environment.
Description of the drawings
Present invention will be further explained below with reference to the attached drawings and examples.
Fig. 1 is the structural schematic diagram of the convolutional neural networks of the present invention;
Fig. 2 is convolutional neural networks structure chart of the present invention for the extraction of image further feature;
Fig. 3 is the schematic diagram of present invention target retrieval MAP value on ImageNet-100 databases;
Fig. 4 is schematic diagram of the training time of the invention with Hash code bit number r change curves;
Fig. 5 is the schematic diagram of present invention target retrieval MAP value on Caltech-256 databases;
Fig. 6 is the schematic diagram of present invention Precision-Recall curves on Caltech-256 databases;
Specific implementation mode
Fig. 1 is the structural schematic diagram of the convolutional neural networks of the present invention, and Fig. 2 is that the present invention is extracted for image further feature
Convolutional neural networks structure chart, Fig. 3 be the present invention on ImageNet-100 databases target retrieval MAP value schematic diagram,
Fig. 4 is schematic diagram of the training time of the invention with Hash code bit number r change curves, and Fig. 5 is the present invention in Caltech-256 data
The schematic diagram of target retrieval MAP value on library, Fig. 6 are that present invention Precision-Recall on Caltech-256 databases is bent
The schematic diagram of line.
First, convolutional neural networks are introduced to learn training image, is implicitly learnt using its special networks structure
The high-order for obtaining image data indicates, generates the further feature with stronger distinction and ability to express;
Secondly, introducing supervision core hash method (Kernel-Based Supervised Hashing, KSH) enhances to line
Property can not divided data resolving power, while being proposed using the equivalence relation of Hash codes inner product and Hamming distance simpler, effective
Object function, and the affinity information of combined training image exercises supervision study to dimensional images feature, and generates and compact
Hash codes;
Finally, trained hash function constructs image index for utilization, realizes the efficient inspection to large-scale image data
Rope.
The target retrieval method based on convolutional neural networks and supervision core Hash utilizes convolutional Neural net first
The input picture size of the further feature of network structure extraction image, convolutional neural networks used is 227 × 227, and it is 4096 to export
× 1 image further feature includes 5 convolutional layers, 3 sub- sample levels altogether.In convolutional layer, the characteristic pattern of preceding layerWith
The convolution kernel K that can learnijConvolution is carried out, the result of convolution generates the characteristic pattern of this layer through nonlinear function f ()Tool
Body form is as follows:
Wherein,For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijCan with it is previous
The one or more features figure of layer determines convolution relation, MjInput feature vector set of graphs is represented, common nonlinear function has f (x)
=tanh (x) and f (x)=(1+e-x)-1, compared with above-mentioned nonlinear function, f (x)=max (0, x) can effectively improve trained effect
Rate.The characteristic pattern size h that convolutional layer generateslIt can be calculated by formula (2):
Wherein, hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel movement step
It is long, ρlTo the columns of preceding layer characteristic pattern edge zero padding when expression convolution algorithm.Here, each layer convolution kernel size Z={ z1=11,
z2=5, z3=z4=z5=3 }, moving step length Λ={ λ1=4, λ2=λ3=λ4=λ5=1 }, characteristic pattern edge zero padding columns P=
{ρ1=0, ρ2=2, ρ3=ρ4=ρ5=1 }.In sub-sampling layer, document shows, relative to traditional non-overlapping sampling, to use overlapping
Sampling can not only improve the accuracy of feature, be also prevented from the training stage and over-fitting occur, therefore, use overlap sampling here
Method carries out maximum value sampling to characteristic pattern, and sampling area is 3 × 3, and sampling step length is 2 pixels.
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
(1) the propagated forward stage.Sample (X, a Y are chosen from training samplep), X is from input layer through converting biography step by step
It is sent to output layer, calculates corresponding reality output:
Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n)) (3)
(2) back-propagating stage, also referred to as error propagation stage.Calculate reality output OpY is exported with corresponding idealpError:
By error EpReversely successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized,
As overall error E≤ε, the training of the batch training sample is completed.After the completion of the training of all batches, by image input convolution god
Through in network, image data step by step by each network layer after, can be obtained the further feature of image in output end.
The target retrieval method based on convolutional neural networks and supervision core Hash, schemes in measurement
As between apart from when, give Hash codes dimension r, then need r coefficient vector a1,…,arConstruct hash functionThe label information of training image can pass through the semantic dependency and space length of image
It obtains, lable (xi,xj)=1 indicates image xi,xjIt is similar;Conversely, lable (xi,xjThe representative image x of)=- 1i,xjDifference
It is very big.To describe label image collection χl={ x1,…,xlIn correlation between element, define check matrix
Wherein, lable (xi,xi) ≡ 1, Sii≡ 1, Sij=0 indicates image xi,xjBetween similitude it is uncertain.For enhancing
The separating capacity of Hash codes so that the similitude between image can be efficiently judged in Hamming space, it should be as possible so that image
xi,xjHamming distance Dh(xi,xj) meet:
Due to Hamming distance calculation formula Dh(xi,xj)=| k | hk(xi)≠hk(xj), 1≤k≤r } | form is complicated, very
Hardly possible directly optimizes it, therefore utilizes inner product of vectors operation to calculate Hash codes distance herein.The Hash codes code of image xr
(x)=[h1(x),…,hr(x)]∈{1,-1}1×r, then image xi,xjDistance calculate as shown in formula (11):
D(xi,xj)=coder(xi)·coder(xj)
=| k | hk(xi)=hk(xj),1≤k≤r}|-|{k|hk(xi)≠hk(xj),1≤k≤r}|
=r-2 | and k | hk(xi)≠hk(xj),1≤k≤r}|
=r-2Dh(xi,xj) (7)
Formula (7) is shown to be consistent by Hash codes inner product operation and Hamming distance operation, and D (xi,xj)∈[-r,
R], to D (xi,xj) obtain after normalizationTo make similar matrixWith check matrix S
Distance is minimum, objective function:
Wherein,Matrix F robenius norms are sought in expression,For label image collection χlKazakhstan
Uncommon code matrix.Sgn () is generalized to matrix form, according to formula (3) HlIt can be expressed as:
Wherein,By HlSubstitution formula (8)
Compared with BRE and MLH, object function Γ (A) calculates similitude by inner product, is built to parameter A
Mould is more intuitive.It is assumed that at the t=k moment, it is known that vectorIt needs to estimate ak, define matrixWherein R0=rS then can minimize formula (11) by greedy algorithm and gradually estimate
ak:
Remove constant term, more succinct object function can be obtained:
Since sgn (x) functions in object function makeDiscontinuously, andNor convex function, it is difficult to directly
It connects pairIt minimizes, when | x | when > 6, continuous functionIt can approximation sgn (x), therefore profit well
WithSgn (x) is replaced, then approximate objective functionAs shown in formula (14):
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Wherein,⊙ indicates Hadamard inner product operations.It is smoothed that treatedIt is not convex function, globally optimal solution can not be acquired, in order to accelerateConvergence, the present invention utilize the spectral analysis method in spectrum Hash
Generate initial valueGradient search process is further accelerated, specific implementation step is as follows:
It supervises core hash method and realizes step:
Input:Training image collectionLabel image collectionAnd check matrix
Kernel function κHash code bit number r participates in trained sample number m (< l).
Pretreatment:M images are randomly selected from χ, are obtained
Training:Initialize R0=rS, Tmax=500;
For k=1 ..., r do
It is obtained using spectral analysis methodIt willSubstitute into object functionIn, pass through gradient
Descent method is calculated
end if
Rk←Rk-1-h*(h*)T
end for
Coding:For i=1 ..., n do
Output:Hash functionWith Hash table H={ coder(xi)|i∈[1,
n]}。
After obtaining hash function H and Hash table H, Hash mapping is carried out to the further feature of query image and obtains coder
(xq), calculate coder(xq) at a distance from Hash codes, the closer image of layback is as retrieval result in Hash table H.
Embodiment one:Target retrieval method of the present embodiment based on convolutional neural networks and supervision core Hash.First, it introduces
Convolutional neural networks learn training image, implicitly learn to obtain the high-order of image data using its special networks structure
It indicates, generates the further feature with stronger distinction and ability to express;Then, supervision core hash method (Kernel- is introduced
Based Supervised Hashing, KSH) enhancing utilizes Hash codes inner product to the resolving powers of linearly inseparable data
Simpler, effective object function, and the affinity information pair of combined training image are proposed with the equivalence relation of Hamming distance
Dimensional images feature exercises supervision study, and generates the Hash codes compacted;Finally, utilization trained hash function structural map
As index, the efficient retrieval to large-scale image data is realized.
Compared with traditional target retrieval method, the independent learning ability of characteristics of image is effectively enhanced, and can
Hash codes of compacting are generated using supervision core Hash, time overhead is reduced, enhances the practicability under large-scale data.
Embodiment two:Referring to Fig. 1, Fig. 2, the target retrieval based on convolutional neural networks and supervision core Hash of the present embodiment
Method generates the further feature of image data using following step:
First, the further feature of convolutional neural networks structure extraction image, the input figure of convolutional neural networks used are utilized
Picture size is 227 × 227, exports the image further feature for 4096 × 1, includes 5 convolutional layers, 3 sub- sample levels altogether.
Convolutional layer, the characteristic pattern of preceding layerWith the convolution kernel K that can learnijConvolution is carried out, the result of convolution is through nonlinear function f
() generates the characteristic pattern of this layerConcrete form is as follows:
Wherein,For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijCan with it is previous
The one or more features figure of layer determines convolution relation, MjInput feature vector set of graphs is represented, common nonlinear function has f (x)
=tanh (x) and f (x)=(1+e-x)-1, compared with above-mentioned nonlinear function, f (x)=max (0, x) can effectively improve trained effect
Rate.The characteristic pattern size h that convolutional layer generateslIt can be calculated by formula (2):
Wherein, hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel movement step
It is long, ρlTo the columns of preceding layer characteristic pattern edge zero padding when expression convolution algorithm.Here, each layer convolution kernel size Z={ z1=11,
z2=5, z3=z4=z5=3 }, moving step length Λ={ λ1=4, λ2=λ3=λ4=λ5=1 }, characteristic pattern edge zero padding columns P=
{ρ1=0, ρ2=2, ρ3=ρ4=ρ5=1 }.In sub-sampling layer, document shows, relative to traditional non-overlapping sampling, to use overlapping
Sampling can not only improve the accuracy of feature, be also prevented from the training stage and over-fitting occur, therefore, use overlap sampling here
Method carries out maximum value sampling to characteristic pattern, and sampling area is 3 × 3, and sampling step length is 2 pixels.
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
(1) the propagated forward stage.Sample (X, a Y are chosen from training samplep), X is from input layer through converting biography step by step
It is sent to output layer, calculates corresponding reality output:
Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n)) (3)
(2) back-propagating stage, also referred to as error propagation stage.Calculate reality output OpY is exported with corresponding idealpError:
By error EpReversely successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized,
As overall error E≤ε, the training of the batch training sample is completed.After the completion of the training of all batches, by image input convolution god
Through in network, image data step by step by each network layer after, can be obtained the further feature of image in output end.
Between measuring image apart from when, give Hash codes dimension r, then need r coefficient vector a1,…,arConstruction is breathed out
Uncommon functionThe label information of training image can pass through the semantic dependency and sky of image
Between distance obtain, lable (xi,xj)=1 indicates image xi,xjIt is similar;Conversely, lable (xi,xjThe representative image of)=- 1
xi,xjIt is widely different.To describe label image collection χl={ x1,…,xlIn correlation between element, define check matrix
Wherein, lable (xi,xi) ≡ 1, Sii≡ 1, Sij=0 indicates image xi,xjBetween similitude it is uncertain.For enhancing
The separating capacity of Hash codes so that the similitude between image can be efficiently judged in Hamming space, it should be as possible so that image
xi,xjHamming distance Dh(xi,xj) meet:
Due to Hamming distance calculation formula Dh(xi,xj)=| k | hk(xi)≠hk(xj), 1≤k≤r } | form is complicated, very
Hardly possible directly optimizes it, therefore utilizes inner product of vectors operation to calculate Hash codes distance herein.The Hash codes code of image xr
(x)=[h1(x),…,hr(x)]∈{1,-1}1×r, then image xi,xjDistance calculate as shown in formula (11):
D(xi,xj)=coder(xi)·coder(xj)
=| k | hk(xi)=hk(xj),1≤k≤r}|-|{k|hk(xi)≠hk(xj),1≤k≤r}|
=r-2 | and k | hk(xi)≠hk(xj),1≤k≤r}|
=r-2Dh(xi,xj) (7)
Formula (7) is shown to be consistent by Hash codes inner product operation and Hamming distance operation, and D (xi,xj)∈[-r,
R], to D (xi,xj) obtain after normalizationTo make similar matrixWith check matrix S
Distance is minimum, objective function:
Wherein,Matrix F robenius norms are sought in expression,For label image collection χlKazakhstan
Uncommon code matrix.Sgn () is generalized to matrix form, according to formula (3) HlIt can be expressed as:
Wherein,By HlSubstitution formula (8)
Compared with BRE and MLH, object function Γ (A) calculates similitude by inner product, more intuitive to parameter A modelings.It is false
It is scheduled on the t=k moment, it is known that vectorIt needs to estimate ak, define matrix
Wherein R0=rS then can minimize formula (11) by greedy algorithm and gradually estimate ak:
Remove constant term, more succinct object function can be obtained:
Since sgn (x) functions in object function makeDiscontinuously, andNor convex function, it is difficult to directly
It connects pairIt minimizes, when | x | when > 6, continuous functionIt can approximation sgn (x), therefore profit well
WithSgn (x) is replaced, then approximate objective functionAs shown in formula (14):
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Wherein,⊙ indicates Hadamard inner product operations.It is smoothed that treatedIt is not convex function, globally optimal solution can not be acquired, in order to accelerateConvergence, the present invention utilize the spectral analysis method in spectrum Hash
Generate initial valueGradient search process is further accelerated, after obtaining hash function H and Hash table H, to the further feature of query image
It carries out Hash mapping and obtains coder(xq), calculate coder(xq) at a distance from Hash codes, layback is closer in Hash table H
Image is as retrieval result.
Experimental result and analysis
Here, using being assessed the present invention on ImageNet-1000 and Caltech-256 image sets,
ImageNet-1000 image sets are a subsets of ImageNet image sets, are large scale visual identity contest (Large
Scale Visual Recognition Challenge, LSVRC) evaluation and test data set, including 1000 classifications total 1,200,000
Open image;Caltech-256 image sets are frequently-used data collection in target classification task, including total 30608 figures of 256 classifications
Picture, wherein including at least 80 images in each classification.Experimental Hardware be configured in save as 6G GPU equipment GTX Titan and
Intel Xeon CPU, the interior server for saving as 16G.Target retrieval performance indicator uses precision ratio and recall ratio, and definition is such as
Under:
The influence of parameter
First, core hash method (abbreviation KSH) retrieval performance is supervised with the variation of Hash code bit number r, this hair in order to verify
It is bright to be tested on ImageNet-1000 image sets, and be compared with some current mainstream hash methods, including
LSH(Locality Sensitive Hashing)、SKLSH(LSH with Shift-Invariant Kernels)、SH
(Spetral Hashing)、DSH(Density Sensitive Hashing)、PCA-ITQ(Iterative
Quantization of PCA), the methods of BRE (Binary Reconstructive Embedding).Experiment first from
50 classes are randomly selected in ImageNet-1000 image sets, and to this 50 class image zooming-out GIST feature;Then, from each class
Training set of the feature (feature for amounting to 50000 images) of 1000 images as supervised training hash function is randomly selected,
Remaining image is as inquiry use-case;Finally, it introduces hash method to be retrieved, it is as shown in Figure 3 to obtain experimental result:
From figure 3, it can be seen that with the increase of Hash code bit number r, the MAP value of each method increases, however, working as r
After increasing to certain value, MAP value increase amplitude, which tapers into, to be tended to be saturated.The target retrieval MAP value for comparing each hash method can
Know, retrieved using the present invention (KSH) has better performance compared with other main stream approach, this is because unsupervised hash method
(such as LSH, SH, DSH, PCA-ITQ etc.) and supervision hash method BRE do not utilize the semantic information construction of image to breathe out well
Uncommon function, causes retrieval performance relatively low;And KSH of the present invention introduces Kernel hash function and strengthens to linearly inseparable number
According to resolution capability, hash function is trained in combination with the affinity information of image, generates the Hash that more compacts
Code, to improve target retrieval performance.
Experiment is again compared the present invention and the hash function training time consumption of current main-stream hash method, specifically
As shown in Figure 4.It is not difficult to find out from Fig. 4, no prison fewer than supervision hash method using the time loss of unsupervised hash method
It superintends and directs hash method mostly and is to retain the position sensing of primitive character as optimization aim, and it is with image language to supervise hash method
For adopted neighbor information as supervision message, searching process is increasingly complex, compares unsupervised hash method time overhead bigger.
Experimental performance is analyzed
To verify the effective of the target retrieval method (abbreviation CNN+KSH) based on convolutional neural networks and supervision core Hash
Property, 50 class images have been randomly selected in ImageNet-1000 image sets is tested to obtain table 1.It is not difficult to find out from table 1,
Based on the CNN MAP values retrieved of extraction further feature than the MAP value retrieved based on global GIST features be higher by 10% with
On, illustrate that the image further feature using CNN extractions has stronger distinction and ability to express, this is because GIST features carry
It takes step to fix, does not have independent learning ability, so that its image expression limited ability, and CNN can imitate brain processing
The pattern of data carries out feature extraction to image, and the network structure of its deep layer can be excavated effectively in image and be closed implicit
System, enhances the image expression ability of feature.Wherein, the online retrieving time of (CNN+KSH) of the invention is 1.775 × 10-4Second,
It is suitable with other main stream approach, and MAP value has reached 40.79%, retrieval performance is apparently higher than other methods, therefore the present invention exists
There is stronger applicability under big data environment.
The target retrieval MAP value of 1 distinct methods of table and retrieval time comparison (64bits)
Comparison diagram 5, Fig. 6 are it is found that the present invention (CNN+KSH) MAP value under equal length Hash codes coding is above other masters
Stream method, and under conditions of ensureing identical precision ratio, CNN+KSH can reach recall ratio more higher than other methods.Document
In method be all to be re-introduced into after image zooming-out GIST features hash method construction index, although GIST features pass through it is multiple dimensioned
Multi-direction Gabor filter group is to having obtained global structure and spatial context information after image filtering, but GIST feature grains
Degree is more coarse, and lacks independent learning ability, limits its image expression ability, and the hash method in these documents
It only considers how that data characteristics is combined to construct hash function, there is no the semantic informations for utilizing image well, lead to its mesh
It is relatively low to mark retrieval performance, it is difficult to be adapted to extensive target retrieval;CNN+KSH methods utilize ImageNet-1000 image sets pair
CNN is trained, and a large amount of training image makes the training of CNN model parameters more abundant, can effectively excavate in image hidden
Containing relationship, the image expression ability of feature is effectively enhanced, and KSH supervises hash function using image similarity information
Supervise and instruct white silk, while searching process is accelerated using greedy algorithm and gradient descent method so that the present invention is better than current main-stream method.It is real
It tests result and also demonstrates the retrieval performance of of the invention (CNN+KSH) on Caltech-256 image sets and be substantially better than other methods.
Finally for different query images, the present invention compared to other methods can retrieve to obtain more with the relevant figure of query image
Picture.
1.1 convolutional neural networks
Convolutional neural networks (Convolutional Neural Network, CNN) are put forward for the first time by Fukushima,
It is first and really successfully trains the learning algorithm of multitiered network structure, and is widely used in solving how to extract study image
The further feature problem of data, basic thought are using the local sensing region of image as the input of network, and information is again successively
It is transferred to different layers, every layer is gone to obtain to translation, rotation and scale notable with invariance by a digital filter
Feature.
As shown in Figure 1, the multilayer depth nerve net that CNN is made of multiple convolutional layers (C layers) and sub-sampling layer (S layers)
Network will generally be followed by sub-sampling layer after convolutional layer, and every layer is all made of multiple two dimensional surfaces, and network structure can be by scheming
Described by 1.Convolutional layer is also referred to as feature extraction layer, is made of multiple characteristic patterns, wherein each characteristic pattern is by multiple independent nerves
Member composition, different characteristic patterns realize the extraction to last layer data different characteristic.All neurons of same characteristic plane are total
One or more connection weight is enjoyed, drastically reduces the number of parameters for needing training, and reduce the complexity of CNN.Son
Sample level samples subregion in characteristic pattern, realizes to characteristic pattern high dimensional feature dimensionality reduction, and it is existing to prevent " dimension disaster "
As this distinctive structure of feature extraction twice makes feature have invariance, Er Qieqi to transformation such as translation, scaling, deformation
The network structure of deep layer can be excavated effectively in image in implication relation, the ability to express of Enhanced feature.
1.2 position sensing core Hash (KLSH)
It is the high dimensional data for enhancing hash function to linearly inseparableResolution capability, Kulis etc.
Utilize kernel function κ:Build hash function h:Mapping is carried out to high dimensional data and generates Hash codes,
Hash function concrete form is shown below:
Wherein,x(1),…,x(n)It is the m sample randomly selected from χ, in order to realize quick Hash
Mapping, m are the constants much smaller than n.Hash function h (x) is in addition to meeting low-dimensional Hamming space similar to original higher dimensional space one
Outside cause property, it also should ensure that the Hash codes of generation are balanced, i.e. hash function h (x) should meet
Then biasThe value of b, which is entered formula (18), to be obtained:
Wherein, a=[a1,…am]T, It is shown in mapping matrix such as formula (21):
Wherein,It can be by precalculating to obtain, coefficient vector a is the m obtained by random sampling
Dimensional vector.The present invention obtains coefficient vector a using the study that exercises supervision of the correlation information of training data, and construction is associated with the data
Hash function, enhancing generate Hash codes distinction, improve retrieval precision ratio.
It is enlightenment with above-mentioned desirable embodiment according to the present invention, through the above description, relevant staff is complete
Various changes and amendments can be carried out without departing from the scope of the technological thought of the present invention' entirely.The technology of this invention
Property range is not limited to the contents of the specification, it is necessary to determine its technical scope according to right.
Claims (3)
1. a kind of target retrieval method based on convolutional neural networks and supervision core Hash, characterized in that include the following steps:
(1) convolutional neural networks are introduced to learn training image, implicitly learns to obtain figure using its special networks structure
As the high-order expression of data, further feature is generated;
(2) introduce supervision core hash method of the enhancing to the resolving powers of linearly inseparable data, at the same using Hash codes inner product with
The equivalence relation of Hamming distance proposes object function, and the affinity information of combined training image supervises dimensional images feature
Educational inspector practises, and generates Hash codes;
(3) image index is constructed using trained hash function, realizes the retrieval to large-scale image data.
2. the target retrieval method according to claim 1 based on convolutional neural networks and supervision core Hash, feature exist
In:The input picture size of the convolutional neural networks is 227 × 227, exports the image further feature for 4096 × 1, altogether
Including 5 convolutional layers, 3 sub- sample levels;In the characteristic pattern of convolutional layer preceding layerWith the convolution kernel K that can learnijIt carries out
Convolution, the result of convolution generate the characteristic pattern of this layer through nonlinear function f ()Concrete form is as follows:
Formula one:
For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijIt can be with one of preceding layer
Or multiple characteristic patterns determine convolution relation, MjInput feature vector set of graphs is represented, common nonlinear function has f (x)=tanh (x)
With f (x)=(1+e-x)-1, the characteristic pattern size h of convolutional layer generationlIt can be calculated by formula two:
Formula two:
hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel moving step length, ρlIt indicates
To the columns of preceding layer characteristic pattern edge zero padding when convolution algorithm;Each layer convolution kernel size Z={ z1=11, z2=5, z3=z4=
z5=3 }, moving step length Λ={ λ1=4, λ2=λ3=λ4=λ5=1 }, characteristic pattern edge zero padding columns P={ ρ1=0, ρ2=2,
ρ3=ρ4=ρ5=1 };Sub-sampling layer carries out maximum value sampling using overlap sampling method to characteristic pattern, and sampling area is 3 × 3,
Sampling step length is 2 pixels;
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
The propagated forward stage:Sample (X, a Y are chosen from training samplep), X is transmitted to output from input layer through converting step by step
Layer, calculates corresponding reality output:
Formula three:Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n))
The back-propagating stage:The stage is the error propagation stage, calculates reality output OpY is exported with corresponding idealpError:
Formula four:
By error EpReversed successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized, and is missed when total
When poor E≤ε, the training of the batch training sample is completed, after the completion of the training of all batches, image is inputted into convolutional neural networks
In, image data step by step by each network layer after, can be obtained the further feature of image in output end.
3. the target retrieval method according to claim 1 based on convolutional neural networks and supervision core Hash, feature exist
In:Measure image between apart from when, give the Hash codes dimension r, then need r coefficient vector a1,…,arConstruct Hash
FunctionThe label information of training image can be related by the semanteme of image
Property and space length obtain, be defined as description label image collection χl={ x1,…,xlIn correlation between element supervision
Matrix
Formula five:
Make image xi,xjHamming distance Dh(xi,xj) meet:
Formula six:
Hash codes distance, the Hash codes code of image x are calculated using inner product of vectors operationr(x)=[h1(x),…,hr(x)]∈
{1,-1}1×r, then image xi,xjDistance calculate as shown in formula 11:
Formula seven:
D(xi,xj) ∈ [- r, r], to D (xi,xj) obtain after normalizationDefinition is so that similar matrixThe minimum object function with check matrix S distances:
Formula eight:
Matrix F robenius norms are sought in expression,For label image collection χlHash codes
Sgn () is generalized to matrix form by matrix, according to formula (3) HlIt can be expressed as:
Formula nine:
By HlSubstitute into formula eight
Formula ten:
Formula 11:
Define matrixWherein R0=rS can then pass through greediness
Algorithmic minimizing formula (11) gradually estimates ak:
Formula 12:
Remove constant term, more succinct object function can be obtained:
Formula 13:
WithSgn (x) is replaced, then approximate objective functionAs shown in formula 14:
Formula 14:
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Formula 15:
Hadamard inner product operations are indicated, to accelerateConvergence and use compose Hash in
Spectral analysis method generate initial valueGradient search process is further accelerated, after obtaining hash function H and Hash table H, to inquiry
The further feature of image carries out Hash mapping and obtains coder(xq), calculate coder(xq) in Hash table H at a distance from Hash codes,
The closer image of layback is as retrieval result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810157751.2A CN108304573A (en) | 2018-02-24 | 2018-02-24 | Target retrieval method based on convolutional neural networks and supervision core Hash |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810157751.2A CN108304573A (en) | 2018-02-24 | 2018-02-24 | Target retrieval method based on convolutional neural networks and supervision core Hash |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108304573A true CN108304573A (en) | 2018-07-20 |
Family
ID=62848988
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810157751.2A Pending CN108304573A (en) | 2018-02-24 | 2018-02-24 | Target retrieval method based on convolutional neural networks and supervision core Hash |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108304573A (en) |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109241313A (en) * | 2018-08-14 | 2019-01-18 | 大连大学 | A kind of image search method based on the study of high-order depth Hash |
CN109800314A (en) * | 2019-01-03 | 2019-05-24 | 上海大学 | A method of generating the Hash codes for being used for image retrieval using depth convolutional network |
CN109885716A (en) * | 2019-02-18 | 2019-06-14 | 成都快眼科技有限公司 | The image search method of discrete Hash is supervised based on heterogeneous multi-task learning depth |
CN109933682A (en) * | 2019-01-11 | 2019-06-25 | 上海交通大学 | A kind of image Hash search method and system based on semanteme in conjunction with content information |
CN109947963A (en) * | 2019-03-27 | 2019-06-28 | 山东大学 | A kind of multiple dimensioned Hash search method based on deep learning |
CN110083734A (en) * | 2019-04-15 | 2019-08-02 | 中南大学 | Semi-supervised image search method based on autoencoder network and robust core Hash |
CN111628866A (en) * | 2020-05-22 | 2020-09-04 | 深圳前海微众银行股份有限公司 | Neural network verification method, device and equipment and readable storage medium |
CN112861976A (en) * | 2021-02-11 | 2021-05-28 | 温州大学 | Sensitive image identification method based on twin graph convolution hash network |
CN113011359A (en) * | 2021-03-26 | 2021-06-22 | 浙江大学 | Method for simultaneously detecting plane structure and generating plane description based on image and application |
CN113343025A (en) * | 2021-08-05 | 2021-09-03 | 中南大学 | Sparse attack resisting method based on weighted gradient Hash activation thermodynamic diagram |
CN113469695A (en) * | 2020-03-30 | 2021-10-01 | 同济大学 | Electronic fraud transaction identification method, system and device based on kernel supervision Hash model |
CN114189351A (en) * | 2021-10-25 | 2022-03-15 | 山东师范大学 | Dense image retrieval method and system based on CNN and signcryption technology |
CN114722208A (en) * | 2022-06-08 | 2022-07-08 | 成都健康医联信息产业有限公司 | Automatic classification and safety level grading method for health medical texts |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106503106A (en) * | 2016-10-17 | 2017-03-15 | 北京工业大学 | A kind of image hash index construction method based on deep learning |
-
2018
- 2018-02-24 CN CN201810157751.2A patent/CN108304573A/en active Pending
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106503106A (en) * | 2016-10-17 | 2017-03-15 | 北京工业大学 | A kind of image hash index construction method based on deep learning |
Non-Patent Citations (1)
Title |
---|
柯圣财 等: "基于卷积神经网络和监督核哈希的图像检索方法", 《电子学报》 * |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109241313A (en) * | 2018-08-14 | 2019-01-18 | 大连大学 | A kind of image search method based on the study of high-order depth Hash |
CN109241313B (en) * | 2018-08-14 | 2021-11-02 | 大连大学 | Image retrieval method based on high-order deep hash learning |
CN109800314A (en) * | 2019-01-03 | 2019-05-24 | 上海大学 | A method of generating the Hash codes for being used for image retrieval using depth convolutional network |
CN109933682A (en) * | 2019-01-11 | 2019-06-25 | 上海交通大学 | A kind of image Hash search method and system based on semanteme in conjunction with content information |
CN109885716A (en) * | 2019-02-18 | 2019-06-14 | 成都快眼科技有限公司 | The image search method of discrete Hash is supervised based on heterogeneous multi-task learning depth |
CN109947963A (en) * | 2019-03-27 | 2019-06-28 | 山东大学 | A kind of multiple dimensioned Hash search method based on deep learning |
CN110083734A (en) * | 2019-04-15 | 2019-08-02 | 中南大学 | Semi-supervised image search method based on autoencoder network and robust core Hash |
CN110083734B (en) * | 2019-04-15 | 2024-05-03 | 中南大学 | Semi-supervised image retrieval method based on self-coding network and robust kernel hash |
CN113469695A (en) * | 2020-03-30 | 2021-10-01 | 同济大学 | Electronic fraud transaction identification method, system and device based on kernel supervision Hash model |
CN113469695B (en) * | 2020-03-30 | 2023-06-30 | 同济大学 | Electronic fraud transaction identification method, system and device based on kernel supervision hash model |
CN111628866A (en) * | 2020-05-22 | 2020-09-04 | 深圳前海微众银行股份有限公司 | Neural network verification method, device and equipment and readable storage medium |
WO2021233183A1 (en) * | 2020-05-22 | 2021-11-25 | 深圳前海微众银行股份有限公司 | Neural network verification method, apparatus and device, and readable storage medium |
CN112861976B (en) * | 2021-02-11 | 2024-01-12 | 温州大学 | Sensitive image identification method based on twin graph convolution hash network |
CN112861976A (en) * | 2021-02-11 | 2021-05-28 | 温州大学 | Sensitive image identification method based on twin graph convolution hash network |
CN113011359B (en) * | 2021-03-26 | 2023-10-24 | 浙江大学 | Method for simultaneously detecting plane structure and generating plane description based on image and application |
CN113011359A (en) * | 2021-03-26 | 2021-06-22 | 浙江大学 | Method for simultaneously detecting plane structure and generating plane description based on image and application |
CN113343025A (en) * | 2021-08-05 | 2021-09-03 | 中南大学 | Sparse attack resisting method based on weighted gradient Hash activation thermodynamic diagram |
CN113343025B (en) * | 2021-08-05 | 2021-11-02 | 中南大学 | Sparse attack resisting method based on weighted gradient Hash activation thermodynamic diagram |
CN114189351A (en) * | 2021-10-25 | 2022-03-15 | 山东师范大学 | Dense image retrieval method and system based on CNN and signcryption technology |
CN114189351B (en) * | 2021-10-25 | 2024-02-23 | 山东师范大学 | Dense state image retrieval method and system based on CNN and signature technology |
CN114722208A (en) * | 2022-06-08 | 2022-07-08 | 成都健康医联信息产业有限公司 | Automatic classification and safety level grading method for health medical texts |
CN114722208B (en) * | 2022-06-08 | 2022-11-01 | 成都健康医联信息产业有限公司 | Automatic classification and safety level grading method for health medical texts |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108304573A (en) | Target retrieval method based on convolutional neural networks and supervision core Hash | |
Cai et al. | Medical image retrieval based on convolutional neural network and supervised hashing | |
Wei Tan et al. | Deep learning for plant species classification using leaf vein morphometric | |
Liu et al. | Learning spatio-temporal representations for action recognition: A genetic programming approach | |
Shao et al. | Feature learning for image classification via multiobjective genetic programming | |
Han et al. | Matchnet: Unifying feature and metric learning for patch-based matching | |
CN106326288B (en) | Image search method and device | |
Liu et al. | A spark-based parallel fuzzy $ c $-Means segmentation algorithm for agricultural image Big Data | |
Xiang et al. | Fabric image retrieval system using hierarchical search based on deep convolutional neural network | |
CN109918528A (en) | A kind of compact Hash code learning method based on semanteme protection | |
CN111339343A (en) | Image retrieval method, device, storage medium and equipment | |
CN110598022B (en) | Image retrieval system and method based on robust deep hash network | |
Nguyen et al. | Deep learning for metagenomic data: using 2d embeddings and convolutional neural networks | |
Wu et al. | Plant leaf identification based on shape and convolutional features | |
CN112883931A (en) | Real-time true and false motion judgment method based on long and short term memory network | |
Singh | Content-based image retrieval using deep learning | |
Lumini et al. | Ocean ecosystems plankton classification | |
Abdel-Nabi et al. | Content based image retrieval approach using deep learning | |
Xu et al. | Weakly supervised facial expression recognition via transferred DAL-CNN and active incremental learning | |
Li et al. | An improved lightweight network architecture for identifying tobacco leaf maturity based on Deep learning | |
Ibrahim et al. | Improving Date Fruit Classification Using CycleGAN-Generated Dataset. | |
Preyanka Lakshme et al. | A review based on machine learning for feature selection and feature extraction | |
Prasetya et al. | Indonesian food items labeling for tourism information using Convolution Neural Network | |
Vasudevan et al. | A Hybrid Approach for Plant Disease Detection Using E-GAN and CapsNet. | |
Zhang et al. | Improved image retrieval algorithm of GoogLeNet neural network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180720 |
|
RJ01 | Rejection of invention patent application after publication |