CN108304573A - Target retrieval method based on convolutional neural networks and supervision core Hash - Google Patents

Target retrieval method based on convolutional neural networks and supervision core Hash Download PDF

Info

Publication number
CN108304573A
CN108304573A CN201810157751.2A CN201810157751A CN108304573A CN 108304573 A CN108304573 A CN 108304573A CN 201810157751 A CN201810157751 A CN 201810157751A CN 108304573 A CN108304573 A CN 108304573A
Authority
CN
China
Prior art keywords
image
hash
formula
layer
convolutional neural
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810157751.2A
Other languages
Chinese (zh)
Inventor
李弼程
赵永威
朱彩英
陈良浩
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Test Joint Space Big Data Application Research Center Ltd Co
Original Assignee
Jiangsu Test Joint Space Big Data Application Research Center Ltd Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Test Joint Space Big Data Application Research Center Ltd Co filed Critical Jiangsu Test Joint Space Big Data Application Research Center Ltd Co
Priority to CN201810157751.2A priority Critical patent/CN108304573A/en
Publication of CN108304573A publication Critical patent/CN108304573A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/51Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Library & Information Science (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Biomedical Technology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to search method field, it is based especially on convolutional neural networks and supervises the target retrieval method of core Hash.The search method includes:(1) it introduces convolutional neural networks to learn training image, the high-order for implicitly being learnt to obtain image data using its special networks structure is indicated, further feature is generated;(2) supervision core hash method of the enhancing to the resolving power of linearly inseparable data is introduced, simultaneously object function is proposed using the equivalence relation of Hash codes inner product and Hamming distance, and the affinity information of combined training image exercises supervision study to dimensional images feature, and generate Hash codes;(3) image index is constructed using trained hash function, realizes the retrieval to large-scale image data.The present invention greatly improves target retrieval efficiency by the target retrieval method based on convolutional neural networks and supervision core Hash, enhances the practicability under big data environment.

Description

Target retrieval method based on convolutional neural networks and supervision core Hash
Technical field
The present invention relates to search method field, it is based especially on convolutional neural networks and supervises the target retrieval side of core Hash Method.
Background technology
With the arrival in big data epoch, internet video image resource rapidly increases, how to extensive video, image Concern target in resource is fast and effeciently retrieved urgently to be resolved hurrily to meet user demand.Although local feature region goes out It is existing, such as SIFT (Scale-Invariant Feature Transform), HOG (Histogram of Orientated ) etc. Gradients excellent performance is shown in image processing field, but generate these and describe the fixed coding step of the period of the day from 11 p.m. to 1 a.m So that description lacks learning ability, its picture material ability to express is limited, it is difficult to various image data is adapted to, certain Large-scale image target retrieval performance is reduced in degree.The inherent implication relation of great amount of images data in order to obtain generates more With distinction and representative feature, deep learning (Deep Learning) is applied to image procossing by the scholars such as Hinton In field, new approaches are provided to extract significantly more efficient characteristics of image.
Depth confidence network (Deep Belief Network, DBN) top layer uses three rank Boltzmann machine (Boltzmann Machine, BM), improved DBN is used for objective feature extraction, this feature is rotationally-varying to target to have higher Shandong Stick.In addition, researcher construct convolution depth confidence network (Convolutional Deep Belief Network, CDBN), learn effective high-order character representation in the natural image that can be never marked using CDBN.And by convolutional Neural SPP (Spatial are added in the convolutional layer of network (Convolutional Neural Network, CNN) and full articulamentum Pyramid Pooling) layer, directly different size image is learnt and generates Analysis On Multi-scale Features.
But deep learning generate characteristics of image dimension it is higher, there are problems that dimension disaster, when image data scale compared with When big, retrieval rate will be made drastically by carrying out retrieval according to traditional nearest _neighbor retrieval method (such as R-tree, KD-tree) Decline, it is difficult to be suitable for large-scale data.Extensive high dimensional image is effectively retrieved to realize, researcher proposes Approximate KNN search strategy (Approximate Nearest Neighbor, ANN).Wherein, salted hash Salted (Hashing) is The main stream approach of approximate KNN search problem is solved, thought is by dimensional images Feature Mapping using hash function race to low In dimension space, while making distance is closer in former space point is mapped to after lower dimensional space still to keep closer distance.LSH and its The hash function of innovatory algorithm construction is all unrelated with data, and in recent years, how researchers are in conjunction with data characteristics structure It makes hash function that is effective, compacting and proposes many algorithms.Hash method (Spetral Hashing, SH) is composed first to similar The Laplacian Matrix characteristic value and feature vector of figure are analyzed, then by the condition of relaxing the restriction, will be to image feature vector The dimensionality reduction problem that encoded question is converted to Laplce's characteristic pattern solves, this method rely on data itself generate index than with Machine generates hash function method and reaches higher accuracy rate.But unsupervised method does not consider the semantic information of image, And user is often more likely to the semantic information of retrieval result.For this purpose, researcher proposes semi-supervised hash method (Semi- Supervised Hashing, SSH).Researchers also proposed some full supervision and breathe out on the basis of semi-supervised learning method Uncommon method, such as SH (Semantic Hashing), BRE (Binary Reconstructive Embedding), MLH (Minimal Loss Hashing) etc., full hash method of supervising can reach higher accuracy rate compared to non-supervisory method, but The problems such as to be that there are optimization process complex, training effectiveness is low, the application that this severely limits it on large-scale dataset.
Invention content
The technical problem to be solved by the present invention is to:In order to solve existing unsupervised method, there is no the languages for considering image The deficiency of adopted information, the target retrieval method based on convolutional neural networks and supervision core Hash that the present invention provides a kind of, passes through Target retrieval method based on convolutional neural networks and supervision core Hash can utilize convolutional neural networks to Large Scale Graphs As data characteristics progress autonomous learning, enhance the ability to express of characteristics of image.Secondly, by supervision core hash method to dimensional images Further feature exercises supervision study, and high dimensional feature is mapped in low-dimensional Hamming space, generates the Hash codes compacted, greatly Target retrieval efficiency is improved, the practicability under big data environment is enhanced.
The technical solution adopted by the present invention to solve the technical problems is:
A kind of target retrieval method based on convolutional neural networks and supervision core Hash, includes the following steps:
(1) convolutional neural networks are introduced to learn training image, is implicitly learnt using its special networks structure High-order to image data indicates, generates further feature;
(2) supervision core hash method of the enhancing to the resolving power of linearly inseparable data is introduced, while using in Hash codes Product and the equivalence relation of Hamming distance propose object function, and the affinity information of combined training image to dimensional images feature into Row supervised learning, and generate Hash codes;
(3) image index is constructed using trained hash function, realizes the retrieval to large-scale image data.
Specifically, the input picture size of the convolutional neural networks is 227 × 227, is exported deep for 4096 × 1 image Layer feature includes 5 convolutional layers, 3 sub- sample levels altogether;In the characteristic pattern of convolutional layer preceding layerWith the volume that can learn Product core KijConvolution is carried out, the result of convolution generates the characteristic pattern of this layer through nonlinear function f ()Concrete form is such as Under:
Formula one:
For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijIt can be with preceding layer One or more features figure determines convolution relation, MjRepresent input feature vector set of graphs, common nonlinear function have f (x)= Tanh (x) and f (x)=(1+e-x)-1, the characteristic pattern size h of convolutional layer generationlIt can be calculated by formula two:
Formula two:
hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel moving step length, ρl To the columns of preceding layer characteristic pattern edge zero padding when expression convolution algorithm;Each layer convolution kernel size Z={ z1=11, z2=5, z3= z4=z5=3 }, moving step length Λ={ λ1=4, λ2345=1 }, characteristic pattern edge zero padding columns P={ ρ1=0, ρ2 =2, ρ345=1 };Sub-sampling layer carries out maximum value sampling, sampling area 3 using overlap sampling method to characteristic pattern × 3, sampling step length is 2 pixels;
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
The propagated forward stage:Sample (X, a Y are chosen from training samplep), X is transmitted to from input layer through converting step by step Output layer calculates corresponding reality output:
Formula three:Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n))
The back-propagating stage:The stage is the error propagation stage, calculates reality output OpY is exported with corresponding idealp
Error:
Formula four:
By error EpReversely successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized, As overall error E≤ε, the training of the batch training sample is completed, after the completion of the training of all batches, by image input convolution god Through in network, image data step by step by each network layer after, can be obtained the further feature of image in output end.
Specifically, measure image between apart from when, give the Hash codes dimension r, then need r coefficient vector a1,…,arConstruct hash functionThe label information of training image can lead to The semantic dependency and space length for crossing image obtain, and are defined as description label image collection χl={ x1,…,xlIn between element Correlation check matrix
Formula five:
Make image xi,xjHamming distance Dh(xi,xj) meet:
Formula six:
Hash codes distance, the Hash codes code of image x are calculated using inner product of vectors operationr(x)=[h1(x),…,hr(x)] ∈{1,-1}1×r, then image xi,xjDistance calculate as shown in formula 11:
D(xi,xj)=coder(xi)·coder(xj)
=| k | hk(xi)=hk(xj),1≤k≤r}|-|{k|hk(xi)≠hk(xj),1≤k≤r}
=r-2 | and k | hk(xi)≠hk(xj),1≤k≤r}|
Formula seven:=r-2Dh(xi,xj)
D(xi,xj) ∈ [- r, r], to D (xi,xj) obtain after normalizationDefinition is so that similar MatrixThe minimum object function with check matrix S distances:
Formula eight:
Matrix F robenius norms are sought in expression,For the Kazakhstan of label image collection χ l Uncommon code matrix, is generalized to matrix form, according to formula (3) H by sgn ()lIt can be expressed as:
Formula nine:
Hl is substituted into public Formula eight
Formula ten:
Formula 11:
Define matrixWherein R0=rS can then lead to It crosses greedy algorithm minimum formula (11) and gradually estimates ak
Formula 12:
Remove constant term, more succinct object function can be obtained:
Formula 13:
WithSgn (x) is replaced, then approximate objective functionAs shown in formula 14:
Formula 14:
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Formula 15:
⊙ indicates Hadamard inner product operations, to accelerateIt restrains and is breathed out with spectrum Spectral analysis method in uncommon generates initial valueGradient search process is further accelerated, it is right after obtaining hash function H and Hash table H The further feature of query image carries out Hash mapping and obtains coder(xq), calculate coder(xq) with Hash table H in Hash codes away from From the closer image of layback is as retrieval result.
The beneficial effects of the invention are as follows:The target based on convolutional neural networks and supervision core Hash that the present invention provides a kind of Search method can utilize convolutional Neural by the target retrieval method based on convolutional neural networks and supervision core Hash Network carries out autonomous learning to large-scale image data feature, enhances the ability to express of characteristics of image.Secondly, by supervision core Hash Method exercises supervision study to dimensional images further feature, and high dimensional feature is mapped in low-dimensional Hamming space, and generation is compacted Hash codes, greatly improve target retrieval efficiency, enhance the practicability under big data environment.
Description of the drawings
Present invention will be further explained below with reference to the attached drawings and examples.
Fig. 1 is the structural schematic diagram of the convolutional neural networks of the present invention;
Fig. 2 is convolutional neural networks structure chart of the present invention for the extraction of image further feature;
Fig. 3 is the schematic diagram of present invention target retrieval MAP value on ImageNet-100 databases;
Fig. 4 is schematic diagram of the training time of the invention with Hash code bit number r change curves;
Fig. 5 is the schematic diagram of present invention target retrieval MAP value on Caltech-256 databases;
Fig. 6 is the schematic diagram of present invention Precision-Recall curves on Caltech-256 databases;
Specific implementation mode
Fig. 1 is the structural schematic diagram of the convolutional neural networks of the present invention, and Fig. 2 is that the present invention is extracted for image further feature Convolutional neural networks structure chart, Fig. 3 be the present invention on ImageNet-100 databases target retrieval MAP value schematic diagram, Fig. 4 is schematic diagram of the training time of the invention with Hash code bit number r change curves, and Fig. 5 is the present invention in Caltech-256 data The schematic diagram of target retrieval MAP value on library, Fig. 6 are that present invention Precision-Recall on Caltech-256 databases is bent The schematic diagram of line.
First, convolutional neural networks are introduced to learn training image, is implicitly learnt using its special networks structure The high-order for obtaining image data indicates, generates the further feature with stronger distinction and ability to express;
Secondly, introducing supervision core hash method (Kernel-Based Supervised Hashing, KSH) enhances to line Property can not divided data resolving power, while being proposed using the equivalence relation of Hash codes inner product and Hamming distance simpler, effective Object function, and the affinity information of combined training image exercises supervision study to dimensional images feature, and generates and compact Hash codes;
Finally, trained hash function constructs image index for utilization, realizes the efficient inspection to large-scale image data Rope.
The target retrieval method based on convolutional neural networks and supervision core Hash utilizes convolutional Neural net first The input picture size of the further feature of network structure extraction image, convolutional neural networks used is 227 × 227, and it is 4096 to export × 1 image further feature includes 5 convolutional layers, 3 sub- sample levels altogether.In convolutional layer, the characteristic pattern of preceding layerWith The convolution kernel K that can learnijConvolution is carried out, the result of convolution generates the characteristic pattern of this layer through nonlinear function f ()Tool Body form is as follows:
Wherein,For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijCan with it is previous The one or more features figure of layer determines convolution relation, MjInput feature vector set of graphs is represented, common nonlinear function has f (x) =tanh (x) and f (x)=(1+e-x)-1, compared with above-mentioned nonlinear function, f (x)=max (0, x) can effectively improve trained effect Rate.The characteristic pattern size h that convolutional layer generateslIt can be calculated by formula (2):
Wherein, hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel movement step It is long, ρlTo the columns of preceding layer characteristic pattern edge zero padding when expression convolution algorithm.Here, each layer convolution kernel size Z={ z1=11, z2=5, z3=z4=z5=3 }, moving step length Λ={ λ1=4, λ2345=1 }, characteristic pattern edge zero padding columns P= {ρ1=0, ρ2=2, ρ345=1 }.In sub-sampling layer, document shows, relative to traditional non-overlapping sampling, to use overlapping Sampling can not only improve the accuracy of feature, be also prevented from the training stage and over-fitting occur, therefore, use overlap sampling here Method carries out maximum value sampling to characteristic pattern, and sampling area is 3 × 3, and sampling step length is 2 pixels.
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
(1) the propagated forward stage.Sample (X, a Y are chosen from training samplep), X is from input layer through converting biography step by step It is sent to output layer, calculates corresponding reality output:
Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n)) (3)
(2) back-propagating stage, also referred to as error propagation stage.Calculate reality output OpY is exported with corresponding idealpError:
By error EpReversely successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized, As overall error E≤ε, the training of the batch training sample is completed.After the completion of the training of all batches, by image input convolution god Through in network, image data step by step by each network layer after, can be obtained the further feature of image in output end.
The target retrieval method based on convolutional neural networks and supervision core Hash, schemes in measurement As between apart from when, give Hash codes dimension r, then need r coefficient vector a1,…,arConstruct hash functionThe label information of training image can pass through the semantic dependency and space length of image It obtains, lable (xi,xj)=1 indicates image xi,xjIt is similar;Conversely, lable (xi,xjThe representative image x of)=- 1i,xjDifference It is very big.To describe label image collection χl={ x1,…,xlIn correlation between element, define check matrix
Wherein, lable (xi,xi) ≡ 1, Sii≡ 1, Sij=0 indicates image xi,xjBetween similitude it is uncertain.For enhancing The separating capacity of Hash codes so that the similitude between image can be efficiently judged in Hamming space, it should be as possible so that image xi,xjHamming distance Dh(xi,xj) meet:
Due to Hamming distance calculation formula Dh(xi,xj)=| k | hk(xi)≠hk(xj), 1≤k≤r } | form is complicated, very Hardly possible directly optimizes it, therefore utilizes inner product of vectors operation to calculate Hash codes distance herein.The Hash codes code of image xr (x)=[h1(x),…,hr(x)]∈{1,-1}1×r, then image xi,xjDistance calculate as shown in formula (11):
D(xi,xj)=coder(xi)·coder(xj)
=| k | hk(xi)=hk(xj),1≤k≤r}|-|{k|hk(xi)≠hk(xj),1≤k≤r}|
=r-2 | and k | hk(xi)≠hk(xj),1≤k≤r}|
=r-2Dh(xi,xj) (7)
Formula (7) is shown to be consistent by Hash codes inner product operation and Hamming distance operation, and D (xi,xj)∈[-r, R], to D (xi,xj) obtain after normalizationTo make similar matrixWith check matrix S Distance is minimum, objective function:
Wherein,Matrix F robenius norms are sought in expression,For label image collection χlKazakhstan Uncommon code matrix.Sgn () is generalized to matrix form, according to formula (3) HlIt can be expressed as:
Wherein,By HlSubstitution formula (8)
Compared with BRE and MLH, object function Γ (A) calculates similitude by inner product, is built to parameter A Mould is more intuitive.It is assumed that at the t=k moment, it is known that vectorIt needs to estimate ak, define matrixWherein R0=rS then can minimize formula (11) by greedy algorithm and gradually estimate ak
Remove constant term, more succinct object function can be obtained:
Since sgn (x) functions in object function makeDiscontinuously, andNor convex function, it is difficult to directly It connects pairIt minimizes, when | x | when > 6, continuous functionIt can approximation sgn (x), therefore profit well WithSgn (x) is replaced, then approximate objective functionAs shown in formula (14):
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Wherein,⊙ indicates Hadamard inner product operations.It is smoothed that treatedIt is not convex function, globally optimal solution can not be acquired, in order to accelerateConvergence, the present invention utilize the spectral analysis method in spectrum Hash Generate initial valueGradient search process is further accelerated, specific implementation step is as follows:
It supervises core hash method and realizes step:
Input:Training image collectionLabel image collectionAnd check matrix
Kernel function κHash code bit number r participates in trained sample number m (< l).
Pretreatment:M images are randomly selected from χ, are obtained
Training:Initialize R0=rS, Tmax=500;
For k=1 ..., r do
It is obtained using spectral analysis methodIt willSubstitute into object functionIn, pass through gradient Descent method is calculated
end if
Rk←Rk-1-h*(h*)T
end for
Coding:For i=1 ..., n do
Output:Hash functionWith Hash table H={ coder(xi)|i∈[1, n]}。
After obtaining hash function H and Hash table H, Hash mapping is carried out to the further feature of query image and obtains coder (xq), calculate coder(xq) at a distance from Hash codes, the closer image of layback is as retrieval result in Hash table H.
Embodiment one:Target retrieval method of the present embodiment based on convolutional neural networks and supervision core Hash.First, it introduces Convolutional neural networks learn training image, implicitly learn to obtain the high-order of image data using its special networks structure It indicates, generates the further feature with stronger distinction and ability to express;Then, supervision core hash method (Kernel- is introduced Based Supervised Hashing, KSH) enhancing utilizes Hash codes inner product to the resolving powers of linearly inseparable data Simpler, effective object function, and the affinity information pair of combined training image are proposed with the equivalence relation of Hamming distance Dimensional images feature exercises supervision study, and generates the Hash codes compacted;Finally, utilization trained hash function structural map As index, the efficient retrieval to large-scale image data is realized.
Compared with traditional target retrieval method, the independent learning ability of characteristics of image is effectively enhanced, and can Hash codes of compacting are generated using supervision core Hash, time overhead is reduced, enhances the practicability under large-scale data.
Embodiment two:Referring to Fig. 1, Fig. 2, the target retrieval based on convolutional neural networks and supervision core Hash of the present embodiment Method generates the further feature of image data using following step:
First, the further feature of convolutional neural networks structure extraction image, the input figure of convolutional neural networks used are utilized Picture size is 227 × 227, exports the image further feature for 4096 × 1, includes 5 convolutional layers, 3 sub- sample levels altogether. Convolutional layer, the characteristic pattern of preceding layerWith the convolution kernel K that can learnijConvolution is carried out, the result of convolution is through nonlinear function f () generates the characteristic pattern of this layerConcrete form is as follows:
Wherein,For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijCan with it is previous The one or more features figure of layer determines convolution relation, MjInput feature vector set of graphs is represented, common nonlinear function has f (x) =tanh (x) and f (x)=(1+e-x)-1, compared with above-mentioned nonlinear function, f (x)=max (0, x) can effectively improve trained effect Rate.The characteristic pattern size h that convolutional layer generateslIt can be calculated by formula (2):
Wherein, hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel movement step It is long, ρlTo the columns of preceding layer characteristic pattern edge zero padding when expression convolution algorithm.Here, each layer convolution kernel size Z={ z1=11, z2=5, z3=z4=z5=3 }, moving step length Λ={ λ1=4, λ2345=1 }, characteristic pattern edge zero padding columns P= {ρ1=0, ρ2=2, ρ345=1 }.In sub-sampling layer, document shows, relative to traditional non-overlapping sampling, to use overlapping Sampling can not only improve the accuracy of feature, be also prevented from the training stage and over-fitting occur, therefore, use overlap sampling here Method carries out maximum value sampling to characteristic pattern, and sampling area is 3 × 3, and sampling step length is 2 pixels.
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
(1) the propagated forward stage.Sample (X, a Y are chosen from training samplep), X is from input layer through converting biography step by step It is sent to output layer, calculates corresponding reality output:
Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n)) (3)
(2) back-propagating stage, also referred to as error propagation stage.Calculate reality output OpY is exported with corresponding idealpError:
By error EpReversely successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized, As overall error E≤ε, the training of the batch training sample is completed.After the completion of the training of all batches, by image input convolution god Through in network, image data step by step by each network layer after, can be obtained the further feature of image in output end.
Between measuring image apart from when, give Hash codes dimension r, then need r coefficient vector a1,…,arConstruction is breathed out Uncommon functionThe label information of training image can pass through the semantic dependency and sky of image Between distance obtain, lable (xi,xj)=1 indicates image xi,xjIt is similar;Conversely, lable (xi,xjThe representative image of)=- 1 xi,xjIt is widely different.To describe label image collection χl={ x1,…,xlIn correlation between element, define check matrix
Wherein, lable (xi,xi) ≡ 1, Sii≡ 1, Sij=0 indicates image xi,xjBetween similitude it is uncertain.For enhancing The separating capacity of Hash codes so that the similitude between image can be efficiently judged in Hamming space, it should be as possible so that image xi,xjHamming distance Dh(xi,xj) meet:
Due to Hamming distance calculation formula Dh(xi,xj)=| k | hk(xi)≠hk(xj), 1≤k≤r } | form is complicated, very Hardly possible directly optimizes it, therefore utilizes inner product of vectors operation to calculate Hash codes distance herein.The Hash codes code of image xr (x)=[h1(x),…,hr(x)]∈{1,-1}1×r, then image xi,xjDistance calculate as shown in formula (11):
D(xi,xj)=coder(xi)·coder(xj)
=| k | hk(xi)=hk(xj),1≤k≤r}|-|{k|hk(xi)≠hk(xj),1≤k≤r}|
=r-2 | and k | hk(xi)≠hk(xj),1≤k≤r}|
=r-2Dh(xi,xj) (7)
Formula (7) is shown to be consistent by Hash codes inner product operation and Hamming distance operation, and D (xi,xj)∈[-r, R], to D (xi,xj) obtain after normalizationTo make similar matrixWith check matrix S Distance is minimum, objective function:
Wherein,Matrix F robenius norms are sought in expression,For label image collection χlKazakhstan Uncommon code matrix.Sgn () is generalized to matrix form, according to formula (3) HlIt can be expressed as:
Wherein,By HlSubstitution formula (8)
Compared with BRE and MLH, object function Γ (A) calculates similitude by inner product, more intuitive to parameter A modelings.It is false It is scheduled on the t=k moment, it is known that vectorIt needs to estimate ak, define matrix Wherein R0=rS then can minimize formula (11) by greedy algorithm and gradually estimate ak
Remove constant term, more succinct object function can be obtained:
Since sgn (x) functions in object function makeDiscontinuously, andNor convex function, it is difficult to directly It connects pairIt minimizes, when | x | when > 6, continuous functionIt can approximation sgn (x), therefore profit well WithSgn (x) is replaced, then approximate objective functionAs shown in formula (14):
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Wherein,⊙ indicates Hadamard inner product operations.It is smoothed that treatedIt is not convex function, globally optimal solution can not be acquired, in order to accelerateConvergence, the present invention utilize the spectral analysis method in spectrum Hash Generate initial valueGradient search process is further accelerated, after obtaining hash function H and Hash table H, to the further feature of query image It carries out Hash mapping and obtains coder(xq), calculate coder(xq) at a distance from Hash codes, layback is closer in Hash table H Image is as retrieval result.
Experimental result and analysis
Here, using being assessed the present invention on ImageNet-1000 and Caltech-256 image sets, ImageNet-1000 image sets are a subsets of ImageNet image sets, are large scale visual identity contest (Large Scale Visual Recognition Challenge, LSVRC) evaluation and test data set, including 1000 classifications total 1,200,000 Open image;Caltech-256 image sets are frequently-used data collection in target classification task, including total 30608 figures of 256 classifications Picture, wherein including at least 80 images in each classification.Experimental Hardware be configured in save as 6G GPU equipment GTX Titan and Intel Xeon CPU, the interior server for saving as 16G.Target retrieval performance indicator uses precision ratio and recall ratio, and definition is such as Under:
The influence of parameter
First, core hash method (abbreviation KSH) retrieval performance is supervised with the variation of Hash code bit number r, this hair in order to verify It is bright to be tested on ImageNet-1000 image sets, and be compared with some current mainstream hash methods, including LSH(Locality Sensitive Hashing)、SKLSH(LSH with Shift-Invariant Kernels)、SH (Spetral Hashing)、DSH(Density Sensitive Hashing)、PCA-ITQ(Iterative Quantization of PCA), the methods of BRE (Binary Reconstructive Embedding).Experiment first from 50 classes are randomly selected in ImageNet-1000 image sets, and to this 50 class image zooming-out GIST feature;Then, from each class Training set of the feature (feature for amounting to 50000 images) of 1000 images as supervised training hash function is randomly selected, Remaining image is as inquiry use-case;Finally, it introduces hash method to be retrieved, it is as shown in Figure 3 to obtain experimental result:
From figure 3, it can be seen that with the increase of Hash code bit number r, the MAP value of each method increases, however, working as r After increasing to certain value, MAP value increase amplitude, which tapers into, to be tended to be saturated.The target retrieval MAP value for comparing each hash method can Know, retrieved using the present invention (KSH) has better performance compared with other main stream approach, this is because unsupervised hash method (such as LSH, SH, DSH, PCA-ITQ etc.) and supervision hash method BRE do not utilize the semantic information construction of image to breathe out well Uncommon function, causes retrieval performance relatively low;And KSH of the present invention introduces Kernel hash function and strengthens to linearly inseparable number According to resolution capability, hash function is trained in combination with the affinity information of image, generates the Hash that more compacts Code, to improve target retrieval performance.
Experiment is again compared the present invention and the hash function training time consumption of current main-stream hash method, specifically As shown in Figure 4.It is not difficult to find out from Fig. 4, no prison fewer than supervision hash method using the time loss of unsupervised hash method It superintends and directs hash method mostly and is to retain the position sensing of primitive character as optimization aim, and it is with image language to supervise hash method For adopted neighbor information as supervision message, searching process is increasingly complex, compares unsupervised hash method time overhead bigger.
Experimental performance is analyzed
To verify the effective of the target retrieval method (abbreviation CNN+KSH) based on convolutional neural networks and supervision core Hash Property, 50 class images have been randomly selected in ImageNet-1000 image sets is tested to obtain table 1.It is not difficult to find out from table 1, Based on the CNN MAP values retrieved of extraction further feature than the MAP value retrieved based on global GIST features be higher by 10% with On, illustrate that the image further feature using CNN extractions has stronger distinction and ability to express, this is because GIST features carry It takes step to fix, does not have independent learning ability, so that its image expression limited ability, and CNN can imitate brain processing The pattern of data carries out feature extraction to image, and the network structure of its deep layer can be excavated effectively in image and be closed implicit System, enhances the image expression ability of feature.Wherein, the online retrieving time of (CNN+KSH) of the invention is 1.775 × 10-4Second, It is suitable with other main stream approach, and MAP value has reached 40.79%, retrieval performance is apparently higher than other methods, therefore the present invention exists There is stronger applicability under big data environment.
The target retrieval MAP value of 1 distinct methods of table and retrieval time comparison (64bits)
Comparison diagram 5, Fig. 6 are it is found that the present invention (CNN+KSH) MAP value under equal length Hash codes coding is above other masters Stream method, and under conditions of ensureing identical precision ratio, CNN+KSH can reach recall ratio more higher than other methods.Document In method be all to be re-introduced into after image zooming-out GIST features hash method construction index, although GIST features pass through it is multiple dimensioned Multi-direction Gabor filter group is to having obtained global structure and spatial context information after image filtering, but GIST feature grains Degree is more coarse, and lacks independent learning ability, limits its image expression ability, and the hash method in these documents It only considers how that data characteristics is combined to construct hash function, there is no the semantic informations for utilizing image well, lead to its mesh It is relatively low to mark retrieval performance, it is difficult to be adapted to extensive target retrieval;CNN+KSH methods utilize ImageNet-1000 image sets pair CNN is trained, and a large amount of training image makes the training of CNN model parameters more abundant, can effectively excavate in image hidden Containing relationship, the image expression ability of feature is effectively enhanced, and KSH supervises hash function using image similarity information Supervise and instruct white silk, while searching process is accelerated using greedy algorithm and gradient descent method so that the present invention is better than current main-stream method.It is real It tests result and also demonstrates the retrieval performance of of the invention (CNN+KSH) on Caltech-256 image sets and be substantially better than other methods. Finally for different query images, the present invention compared to other methods can retrieve to obtain more with the relevant figure of query image Picture.
1.1 convolutional neural networks
Convolutional neural networks (Convolutional Neural Network, CNN) are put forward for the first time by Fukushima, It is first and really successfully trains the learning algorithm of multitiered network structure, and is widely used in solving how to extract study image The further feature problem of data, basic thought are using the local sensing region of image as the input of network, and information is again successively It is transferred to different layers, every layer is gone to obtain to translation, rotation and scale notable with invariance by a digital filter Feature.
As shown in Figure 1, the multilayer depth nerve net that CNN is made of multiple convolutional layers (C layers) and sub-sampling layer (S layers) Network will generally be followed by sub-sampling layer after convolutional layer, and every layer is all made of multiple two dimensional surfaces, and network structure can be by scheming Described by 1.Convolutional layer is also referred to as feature extraction layer, is made of multiple characteristic patterns, wherein each characteristic pattern is by multiple independent nerves Member composition, different characteristic patterns realize the extraction to last layer data different characteristic.All neurons of same characteristic plane are total One or more connection weight is enjoyed, drastically reduces the number of parameters for needing training, and reduce the complexity of CNN.Son Sample level samples subregion in characteristic pattern, realizes to characteristic pattern high dimensional feature dimensionality reduction, and it is existing to prevent " dimension disaster " As this distinctive structure of feature extraction twice makes feature have invariance, Er Qieqi to transformation such as translation, scaling, deformation The network structure of deep layer can be excavated effectively in image in implication relation, the ability to express of Enhanced feature.
1.2 position sensing core Hash (KLSH)
It is the high dimensional data for enhancing hash function to linearly inseparableResolution capability, Kulis etc. Utilize kernel function κ:Build hash function h:Mapping is carried out to high dimensional data and generates Hash codes, Hash function concrete form is shown below:
Wherein,x(1),…,x(n)It is the m sample randomly selected from χ, in order to realize quick Hash Mapping, m are the constants much smaller than n.Hash function h (x) is in addition to meeting low-dimensional Hamming space similar to original higher dimensional space one Outside cause property, it also should ensure that the Hash codes of generation are balanced, i.e. hash function h (x) should meet
Then biasThe value of b, which is entered formula (18), to be obtained:
Wherein, a=[a1,…am]T, It is shown in mapping matrix such as formula (21):
Wherein,It can be by precalculating to obtain, coefficient vector a is the m obtained by random sampling Dimensional vector.The present invention obtains coefficient vector a using the study that exercises supervision of the correlation information of training data, and construction is associated with the data Hash function, enhancing generate Hash codes distinction, improve retrieval precision ratio.
It is enlightenment with above-mentioned desirable embodiment according to the present invention, through the above description, relevant staff is complete Various changes and amendments can be carried out without departing from the scope of the technological thought of the present invention' entirely.The technology of this invention Property range is not limited to the contents of the specification, it is necessary to determine its technical scope according to right.

Claims (3)

1. a kind of target retrieval method based on convolutional neural networks and supervision core Hash, characterized in that include the following steps:
(1) convolutional neural networks are introduced to learn training image, implicitly learns to obtain figure using its special networks structure As the high-order expression of data, further feature is generated;
(2) introduce supervision core hash method of the enhancing to the resolving powers of linearly inseparable data, at the same using Hash codes inner product with The equivalence relation of Hamming distance proposes object function, and the affinity information of combined training image supervises dimensional images feature Educational inspector practises, and generates Hash codes;
(3) image index is constructed using trained hash function, realizes the retrieval to large-scale image data.
2. the target retrieval method according to claim 1 based on convolutional neural networks and supervision core Hash, feature exist In:The input picture size of the convolutional neural networks is 227 × 227, exports the image further feature for 4096 × 1, altogether Including 5 convolutional layers, 3 sub- sample levels;In the characteristic pattern of convolutional layer preceding layerWith the convolution kernel K that can learnijIt carries out Convolution, the result of convolution generate the characteristic pattern of this layer through nonlinear function f ()Concrete form is as follows:
Formula one:
For first of convolutional layer ClOutput,Represent convolution algorithm, bjFor biasing, convolution kernel KijIt can be with one of preceding layer Or multiple characteristic patterns determine convolution relation, MjInput feature vector set of graphs is represented, common nonlinear function has f (x)=tanh (x) With f (x)=(1+e-x)-1, the characteristic pattern size h of convolutional layer generationlIt can be calculated by formula two:
Formula two:
hl-1For the size of l-1 layers of characteristic pattern, zlIndicate the size of l layers of convolution kernel, λlIt is convolution kernel moving step length, ρlIt indicates To the columns of preceding layer characteristic pattern edge zero padding when convolution algorithm;Each layer convolution kernel size Z={ z1=11, z2=5, z3=z4= z5=3 }, moving step length Λ={ λ1=4, λ2345=1 }, characteristic pattern edge zero padding columns P={ ρ1=0, ρ2=2, ρ345=1 };Sub-sampling layer carries out maximum value sampling using overlap sampling method to characteristic pattern, and sampling area is 3 × 3, Sampling step length is 2 pixels;
The training of convolutional neural networks is mainly divided to two stages of propagated forward and back-propagating:
The propagated forward stage:Sample (X, a Y are chosen from training samplep), X is transmitted to output from input layer through converting step by step Layer, calculates corresponding reality output:
Formula three:Op=Fn(…(F2(F1(XpW(1))W(2))…)W(n))
The back-propagating stage:The stage is the error propagation stage, calculates reality output OpY is exported with corresponding idealpError:
Formula four:
By error EpReversed successively pusher obtains the error of each layer, and adjusts neuron weights by error approach is minimized, and is missed when total When poor E≤ε, the training of the batch training sample is completed, after the completion of the training of all batches, image is inputted into convolutional neural networks In, image data step by step by each network layer after, can be obtained the further feature of image in output end.
3. the target retrieval method according to claim 1 based on convolutional neural networks and supervision core Hash, feature exist In:Measure image between apart from when, give the Hash codes dimension r, then need r coefficient vector a1,…,arConstruct Hash FunctionThe label information of training image can be related by the semanteme of image Property and space length obtain, be defined as description label image collection χl={ x1,…,xlIn correlation between element supervision Matrix
Formula five:
Make image xi,xjHamming distance Dh(xi,xj) meet:
Formula six:
Hash codes distance, the Hash codes code of image x are calculated using inner product of vectors operationr(x)=[h1(x),…,hr(x)]∈ {1,-1}1×r, then image xi,xjDistance calculate as shown in formula 11:
Formula seven:
D(xi,xj) ∈ [- r, r], to D (xi,xj) obtain after normalizationDefinition is so that similar matrixThe minimum object function with check matrix S distances:
Formula eight:
Matrix F robenius norms are sought in expression,For label image collection χlHash codes Sgn () is generalized to matrix form by matrix, according to formula (3) HlIt can be expressed as:
Formula nine:
By HlSubstitute into formula eight
Formula ten:
Formula 11:
Define matrixWherein R0=rS can then pass through greediness Algorithmic minimizing formula (11) gradually estimates ak
Formula 12:
Remove constant term, more succinct object function can be obtained:
Formula 13:
WithSgn (x) is replaced, then approximate objective functionAs shown in formula 14:
Formula 14:
Gradient descent method pair can be passed throughIt minimizes,About akGradient is asked to obtain:
Formula 15:
Hadamard inner product operations are indicated, to accelerateConvergence and use compose Hash in Spectral analysis method generate initial valueGradient search process is further accelerated, after obtaining hash function H and Hash table H, to inquiry The further feature of image carries out Hash mapping and obtains coder(xq), calculate coder(xq) in Hash table H at a distance from Hash codes, The closer image of layback is as retrieval result.
CN201810157751.2A 2018-02-24 2018-02-24 Target retrieval method based on convolutional neural networks and supervision core Hash Pending CN108304573A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810157751.2A CN108304573A (en) 2018-02-24 2018-02-24 Target retrieval method based on convolutional neural networks and supervision core Hash

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810157751.2A CN108304573A (en) 2018-02-24 2018-02-24 Target retrieval method based on convolutional neural networks and supervision core Hash

Publications (1)

Publication Number Publication Date
CN108304573A true CN108304573A (en) 2018-07-20

Family

ID=62848988

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810157751.2A Pending CN108304573A (en) 2018-02-24 2018-02-24 Target retrieval method based on convolutional neural networks and supervision core Hash

Country Status (1)

Country Link
CN (1) CN108304573A (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109241313A (en) * 2018-08-14 2019-01-18 大连大学 A kind of image search method based on the study of high-order depth Hash
CN109800314A (en) * 2019-01-03 2019-05-24 上海大学 A method of generating the Hash codes for being used for image retrieval using depth convolutional network
CN109885716A (en) * 2019-02-18 2019-06-14 成都快眼科技有限公司 The image search method of discrete Hash is supervised based on heterogeneous multi-task learning depth
CN109933682A (en) * 2019-01-11 2019-06-25 上海交通大学 A kind of image Hash search method and system based on semanteme in conjunction with content information
CN109947963A (en) * 2019-03-27 2019-06-28 山东大学 A kind of multiple dimensioned Hash search method based on deep learning
CN110083734A (en) * 2019-04-15 2019-08-02 中南大学 Semi-supervised image search method based on autoencoder network and robust core Hash
CN111628866A (en) * 2020-05-22 2020-09-04 深圳前海微众银行股份有限公司 Neural network verification method, device and equipment and readable storage medium
CN112861976A (en) * 2021-02-11 2021-05-28 温州大学 Sensitive image identification method based on twin graph convolution hash network
CN113011359A (en) * 2021-03-26 2021-06-22 浙江大学 Method for simultaneously detecting plane structure and generating plane description based on image and application
CN113343025A (en) * 2021-08-05 2021-09-03 中南大学 Sparse attack resisting method based on weighted gradient Hash activation thermodynamic diagram
CN113469695A (en) * 2020-03-30 2021-10-01 同济大学 Electronic fraud transaction identification method, system and device based on kernel supervision Hash model
CN114189351A (en) * 2021-10-25 2022-03-15 山东师范大学 Dense image retrieval method and system based on CNN and signcryption technology
CN114722208A (en) * 2022-06-08 2022-07-08 成都健康医联信息产业有限公司 Automatic classification and safety level grading method for health medical texts

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106503106A (en) * 2016-10-17 2017-03-15 北京工业大学 A kind of image hash index construction method based on deep learning

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106503106A (en) * 2016-10-17 2017-03-15 北京工业大学 A kind of image hash index construction method based on deep learning

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
柯圣财 等: "基于卷积神经网络和监督核哈希的图像检索方法", 《电子学报》 *

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109241313A (en) * 2018-08-14 2019-01-18 大连大学 A kind of image search method based on the study of high-order depth Hash
CN109241313B (en) * 2018-08-14 2021-11-02 大连大学 Image retrieval method based on high-order deep hash learning
CN109800314A (en) * 2019-01-03 2019-05-24 上海大学 A method of generating the Hash codes for being used for image retrieval using depth convolutional network
CN109933682A (en) * 2019-01-11 2019-06-25 上海交通大学 A kind of image Hash search method and system based on semanteme in conjunction with content information
CN109885716A (en) * 2019-02-18 2019-06-14 成都快眼科技有限公司 The image search method of discrete Hash is supervised based on heterogeneous multi-task learning depth
CN109947963A (en) * 2019-03-27 2019-06-28 山东大学 A kind of multiple dimensioned Hash search method based on deep learning
CN110083734A (en) * 2019-04-15 2019-08-02 中南大学 Semi-supervised image search method based on autoencoder network and robust core Hash
CN110083734B (en) * 2019-04-15 2024-05-03 中南大学 Semi-supervised image retrieval method based on self-coding network and robust kernel hash
CN113469695A (en) * 2020-03-30 2021-10-01 同济大学 Electronic fraud transaction identification method, system and device based on kernel supervision Hash model
CN113469695B (en) * 2020-03-30 2023-06-30 同济大学 Electronic fraud transaction identification method, system and device based on kernel supervision hash model
CN111628866A (en) * 2020-05-22 2020-09-04 深圳前海微众银行股份有限公司 Neural network verification method, device and equipment and readable storage medium
WO2021233183A1 (en) * 2020-05-22 2021-11-25 深圳前海微众银行股份有限公司 Neural network verification method, apparatus and device, and readable storage medium
CN112861976B (en) * 2021-02-11 2024-01-12 温州大学 Sensitive image identification method based on twin graph convolution hash network
CN112861976A (en) * 2021-02-11 2021-05-28 温州大学 Sensitive image identification method based on twin graph convolution hash network
CN113011359B (en) * 2021-03-26 2023-10-24 浙江大学 Method for simultaneously detecting plane structure and generating plane description based on image and application
CN113011359A (en) * 2021-03-26 2021-06-22 浙江大学 Method for simultaneously detecting plane structure and generating plane description based on image and application
CN113343025A (en) * 2021-08-05 2021-09-03 中南大学 Sparse attack resisting method based on weighted gradient Hash activation thermodynamic diagram
CN113343025B (en) * 2021-08-05 2021-11-02 中南大学 Sparse attack resisting method based on weighted gradient Hash activation thermodynamic diagram
CN114189351A (en) * 2021-10-25 2022-03-15 山东师范大学 Dense image retrieval method and system based on CNN and signcryption technology
CN114189351B (en) * 2021-10-25 2024-02-23 山东师范大学 Dense state image retrieval method and system based on CNN and signature technology
CN114722208A (en) * 2022-06-08 2022-07-08 成都健康医联信息产业有限公司 Automatic classification and safety level grading method for health medical texts
CN114722208B (en) * 2022-06-08 2022-11-01 成都健康医联信息产业有限公司 Automatic classification and safety level grading method for health medical texts

Similar Documents

Publication Publication Date Title
CN108304573A (en) Target retrieval method based on convolutional neural networks and supervision core Hash
Cai et al. Medical image retrieval based on convolutional neural network and supervised hashing
Wei Tan et al. Deep learning for plant species classification using leaf vein morphometric
Liu et al. Learning spatio-temporal representations for action recognition: A genetic programming approach
Shao et al. Feature learning for image classification via multiobjective genetic programming
Han et al. Matchnet: Unifying feature and metric learning for patch-based matching
CN106326288B (en) Image search method and device
Liu et al. A spark-based parallel fuzzy $ c $-Means segmentation algorithm for agricultural image Big Data
Xiang et al. Fabric image retrieval system using hierarchical search based on deep convolutional neural network
CN109918528A (en) A kind of compact Hash code learning method based on semanteme protection
CN111339343A (en) Image retrieval method, device, storage medium and equipment
CN110598022B (en) Image retrieval system and method based on robust deep hash network
Nguyen et al. Deep learning for metagenomic data: using 2d embeddings and convolutional neural networks
Wu et al. Plant leaf identification based on shape and convolutional features
CN112883931A (en) Real-time true and false motion judgment method based on long and short term memory network
Singh Content-based image retrieval using deep learning
Lumini et al. Ocean ecosystems plankton classification
Abdel-Nabi et al. Content based image retrieval approach using deep learning
Xu et al. Weakly supervised facial expression recognition via transferred DAL-CNN and active incremental learning
Li et al. An improved lightweight network architecture for identifying tobacco leaf maturity based on Deep learning
Ibrahim et al. Improving Date Fruit Classification Using CycleGAN-Generated Dataset.
Preyanka Lakshme et al. A review based on machine learning for feature selection and feature extraction
Prasetya et al. Indonesian food items labeling for tourism information using Convolution Neural Network
Vasudevan et al. A Hybrid Approach for Plant Disease Detection Using E-GAN and CapsNet.
Zhang et al. Improved image retrieval algorithm of GoogLeNet neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20180720

RJ01 Rejection of invention patent application after publication