CN108108806A - Convolutional neural networks initial method based on the extraction of pre-training model filter - Google Patents
Convolutional neural networks initial method based on the extraction of pre-training model filter Download PDFInfo
- Publication number
- CN108108806A CN108108806A CN201711335174.3A CN201711335174A CN108108806A CN 108108806 A CN108108806 A CN 108108806A CN 201711335174 A CN201711335174 A CN 201711335174A CN 108108806 A CN108108806 A CN 108108806A
- Authority
- CN
- China
- Prior art keywords
- mrow
- msub
- filter
- msubsup
- model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Processing (AREA)
Abstract
The present invention provides a kind of convolutional neural networks initial methods based on the extraction of pre-training model filter, it is related to technical field of video processing, the present invention utilizes minimum entropy loss and minimal reconstruction error approach, extract pre-training model median filter parameter, to initialized target Task Network model, realization meets actual application problem medium or small scale network initialization matter.The present invention is due to the use of minimum entropy loss and minimal linear reconstructing method, the extraction filter parameter from pre-training model, goal task network model is initialized, it is consistent with pre-training network structure present invention does not require goal task network structure, goal task can be made to meet memory overhead and calculating speed requirement in actual application problem according to practical application flexible design network structure.
Description
Technical field
The present invention relates to technical field of video processing, especially a kind of convolutional neural networks initial method.
Background technology
Depth convolutional neural networks (ConvolutionalNeuralNetwork, CNN) are by learning multilayered nonlinear net
Network structure using approaching for simple network structure function to achieve the objective, and then can be obtained from primary data sample focusing study
Obtain the character representation of sample data.Have benefited from mass data, depth convolutional neural networks are in recent years in artificial intelligence and machine
One of important breakthrough that learning areas obtain, it obtained in terms of graphical analysis, speech recognition and natural language processing it is huge into
Work(.
Due to often only having low volume data in actual application problem, the CNN models that training obtains have over-fitting situation,
Its model generalization ability is weaker, the poor performance on goal task.A kind of effective solution is right using good strategy
Network model initializes.Conventional method carries out convolutional network median filter parameter initial usually by being sampled to Gaussian Profile
Change.As network structure extends in breadth and depth, Gaussian Profile initial method is difficult to meet complex network structures requirement.
To solve the above problems, Lee and Sermanet et al. research and utilization supervised learning or unsupervised learning method, in CNN networks
Convolutional layer successively training initialization.Since correlation technique needs the additional training time, simultaneously because convolutional layer training is local most
Excellent problem, above-mentioned initial method are not used widely in practical applications.Girshick et al. proposes to utilize pre-training
The method of model initialization goal task network model.The pre-training model that training obtains on large-scale dataset has one
Fixed character representation ability and generalization ability.Pre-training model is used for the initialization of goal task network model, CNN network moulds
Type can be outstanding completion goal task.However, this limited to using pre-training model initialization method with following aspect.It is first
First, it is consistent with goal task network structure using pre-training model needs pre-training prototype network structure, as filtered in convolutional layer
Device quantity, wave filter size, step-length, this causes network model can not be according to goal task flexible design network structure.Secondly,
Since pre-training prototype network structure and scale is usually larger, goal task content expense and calculating speed will in actual application problem
Ask higher, therefore large scale network structural model can not be adapted to the goal task in actual application problem.How profit is studied
With pre-training network model, concrete application problem medium or small scale network is initialized, meets content expense in practical application
And calculating speed, it has a very important significance.
The content of the invention
For overcome the deficiencies in the prior art, the present invention provides a kind of convolution god based on the extraction of pre-training model filter
Through netinit method.The object of the invention is convolutional neural networks initialization of the design based on the extraction of pre-training model filter
Method using minimum entropy loss and minimal reconstruction error approach, extracts pre-training model median filter parameter, to initialize mesh
Task Network model is marked, realization meets actual application problem medium or small scale network initialization matter.
The technical solution adopted by the present invention to solve the technical problems comprises the following steps:
The first step:CNN network structures are designed for goal task;
Second step:Select pre-training network model;
3rd step:According to goal task network structure, extracted using two methods of minimum entropy loss or minimal reconstruction error
Pre-training model median filter parameter;Define Fs={ fi}M, FsRepresent M wave filter in pre-training model, wave filter size is ks
×ks, goal task network structure has N number of wave filter Ft, wave filter size is kt×kt, it is assumed that wave filter FsWith wave filter FtPlace
In identical convolutional layer depth;
A) the filter parameter extracting method based on minimum entropy loss
Utilize gauss hybrid models (GaussianMixtureModel, GMM) description pre-training model filter probability point
Cloth, filter set FsGaussian Mixture distribution g (Fs) it is expressed as form:
In formula (1), C represents Gauss model number in mixed model, gi(Fs) represent i-th of Gauss model,It represents
The prior probability (PriorProbability) of corresponding i-th of Gauss model, makes Fs(i) filter set F is representedsMiddle removal i-th
The filter set formed after wave filter represents mixed model probability distribution g using Kullback-Leibler (KL) is scattered
(Fs) and g (Fs(i)) cross entropy (RelativeEntropy) between is following form:
Wherein Filter set F is represented respectivelysIn j-th and k-th Gauss model prior probability,Table
Show filter set Fs(i) prior probability of l-th of Gauss model, D inkl(gi(Fs)||gk(Fs)) represent filter set FsIn
I-th of Gauss model gi(Fs) and filter set FsIn k-th of Gauss model gk(Fs) KL disperse, Dkl(gi(Fs)||gl
(Fs(i)) represent filter set FsIn i-th of Gauss model gi(Fs) and filter set Fs(i)In l-th of Gauss model gl
(Fs(i)) KL disperse;
Finally, the wave filter of N number of cross entropy minimum is extracted using formula (3):
Wave filter is extracted in a manner that nothing is put back to from FsMiddle extraction filter;
B) the filter parameter extracting method based on minimal reconstruction error
In minimal reconstruction error approach, filter set FsMedian filter fiUse filter set FtMedian filter f 'j
Linear reconstruction expression, i.e.,Wherein weight factorIt is N number of for scalarForm weight vectors Wi, wave filter
Set FsThe reconstructed error of all wave filters is expressed as:
In formula (4), all weight factorsForm weight matrix W, and W ∈ RM×N, N≤M, use L1 normal forms constraint power
Weight matrix W extracts problem to wave filter, is converted to the solution of following optimization problem:
Wherein, regular terms | W |1For L1 normal form representations, using L1 regularization least square methods (L1-regularized
LeastSquares) method solves formula (5), and γ is weight parameter, adjusts the proportionality coefficient of regular terms;
4th step:According to goal task network structure median filter size parameter kt×kt, in the pre-training model of extraction
Size is ks×ksWave filter carry out bilinear interpolation, the wave filter F in goal task networktWith the pre-training mould extracted
Type median filter FsWith same size and scale, with wave filter FsIn parameter to wave filter FtIn relevant parameter assigned
Value, you can realize the initialization to goal task network.
The beneficial effects of the present invention are due to the use of minimum entropy loss and minimal linear reconstructing method, from pre-training model
Middle extraction filter parameter, initializes goal task network model, present invention does not require goal task network structure and
Pre-training network structure is consistent, and goal task can be made to meet actual application problem according to practical application flexible design network structure
Middle memory overhead and calculating speed requirement.
Description of the drawings
Fig. 1 is the flow chart of the present invention.
Specific embodiment
The present invention is further described with reference to the accompanying drawings and examples.
Embodiment of the method, for goal task, is selected with CIFAR10, CIFAR100, SVHN and STL10 classification task data set
GoogleNet, CaffeNet and the VGG16 that training obtains on ImageNet are pre-training model, using minimum entropy loss and
Minimal reconstruction error approach extraction filter parameter, initializes goal task network model, with using Gaussian Profile
Random initializtion method compares, and investigates goal task network model classification accuracy (TestingError) and network model training
Convergence rate (normalizedAUC).The method of the present invention workflow is as shown in Figure 1.
As shown in Figure 1, the present invention comprises the steps of:
The first step:CNN network structures are designed for goal task;
The present invention designs CNN network structures for CIFAR10, CIFAR100, SVHN and STL10 classification task, and target is appointed
Network structure of being engaged in is as shown in table 1.
Table 1 CIFAR10/100, SVHN, STL10 goal task network structure
Second step:Select pre-training network model;
It is pre-training to select the network models such as GoogleNet, CaffeNet and the VGG16 that training obtains on ImageNet
Model;
3rd step:According to goal task network structure, extracted using two methods of minimum entropy loss or minimal reconstruction error
Pre-training model median filter parameter;Define Fs={ fi}M, FsRepresent M wave filter in pre-training model, wave filter size is ks
×ks, goal task network structure has N number of wave filter Ft, wave filter size is kt×kt, it is assumed that wave filter FsWith wave filter FtPlace
In identical convolutional layer depth;
A) the filter parameter extracting method based on minimum entropy loss
Utilize gauss hybrid models (GaussianMixtureModel, GMM) description pre-training model filter probability point
Cloth, filter set FsGaussian Mixture distribution g (Fs) it is expressed as form:
In formula (1), C represents Gauss model number in mixed model, gi(Fs) represent i-th of Gauss model,It represents
The prior probability (PriorProbability) of corresponding i-th of Gauss model, makes Fs(i)Represent filter set FsMiddle removal i-th
The filter set formed after wave filter represents mixed model probability distribution g using Kullback-Leibler (KL) is scattered
(Fs) and g (Fs(i)) between cross entropy (RelativeEntropy) be following form:
Wherein Filter set F is represented respectivelysIn j-th and k-th Gauss model prior probability,
Represent filter set Fs(i)In l-th of Gauss model prior probability, Dkl(gi(Fs)||gk(Fs)) represent filter set Fs
In i-th of Gauss model gi(Fs) and filter set FsIn k-th of Gauss model gk(Fs) KL disperse, Dkl(gi(Fs)||gl
(Fs(i)) represent filter set FsIn i-th of Gauss model gi(Fs) and filter set Fs(i)In l-th of Gauss model gl
(Fs(i)) KL disperse;
Finally, the wave filter of N number of cross entropy minimum is extracted using formula (3):
Wave filter is extracted in a manner that nothing is put back to from FsMiddle extraction filter;
In formula (2), Dkl(g(Fs)||g(Fs(i))) it is smaller when represent wave filter fiTo filter set FsProbability distribution tribute
It offers less.For the present invention by deleting the smaller wave filter of cross entropy, the larger wave filter of reservation cross entropy is larger using cross entropy
The parameter of wave filter, to initializing for goal task network model.
B) the filter parameter extracting method based on minimal reconstruction error
In minimal reconstruction error approach, filter set FsMedian filter fiUse filter set FtMedian filter f 'j
Linear reconstruction expression, i.e.,Wherein weight factorIt is N number of for scalarForm weight vectors Wi, wave filter
Set FsThe reconstructed error of all wave filters is expressed as:
In formula (4), all weight factorsForm weight matrix W, and W ∈ RM×N, since objective network scale is usually small
In pre-training prototype network structure and scale, thus N≤M, it is openness to strengthen weight matrix W, use L1 normal forms constraint weight square
Battle array W extracts problem to wave filter, is converted to the solution of following optimization problem:
Wherein, regular terms | W |1For L1 normal form representations, using L1 regularization least square methods (L1-regularized
LeastSquares) method solves formula (5), and γ is weight parameter, adjusts the proportionality coefficient of regular terms, γ is taken in the present invention
=0.4;
4th step:According to goal task network structure median filter size parameter kt×kt, in the pre-training model of extraction
Size is ks×ksWave filter carry out bilinear interpolation, the wave filter F in goal task networktWith the pre-training mould extracted
Type median filter FsWith same size and scale, with wave filter FsIn parameter to wave filter FtIn relevant parameter assigned
Value, you can realize the initialization to goal task network.
It is referred to using symbol M EL G, MEL C, MEL V using minimum entropy loss method in pre-training model GoogleNet,
CaffeNet, VGG16 extraction filter parameter initialization are referred to using symbol M RE@G, MRE@C, MRE@V6 and use minimal reconstruction
Error approach is in pre-training model GoogleNet, CaffeNet, VGG16 extraction filter parameter initialization.Different initialization sides
Method classification accuracy (TestingError) and convergence rate on goal task CIFAR10, CIFAR100, SVHN and STL10
(normalizedAUC) index is as shown in table 2.
Table 2 different initial method classification accuracy and convergence rates on goal task
As shown in table 2, the present invention extracts filter using minimum entropy loss and minimal reconstruction error approach from pre-training model
The method of ripple device parameters on target Task Network initialization, compares the method using Gaussian Profile random initializtion, accurate in classification
Best test error has been taken in true rate, has also there is faster convergence rate on model training.It is pointed out that different is pre-
Training pattern shows difference, it is necessary to select different models as goal task according to specific tasks on different goal tasks
The pre-training model of netinit.
Claims (1)
1. a kind of convolutional neural networks initial method based on the extraction of pre-training model filter, it is characterised in that including following
Step:
The first step:CNN network structures are designed for goal task;
Second step:Select pre-training network model;
3rd step:According to goal task network structure, two methods of minimum entropy loss or the pre- instruction of minimal reconstruction error extraction are used
Practice model median filter parameter;Define Fs={ fi}M, FsRepresent M wave filter in pre-training model, wave filter size is ks×
ks, goal task network structure has N number of wave filter Ft, wave filter size is kt×kt, it is assumed that wave filter FsWith wave filter FtIt is in
Identical convolutional layer depth;
A) the filter parameter extracting method based on minimum entropy loss
Pre-training model filter probability distribution, filter set F are described using gauss hybrid modelssGaussian Mixture distribution g (Fs)
It is expressed as form:
<mrow>
<mi>g</mi>
<mrow>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>=</mo>
<msubsup>
<mrow>
<mo>{</mo>
<msubsup>
<mi>&pi;</mi>
<mi>i</mi>
<mi>s</mi>
</msubsup>
<mo>,</mo>
<msub>
<mi>g</mi>
<mi>i</mi>
</msub>
<mrow>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>}</mo>
</mrow>
<mrow>
<mi>i</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>C</mi>
</msubsup>
</mrow>
In formula (1), C represents Gauss model number in mixed model, gi(Fs) represent i-th of Gauss model,Represent corresponding the
The prior probability of i Gauss model, makes Fs(i) filter set F is representedsThe wave filter collection formed after the i-th wave filter of middle removal
It closes, disperses to represent mixed model probability distribution g (F using Kullback-Leiblers) and g (Fs(i)) between cross entropy for such as
Lower form:
<mrow>
<msub>
<mi>D</mi>
<mrow>
<mi>k</mi>
<mi>l</mi>
</mrow>
</msub>
<mrow>
<mo>(</mo>
<mi>g</mi>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
<mo>|</mo>
<mo>|</mo>
<mi>g</mi>
<mo>(</mo>
<msub>
<mi>F</mi>
<mrow>
<mi>s</mi>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
</mrow>
</msub>
<mo>)</mo>
<mo>)</mo>
</mrow>
<mo>=</mo>
<munder>
<mo>&Sigma;</mo>
<mi>j</mi>
</munder>
<msubsup>
<mi>&pi;</mi>
<mi>j</mi>
<mi>s</mi>
</msubsup>
<mi>l</mi>
<mi>o</mi>
<mi>g</mi>
<mfrac>
<mrow>
<msub>
<mi>&Sigma;</mi>
<mi>k</mi>
</msub>
<msubsup>
<mi>&pi;</mi>
<mi>k</mi>
<mi>s</mi>
</msubsup>
<mi>exp</mi>
<mo>{</mo>
<mo>-</mo>
<msub>
<mi>D</mi>
<mrow>
<mi>k</mi>
<mi>l</mi>
</mrow>
</msub>
<mrow>
<mo>(</mo>
<msub>
<mi>g</mi>
<mi>i</mi>
</msub>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>|</mo>
<mo>|</mo>
<msub>
<mi>g</mi>
<mi>k</mi>
</msub>
<mrow>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>)</mo>
<mo>}</mo>
</mrow>
<mrow>
<msub>
<mi>&Sigma;</mi>
<mi>l</mi>
</msub>
<msubsup>
<mi>&pi;</mi>
<mi>l</mi>
<mrow>
<mi>s</mi>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
</mrow>
</msubsup>
<mi>exp</mi>
<mo>{</mo>
<mo>-</mo>
<msub>
<mi>D</mi>
<mrow>
<mi>k</mi>
<mi>l</mi>
</mrow>
</msub>
<mrow>
<mo>(</mo>
<msub>
<mi>g</mi>
<mi>i</mi>
</msub>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>|</mo>
<mo>|</mo>
<msub>
<mi>g</mi>
<mi>k</mi>
</msub>
<mrow>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
</mrow>
<mo>)</mo>
<mo>}</mo>
</mrow>
</mfrac>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>2</mn>
<mo>)</mo>
</mrow>
</mrow>
WhereinFilter set F is represented respectivelysIn j-th and k-th Gauss model prior probability,Represent filter
Ripple device set Fs(i) prior probability of l-th of Gauss model, D inkl(gi(Fs)||gk(Fs)) represent filter set FsIn i-th
A Gauss model gi(Fs) and filter set FsIn k-th of Gauss model gk(Fs) KL disperse, Dkl(gi(Fs)||gl(Fs(i))
Represent filter set FsIn i-th of Gauss model gi(Fs) and filter set Fs(i)In l-th of Gauss model gl(Fs(i))
KL disperses;
Finally, the wave filter of N number of cross entropy minimum is extracted using formula (3):
<mrow>
<msub>
<mi>F</mi>
<mi>t</mi>
</msub>
<mo>=</mo>
<mi>arg</mi>
<munder>
<mi>max</mi>
<mrow>
<mn>1</mn>
<mo>,</mo>
<mn>...</mn>
<mo>,</mo>
<mi>N</mi>
</mrow>
</munder>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>N</mi>
</munderover>
<msub>
<mi>D</mi>
<mrow>
<mi>k</mi>
<mi>l</mi>
</mrow>
</msub>
<mrow>
<mo>(</mo>
<mi>g</mi>
<mo>(</mo>
<msub>
<mi>F</mi>
<mi>s</mi>
</msub>
<mo>)</mo>
<mo>|</mo>
<mo>|</mo>
<mi>g</mi>
<mo>(</mo>
<msub>
<mi>F</mi>
<mrow>
<mi>s</mi>
<mrow>
<mo>(</mo>
<mi>i</mi>
<mo>)</mo>
</mrow>
</mrow>
</msub>
<mo>)</mo>
<mo>)</mo>
</mrow>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>3</mn>
<mo>)</mo>
</mrow>
</mrow>
Wave filter is extracted in a manner that nothing is put back to from FsMiddle extraction filter;
B) the filter parameter extracting method based on minimal reconstruction error
In minimal reconstruction error approach, filter set FsMedian filter fiUse filter set FtMedian filter fj' linear
Reconstruct expression, i.e.,Wherein weight factorIt is N number of for scalarForm weight vectors Wi, filter set
FsThe reconstructed error of all wave filters is expressed as:
<mrow>
<mi>E</mi>
<mo>=</mo>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>M</mi>
</munderover>
<mo>|</mo>
<msub>
<mi>f</mi>
<mi>i</mi>
</msub>
<mo>-</mo>
<munderover>
<mo>&Sigma;</mo>
<mi>j</mi>
<mi>N</mi>
</munderover>
<msubsup>
<mi>w</mi>
<mi>j</mi>
<mi>i</mi>
</msubsup>
<msub>
<msup>
<mi>f</mi>
<mo>&prime;</mo>
</msup>
<mi>j</mi>
</msub>
<msubsup>
<mo>|</mo>
<mn>2</mn>
<mn>2</mn>
</msubsup>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>4</mn>
<mo>)</mo>
</mrow>
</mrow>
In formula (4), all weight factorsForm weight matrix W, and W ∈ RM×N, N≤M, use L1 normal forms constraint weight square
Battle array W extracts problem to wave filter, is converted to the solution of following optimization problem:
<mrow>
<mi>min</mi>
<munderover>
<mo>&Sigma;</mo>
<mrow>
<mi>i</mi>
<mo>=</mo>
<mn>1</mn>
</mrow>
<mi>M</mi>
</munderover>
<mo>|</mo>
<msub>
<mi>f</mi>
<mi>i</mi>
</msub>
<mo>-</mo>
<munderover>
<mo>&Sigma;</mo>
<mi>j</mi>
<mi>N</mi>
</munderover>
<msubsup>
<mi>w</mi>
<mi>j</mi>
<mi>i</mi>
</msubsup>
<msub>
<msup>
<mi>f</mi>
<mo>&prime;</mo>
</msup>
<mi>j</mi>
</msub>
<msubsup>
<mo>|</mo>
<mn>2</mn>
<mn>2</mn>
</msubsup>
<mo>+</mo>
<mi>&gamma;</mi>
<mo>|</mo>
<mi>W</mi>
<msub>
<mo>|</mo>
<mn>1</mn>
</msub>
<mo>-</mo>
<mo>-</mo>
<mo>-</mo>
<mrow>
<mo>(</mo>
<mn>5</mn>
<mo>)</mo>
</mrow>
</mrow>
Wherein, regular terms | W |1For L1 normal form representations, formula (5) is solved using L1 regularization least squares method, γ is power
Weight parameter adjusts the proportionality coefficient of regular terms;
4th step:According to goal task network structure median filter size parameter kt×kt, to size in the pre-training model of extraction
For ks×ksWave filter carry out bilinear interpolation, the wave filter F in goal task networktWith in the pre-training model that extracts
Wave filter FsWith same size and scale, with wave filter FsIn parameter to wave filter FtIn relevant parameter carry out assignment, i.e.,
The initialization to goal task network can be achieved.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711335174.3A CN108108806A (en) | 2017-12-14 | 2017-12-14 | Convolutional neural networks initial method based on the extraction of pre-training model filter |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711335174.3A CN108108806A (en) | 2017-12-14 | 2017-12-14 | Convolutional neural networks initial method based on the extraction of pre-training model filter |
Publications (1)
Publication Number | Publication Date |
---|---|
CN108108806A true CN108108806A (en) | 2018-06-01 |
Family
ID=62215927
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711335174.3A Pending CN108108806A (en) | 2017-12-14 | 2017-12-14 | Convolutional neural networks initial method based on the extraction of pre-training model filter |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108108806A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109472360A (en) * | 2018-10-30 | 2019-03-15 | 北京地平线机器人技术研发有限公司 | Update method, updating device and the electronic equipment of neural network |
CN110766044A (en) * | 2019-09-11 | 2020-02-07 | 浙江大学 | Neural network training method based on Gaussian process prior guidance |
-
2017
- 2017-12-14 CN CN201711335174.3A patent/CN108108806A/en active Pending
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109472360A (en) * | 2018-10-30 | 2019-03-15 | 北京地平线机器人技术研发有限公司 | Update method, updating device and the electronic equipment of neural network |
US11328180B2 (en) | 2018-10-30 | 2022-05-10 | Beijing Horizon Robotics Technology Research And Development Co., Ltd. | Method for updating neural network and electronic device |
CN110766044A (en) * | 2019-09-11 | 2020-02-07 | 浙江大学 | Neural network training method based on Gaussian process prior guidance |
CN110766044B (en) * | 2019-09-11 | 2021-10-26 | 浙江大学 | Neural network training method based on Gaussian process prior guidance |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108416755A (en) | A kind of image de-noising method and system based on deep learning | |
CN108304826A (en) | Facial expression recognizing method based on convolutional neural networks | |
CN105787557B (en) | A kind of deep-neural-network construction design method of computer intelligence identification | |
CN106651830A (en) | Image quality test method based on parallel convolutional neural network | |
CN109949255A (en) | Image rebuilding method and equipment | |
CN109829541A (en) | Deep neural network incremental training method and system based on learning automaton | |
CN107633486A (en) | Structure Magnetic Resonance Image Denoising based on three-dimensional full convolutional neural networks | |
CN107563422A (en) | A kind of polarization SAR sorting technique based on semi-supervised convolutional neural networks | |
CN108961245A (en) | Picture quality classification method based on binary channels depth parallel-convolution network | |
CN106845529A (en) | Image feature recognition methods based on many visual field convolutional neural networks | |
CN106874956A (en) | The construction method of image classification convolutional neural networks structure | |
CN107833183A (en) | A kind of satellite image based on multitask deep neural network while super-resolution and the method for coloring | |
CN109146162B (en) | A kind of probability wind speed forecasting method based on integrated Recognition with Recurrent Neural Network | |
CN107609638A (en) | A kind of method based on line decoder and interpolation sampling optimization convolutional neural networks | |
CN107358293A (en) | A kind of neural network training method and device | |
CN107316054A (en) | Non-standard character recognition methods based on convolutional neural networks and SVMs | |
CN110097075A (en) | Ocean mesoscale eddy classifying identification method based on deep learning | |
CN106991374A (en) | Handwritten Digit Recognition method based on convolutional neural networks and random forest | |
CN108053398A (en) | A kind of melanoma automatic testing method of semi-supervised feature learning | |
CN107798697A (en) | A kind of medical image registration method based on convolutional neural networks, system and electronic equipment | |
CN107516128A (en) | A kind of flowers recognition methods of the convolutional neural networks based on ReLU activation primitives | |
CN108573284A (en) | Deep learning facial image extending method based on orthogonal experiment analysis | |
CN110363239A (en) | Multi-mode data-oriented hand sample machine learning method, system and medium | |
CN109325513B (en) | Image classification network training method based on massive single-class images | |
CN107133640A (en) | Image classification method based on topography's block description and Fei Sheer vectors |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180601 |