CN110766044B - Neural network training method based on Gaussian process prior guidance - Google Patents
Neural network training method based on Gaussian process prior guidance Download PDFInfo
- Publication number
- CN110766044B CN110766044B CN201910858834.9A CN201910858834A CN110766044B CN 110766044 B CN110766044 B CN 110766044B CN 201910858834 A CN201910858834 A CN 201910858834A CN 110766044 B CN110766044 B CN 110766044B
- Authority
- CN
- China
- Prior art keywords
- training
- batch
- neural network
- samples
- period
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a neural network training method based on Gaussian process prior guidance, which is used for improving the training process of a neural network to obtain better training effect. The method specifically comprises the following steps: s1, acquiring a data set for neural network training, selecting a representative set for modeling prior knowledge, and defining an algorithm target; s2, carrying out a training process of iterative learning on the neural network model in batches in one period, and sequentially executing steps S21-S24 in each iterative batch; s3, after the training process of the current period is finished, verifying the neural network model by using a verification set to obtain the error rate of the verification set of the current model; and S4, continuously repeating the steps S2 and S3 to carry out a multi-period training process on the neural network model until the model converges. The neural network training method based on the Gaussian process prior guidance can effectively improve the training effectiveness in tasks, improves the network learning ability and learning quality, and has good application value.
Description
Technical Field
The invention belongs to the field of computer vision, and particularly relates to a neural network training method based on Gaussian process prior guidance.
Background
Image classification is a task of distinguishing between different classes of pictures in a data set. At present, the mainstream solution on the task of image classification is to train a convolutional neural network to solve the problem, and the training method generally adopts a random gradient descent method. In recent years, as the progress rate of network architecture is gradually reduced, improvement of training strategies is increasingly important. To this end, the present invention recognizes that it is desirable to provide as sophisticated and efficient supervised information as possible to train a given model better in supervised learning such as image classification. The data set provides a label, but the inherent label only represents the classification result of the picture and does not show the relationship between the picture and other categories. On the basis of utilizing the inherent labels of the data set, the invention introduces the soft labels representing the probability distribution of the image classification result by random process modeling, and combines the soft labels with the inherent labels of the data set for use, thereby improving the effectiveness of the training method.
Disclosure of Invention
In order to solve the problems, the invention provides a neural network training method based on Gaussian process prior guidance. The method is based on deep learning and random processes, the Gaussian process in the random process is used for modeling the correlation between images, the model is used for giving a 'soft label' to each training sample, the soft label and the inherent label of a data set are used for guiding the training process, and therefore the trained model is more accurate and robust.
In order to achieve the purpose, the technical scheme of the invention is as follows:
a neural network training method based on Gaussian process prior guidance comprises the following steps:
s1, acquiring a data set for neural network training, selecting a representative set for modeling prior knowledge, and defining an algorithm target;
s2, carrying out a training process of batch iterative learning in one period (epoch) on the neural network model, and sequentially executing steps S21-S24 in each iterative batch (batch):
s21, before the current iteration batch starts, carrying out combined modeling on samples in the representative set and training samples of the batch to obtain related prior knowledge;
s22, starting the current iteration batch learning process, and calculating the soft label of the batch of training samples according to the representative set and the batch of training samples; after the forward propagation process of the batch of training samples is carried out, the loss functions of the network output and the inherent labels of the batch of training samples are calculatedAnd loss functions of inherent labels and soft labels of the training samples of the batch
S24, making a total loss functionAnd toIs counter-propagating, whereinIn part andsome of which are used to optimize all parameters of the neural network,part of the convolutional layer parameters are only used for optimizing the neural network;
s3, after the training process of the current period is finished, verifying the neural network model by using a verification set to obtain the error rate of the verification set of the current model;
and S4, continuously repeating the steps S2 and S3 to carry out a multi-period training process on the neural network model until the model converges.
Based on the scheme, the steps can be realized in the following modes:
the representative set in step S1 is a set containing a plurality of images of different types, and the method for constructing the representative set includes:
first, the number of categories for the entire data set is evaluated:
when the category number of the data set is less than 50 classes, taking 50 pictures from each class of images, and then taking the pictures taken from all the classes as a representative set;
when the category number of the data set is more than or equal to 50 categories, taking 100 pictures from each category of images, and then taking the pictures taken from all the categories as a representative set;
In step S21, the specific steps of performing joint modeling on the samples in the representative set and the training samples of the batch and obtaining the related prior knowledge include:
s211, carrying out feature extraction on samples in the representative set and training samples of the batch by using convolutional layer parameters of the initial neural network model in each training process to obtain feature vectors of all samples;
s212, jointly modeling all samples in the representative set and samples to be predicted into a Gaussian process:
wherein the content of the first and second substances,a representative set is represented that represents a set of representations,to represent the set of feature vectors for all picture samples in a set,a set of feature vectors representing all samples in the set;is the label of the sample to be predicted, hbIs the feature vector of the sample to be predicted; k (·, ·) represents a covariance matrix, and an RBF kernel function is used for calculation, wherein the calculation general formula of the RBF kernel function is as follows:
wherein r is2(a, b) represents a second order Euclidean distance between a and b, and l is a characteristic length; -
The specific implementation method of step S22 is:
s221, calculating according to the batch of training samplesK(hb,hb)、Performing a gaussian process regression algorithm using the gaussian process constructed in step S21, andand (3) predicting:
wherein g ismAnd gvRespectively, the predicted mean and variance;
s222, carrying out forward propagation on the network, and calculating the current output h (x) of the networki) And using h (x)i) And gm、gvComputing loss function network output h (x)i) And the inherent label y of the training sample of the batchiLoss function ofAnd the inherent label y of the training sample of the batchiAnd a soft label gm(xi) Loss function of
WhereinAndboth represent cross entropy calculations, and the calculation formula for the parameters α and γ is:
u is the error rate of the validation set during the previous training period, and the initial value during the first training period isC is the category number of the data set;for the last iteration batch in the training process of this periodAbsolute value of, in the first iteration batchAll initial values of (1).
In step S23, the network outputs h (x)i) And the soft label g of the training sample of the batchm(xi) Loss function ofThe calculation formula of (2) is as follows:
wherein:the relative entropy calculation is expressed, and the calculation formula of the parameter beta is as follows;
u is the error rate of the verification set in the previous training process, and the initial value isC is the number of categories;for the last iteration batch in the training process of this periodAbsolute value of, in the first iteration batchAll initial values of (1).
In step S24, the three term loss functions obtained in steps S22 and S23 are added to form a total loss functionTo pairOptimizing to achieve the algorithm goal, whereinIn part andsome of which are used to optimize all parameters of the neural network,part of which is used only to optimize convolutional layer parameters of the neural network.
In step S3, after all iterations of the training process in this period are completed, the pictures of the verification set are sequentially passed through the network by using the current network, and then the error rate of the prediction result given by the current network is calculated.
Compared with the prior art, the invention has the following beneficial effects:
firstly, the neural network training method based on the prior guidance of the Gaussian process provides a solution to the problem that only a small part of samples can be sampled simultaneously for training and global information cannot be considered in the random gradient descent method commonly used in the current deep learning, and the performance of the trained network can be improved by effectively solving the problem.
Secondly, the representative set sampling method of the invention can play a role in data sets with different sizes based on different characteristics of different data sets.
Finally, compared with the traditional one-item loss function, the three-item loss function provided by the invention can enable the model to consider information brought by different labels and global information contained in the 'soft label', so that the training of the model is more perfect.
The neural network training method based on the Gaussian process prior guidance can effectively improve the training effectiveness in tasks, improves the network learning ability and learning quality, and has good application value.
Drawings
FIG. 1 is a schematic flow diagram of the present invention;
FIG. 2 is a comparison of training set error rates of the ResNet20 network on a CIFAR-100 dataset in an example;
FIG. 3 is a comparison of validation set error rates of the ResNet20 network on a CIFAR-100 dataset in an example.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
On the contrary, the invention is intended to cover alternatives, modifications, equivalents and alternatives which may be included within the spirit and scope of the invention as defined by the appended claims. Furthermore, in the following detailed description of the present invention, certain specific details are set forth in order to provide a better understanding of the present invention. It will be apparent to one skilled in the art that the present invention may be practiced without these specific details.
As shown in fig. 1, a neural network training method based on a gaussian process prior guidance is characterized by comprising the following steps:
s1, acquiring a data set for neural network training, and defining a neural network model structure to be trained. A representative set for modeling prior knowledge is selected from the data set. The representative set in the step is a set comprising a plurality of images of different types, and the construction method of the representative set comprises the following steps:
first, the number of categories for the entire data set is evaluated:
when the category number of the data set is less than 50 classes, taking 50 pictures from each class of images, and then taking the pictures taken from all the classes as a representative set;
when the category number of the data set is more than or equal to 50 categories, taking 100 pictures from each category of images, and then taking the pictures taken from all the categories as a representative set;
S2, carrying out a training process of iterative learning on the neural network model in batches in one period, and sequentially executing steps S21-S24 in each iterative batch:
and S21, before the current iteration batch starts, carrying out combined modeling on the samples in the representative set and the training samples of the batch and obtaining related prior knowledge. In this embodiment, the specific implementation step in step S21 includes:
s211, carrying out feature extraction on samples in the representative set and training samples of the batch by using convolutional layer parameters of the initial neural network model in each training process to obtain feature vectors of all samples;
s212, supposing that the representative set isAll picture sample sets in the representative set areThe set of intrinsic labels representing all picture samples in the set isX hereiIs a picture, yiIs a label of the picture. Defining: to represent all samples in the setThe function f () is the structure of the neural network model except the fully connected layer, i.e. all convolutional layer parameters of the neural network model, and is used to extract the feature vector of the sample。Is the label of the sample to be predicted, hbIs the feature vector of the sample to be predicted. The function h () represents a fully connected layer.
Jointly modeling all samples in the representative set and the samples to be predicted into a Gaussian process:
wherein, K (·,) represents a covariance matrix, which is calculated using an RBF kernel, the calculation formula of the RBF kernel is:
wherein r is2(a, b) represents the second order Euclidean distance between a and b, and l is the characteristic length. It should be noted that, when one of the two terms a and b is a matrix and the other is a vector, we need to first extend the vector to the same dimension as the matrix and then calculate the euclidean distance.
S22, starting a current iteration batch learning process, and firstly calculating soft labels of training samples of a batch according to a representative set and the training samples of the batch under the guidance of prior knowledge; after the forward propagation process of the batch of training samples is carried out, the loss functions of the network output and the inherent labels of the batch of training samples are calculatedAnd inherent label and soft label of the training sample of the batchLoss function of label
In this embodiment, the specific implementation method of step S22 is as follows:
s221, calculating according to the batch of training samplesK(hb,hb)、Performing a Gaussian process regression algorithm using the Gaussian process constructed in step S21, assuming thatSo as toThe predictions of (1) are:
wherein g ismAnd gvRespectively, the predicted mean and variance;
s222, carrying out forward propagation on the network similarly to the traditional deep learning process, and calculating the current output h (x) of the networki) And using h (x)i) And gm、gvComputing loss function network output h (x)i) And the inherent label y of the training sample of the batchiLoss function ofAnd the inherent label y of the training sample of the batchiAnd a soft label gm(xi) Loss function of
It should be noted that h (x)i) Represents the batch of training samples xiIs output from the network, and gm(xi) Represents the batch of training samples xiIs taken as the feature vector ofbInputting the average value g predicted after the formula in S221m。
In addition, the first and second substrates are,andboth represent cross entropy calculations, and the calculation formula for the parameters α and γ is:
u is the error rate of the validation set in the previous training process, and since the error rate of the validation set in the previous training process does not exist in the first training process, the initial value of u in the first training process is directly adopted asC is the number of categories of the data set.For the last iteration batch in the training process of this periodSince there is no previous iteration batch for the first iteration batch, the absolute value of (c) in the first iteration batchAll initial values of (1).
In this embodiment, in step S23, the network outputs h (x)i) And the soft label g of the training sample of the batchm(xi) Loss function ofThe calculation formula of (2) is as follows:
wherein:the relative entropy calculation is expressed, and the calculation formula of the parameter beta is as follows;
u is the error rate of the validation set in the previous training process, and the initial value in the same first training process isC is the number of categories;for the last iteration batch in the training process of this periodAbsolute value of (a), in the same first iteration batchAll initial values of (1).
S24, making a total loss functionAnd toIs counter-propagating, whereinIn part andsome of which are used to optimize all parameters of the neural network,part of which is used only to optimize convolutional layer parameters of the neural network. In this step, the three loss functions obtained in steps S22 and S23 are added to form a total loss functionTo pairAnd optimizing to achieve the algorithm target.
And S3, after the training process of the current period is finished, verifying the neural network model by using the verification set to obtain the error rate of the verification set of the current model.
In the specific calculation of the embodiment, after all iterations (iteration) of the current epoch training process are completed, the pictures of the verification set are sequentially passed through the network by using the current network, and then the error rate of the prediction result given by the current network is calculated. Whether the model converges may be determined based on whether the validation set error rate is below a threshold. If the model converges, the training of the neural network is ended, and if the model does not converge, the step S4 is executed continuously.
And S4, continuously repeating the steps S2 and S3 to carry out a multi-period training process on the neural network model until the model converges.
The above-described method is applied to specific examples so that those skilled in the art can better understand the effects of the present invention.
Examples
The implementation method of this embodiment is as described above, and specific steps are not elaborated, and the effect is shown only for case data. The invention is implemented on three data sets with truth labels based on a ResNet network, and the three data sets are respectively as follows:
cifar10 dataset
Cifar100 dataset
Tiny-ImageNet dataset
This example performed a set of experiments on each selected data set, comparing the general SGD optimization method with the method of the present invention.
The results of the experiments of this example are shown in Table 1 for comparison of accuracy. The data in the figure shows the average performance of 5 tests of the invention on the relevant data set, GPGL in the table is the neural network training method (Gaussian Process Guided Learning) based on the prior guidance of the Gaussian Process
TABLE 1 comparison of accuracy of experimental results
In addition, fig. 2 shows the comparison of training set error rates on a CIFAR-100 data set between the general SGD optimization method and the method of the present invention; fig. 3 is a comparison of the common SGD optimization method and the verification set error rate of the method of the present invention on a CIFAR-100 data set, which can visually indicate that the method of the present invention can improve the performance of the trained network compared to the conventional method.
Through the technical scheme, the neural network training method based on the Gaussian process prior guidance is implemented. The method can model the relationship between different types of pictures on various real image data, thereby helping the convolutional neural network training to be better.
The above description is only exemplary of the present invention and should not be construed as limiting the present invention, and any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.
Claims (1)
1. A neural network training method based on Gaussian process prior guidance is characterized by comprising the following steps:
s1, acquiring a data set for neural network training, selecting a representative set for modeling prior knowledge, and defining an algorithm target;
s2, carrying out a training process of iterative learning on the neural network model in batches in one period, and sequentially executing steps S21-S24 in each iterative batch:
s21, before the current iteration batch starts, carrying out combined modeling on samples in the representative set and training samples of the batch to obtain related prior knowledge;
s22, starting the current iteration batch learning process, and calculating the soft label of the batch of training samples according to the representative set and the batch of training samples; after the forward propagation process of the batch of training samples is carried out, the loss functions of the network output and the inherent labels of the batch of training samples are calculatedAnd loss functions of inherent labels and soft labels of the training samples of the batch
S24, making a total loss functionAnd toIs counter-propagating, whereinIn part andsome of which are used to optimize all parameters of the neural network,part of the convolutional layer parameters are only used for optimizing the neural network;
s3, after the training process of the current period is finished, verifying the neural network model by using a verification set to obtain the error rate of the verification set of the current model;
s4, continuously repeating the steps S2 and S3 to carry out a multi-period training process on the neural network model until the model converges;
the representative set in step S1 is a set containing a plurality of images of different types, and the method for constructing the representative set includes:
first, the number of categories for the entire data set is evaluated:
when the category number of the data set is less than 50 classes, taking 50 pictures from each class of images, and then taking the pictures taken from all the classes as a representative set;
when the category number of the data set is more than or equal to 50 categories, taking 100 pictures from each category of images, and then taking the pictures taken from all the categories as a representative set;
in step S21, the specific steps of performing joint modeling on the samples in the representative set and the training samples of the batch and obtaining the related prior knowledge include:
s211, carrying out feature extraction on samples in the representative set and training samples of the batch by using convolutional layer parameters of the initial neural network model in each training process to obtain feature vectors of all samples;
s212, jointly modeling all samples in the representative set and samples to be predicted into a Gaussian process:
wherein the content of the first and second substances,a representative set is represented that represents a set of representations,to represent the set of feature vectors for all picture samples in a set,a set of feature vectors representing all samples in the set;is the label of the sample to be predicted, hbIs the feature vector of the sample to be predicted; k (·, ·) represents a covariance matrix, and an RBF kernel function is used for calculation, wherein the calculation general formula of the RBF kernel function is as follows:
wherein r is2(a, b) represents a second order Euclidean distance between a and b, and l is a characteristic length;
The specific implementation method of step S22 is:
s221, calculating according to the batch of training samplesK(hb,hb)、Performing a gaussian process regression algorithm using the gaussian process constructed in step S21, andand (3) predicting:
wherein g ismAnd gvRespectively, the predicted mean and variance;
s222, carrying out forward propagation on the network, and calculating the current output h (x) of the networki) And using h (x)i) And gm、gvComputing loss function network output h (x)i) And the inherent label y of the training sample of the batchiLoss function ofAnd inherent mark of the batch of training samplesSign yiAnd a soft label gm(xi) Loss function of
WhereinAndboth represent cross entropy calculations, and the calculation formula for the parameters α and γ is:
u is the error rate of the validation set during the previous training period, and the initial value during the first training period isC is the category number of the data set;for the last iteration batch in the training process of this periodAbsolute value of, in the first iteration batchThe initial values of (a) are all 1;
in step S23, the network outputs h (x)i) And the soft label g of the training sample of the batchm(xi) Loss function ofThe calculation formula of (2) is as follows:
wherein:the relative entropy calculation is expressed, and the calculation formula of the parameter beta is as follows;
u is the error rate of the verification set in the previous training process, and the initial value isC is the number of categories;for the last iteration batch in the training process of this periodAbsolute value of, in the first iteration batchThe initial values of (a) are all 1;
in step S24, the three term loss functions obtained in steps S22 and S23 are added to form a total loss functionTo pairOptimizing to achieve the algorithm goal, whereinIn part andsome of which are used to optimize all parameters of the neural network,part of the convolutional layer parameters are only used for optimizing the neural network;
in step S3, after all iterations of the training process in this period are completed, the pictures of the verification set are sequentially passed through the network by using the current network, and then the error rate of the prediction result given by the current network is calculated.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910858834.9A CN110766044B (en) | 2019-09-11 | 2019-09-11 | Neural network training method based on Gaussian process prior guidance |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910858834.9A CN110766044B (en) | 2019-09-11 | 2019-09-11 | Neural network training method based on Gaussian process prior guidance |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110766044A CN110766044A (en) | 2020-02-07 |
CN110766044B true CN110766044B (en) | 2021-10-26 |
Family
ID=69329421
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910858834.9A Active CN110766044B (en) | 2019-09-11 | 2019-09-11 | Neural network training method based on Gaussian process prior guidance |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110766044B (en) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111931801B (en) * | 2020-05-28 | 2024-03-12 | 浙江大学 | Dynamic route network learning method based on path diversity and consistency |
CN111639759A (en) * | 2020-06-01 | 2020-09-08 | 深圳前海微众银行股份有限公司 | Neural network model protection method, device, equipment and readable storage medium |
CN111667016B (en) * | 2020-06-12 | 2023-01-06 | 中国电子科技集团公司第三十六研究所 | Incremental information classification method based on prototype |
CN111860660A (en) * | 2020-07-24 | 2020-10-30 | 辽宁工程技术大学 | Small sample learning garbage classification method based on improved Gaussian network |
CN111860424B (en) * | 2020-07-30 | 2023-07-11 | 厦门熵基科技有限公司 | Training method and device for visible light palm recognition model |
CN112380631B (en) * | 2020-12-02 | 2023-02-14 | 黑龙江科技大学 | Novel iterative hybrid test method based on neural network |
CN112633503B (en) * | 2020-12-16 | 2023-08-22 | 浙江大学 | Tool variable generation and handwritten number recognition method and device based on neural network |
CN112614550B (en) * | 2020-12-17 | 2024-03-15 | 华东理工大学 | Molecular sieve X-ray diffraction spectrum peak position prediction method based on neural network |
CN112884059B (en) * | 2021-03-09 | 2022-07-05 | 电子科技大学 | Small sample radar working mode classification method fusing priori knowledge |
CN113435641B (en) * | 2021-06-24 | 2023-03-07 | 布瑞克农业大数据科技集团有限公司 | Full-automatic management method and system for agricultural products and storage medium |
CN114463602B (en) * | 2022-04-12 | 2022-07-08 | 北京云恒科技研究院有限公司 | Target identification data processing method based on big data |
CN116127371B (en) * | 2022-12-06 | 2023-09-08 | 东北林业大学 | Multi-user model joint iteration method integrating prior distribution and homomorphic chaotic encryption |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108108806A (en) * | 2017-12-14 | 2018-06-01 | 西北工业大学 | Convolutional neural networks initial method based on the extraction of pre-training model filter |
WO2018184222A1 (en) * | 2017-04-07 | 2018-10-11 | Intel Corporation | Methods and systems using improved training and learning for deep neural networks |
CN110020718A (en) * | 2019-03-14 | 2019-07-16 | 上海交通大学 | The layer-by-layer neural networks pruning method and system inferred based on variation |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8666148B2 (en) * | 2010-06-03 | 2014-03-04 | Adobe Systems Incorporated | Image adjustment |
-
2019
- 2019-09-11 CN CN201910858834.9A patent/CN110766044B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018184222A1 (en) * | 2017-04-07 | 2018-10-11 | Intel Corporation | Methods and systems using improved training and learning for deep neural networks |
CN108108806A (en) * | 2017-12-14 | 2018-06-01 | 西北工业大学 | Convolutional neural networks initial method based on the extraction of pre-training model filter |
CN110020718A (en) * | 2019-03-14 | 2019-07-16 | 上海交通大学 | The layer-by-layer neural networks pruning method and system inferred based on variation |
Also Published As
Publication number | Publication date |
---|---|
CN110766044A (en) | 2020-02-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110766044B (en) | Neural network training method based on Gaussian process prior guidance | |
CN108536784B (en) | Comment information sentiment analysis method and device, computer storage medium and server | |
CN111126488A (en) | Image identification method based on double attention | |
CN113326731A (en) | Cross-domain pedestrian re-identification algorithm based on momentum network guidance | |
CN109508686B (en) | Human behavior recognition method based on hierarchical feature subspace learning | |
CN113128671B (en) | Service demand dynamic prediction method and system based on multi-mode machine learning | |
CN110033089B (en) | Method and system for optimizing parameters of handwritten digital image recognition deep neural network based on distributed estimation algorithm | |
CN112740200B (en) | Systems and methods for end-to-end deep reinforcement learning based on coreference resolution | |
CN114936639A (en) | Progressive confrontation training method and device | |
CN112710310A (en) | Visual language indoor navigation method, system, terminal and application | |
CN113902129A (en) | Multi-mode unified intelligent learning diagnosis modeling method, system, medium and terminal | |
CN113011532A (en) | Classification model training method and device, computing equipment and storage medium | |
CN107240100B (en) | Image segmentation method and system based on genetic algorithm | |
CN114444727B (en) | Living body detection method and device, electronic model and storage medium | |
CN111161238A (en) | Image quality evaluation method and device, electronic device, and storage medium | |
Bai et al. | A unified deep learning model for protein structure prediction | |
CN110674860A (en) | Feature selection method based on neighborhood search strategy, storage medium and terminal | |
Zhu et al. | Fast Adaptive Character Animation Synthesis Based on Greedy Algorithm | |
CN115329863A (en) | Novel linear rectification gradient balance loss function classification method and system | |
Calderhead et al. | Sparse approximate manifolds for differential geometric mcmc | |
CN114936890A (en) | Counter-fact fairness recommendation method based on inverse tendency weighting method | |
CN114154582A (en) | Deep reinforcement learning method based on environment dynamic decomposition model | |
CN111860556A (en) | Model processing method and device and storage medium | |
Pour et al. | Optimal Bayesian feature selection with missing data | |
WO2021215261A1 (en) | Information processing method, information processing device, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |