CN113052261A - Image classification loss function design method based on cosine space optimization - Google Patents
Image classification loss function design method based on cosine space optimization Download PDFInfo
- Publication number
- CN113052261A CN113052261A CN202110434753.3A CN202110434753A CN113052261A CN 113052261 A CN113052261 A CN 113052261A CN 202110434753 A CN202110434753 A CN 202110434753A CN 113052261 A CN113052261 A CN 113052261A
- Authority
- CN
- China
- Prior art keywords
- loss function
- class
- expressed
- batch
- loss
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 26
- 238000005457 optimization Methods 0.000 title claims abstract description 14
- 238000013461 design Methods 0.000 title claims description 4
- 238000012549 training Methods 0.000 claims abstract description 34
- 239000013598 vector Substances 0.000 claims abstract description 9
- 230000006870 function Effects 0.000 claims description 110
- 238000013136 deep learning model Methods 0.000 claims description 13
- 238000005056 compaction Methods 0.000 claims description 3
- 239000000654 additive Substances 0.000 abstract 1
- 230000000996 additive effect Effects 0.000 abstract 1
- 230000001276 controlling effect Effects 0.000 abstract 1
- 230000001105 regulatory effect Effects 0.000 abstract 1
- 238000007635 classification algorithm Methods 0.000 description 5
- 238000004422 calculation algorithm Methods 0.000 description 4
- 101100153586 Caenorhabditis elegans top-1 gene Proteins 0.000 description 1
- 101100370075 Mus musculus Top1 gene Proteins 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Molecular Biology (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a method for designing an image classification loss function based on cosine space optimization, which provides a loss function capable of actively regulating and controlling the intra-class distance and the inter-class distance of image classification simultaneously based on common Additive Margin Softmax optimization. According to the method, an AM-Softmax loss function is adopted in the first half stage of model training, the inter-class distance is pulled, the intra-class centers which can be dynamically adjusted along with a training batch are added in the second half stage of the training, the feature vectors of objects in the same class are further compacted, and the cosine distances between the feature vectors of objects in different classes are pulled, so that the model can be converged more quickly, similar classes can be fully distinguished, and the performance of the model can be further improved.
Description
Technical Field
The invention relates to the field of computer vision and artificial intelligence, in particular to a method for designing an image classification loss function based on cosine space optimization.
Background
The target classification algorithm is a basic and important research field in computer vision, and the current field of target classification algorithms can solve most of simple classification problems along with the development of deep learning technology, namely, a series of CNN models such as ResNet, GoogleNet and EfficientNet emerge after AlexNet, the performance of ImageNet data sets is continuously refreshed, and the Top-1 precision of the current optimal image classification algorithm on the ImageNet data sets reaches 84.4% along with the adoption of more complex network structures and the introduction of deep residual error connection, however, the algorithm models usually have huge parameter quantity and calculation complexity. For the edge mobile scene needing to deploy the algorithm, due to the problems of limited memory and limited computing power, a large-scale network is difficult to use, and therefore the demand of a light intelligent network is gradually increased. In order to improve the precision of the lightweight image classification algorithm, the precision of the network can be effectively improved by optimizing the loss function under the condition of not increasing the complexity of the model and the data volume, and the method is an effective scheme for solving the problem of low precision of the lightweight image classification network.
The loss function commonly used in the field of image classification is a sigmoid cross entropy loss function and a softmax cross entropy loss function, the supervision capability of the loss function is limited, the classification result of an object with larger difference can only be pulled open in a Euclidean space, and the classification capability is weaker for the class with larger similarity, so that the loss function in the field of face recognition is introduced, and the image classification is further optimized to achieve the better effect of the image classification, and the loss function has higher practical value.
Disclosure of Invention
In view of the above, the present invention provides a method for designing an image classification loss function based on cosine space optimization, where the loss function designed by the method can improve the precision of a classification algorithm without increasing the number of network parameters and the amount of training data, and is suitable for a light intelligent network.
In order to achieve the purpose, the invention adopts the following technical scheme:
a design method of an image classification loss function based on cosine space optimization comprises the following steps:
step S1, acquiring a data set, setting hyper-parameters and initializing a deep learning model;
step S2, carrying out multi-batch iterative training on the deep learning model, and sequentially executing steps S21-S23 in each iterative batch;
step S21, calculating the intra-class center of each class of object in the current iteration batch according to the feature vector obtained by the deep learning model in the forward propagation process, and cumulatively updating the intra-class center;
step S22, calculating a cross entropy loss function value and an inter-class loss function value of the current iteration batch;
step S23, judging whether the current iteration batch reaches the preset batch number N;
if not, calculating a first total loss function of the current iteration batch, performing gradient back propagation on the first total loss function, updating model parameters, and returning to the step S2 for a new round of training;
if yes, calculating an intra-class loss function value of the current iteration batch, calculating a second total loss function value by combining the first total loss function and the intra-class loss function value, performing gradient back propagation on the second total loss function value, updating model parameters, and entering step S3;
step S3, judging whether the deep learning model converges,
if not, returning to the step S2 to repeat the iterative training until the model converges;
if the convergence is reached, the model is output.
Further, in the step S1, the hyper-parameter includes: a weighting coefficient α, a weighting coefficient β, a weighting coefficient γ, and a compaction coefficient ∈ and satisfies: gamma > alpha > beta;
the hyper-parameter further comprises the number of batches N in step 23, the N e (0, Epoch)end) Epoch in the formulaendRepresenting the last training batch.
Further, the expression of the first total loss function is:
Loss=αLoss1+γLosscross-entropy (1)
in the formula (1), α and γ are expressed as weighting coefficients, Loss1Expressed as the inter-class loss function; losscross-entropyExpressed as the cross entropy loss function.
Further, the expression of the inter-class loss function is:
in formula (2), n is the number of samples in a batch, s is the scaling factor, and m is the distance of decision boundaries of different classes in cosine space, cosExpressed as i samples in their corresponding category yiC is expressed as the total number of classes, cos θjRepresenting the projection of the i sample on the other class j.
Further, the expression of the second total loss function is:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy (3)
in the formula (3), α, β and γ are all expressed as weighting coefficients, Loss1Expressed as the Loss function between classes, Losscross-entropyExpressed as the cross entropy Loss function, Loss2Expressed as the intra-class loss function, epsilon is a compact coefficient of a determined class in a cosine space, Truc (x) is expressed as a piecewise function, and the expression is as follows:
further, the expression of the intra-class loss function is as follows:
in the formula (4), CbExpressed as the number of categories in a batch, niExpressed as the number of samples of the ith category in a batch,represented as the projection of the jth sample in the ith class on its corresponding class i, ciDenoted as the intra-class center.
The invention has the beneficial effects that:
the invention adopts the AM-Softmax loss function in the first half stage of model training, pulls open the inter-class distance, adds the intra-class center which can be dynamically adjusted along with the training batch in the second half stage of the training, further compacts the characteristic vector of the object in the same class, and simultaneously pulls open the cosine distance between the characteristic vectors of the objects in different classes, thereby leading the model to be more quickly converged, fully distinguishing similar classes and further improving the performance of the model.
Drawings
Fig. 1 is a schematic flow chart of a method for designing an image classification loss function based on cosine space optimization according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1
The implementation provides an image classification loss function based on cosine space optimization, wherein the loss function consists of three parts, and a specific expression is shown as the following formula:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy
in the formula, alpha is the weighting coefficient of the loss function between classes, beta is the weighting coefficient of the loss function in the classes, gamma is the weighting coefficient of the cross entropy loss function, and the proportion of gamma is more than alpha and more than beta is adopted; loss1Expressed as an inter-class Loss function, Losscross-entropyExpressed as cross-entropy Loss function, Loss2Expressed as an intra-class loss function;
the hyper-parameter beta is a piecewise function, when the batch of the algorithm model training is smaller than the hyper-parameter N, the value of beta is 0, and when the training batch exceeds N, the value of beta is larger than 0.
And the value of the super-parameter N is determined according to the convergence condition of the model, and after the model has converged, beta is started, namely the intra-class loss function is used for further optimizing the network, so that the intra-class loss function does not influence the convergence of the network when the intra-class center is unstable in the early stage of model training.
Specifically, in this embodiment, the expression of the truc (x) function is:
the function is mainly used for limiting the influence of the intra-class loss function on the overall loss function, the compaction of the intra-class loss function is controlled through the hyper-parameter epsilon, and the problem of model overfitting caused by the fact that the intra-class loss function is too compact is solved, so that the generalization capability of the model can be effectively improved, overfitting of the model to a training set picture is reduced, and the robustness of an algorithm model is improved.
Specifically, in this embodiment, the Loss is described abovecross-entropyIs a common cross entropy loss function and is a main function for calculating loss.
Specifically, in this embodiment, the Loss is described above1Is an inter-class loss function, namely an AM-Softmax function, and the expression of the inter-class loss function is as follows:
where n is the number of samples in a batch, the hyper-parameter is the scaling factor, and the hyper-parameter m is the distance of decision boundaries of different classes in cosine space, cosFor i samples in their corresponding category yiC is the total number of categories. The loss function is used as auxiliary supervision, the decision boundaries of different types of samples are guided to be separated as much as possible in a cosine space, and the classification precision is effectively improved.
Specifically, in this embodiment, the Loss function Loss in class mentioned above2The method is mainly used for actively constraining the distribution of the features of the objects in the same category in the cosine space, and compared with the original AM-Softmax loss function, the method can help the features of the objects in the same category to be effectively distributed in the decision boundary of the category more effectively, so that the network model is further converged.
More specifically, in the present embodiment, the Loss-in-class function Loss is described above2The expression of (a) is:
in the formula, CbIs the number of categories, n, within a batchiThe number of samples of the ith category in a batch,for the projection of the jth sample in the ith class on its corresponding class i, ciIs the intra-class center of the ith category.
It should be noted that the intra-class loss function adopted in this implementation requires that the network continuously update the intra-class center c in the cosine space with the training batch in the training processiI.e. cumulatively iterates over all training samples of the previous epoch. It is used as its computing Loss only when the next epoch is greater than N, i.e. β > 02Is located in the center of the class.
Example 2
The embodiment provides a method for designing an image classification loss function based on cosine space optimization, which comprises the following steps:
step S1, acquiring a data set, setting hyper-parameters and initializing a deep learning model;
specifically, in this embodiment, the above-mentioned hyper-parameters include: the weighting coefficient alpha, the weighting coefficient beta, the weighting coefficient gamma, the compact coefficient epsilon and the batch number N of the iterative training satisfy the following conditions: gamma > alpha > beta.
S2, inputting the acquired data set into the initialized deep learning model, performing multi-batch iterative training on the deep learning model, and sequentially executing steps S21-S23 in each iterative batch;
step S21, calculating the in-class center c of each class of object in the current iteration batch according to the feature vector obtained by the deep learning model in the forward propagation processiAnd cumulatively updating the class center ci;
Step S22, calculating a cross entropy loss function value and an inter-class loss function value of the current iteration batch;
specifically, the above-mentioned inter-class loss function, i.e. AM-Softmax function, has the expression:
where n is the number of samples in a batch, the hyper-parameter is the scaling factor, and the hyper-parameter m is the distance of decision boundaries of different classes in cosine space, cosFor i samples in their corresponding category yiC is the total number of categories. The loss function is used as auxiliary supervision, the decision boundaries of different types of samples are guided to be separated as much as possible in a cosine space, the classification precision is effectively improved, and cos theta isjRepresenting the projection of the i sample on the other class j.
Step (ii) ofS23, judging whether the current iteration batch reaches the preset number N, wherein N belongs to (0, Epoch)end) Epoch in the formulaendRepresenting the last training batch, the number N being the number N of iterative training batches determined in step S1;
if not, calculating a first total loss function of the current iteration batch, performing gradient back propagation on the first total loss function, updating model parameters, and returning to the step S2 for a new round of training;
if yes, calculating an intra-class loss function value of the current iteration batch, calculating a second total loss function value by combining the first total loss function and the intra-class loss function value, performing gradient back propagation on the second total loss function value, updating model parameters, and entering step S3;
specifically, in this embodiment, the expression of the first total loss function is:
Loss=αLoss1+γLosscross-entropy
in the formula, α and γ are expressed as weighting coefficients, Loss1Expressed as the above-mentioned inter-class loss function; losscross-entropyExpressed as the cross entropy loss function described above.
The expression of the loss-in-class function is:
in the formula, CbExpressed as the number of categories in a batch, niExpressed as the number of samples of the ith category in a batch,represented as the projection of the jth sample in the ith class on its corresponding class i, ciDenoted as the intra-class center.
The expression of the second total loss function is:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy
in the formula, alpha, beta and gamma are all expressed as weighting coefficients, Loss1Expressed as the Loss function between classes, Losscross-entropyExpressed as the cross entropy Loss function, Loss2The method is expressed as an intra-class loss function, epsilon is a compact coefficient of a determined class in a cosine space, Truc (x) is expressed as a piecewise function, and the expression is as follows:
step S3, judging whether the deep learning model converges,
if not, returning to the step S2 to repeat the iterative training until the model converges;
if the convergence is reached, the model is output.
It should be noted that the Loss-in-class function Loss used in this embodiment2The network is required to continuously update the class center c in the cosine space with the training batch in the training processiI.e. cumulatively iterates over all training samples of the previous epoch. It is used as its computing Loss only when the next epoch is greater than N, i.e. β > 02Is located in the center of the class.
In summary, the method for designing the image classification loss function based on the cosine space optimization comprises two stages, wherein a hyper-parameter N epsilon (0, Epoch) is usedend) Is cut off in the form of an EpochendRepresenting the last training batch. Wherein the first stage uses the cross entropy Loss function Losscross-entropyLoss function Loss between classes1Calculating Loss, entering a second stage when the training batch exceeds N, and increasing Loss function Loss in class2。
The AM-Softmax loss function is adopted in the first stage, the inter-class distance is pulled, the intra-class center which can be dynamically adjusted along with the training batch is added in the second stage, the feature vectors of the objects in the same class are further compacted, and meanwhile, the cosine distances between the feature vectors of the objects in different classes are pulled, so that the model can be converged more quickly, the similar classes can be fully distinguished, and the performance of the model can be further improved.
The invention is not described in detail, but is well known to those skilled in the art.
The foregoing detailed description of the preferred embodiments of the invention has been presented. It should be understood that numerous modifications and variations could be devised by those skilled in the art in light of the present teachings without departing from the inventive concepts. Therefore, the technical solutions available to those skilled in the art through logic analysis, reasoning and limited experiments based on the prior art according to the concept of the present invention should be within the scope of protection defined by the claims.
Claims (6)
1. A design method of an image classification loss function based on cosine space optimization is characterized by comprising the following steps:
step S1, acquiring a data set, setting hyper-parameters and initializing a deep learning model;
step S2, carrying out multi-batch iterative training on the deep learning model, and sequentially executing steps S21-S23 in each iterative batch;
step S21, calculating the intra-class center of each class of object in the current iteration batch according to the feature vector obtained by the deep learning model in the forward propagation process, and cumulatively updating the intra-class center;
step S22, calculating a cross entropy loss function value and an inter-class loss function value of the current iteration batch;
step S23, judging whether the current iteration batch reaches the preset batch number N;
if not, calculating a first total loss function of the current iteration batch, performing gradient back propagation on the first total loss function, updating model parameters, and returning to the step S2 for a new round of training;
if yes, calculating an intra-class loss function value of the current iteration batch, calculating a second total loss function value by combining the first total loss function and the intra-class loss function value, performing gradient back propagation on the second total loss function value, updating model parameters, and entering step S3;
step S3, judging whether the deep learning model converges,
if not, returning to the step S2 to repeat the iterative training until the model converges;
if the convergence is reached, the model is output.
2. The method for designing an image classification loss function based on cosine space optimization as claimed in claim 1, wherein in the step S1, the hyper-parameter comprises: a weighting coefficient α, a weighting coefficient β, a weighting coefficient γ, and a compaction coefficient ∈ and satisfies: gamma > alpha > beta;
the hyper-parameter further comprises the number of batches N in step 23, the N e (0, Epoch)end) Epoch in the formulaendRepresenting the last training batch.
3. The method for designing an image classification loss function based on cosine space optimization according to claim 2, wherein the expression of the first total loss function is as follows:
Loss=αLoss1+γLosscross-entropy (1)
in the formula (1), α and γ are expressed as weighting coefficients, Loss1Expressed as the inter-class loss function; losscross-entropyExpressed as the cross entropy loss function.
4. The method according to claim 3, wherein the inter-class loss function is expressed as:
in equation (2), n is expressed as the number of samples in a batch, s is expressed as a scaling factor, m is expressed as the distance of decision boundaries of different classes in cosine space,expressed as i samples in their corresponding category yiC is expressed as the total number of classes, cos θjRepresenting the projection of the i sample on the other class j.
5. The method as claimed in claim 4, wherein the second total loss function is expressed as:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy (3)
in the formula (3), α, β and γ are all expressed as weighting coefficients, Loss1Expressed as the Loss function between classes, Losscross-entropyExpressed as the cross entropy Loss function, Loss2Expressed as the intra-class loss function, epsilon is a compact coefficient of a determined class in a cosine space, Truc (x) is expressed as a piecewise function, and the expression is as follows:
6. the method according to claim 5, wherein the intra-class loss function is expressed as:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110434753.3A CN113052261B (en) | 2021-04-22 | 2021-04-22 | Design method of image classification loss function based on cosine space optimization |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110434753.3A CN113052261B (en) | 2021-04-22 | 2021-04-22 | Design method of image classification loss function based on cosine space optimization |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113052261A true CN113052261A (en) | 2021-06-29 |
CN113052261B CN113052261B (en) | 2024-05-31 |
Family
ID=76520142
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110434753.3A Active CN113052261B (en) | 2021-04-22 | 2021-04-22 | Design method of image classification loss function based on cosine space optimization |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113052261B (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113705647A (en) * | 2021-08-19 | 2021-11-26 | 电子科技大学 | Dynamic interval-based dual semantic feature extraction method |
CN113763501A (en) * | 2021-09-08 | 2021-12-07 | 上海壁仞智能科技有限公司 | Iteration method of image reconstruction model and image reconstruction method |
CN116310648A (en) * | 2023-03-23 | 2023-06-23 | 北京的卢铭视科技有限公司 | Model training method, face recognition method, electronic device and storage medium |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109903774A (en) * | 2019-04-12 | 2019-06-18 | 南京大学 | A kind of method for recognizing sound-groove based on angle separation loss function |
CN110222841A (en) * | 2019-06-17 | 2019-09-10 | 苏州思必驰信息科技有限公司 | Neural network training method and device based on spacing loss function |
CN112613552A (en) * | 2020-12-18 | 2021-04-06 | 北京工业大学 | Convolutional neural network emotion image classification method combining emotion category attention loss |
-
2021
- 2021-04-22 CN CN202110434753.3A patent/CN113052261B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109903774A (en) * | 2019-04-12 | 2019-06-18 | 南京大学 | A kind of method for recognizing sound-groove based on angle separation loss function |
CN110222841A (en) * | 2019-06-17 | 2019-09-10 | 苏州思必驰信息科技有限公司 | Neural network training method and device based on spacing loss function |
CN112613552A (en) * | 2020-12-18 | 2021-04-06 | 北京工业大学 | Convolutional neural network emotion image classification method combining emotion category attention loss |
Non-Patent Citations (2)
Title |
---|
SUN YIFAN 等: "Circle Loss: A Unified Perspective of Pair Similarity Optimization", 《ARXIV》, 15 June 2020 (2020-06-15), pages 1 - 10 * |
张强 等: "CS-Softmax:一种基于余弦相似性的Softmax损失函数", 《计算机研究与发展》, vol. 59, no. 4, 16 April 2021 (2021-04-16), pages 936 - 949 * |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113705647A (en) * | 2021-08-19 | 2021-11-26 | 电子科技大学 | Dynamic interval-based dual semantic feature extraction method |
CN113705647B (en) * | 2021-08-19 | 2023-04-28 | 电子科技大学 | Dual semantic feature extraction method based on dynamic interval |
CN113763501A (en) * | 2021-09-08 | 2021-12-07 | 上海壁仞智能科技有限公司 | Iteration method of image reconstruction model and image reconstruction method |
CN113763501B (en) * | 2021-09-08 | 2024-02-27 | 上海壁仞智能科技有限公司 | Iterative method of image reconstruction model and image reconstruction method |
CN116310648A (en) * | 2023-03-23 | 2023-06-23 | 北京的卢铭视科技有限公司 | Model training method, face recognition method, electronic device and storage medium |
CN116310648B (en) * | 2023-03-23 | 2023-12-12 | 北京的卢铭视科技有限公司 | Model training method, face recognition method, electronic device and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN113052261B (en) | 2024-05-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113052261A (en) | Image classification loss function design method based on cosine space optimization | |
CN111461322B (en) | Deep neural network model compression method | |
WO2022160771A1 (en) | Method for classifying hyperspectral images on basis of adaptive multi-scale feature extraction model | |
CN111079781B (en) | Lightweight convolutional neural network image recognition method based on low rank and sparse decomposition | |
CN106250939B (en) | Handwritten character recognition method based on FPGA + ARM multilayer convolutional neural network | |
CN108985457B (en) | Deep neural network structure design method inspired by optimization algorithm | |
CN110046252B (en) | Medical text grading method based on attention mechanism neural network and knowledge graph | |
CN114841257B (en) | Small sample target detection method based on self-supervision comparison constraint | |
CN110097060B (en) | Open set identification method for trunk image | |
CN112766399B (en) | Self-adaptive neural network training method for image recognition | |
CN111476346A (en) | Deep learning network architecture based on Newton conjugate gradient method | |
CN109190666B (en) | Flower image classification method based on improved deep neural network | |
CN115631393A (en) | Image processing method based on characteristic pyramid and knowledge guided knowledge distillation | |
CN113837376A (en) | Neural network pruning method based on dynamic coding convolution kernel fusion | |
CN112597979B (en) | Face recognition method for updating cosine included angle loss function parameters in real time | |
Yang et al. | Triple-GAN with variable fractional order gradient descent method and mish activation function | |
CN111310807B (en) | Feature subspace and affinity matrix joint learning method based on heterogeneous feature joint self-expression | |
CN110288002B (en) | Image classification method based on sparse orthogonal neural network | |
CN115601578A (en) | Multi-view clustering method and system based on self-walking learning and view weighting | |
An | Xception network for weather image recognition based on transfer learning | |
CN114202694A (en) | Small sample remote sensing scene image classification method based on manifold mixed interpolation and contrast learning | |
CN111178174B (en) | Urine formed component image identification method based on deep convolutional neural network | |
CN113112397A (en) | Image style migration method based on style and content decoupling | |
TW202232431A (en) | Training method for adaptively adjusting mini-batch size of neural network wherein the mini-batch size can be gradually adjusted in real time according to the current situation during training so as to enable the neural network model to obtain better accuracy | |
Zhao et al. | An efficient and flexible automatic search algorithm for convolution network architectures |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |