CN113052261A - Image classification loss function design method based on cosine space optimization - Google Patents

Image classification loss function design method based on cosine space optimization Download PDF

Info

Publication number
CN113052261A
CN113052261A CN202110434753.3A CN202110434753A CN113052261A CN 113052261 A CN113052261 A CN 113052261A CN 202110434753 A CN202110434753 A CN 202110434753A CN 113052261 A CN113052261 A CN 113052261A
Authority
CN
China
Prior art keywords
loss function
class
expressed
batch
loss
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110434753.3A
Other languages
Chinese (zh)
Other versions
CN113052261B (en
Inventor
李晨
许虞俊
孙翔
曹悦欣
杜文娟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Southeast University
Original Assignee
Southeast University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Southeast University filed Critical Southeast University
Priority to CN202110434753.3A priority Critical patent/CN113052261B/en
Publication of CN113052261A publication Critical patent/CN113052261A/en
Application granted granted Critical
Publication of CN113052261B publication Critical patent/CN113052261B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method for designing an image classification loss function based on cosine space optimization, which provides a loss function capable of actively regulating and controlling the intra-class distance and the inter-class distance of image classification simultaneously based on common Additive Margin Softmax optimization. According to the method, an AM-Softmax loss function is adopted in the first half stage of model training, the inter-class distance is pulled, the intra-class centers which can be dynamically adjusted along with a training batch are added in the second half stage of the training, the feature vectors of objects in the same class are further compacted, and the cosine distances between the feature vectors of objects in different classes are pulled, so that the model can be converged more quickly, similar classes can be fully distinguished, and the performance of the model can be further improved.

Description

Image classification loss function design method based on cosine space optimization
Technical Field
The invention relates to the field of computer vision and artificial intelligence, in particular to a method for designing an image classification loss function based on cosine space optimization.
Background
The target classification algorithm is a basic and important research field in computer vision, and the current field of target classification algorithms can solve most of simple classification problems along with the development of deep learning technology, namely, a series of CNN models such as ResNet, GoogleNet and EfficientNet emerge after AlexNet, the performance of ImageNet data sets is continuously refreshed, and the Top-1 precision of the current optimal image classification algorithm on the ImageNet data sets reaches 84.4% along with the adoption of more complex network structures and the introduction of deep residual error connection, however, the algorithm models usually have huge parameter quantity and calculation complexity. For the edge mobile scene needing to deploy the algorithm, due to the problems of limited memory and limited computing power, a large-scale network is difficult to use, and therefore the demand of a light intelligent network is gradually increased. In order to improve the precision of the lightweight image classification algorithm, the precision of the network can be effectively improved by optimizing the loss function under the condition of not increasing the complexity of the model and the data volume, and the method is an effective scheme for solving the problem of low precision of the lightweight image classification network.
The loss function commonly used in the field of image classification is a sigmoid cross entropy loss function and a softmax cross entropy loss function, the supervision capability of the loss function is limited, the classification result of an object with larger difference can only be pulled open in a Euclidean space, and the classification capability is weaker for the class with larger similarity, so that the loss function in the field of face recognition is introduced, and the image classification is further optimized to achieve the better effect of the image classification, and the loss function has higher practical value.
Disclosure of Invention
In view of the above, the present invention provides a method for designing an image classification loss function based on cosine space optimization, where the loss function designed by the method can improve the precision of a classification algorithm without increasing the number of network parameters and the amount of training data, and is suitable for a light intelligent network.
In order to achieve the purpose, the invention adopts the following technical scheme:
a design method of an image classification loss function based on cosine space optimization comprises the following steps:
step S1, acquiring a data set, setting hyper-parameters and initializing a deep learning model;
step S2, carrying out multi-batch iterative training on the deep learning model, and sequentially executing steps S21-S23 in each iterative batch;
step S21, calculating the intra-class center of each class of object in the current iteration batch according to the feature vector obtained by the deep learning model in the forward propagation process, and cumulatively updating the intra-class center;
step S22, calculating a cross entropy loss function value and an inter-class loss function value of the current iteration batch;
step S23, judging whether the current iteration batch reaches the preset batch number N;
if not, calculating a first total loss function of the current iteration batch, performing gradient back propagation on the first total loss function, updating model parameters, and returning to the step S2 for a new round of training;
if yes, calculating an intra-class loss function value of the current iteration batch, calculating a second total loss function value by combining the first total loss function and the intra-class loss function value, performing gradient back propagation on the second total loss function value, updating model parameters, and entering step S3;
step S3, judging whether the deep learning model converges,
if not, returning to the step S2 to repeat the iterative training until the model converges;
if the convergence is reached, the model is output.
Further, in the step S1, the hyper-parameter includes: a weighting coefficient α, a weighting coefficient β, a weighting coefficient γ, and a compaction coefficient ∈ and satisfies: gamma > alpha > beta;
the hyper-parameter further comprises the number of batches N in step 23, the N e (0, Epoch)end) Epoch in the formulaendRepresenting the last training batch.
Further, the expression of the first total loss function is:
Loss=αLoss1+γLosscross-entropy (1)
in the formula (1), α and γ are expressed as weighting coefficients, Loss1Expressed as the inter-class loss function; losscross-entropyExpressed as the cross entropy loss function.
Further, the expression of the inter-class loss function is:
Figure BDA0003032453970000021
in formula (2), n is the number of samples in a batch, s is the scaling factor, and m is the distance of decision boundaries of different classes in cosine space, cos
Figure BDA0003032453970000022
Expressed as i samples in their corresponding category yiC is expressed as the total number of classes, cos θjRepresenting the projection of the i sample on the other class j.
Further, the expression of the second total loss function is:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy (3)
in the formula (3), α, β and γ are all expressed as weighting coefficients, Loss1Expressed as the Loss function between classes, Losscross-entropyExpressed as the cross entropy Loss function, Loss2Expressed as the intra-class loss function, epsilon is a compact coefficient of a determined class in a cosine space, Truc (x) is expressed as a piecewise function, and the expression is as follows:
Figure BDA0003032453970000031
further, the expression of the intra-class loss function is as follows:
Figure BDA0003032453970000032
in the formula (4), CbExpressed as the number of categories in a batch, niExpressed as the number of samples of the ith category in a batch,
Figure BDA0003032453970000033
represented as the projection of the jth sample in the ith class on its corresponding class i, ciDenoted as the intra-class center.
The invention has the beneficial effects that:
the invention adopts the AM-Softmax loss function in the first half stage of model training, pulls open the inter-class distance, adds the intra-class center which can be dynamically adjusted along with the training batch in the second half stage of the training, further compacts the characteristic vector of the object in the same class, and simultaneously pulls open the cosine distance between the characteristic vectors of the objects in different classes, thereby leading the model to be more quickly converged, fully distinguishing similar classes and further improving the performance of the model.
Drawings
Fig. 1 is a schematic flow chart of a method for designing an image classification loss function based on cosine space optimization according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1
The implementation provides an image classification loss function based on cosine space optimization, wherein the loss function consists of three parts, and a specific expression is shown as the following formula:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy
in the formula, alpha is the weighting coefficient of the loss function between classes, beta is the weighting coefficient of the loss function in the classes, gamma is the weighting coefficient of the cross entropy loss function, and the proportion of gamma is more than alpha and more than beta is adopted; loss1Expressed as an inter-class Loss function, Losscross-entropyExpressed as cross-entropy Loss function, Loss2Expressed as an intra-class loss function;
the hyper-parameter beta is a piecewise function, when the batch of the algorithm model training is smaller than the hyper-parameter N, the value of beta is 0, and when the training batch exceeds N, the value of beta is larger than 0.
And the value of the super-parameter N is determined according to the convergence condition of the model, and after the model has converged, beta is started, namely the intra-class loss function is used for further optimizing the network, so that the intra-class loss function does not influence the convergence of the network when the intra-class center is unstable in the early stage of model training.
Specifically, in this embodiment, the expression of the truc (x) function is:
Figure BDA0003032453970000041
the function is mainly used for limiting the influence of the intra-class loss function on the overall loss function, the compaction of the intra-class loss function is controlled through the hyper-parameter epsilon, and the problem of model overfitting caused by the fact that the intra-class loss function is too compact is solved, so that the generalization capability of the model can be effectively improved, overfitting of the model to a training set picture is reduced, and the robustness of an algorithm model is improved.
Specifically, in this embodiment, the Loss is described abovecross-entropyIs a common cross entropy loss function and is a main function for calculating loss.
Specifically, in this embodiment, the Loss is described above1Is an inter-class loss function, namely an AM-Softmax function, and the expression of the inter-class loss function is as follows:
Figure BDA0003032453970000042
where n is the number of samples in a batch, the hyper-parameter is the scaling factor, and the hyper-parameter m is the distance of decision boundaries of different classes in cosine space, cos
Figure BDA0003032453970000045
For i samples in their corresponding category yiC is the total number of categories. The loss function is used as auxiliary supervision, the decision boundaries of different types of samples are guided to be separated as much as possible in a cosine space, and the classification precision is effectively improved.
Specifically, in this embodiment, the Loss function Loss in class mentioned above2The method is mainly used for actively constraining the distribution of the features of the objects in the same category in the cosine space, and compared with the original AM-Softmax loss function, the method can help the features of the objects in the same category to be effectively distributed in the decision boundary of the category more effectively, so that the network model is further converged.
More specifically, in the present embodiment, the Loss-in-class function Loss is described above2The expression of (a) is:
Figure BDA0003032453970000043
in the formula, CbIs the number of categories, n, within a batchiThe number of samples of the ith category in a batch,
Figure BDA0003032453970000044
for the projection of the jth sample in the ith class on its corresponding class i, ciIs the intra-class center of the ith category.
It should be noted that the intra-class loss function adopted in this implementation requires that the network continuously update the intra-class center c in the cosine space with the training batch in the training processiI.e. cumulatively iterates over all training samples of the previous epoch. It is used as its computing Loss only when the next epoch is greater than N, i.e. β > 02Is located in the center of the class.
Example 2
The embodiment provides a method for designing an image classification loss function based on cosine space optimization, which comprises the following steps:
step S1, acquiring a data set, setting hyper-parameters and initializing a deep learning model;
specifically, in this embodiment, the above-mentioned hyper-parameters include: the weighting coefficient alpha, the weighting coefficient beta, the weighting coefficient gamma, the compact coefficient epsilon and the batch number N of the iterative training satisfy the following conditions: gamma > alpha > beta.
S2, inputting the acquired data set into the initialized deep learning model, performing multi-batch iterative training on the deep learning model, and sequentially executing steps S21-S23 in each iterative batch;
step S21, calculating the in-class center c of each class of object in the current iteration batch according to the feature vector obtained by the deep learning model in the forward propagation processiAnd cumulatively updating the class center ci
Step S22, calculating a cross entropy loss function value and an inter-class loss function value of the current iteration batch;
specifically, the above-mentioned inter-class loss function, i.e. AM-Softmax function, has the expression:
Figure BDA0003032453970000051
where n is the number of samples in a batch, the hyper-parameter is the scaling factor, and the hyper-parameter m is the distance of decision boundaries of different classes in cosine space, cos
Figure BDA0003032453970000053
For i samples in their corresponding category yiC is the total number of categories. The loss function is used as auxiliary supervision, the decision boundaries of different types of samples are guided to be separated as much as possible in a cosine space, the classification precision is effectively improved, and cos theta isjRepresenting the projection of the i sample on the other class j.
Step (ii) ofS23, judging whether the current iteration batch reaches the preset number N, wherein N belongs to (0, Epoch)end) Epoch in the formulaendRepresenting the last training batch, the number N being the number N of iterative training batches determined in step S1;
if not, calculating a first total loss function of the current iteration batch, performing gradient back propagation on the first total loss function, updating model parameters, and returning to the step S2 for a new round of training;
if yes, calculating an intra-class loss function value of the current iteration batch, calculating a second total loss function value by combining the first total loss function and the intra-class loss function value, performing gradient back propagation on the second total loss function value, updating model parameters, and entering step S3;
specifically, in this embodiment, the expression of the first total loss function is:
Loss=αLoss1+γLosscross-entropy
in the formula, α and γ are expressed as weighting coefficients, Loss1Expressed as the above-mentioned inter-class loss function; losscross-entropyExpressed as the cross entropy loss function described above.
The expression of the loss-in-class function is:
Figure BDA0003032453970000052
in the formula, CbExpressed as the number of categories in a batch, niExpressed as the number of samples of the ith category in a batch,
Figure BDA0003032453970000061
represented as the projection of the jth sample in the ith class on its corresponding class i, ciDenoted as the intra-class center.
The expression of the second total loss function is:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy
in the formula, alpha, beta and gamma are all expressed as weighting coefficients, Loss1Expressed as the Loss function between classes, Losscross-entropyExpressed as the cross entropy Loss function, Loss2The method is expressed as an intra-class loss function, epsilon is a compact coefficient of a determined class in a cosine space, Truc (x) is expressed as a piecewise function, and the expression is as follows:
Figure BDA0003032453970000062
step S3, judging whether the deep learning model converges,
if not, returning to the step S2 to repeat the iterative training until the model converges;
if the convergence is reached, the model is output.
It should be noted that the Loss-in-class function Loss used in this embodiment2The network is required to continuously update the class center c in the cosine space with the training batch in the training processiI.e. cumulatively iterates over all training samples of the previous epoch. It is used as its computing Loss only when the next epoch is greater than N, i.e. β > 02Is located in the center of the class.
In summary, the method for designing the image classification loss function based on the cosine space optimization comprises two stages, wherein a hyper-parameter N epsilon (0, Epoch) is usedend) Is cut off in the form of an EpochendRepresenting the last training batch. Wherein the first stage uses the cross entropy Loss function Losscross-entropyLoss function Loss between classes1Calculating Loss, entering a second stage when the training batch exceeds N, and increasing Loss function Loss in class2
The AM-Softmax loss function is adopted in the first stage, the inter-class distance is pulled, the intra-class center which can be dynamically adjusted along with the training batch is added in the second stage, the feature vectors of the objects in the same class are further compacted, and meanwhile, the cosine distances between the feature vectors of the objects in different classes are pulled, so that the model can be converged more quickly, the similar classes can be fully distinguished, and the performance of the model can be further improved.
The invention is not described in detail, but is well known to those skilled in the art.
The foregoing detailed description of the preferred embodiments of the invention has been presented. It should be understood that numerous modifications and variations could be devised by those skilled in the art in light of the present teachings without departing from the inventive concepts. Therefore, the technical solutions available to those skilled in the art through logic analysis, reasoning and limited experiments based on the prior art according to the concept of the present invention should be within the scope of protection defined by the claims.

Claims (6)

1. A design method of an image classification loss function based on cosine space optimization is characterized by comprising the following steps:
step S1, acquiring a data set, setting hyper-parameters and initializing a deep learning model;
step S2, carrying out multi-batch iterative training on the deep learning model, and sequentially executing steps S21-S23 in each iterative batch;
step S21, calculating the intra-class center of each class of object in the current iteration batch according to the feature vector obtained by the deep learning model in the forward propagation process, and cumulatively updating the intra-class center;
step S22, calculating a cross entropy loss function value and an inter-class loss function value of the current iteration batch;
step S23, judging whether the current iteration batch reaches the preset batch number N;
if not, calculating a first total loss function of the current iteration batch, performing gradient back propagation on the first total loss function, updating model parameters, and returning to the step S2 for a new round of training;
if yes, calculating an intra-class loss function value of the current iteration batch, calculating a second total loss function value by combining the first total loss function and the intra-class loss function value, performing gradient back propagation on the second total loss function value, updating model parameters, and entering step S3;
step S3, judging whether the deep learning model converges,
if not, returning to the step S2 to repeat the iterative training until the model converges;
if the convergence is reached, the model is output.
2. The method for designing an image classification loss function based on cosine space optimization as claimed in claim 1, wherein in the step S1, the hyper-parameter comprises: a weighting coefficient α, a weighting coefficient β, a weighting coefficient γ, and a compaction coefficient ∈ and satisfies: gamma > alpha > beta;
the hyper-parameter further comprises the number of batches N in step 23, the N e (0, Epoch)end) Epoch in the formulaendRepresenting the last training batch.
3. The method for designing an image classification loss function based on cosine space optimization according to claim 2, wherein the expression of the first total loss function is as follows:
Loss=αLoss1+γLosscross-entropy (1)
in the formula (1), α and γ are expressed as weighting coefficients, Loss1Expressed as the inter-class loss function; losscross-entropyExpressed as the cross entropy loss function.
4. The method according to claim 3, wherein the inter-class loss function is expressed as:
Figure FDA0003032453960000021
in equation (2), n is expressed as the number of samples in a batch, s is expressed as a scaling factor, m is expressed as the distance of decision boundaries of different classes in cosine space,
Figure FDA0003032453960000025
expressed as i samples in their corresponding category yiC is expressed as the total number of classes, cos θjRepresenting the projection of the i sample on the other class j.
5. The method as claimed in claim 4, wherein the second total loss function is expressed as:
Loss=αLoss1+β·Truc(Loss2-ε)+γLosscross-entropy (3)
in the formula (3), α, β and γ are all expressed as weighting coefficients, Loss1Expressed as the Loss function between classes, Losscross-entropyExpressed as the cross entropy Loss function, Loss2Expressed as the intra-class loss function, epsilon is a compact coefficient of a determined class in a cosine space, Truc (x) is expressed as a piecewise function, and the expression is as follows:
Figure FDA0003032453960000022
6. the method according to claim 5, wherein the intra-class loss function is expressed as:
Figure FDA0003032453960000023
in the formula (4), CbExpressed as the number of categories in a batch, niExpressed as the number of samples of the ith category in a batch,
Figure FDA0003032453960000024
expressed as the jth sample in the ith class in its correspondenceProjection on class i, ciDenoted as the intra-class center.
CN202110434753.3A 2021-04-22 2021-04-22 Design method of image classification loss function based on cosine space optimization Active CN113052261B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110434753.3A CN113052261B (en) 2021-04-22 2021-04-22 Design method of image classification loss function based on cosine space optimization

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110434753.3A CN113052261B (en) 2021-04-22 2021-04-22 Design method of image classification loss function based on cosine space optimization

Publications (2)

Publication Number Publication Date
CN113052261A true CN113052261A (en) 2021-06-29
CN113052261B CN113052261B (en) 2024-05-31

Family

ID=76520142

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110434753.3A Active CN113052261B (en) 2021-04-22 2021-04-22 Design method of image classification loss function based on cosine space optimization

Country Status (1)

Country Link
CN (1) CN113052261B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113705647A (en) * 2021-08-19 2021-11-26 电子科技大学 Dynamic interval-based dual semantic feature extraction method
CN113763501A (en) * 2021-09-08 2021-12-07 上海壁仞智能科技有限公司 Iteration method of image reconstruction model and image reconstruction method
CN116310648A (en) * 2023-03-23 2023-06-23 北京的卢铭视科技有限公司 Model training method, face recognition method, electronic device and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109903774A (en) * 2019-04-12 2019-06-18 南京大学 A kind of method for recognizing sound-groove based on angle separation loss function
CN110222841A (en) * 2019-06-17 2019-09-10 苏州思必驰信息科技有限公司 Neural network training method and device based on spacing loss function
CN112613552A (en) * 2020-12-18 2021-04-06 北京工业大学 Convolutional neural network emotion image classification method combining emotion category attention loss

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109903774A (en) * 2019-04-12 2019-06-18 南京大学 A kind of method for recognizing sound-groove based on angle separation loss function
CN110222841A (en) * 2019-06-17 2019-09-10 苏州思必驰信息科技有限公司 Neural network training method and device based on spacing loss function
CN112613552A (en) * 2020-12-18 2021-04-06 北京工业大学 Convolutional neural network emotion image classification method combining emotion category attention loss

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
SUN YIFAN 等: "Circle Loss: A Unified Perspective of Pair Similarity Optimization", 《ARXIV》, 15 June 2020 (2020-06-15), pages 1 - 10 *
张强 等: "CS-Softmax:一种基于余弦相似性的Softmax损失函数", 《计算机研究与发展》, vol. 59, no. 4, 16 April 2021 (2021-04-16), pages 936 - 949 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113705647A (en) * 2021-08-19 2021-11-26 电子科技大学 Dynamic interval-based dual semantic feature extraction method
CN113705647B (en) * 2021-08-19 2023-04-28 电子科技大学 Dual semantic feature extraction method based on dynamic interval
CN113763501A (en) * 2021-09-08 2021-12-07 上海壁仞智能科技有限公司 Iteration method of image reconstruction model and image reconstruction method
CN113763501B (en) * 2021-09-08 2024-02-27 上海壁仞智能科技有限公司 Iterative method of image reconstruction model and image reconstruction method
CN116310648A (en) * 2023-03-23 2023-06-23 北京的卢铭视科技有限公司 Model training method, face recognition method, electronic device and storage medium
CN116310648B (en) * 2023-03-23 2023-12-12 北京的卢铭视科技有限公司 Model training method, face recognition method, electronic device and storage medium

Also Published As

Publication number Publication date
CN113052261B (en) 2024-05-31

Similar Documents

Publication Publication Date Title
CN113052261A (en) Image classification loss function design method based on cosine space optimization
CN111461322B (en) Deep neural network model compression method
WO2022160771A1 (en) Method for classifying hyperspectral images on basis of adaptive multi-scale feature extraction model
CN111079781B (en) Lightweight convolutional neural network image recognition method based on low rank and sparse decomposition
CN106250939B (en) Handwritten character recognition method based on FPGA + ARM multilayer convolutional neural network
CN108985457B (en) Deep neural network structure design method inspired by optimization algorithm
CN110046252B (en) Medical text grading method based on attention mechanism neural network and knowledge graph
CN114841257B (en) Small sample target detection method based on self-supervision comparison constraint
CN110097060B (en) Open set identification method for trunk image
CN112766399B (en) Self-adaptive neural network training method for image recognition
CN111476346A (en) Deep learning network architecture based on Newton conjugate gradient method
CN109190666B (en) Flower image classification method based on improved deep neural network
CN115631393A (en) Image processing method based on characteristic pyramid and knowledge guided knowledge distillation
CN113837376A (en) Neural network pruning method based on dynamic coding convolution kernel fusion
CN112597979B (en) Face recognition method for updating cosine included angle loss function parameters in real time
Yang et al. Triple-GAN with variable fractional order gradient descent method and mish activation function
CN111310807B (en) Feature subspace and affinity matrix joint learning method based on heterogeneous feature joint self-expression
CN110288002B (en) Image classification method based on sparse orthogonal neural network
CN115601578A (en) Multi-view clustering method and system based on self-walking learning and view weighting
An Xception network for weather image recognition based on transfer learning
CN114202694A (en) Small sample remote sensing scene image classification method based on manifold mixed interpolation and contrast learning
CN111178174B (en) Urine formed component image identification method based on deep convolutional neural network
CN113112397A (en) Image style migration method based on style and content decoupling
TW202232431A (en) Training method for adaptively adjusting mini-batch size of neural network wherein the mini-batch size can be gradually adjusted in real time according to the current situation during training so as to enable the neural network model to obtain better accuracy
Zhao et al. An efficient and flexible automatic search algorithm for convolution network architectures

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant