CN115797711B - Improved classification method for countermeasure sample based on reconstruction model - Google Patents

Improved classification method for countermeasure sample based on reconstruction model Download PDF

Info

Publication number
CN115797711B
CN115797711B CN202310132811.6A CN202310132811A CN115797711B CN 115797711 B CN115797711 B CN 115797711B CN 202310132811 A CN202310132811 A CN 202310132811A CN 115797711 B CN115797711 B CN 115797711B
Authority
CN
China
Prior art keywords
model
training
classifier
loss function
classification
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202310132811.6A
Other languages
Chinese (zh)
Other versions
CN115797711A (en
Inventor
郭杰龙
魏宪
俞辉
阳帆
李�杰
张剑锋
邵东恒
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Quanzhou Institute of Equipment Manufacturing
Original Assignee
Quanzhou Institute of Equipment Manufacturing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Quanzhou Institute of Equipment Manufacturing filed Critical Quanzhou Institute of Equipment Manufacturing
Priority to CN202310132811.6A priority Critical patent/CN115797711B/en
Publication of CN115797711A publication Critical patent/CN115797711A/en
Application granted granted Critical
Publication of CN115797711B publication Critical patent/CN115797711B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses an improved classification method of an countermeasure sample based on a reconstruction model, which belongs to the field of automatic driving image classification and comprises the following steps: constructing a reconstruction model, and extracting characteristic information of an image dataset based on the reconstruction model; learning an abstract representation from the reconstruction model using a challenge attack, generating a challenge sample based on the generator; constructing a classifier, and training the classifier based on the challenge sample to obtain a trained classifier; and training the countermeasure sample and the trained classifier based on the common loss function in a combined way to obtain a target classification model, and classifying the image data set based on the target classification model to obtain a classification result. The invention can learn the characteristics for assisting classification from the countermeasure sample based on the reconstruction model, thereby improving the model classification precision.

Description

Improved classification method for countermeasure sample based on reconstruction model
Technical Field
The invention belongs to the field of automatic driving image classification, and particularly relates to an improved classification method for an countermeasure sample based on a reconstruction model.
Background
Deep Neural Networks (DNNs) are a powerful learning model that achieves excellent performance in a variety of fields, such as language translation, image recognition, object detection and recognition, and the like. While deep neural networks are susceptible to antagonistic samples, outputting erroneous results with high confidence. In deep learning Szegedy et al originally proposed the concept of a challenge sample, a non-random disturbance that is small on the original image and that may arbitrarily alter the network predictions. Such disturbances are often unrecognizable by the human eye but can mislead the output of the model to be mispredicted. Such a defect has relatively little impact on applications in non-secure systems, but may cause serious practical problems in secure systems. A vulnerability in the image classification detection model, such as in an unmanned system, may be exploited to maliciously cause failure of the vehicle-mounted identification camera detection. The application and development of deep neural networks in real life is severely hampered by significant safety hazards created by the nature of the challenge sample and its portability. Accordingly, challenge samples have received extensive attention and research. At present, most researches on challenge samples take the challenge samples as attacks for destroying the robustness of the model, or propose a new attack method to greatly reduce the classification performance of the classification model, or try to promote the defensive ability of the model to the challenge samples. Whereas analysis of challenge samples from the opposite perspective may contain potential information related to model prediction, studies to aid model prediction classification have been rarely reported.
Disclosure of Invention
Since the challenge sample may also contain potential information related to model prediction, the present invention aims to obtain more accurate classification results by performing model prediction classification by taking the information contained in the challenge sample into consideration.
In order to achieve the above object, the present invention provides the following solutions: a method of improving classification of challenge samples based on a reconstruction model, comprising:
constructing a reconstruction model, and extracting characteristic information of an image dataset based on the reconstruction model; learning an abstract representation from the reconstruction model using a challenge attack, generating a challenge sample based on a generator;
constructing a classifier, and training the classifier based on the challenge sample to obtain a trained classifier;
and training the countermeasure sample and the trained classifier based on the common loss function in a combined way to obtain a target classification model, and classifying the image data set based on the target classification model to obtain a classification result.
Preferably, the reconstruction model includes a self-encoder network structure and a variational self-encoder network structure for transforming the original high-dimensional image
Figure SMS_1
Compressing to a low-dimensional space to obtain a low-dimensional space image, and reconstructing the low-dimensional space image to be the same as the original high-dimensional image +.>
Figure SMS_2
Approximate picture->
Figure SMS_3
Preferably, the reconstructed model of the self-encoder network structure obtains model parameters by minimizing a loss function;
the minimization loss function expression is:
Figure SMS_4
preferably, the reconstruction model loss function of the variation self-encoder network structure comprises two parts of reconstruction loss and KL divergence, and the training is optimized through the following formula:
Figure SMS_5
wherein ,
Figure SMS_9
for image->
Figure SMS_13
Potential variable of low-dimensional space obtained after passing through encoder,/->
Figure SMS_16
For the functional representation of the encoder, < >>
Figure SMS_7
The representation is based on the input +.>
Figure SMS_12
Is>
Figure SMS_18
Conditional probability of->
Figure SMS_19
For the actual data distribution of the potential space, +.>
Figure SMS_6
Representing a constant,/->
Figure SMS_11
Represents the KL divergence, which is used to measure the magnitude of the difference between the two distributions,
Figure SMS_14
i.e. express +.>
Figure SMS_17
and />
Figure SMS_8
KL divergence between two data distributions,
Figure SMS_10
representing a minimization of the loss function->
Figure SMS_15
To complete the training of the VAE model parameters.
Preferably, the abstract representation is learned from the reconstruction model using a challenge attack, the generating of the challenge sample based on the generator comprising,
after training of the reconstruction model is completed, a generator with consistent input and output image sizes is constructed; image is formed
Figure SMS_20
The generator is trained as input until the reconstruction performance of the reconstruction model is destroyed when the loss function value of the reconstruction model is maximized, rendering the reconstruction model antagonistic.
Preferably, the training process has the formula:
Figure SMS_21
wherein ,
Figure SMS_22
representing by maximizing +.>
Figure SMS_26
This loss function, training classifier->
Figure SMS_28
Parameter of->
Figure SMS_24
Representing challenge sample->
Figure SMS_25
Represents the intermediate layer variables of the challenge sample after the dimension reduction of the self-coding model AE or VAE, < >>
Figure SMS_30
A reconstructed image representing the challenge sample; />
Figure SMS_31
Model parameters representing a generator based on generator +.>
Figure SMS_23
Arbitrary clean samples->
Figure SMS_27
Corresponding challenge sample->
Figure SMS_29
Can be directly producedThe method comprises the following steps:
Figure SMS_32
preferably, the classifier is trained based on the challenge sample, and the formula for obtaining the trained classifier is:
Figure SMS_33
wherein ,
Figure SMS_34
representing by minimizing the loss function->
Figure SMS_35
Training to obtain classifier parameters->
Figure SMS_36
;/>
Figure SMS_37
For model parameters of the classifier, +.>
Figure SMS_38
For classification model predictive value, +.>
Figure SMS_39
For clean sample->
Figure SMS_40
Is a true mark of (c).
Preferably, the training of the challenge sample and said trained classifier is jointly based on a common loss function, the process of obtaining a target classification model comprising,
and combining the training processes of the generator and the classifier, performing an end-to-end global training step, and fine-tuning the parameters of the generator and the classifier by using a common loss function.
Preferably, the formula for fine tuning the parameters of the generator and classifier using a common loss function comprises:
Figure SMS_41
wherein ,
Figure SMS_42
representing minimizing the training of the loss function in brackets +.>
Figure SMS_43
and />
Figure SMS_44
Parameter of->
Figure SMS_45
Representing the loss function of AE or VAE, +.>
Figure SMS_46
Representing the deviation of the classifier from the predicted value and the true value of the input image by calculating the cross entropy +.>
Figure SMS_47
Obtained.
The invention discloses the following technical effects:
according to the reconstruction model-based antagonism sample improved classification method, in the first stage, feature information of a data set image is extracted by means of the reconstruction model, and then abstract representation is learned from the reconstruction model by means of antagonism attack. And in the second stage, the information of the beneficial classification is fed back to the classification model through the generated countermeasure sample. And finally, by combining the two stages, constructing an end-to-end global training structure to complete the overall training of the two-stage combined learning paradigm, thereby improving the generalization precision of the classification model. Compared with the existing countermeasure training method, the method provided by the invention realizes higher generalization precision and larger amplitude improvement in image recognition in a plurality of data sets such as MNIST, CIFAR10, CIFAR100 and the like and different classification models. Meanwhile, experiments prove that compared with a countermeasure sample based on a classification model, even a clean sample, the countermeasure sample based on the reconstruction model provided by the invention contains more sufficient characteristic information which is more beneficial to learning and understanding of the classification model.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are needed in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flow chart of a method according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a two-stage joint learning process according to an embodiment of the present invention;
FIG. 3 is a graph of the comparison of a challenge sample and a Gaussian noise sample according to an embodiment of the invention;
fig. 4 is a diagram of classification accuracy of different models under different noise intensities according to an embodiment of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
In order that the above-recited objects, features and advantages of the present invention will become more readily apparent, a more particular description of the invention will be rendered by reference to the appended drawings and appended detailed description.
As shown in fig. 1-2, the present invention provides an improved classification method for an countermeasure sample based on a reconstruction model, which mainly comprises two training parts: the first part of training is separation training, and a training reconstruction model learns characteristic information of a data set, wherein the attack reconstruction model generation countermeasure sample and the training classifier are independently performed. The second part is global training, and the common loss function is used for jointly training the countermeasures and the classifier, so that end-to-end information transfer is realized.
Further, the separation training comprises three parts of training a reconstruction model, generating an countermeasure sample and training a classifier, and each part of training is independently performed.
The reconstruction model employs a network architecture of a self-encoder (AE) and a variational self-encoder (VAE). Both of which can convert the original high-dimensional image
Figure SMS_48
Compressing to low dimensional space, and reconstructing into picture similar to original picture +.>
Figure SMS_49
For the AE model, by minimizing the loss function +.>
Figure SMS_50
Parameters of the model are obtained. For the VAE model, its loss function includes two parts, reconstruction loss and KL divergence, the training is often optimized by the following formula:
Figure SMS_51
wherein ,
Figure SMS_62
for image->
Figure SMS_54
Potential variable of low-dimensional space obtained after passing through encoder,/->
Figure SMS_58
For the functional representation of the encoder, < >>
Figure SMS_61
The representation is based on the input +.>
Figure SMS_64
Is>
Figure SMS_66
Conditional probability of->
Figure SMS_68
Is the actual data distribution of the potential space. />
Figure SMS_60
Represents a constant, obviously when +.>
Figure SMS_63
Loss function of VAE->
Figure SMS_53
Loss function with AE->
Figure SMS_56
And consistent. />
Figure SMS_52
Represents KL divergence for measuring the magnitude of the difference between the two distributions,/L>
Figure SMS_59
I.e. the representation
Figure SMS_65
and />
Figure SMS_67
KL divergence between two data distributions, < >>
Figure SMS_55
Representing a minimization loss function
Figure SMS_57
To complete the training of the VAE model parameters.
After the training of the reconstructed model is completed, an challenge sample is generated for the model, which is based on the generator. By constructing one Generator, both input and output are images of uniform size. Training the images generated by the generator in order to make the images generated by the generator antagonistic
Figure SMS_69
Let the loss function of the reconstruction model +.>
Figure SMS_70
As large as possible.
When the loss function value is maximum, it is shown that the reconstruction performance of the reconstruction model is destroyed, i.e., the reconstruction model is resistant. The training process can be formulated as:
Figure SMS_71
wherein ,
Figure SMS_72
representing by maximizing +.>
Figure SMS_73
This loss function, training classifier->
Figure SMS_74
Parameter of->
Figure SMS_75
Representing challenge sample->
Figure SMS_76
Represents the intermediate layer variables of the challenge sample after the dimension reduction of the self-coding model AE or VAE, < >>
Figure SMS_77
A reconstructed image representing the challenge sample; />
Figure SMS_78
Model parameters representing the generator. Based on generator->
Figure SMS_79
Arbitrary clean samples->
Figure SMS_80
Corresponding challenge sample->
Figure SMS_81
Can directly generate:
Figure SMS_82
this process of reconstructing the model by attack deviates the key structure and basic distribution of the challenge sample generated from the clean sample to obtain more discriminative class information. Thus in the final step of the separation phase, the challenge sample is obtained by training
Figure SMS_83
Training a classifier:
Figure SMS_84
wherein ,
Figure SMS_85
representing by minimizing the loss function->
Figure SMS_86
Training to obtain classifier parameters->
Figure SMS_87
;/>
Figure SMS_88
For model parameters of the classifier, +.>
Figure SMS_89
For classification model predictive value, +.>
Figure SMS_90
For clean sample->
Figure SMS_91
Is a true mark of (c).
Further, after separation training is completed, a joint generator
Figure SMS_92
And classifier->
Figure SMS_93
And (3) performing an end-to-end global training step to realize global optimization and further improve the precision of the classifier. At this stage, the generator and classifier parameters are fine-tuned using a common loss function:
Figure SMS_94
wherein ,
Figure SMS_95
representing minimizing the training of the loss function in brackets +.>
Figure SMS_96
and />
Figure SMS_97
Parameter of->
Figure SMS_98
Representing the loss function of AE or VAE, +.>
Figure SMS_99
Representing the deviation of the classifier from the predicted value and the true value of the input image by calculating the cross entropy +.>
Figure SMS_100
Obtained.
Example 1
The part takes the automatic driving image classification process as an application scene, firstly, the sample obtained by using the generator is a countermeasure sample, and the countermeasure thereof is verified. And secondly, comparison tests prove that the generalization precision of the classification model can be improved only by using the countermeasure sample training generated by the method. Finally, ablation experiments are carried out on the aspects of global training, attack methods and the like.
Experimental setup
Data set: the embodiment adoptsThe dataset is MNIST, CIFAR10, CIFAR100. The MNIST dataset consisted of black and white images of handwritten numbers from 0 to 9, containing 60000 training images and 10000 test images. Wherein the images are all single channels, and the picture size is 28
Figure SMS_101
28. Both the CIFAR10 and CIFAR100 datasets consisted of 60000 sheets of 3 +.>
Figure SMS_102
32/>
Figure SMS_103
32, of which 50000 are training images and 10000 are test images. The pictures in CIFAR10 are real object images in the real world for a total of 10 categories. Whereas CIFAR100 contains 100 categories of pictures, each category containing 600 images. To minimize potential effects, the image pixel values used for the experiments were determined to be from the original range [0,255]Normalized to [0,1]。
And (3) model: the models used in the experiments mainly include a reconstruction model, a classification model and a generator. The reconstruction model mainly employs AE and VAE architecture, where VAE is used in most experiments. And the classification model mainly adopts ResNet-20, which is specially designed for CIFAR data sets. VGG-19 was also used for ablation experiments. The generator used in generating the challenge sample is a generative model based on the VAE structure.
Parameters: model training in experiments is optimized by adopting an Adam optimizer, and the learning rate is adjusted by using a MultiStepLR. In generating the challenge sample, an infinite norm is employed
Figure SMS_104
Limiting the magnitude of the noise immunity. The paper precision adopted in the experiment is the generalization precision obtained on a clean sample, and the training of all classifiers has no model pre-training and data enhancement.
To demonstrate that the samples generated by the above method are resistant, i.e., the samples are destructive to the reconstruction effect of the reconstruction model, two evaluation indexes, namely Structural Similarity (SSIM) and peak signal to noise ratio (PSNR), are introduced. Both are used for measuring the similarity between the reconstructed image and the original image, and the larger the value is, the better the reconstruction effect is.
In the experiment, a sample added with ordinary Gaussian noise is used as a control group, and the difference of the influence of the reconstruction effect of the countermeasure sample and the sample containing ordinary Gaussian noise is compared. The experimental results show that the reconstruction effect of the challenge samples is significantly lower than that of gaussian noise samples, as shown in fig. 3, and this difference increases with increasing noise intensity. The noise intensity of the common Gaussian noise needs to be more than 8 times (0.25) to be equal to the noise intensity of the common Gaussian noise when the noise intensity of the anti-attack effect is 0.03. Thus, the challenge sample is more misleading, i.e. more resistive, to the reconstruction model completing the reconstruction task.
By incorporating the reconstruction model-based challenge sample proposed in the present embodiment
Figure SMS_105
Challenge sample for classification model with mainstream study +.>
Figure SMS_106
And clean sample->
Figure SMS_107
Compared with the trained model, the countermeasure sample provided by the embodiment can improve the precision of the classification model. To make the difference more obvious, the noise intensity of the challenge sample is increased to 0.3. Obviously, challenge samples based on reconstruction model +.>
Figure SMS_108
Noise and sample for classification model>
Figure SMS_109
The difference exists, and the noise of the noise is regular and is in a flat grid shape; whereas the latter noise is chaotic and chapter-free.
As shown in table 1, under different data sets, challenge samples for the classifier
Figure SMS_110
Trained model accuracy is better than using clean samples +.>
Figure SMS_111
Indicating that challenge training reduces the generalization accuracy of the model, which is in substantial agreement with the conclusions of the prior studies. Unlike this, the present embodiment provides a reconstructed model-based challenge sample ∈ ->
Figure SMS_112
The generalization precision of the classification model can be improved, and the MNIST, CIFAR10 and CIFAR100 data sets are respectively improved by 0.06%,1.34% and 0.98% under the condition that data enhancement and model pre-training are not carried out.
TABLE 1
Figure SMS_113
In addition, compared with other classification models trained using challenge samples (e.g., table 2, table 3), the ICRAE method of the present embodiment can achieve the highest classification accuracy of 99.7% on MNIST and the greatest accuracy improvement of 1.34% on CIFAR10 dataset.
TABLE 2
Figure SMS_114
TABLE 3 Table 3
Figure SMS_115
Next, the influence of different factors including attack mode, noise intensity, used model and the like on the experimental result is explored.
In the second step of the training process, different challenge methods, such as the usual fast gradient descent method (FGSM), projection gradient descent method (PGD), and the Generator-based method used in this embodiment, may be used. Wherein FGSM and PGD are both baseIterative attack-countering method by inputting
Figure SMS_116
Add a disturbance->
Figure SMS_119
Obtain challenge sample->
Figure SMS_122
. The specific calculation method of FGSM is that firstly, the loss function is calculated about input +>
Figure SMS_118
Gradient of->
Figure SMS_120
Intensity of attack and gradient->
Figure SMS_121
Is multiplied by the sign of (a) and the resulting disturbance +.>
Figure SMS_123
And->
Figure SMS_117
Adding, and obtaining the countermeasure sample by one step of iteration. The calculation process is formulated as:
Figure SMS_124
wherein ,
Figure SMS_125
representing the amount of change of the loss function +.>
Figure SMS_126
Representing the amount of change of the input image,/->
Figure SMS_127
Indicating the strength of the attack and,
Figure SMS_128
representing calculated +.>
Figure SMS_129
Is a sign of (3).
And PGD is equivalent to repeating the FGSM process multiple times, with multiple iterations. And thus its attack effect is better than FGSM. In contrast to both, the method of deriving the challenge sample with the generator is based on optimization, with the parameters of the generator being trained by means of an optimizer. The classification accuracy corresponding to the challenge samples generated by the different attack methods is compared in table 4, wherein G-s represents the model accuracy after the generator-based and separation training is completed; and G-G represents model accuracy after generator-based and global training. According to the experimental result, the overall training is completed, and the model classification effect is improved to the greatest extent by a generator-based method (namely the ICRAE algorithm proposed by the embodiment). In combination with table 5, the end-to-end global training can further improve the classification precision of the model and realize the global optimization of the feature extraction of the classification information.
TABLE 4 Table 4
Figure SMS_130
Noise intensity
Figure SMS_131
Refers to infinite norm +.>
Figure SMS_132
The method comprises the following steps:
Figure SMS_133
noise intensity
Figure SMS_134
The larger the noise on the challenge sample, the more noticeable the gap from the original image. Table 5 shows the generalization accuracy of the trained classification model at different noise intensities. In different data sets, the precision is along withThe noise increases, and the trend of increasing and then decreasing is shown, and the peak value in each data set appears when the noise intensity is 0.01. At the same time, the general noise intensity is within the interval [0.01,0.05 ]]When training the classification model against the samples, the accuracy is higher than for a model corresponding to a clean sample. This means that too high an anti-noise still damages the feature information in the image, thereby reducing the classification accuracy.
TABLE 5
Figure SMS_135
Further, experiments were also performed using different reconstruction models and classification models, the results of which are shown in fig. 4. In fig. 4, the abscissa indicates noise intensity, and the ordinate indicates classification accuracy. The left graph is the result of using the ResNet model, the right graph is the result of using the VGG model, and the dashed line represents the generalization accuracy obtained by training the classifier using clean samples. Obviously, the noise intensity corresponding to the peak of the classification accuracy is different for different classification models. ResNet corresponds to the highest accuracy at 0.01, while VGG is optimal at a noise strength of 0.03. On the other hand, under the ResNet model, the AE is used as a reconstruction model to improve the final classification effect optimally; while the reconstruction with VAE is better for VGG models. However, in general, the trend of classification accuracy with noise size is substantially consistent under different models, and also at noise intensities of 0.01 to 0.05, training using the challenge sample improves classification model accuracy. This shows that the ICRAE algorithm provided in this embodiment can be applied to different reconstruction models and classification models, so as to improve classification accuracy.
In most of the prior studies, the challenge sample is considered as an image with destroyed characteristic information, so that an excellent classification model cannot be trained. In the embodiment, the challenge sample is researched from a forward direction, a challenge sample lifting and classifying algorithm based on a reconstruction model is provided, a classifier with better performance than that of training a clean sample can be trained by only using the challenge sample, and the improvement of the model generalization precision is realized. This demonstrates that the challenge sample contains sufficient classification information to facilitate further investigation of the challenge sample. The characteristic information in the countermeasure sample can be split on the basis of the characteristic information in the future, and the action mechanism of different characteristic information in the characteristic information on the basis of the characteristic information can be explored, so that a classification model with stronger predictability and stronger robustness can be obtained.
The above embodiments are only illustrative of the preferred embodiments of the present invention and are not intended to limit the scope of the present invention, and various modifications and improvements made by those skilled in the art to the technical solutions of the present invention should fall within the protection scope defined by the claims of the present invention without departing from the design spirit of the present invention.

Claims (1)

1. A method for improving classification of challenge samples based on a reconstruction model, comprising:
in a scene of automatic driving image classification processing, constructing a reconstruction model, and extracting characteristic information of an image dataset based on the reconstruction model; learning an abstract representation from the reconstruction model using a challenge attack, generating a challenge sample based on a generator;
constructing a classifier, and training the classifier based on the challenge sample to obtain a trained classifier;
training a challenge sample and the trained classifier in a combined way based on a common loss function to obtain a target classification model, classifying the image data set based on the target classification model, and obtaining a classification result;
the reconstruction model comprises a self-encoder network structure and a variational self-encoder network structure for transforming the original high-dimensional image
Figure QLYQS_1
Compressing to a low-dimensional space to obtain a low-dimensional space image, and reconstructing the low-dimensional space image to be the same as the original high-dimensional image +.>
Figure QLYQS_2
Approximate picture->
Figure QLYQS_3
The reconstruction model of the self-encoder network structure obtains model parameters by minimizing a loss function;
the minimization loss function expression is:
Figure QLYQS_4
the reconstruction model loss function of the variation self-encoder network structure comprises two parts, namely reconstruction loss and KL divergence, and the training is optimized through the following formula:
Figure QLYQS_6
wherein ,/>
Figure QLYQS_8
For image->
Figure QLYQS_15
Potential variable of low-dimensional space obtained after passing through encoder,/->
Figure QLYQS_9
For the functional representation of the encoder, < >>
Figure QLYQS_12
The representation is based on input
Figure QLYQS_14
Is>
Figure QLYQS_16
Conditional probability of->
Figure QLYQS_5
For the actual data distribution of the potential space, +.>
Figure QLYQS_10
Representing a constant,/->
Figure QLYQS_18
Represents KL divergence for measuring the magnitude of the difference between the two distributions,/L>
Figure QLYQS_19
I.e. express +.>
Figure QLYQS_7
And
Figure QLYQS_11
KL divergence between two data distributions, < >>
Figure QLYQS_13
Representing a minimization of the loss function->
Figure QLYQS_17
Training parameters of the VAE model is completed;
learning an abstract representation from the reconstruction model using a challenge attack, the generator-based process of generating a challenge sample including,
after training of the reconstruction model is completed, a generator with consistent input and output image sizes is constructed; image is formed
Figure QLYQS_20
Training the generator as input until the reconstruction performance of the reconstruction model is destroyed when the loss function value of the reconstruction model is maximum, so that the reconstruction model has antagonism;
the training process has the following formula:
Figure QLYQS_21
wherein ,/>
Figure QLYQS_22
Representation by maximization
Figure QLYQS_23
This loss function, training classifier->
Figure QLYQS_24
Parameter of->
Figure QLYQS_25
Representing challenge sample->
Figure QLYQS_26
Represents the intermediate layer variables of the challenge sample after the dimension reduction of the self-coding model AE or VAE, < >>
Figure QLYQS_27
A reconstructed image representing the challenge sample;
Figure QLYQS_28
model parameters representing a generator based on generator +.>
Figure QLYQS_29
Arbitrary clean samples->
Figure QLYQS_30
Corresponding challenge sample->
Figure QLYQS_31
Can directly generate: />
Figure QLYQS_32
Training the classifier based on the challenge sample to obtain a trained classifier with the following formula:
Figure QLYQS_34
wherein ,/>
Figure QLYQS_37
Representing passing through the mostMinimization of the loss function->
Figure QLYQS_39
Training to obtain classifier parameters->
Figure QLYQS_35
;/>
Figure QLYQS_36
For model parameters of the classifier, +.>
Figure QLYQS_38
For classification model predictive value, +.>
Figure QLYQS_40
For clean samples
Figure QLYQS_33
Is a true mark of (2);
training the challenge sample and the trained classifier jointly based on a common loss function, the process of obtaining a target classification model comprising,
combining the training process of the generator and the classifier, performing an end-to-end global training step, and fine-tuning parameters of the generator and the classifier by using a common loss function;
the formula for fine tuning the parameters of the generator and classifier using a common loss function includes:
Figure QLYQS_41
wherein ,
Figure QLYQS_42
representing minimizing the training of the loss function in brackets +.>
Figure QLYQS_43
and />
Figure QLYQS_44
Parameter of->
Figure QLYQS_45
Representing the loss function of AE or VAE, +.>
Figure QLYQS_46
Representing the deviation of the classifier from the predicted value and the true value of the input image by calculating the cross entropy +.>
Figure QLYQS_47
Obtained. />
CN202310132811.6A 2023-02-20 2023-02-20 Improved classification method for countermeasure sample based on reconstruction model Active CN115797711B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310132811.6A CN115797711B (en) 2023-02-20 2023-02-20 Improved classification method for countermeasure sample based on reconstruction model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310132811.6A CN115797711B (en) 2023-02-20 2023-02-20 Improved classification method for countermeasure sample based on reconstruction model

Publications (2)

Publication Number Publication Date
CN115797711A CN115797711A (en) 2023-03-14
CN115797711B true CN115797711B (en) 2023-04-21

Family

ID=85431017

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310132811.6A Active CN115797711B (en) 2023-02-20 2023-02-20 Improved classification method for countermeasure sample based on reconstruction model

Country Status (1)

Country Link
CN (1) CN115797711B (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114707572A (en) * 2022-02-24 2022-07-05 浙江工业大学 Deep learning sample testing method and device based on loss function sensitivity
CN115641471A (en) * 2022-10-18 2023-01-24 西安交通大学 Countermeasure sample generation method and system based on generation of countermeasure network

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110647645A (en) * 2019-08-06 2020-01-03 厦门大学 Attack image retrieval method based on general disturbance
CN110674836B (en) * 2019-08-06 2024-03-22 厦门大学 Sparse countermeasure sample generation method based on generation network
CN111598805A (en) * 2020-05-13 2020-08-28 华中科技大学 Confrontation sample defense method and system based on VAE-GAN
CN113222002B (en) * 2021-05-07 2024-04-05 西安交通大学 Zero sample classification method based on generative discriminative contrast optimization
CN113283599B (en) * 2021-06-11 2024-03-19 浙江工业大学 Attack resistance defense method based on neuron activation rate
CN113554089B (en) * 2021-07-22 2023-04-18 西安电子科技大学 Image classification countermeasure sample defense method and system and data processing terminal
CN113723564A (en) * 2021-09-14 2021-11-30 燕山大学 Method and system for training defense model of confrontation sample and application of method and system
CN113963213A (en) * 2021-10-27 2022-01-21 上海交通大学 Method and system for removing antagonistic noise aiming at antagonistic sample of deep neural network
CN114881103A (en) * 2022-03-25 2022-08-09 重庆邮电大学 Countermeasure sample detection method and device based on universal disturbance sticker
CN114648642A (en) * 2022-03-30 2022-06-21 京东科技信息技术有限公司 Model training method, image detection method, image classification method and device
CN115294399A (en) * 2022-08-18 2022-11-04 中国人民解放军国防科技大学 Image classification method and device for defending against attack and computer equipment
CN115439719B (en) * 2022-10-27 2023-03-28 泉州装备制造研究所 Deep learning model defense method and model for resisting attack

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114707572A (en) * 2022-02-24 2022-07-05 浙江工业大学 Deep learning sample testing method and device based on loss function sensitivity
CN115641471A (en) * 2022-10-18 2023-01-24 西安交通大学 Countermeasure sample generation method and system based on generation of countermeasure network

Also Published As

Publication number Publication date
CN115797711A (en) 2023-03-14

Similar Documents

Publication Publication Date Title
Tang et al. An automatic cost learning framework for image steganography using deep reinforcement learning
You et al. Adversarial noise layer: Regularize neural network by adding noise
CN113674140B (en) Physical countermeasure sample generation method and system
Guo et al. Robust student network learning
CN111160343A (en) Off-line mathematical formula symbol identification method based on Self-Attention
Zhang et al. Defense against adversarial attacks by reconstructing images
CN112801280B (en) One-dimensional convolution position coding method of visual depth self-adaptive neural network
CN113378949A (en) Dual-generation confrontation learning method based on capsule network and mixed attention
Lv et al. Chinese character CAPTCHA recognition based on convolution neural network
CN115422518A (en) Text verification code identification method based on data-free knowledge distillation
Li et al. Image operation chain detection with machine translation framework
Jalwana et al. Attack to explain deep representation
Zanddizari et al. Generating black-box adversarial examples in sparse domain
Biswas et al. Deepfake detection using 3d-xception net with discrete fourier transformation
WO2021189364A1 (en) Method and device for generating adversarial image, equipment, and readable storage medium
CN115797711B (en) Improved classification method for countermeasure sample based on reconstruction model
CN115481719B (en) Method for defending against attack based on gradient
Yang et al. APE-GAN++: An improved APE-GAN to eliminate adversarial perturbations
Zhou et al. Markov chain based efficient defense against adversarial examples in computer vision
Sun et al. CAMA: Class activation mapping disruptive attack for deep neural networks
Ishii et al. Training deep neural networks with adversarially augmented features for small-scale training datasets
CN113887504B (en) Strong-generalization remote sensing image target identification method
Liang et al. Large-scale image classification using fast svm with deep quasi-linear kernel
Ma et al. DIHBA: Dynamic, invisible and high attack success rate boundary backdoor attack with low poison ratio
Sun et al. Instance-level Trojan Attacks on Visual Question Answering via Adversarial Learning in Neuron Activation Space

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant