CN112465111A - Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining - Google Patents

Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining Download PDF

Info

Publication number
CN112465111A
CN112465111A CN202011289486.7A CN202011289486A CN112465111A CN 112465111 A CN112465111 A CN 112465111A CN 202011289486 A CN202011289486 A CN 202011289486A CN 112465111 A CN112465111 A CN 112465111A
Authority
CN
China
Prior art keywords
network
teacher
image
student
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011289486.7A
Other languages
Chinese (zh)
Inventor
李天波
候亚庆
周东生
张强
魏小鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian University of Technology
Original Assignee
Dalian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dalian University of Technology filed Critical Dalian University of Technology
Priority to CN202011289486.7A priority Critical patent/CN112465111A/en
Publication of CN112465111A publication Critical patent/CN112465111A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Mathematical Physics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Medical Informatics (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the field of image segmentation, and provides a three-dimensional voxel image segmentation method based on knowledge distillation and confrontation training, which is divided into two parts: training a plurality of teacher networks for integrated learning, wherein the plurality of teacher networks are used as prior knowledge to obtain soft targets for training student networks; and (3) carrying out confrontation training on the student network, carrying out supervision training by using soft targets and image labels of the teacher network, carrying out supervision by using a characteristic diagram in the middle of the teacher network, and adding a discriminator to construct GANs loss supervision training. The invention expands the available data using knowledge distillation and performs knowledge migration using knowledge distillation, enabling model compression. For the deep learning model of the three-dimensional voxel image, parameters are saved, and the model is easier to deploy. By combining knowledge distillation and antagonistic training, the regularization effect is more obvious, and the overfitting problem of the three-dimensional voxel image caused by the rare data volume is solved to a certain extent.

Description

Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining
Technical Field
The invention belongs to the field of image segmentation, and relates to a three-dimensional voxel image segmentation method based on knowledge distillation and confrontation training.
Background
With the rapid development of science and technology and imaging technology, three-dimensional images are more and more widely used, especially three-dimensional voxel images. The voxel is the minimum unit of digital data on three-dimensional space segmentation, and the voxel is widely applied to the fields of three-dimensional imaging, scientific data, medical images and the like. In the use scene of the three-dimensional voxel image, a plurality of basic tasks of the three-dimensional voxel image segmentation depending on the foundation are applied. But is very complex and inefficient if the segmentation is performed manually on a three-dimensional voxel image. This is not only because the volume of the three-dimensional voxel image needs to be segmented on each slice, but also the accurate segmentation of each slice requires a professional in the corresponding usage scenario, which results in a large amount of high personnel and labor costs, which is obviously not feasible.
For a three-dimensional voxel image, the characteristics of large single data and small total amount of data are generally provided, which also causes that the influence of the model volume on the three-dimensional voxel image is large, and the model is easy to be over-fitted. For traditional machine learning methods, extracting features is too complex and generally not feasible. The depth learning-based method is mostly applied to two-dimensional natural image segmentation, and if the depth learning-based two-dimensional natural image segmentation method is applied to a three-dimensional voxel image, efficiency and accuracy are low. Because the characteristics of the three-dimensional voxel image and the two-dimensional natural image are completely different, the two-dimensional natural image usually has small single data and large data total amount. Therefore, a new method is needed to be designed to be applied to the three-dimensional voxel image based on the characteristics of the three-dimensional voxel image.
Hinton et al first proposed the concept of knowledge distillation in 2015, the general framework of which mainly includes a teacher network and a student network. Soft goals for the teacher network are introduced as part of the total loss to induce training on the student network and to effect knowledge transfer. Since hard objects are single hot coded, the entropy of the information contained in hard objects is low, while the entropy of the information contained in soft objects is high, and there is a relationship between the different classes. The temperature T is used to smooth the classification results, obtain soft targets with higher entropy of information, and train the model using soft target supervision as part of the loss.
The generation of countermeasure networks (GANs) was proposed by Goodfellow et al in 2014. GANs consist of two neural networks, one being a generator and the other being a discriminator. The generator attempts to capture the data distribution and the discriminator estimates the probability that the sample is from training data rather than generated data. That is, the generator attempts to create an image that looks "natural" and requires that it be as consistent as possible with the original data distribution. The task of the discriminator is to determine whether a given image appears to be "natural". Pauline Luc et al trained semantic segmentation models for the first time in 2016 using GANs and performed pixel-level dense classification on the original images. The method uses a discriminator to distinguish the results of the segmented network from the correct labels, where the loss function combines the traditional multi-class cross-entropy loss and the GANs loss. By using the countermeasure network, the robustness of the model under small-scale data is improved, overfitting is prevented, the robustness of the model is improved, and the target image space is more continuous and smooth.
The use of GANs in semantic segmentation has a very large gap between the true tags and the output results. Since the tags are one-hot coded and the output result is non-one-hot coded, the discriminator can easily and directly judge the tags and the output result, and thus the GANs have a limited effect on semantic segmentation. Why do this result? It is theoretically analyzed that this is due to "unnatural" tags, so tags like soft objects are needed to serve as "natural" data.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention provides a three-dimensional voxel image segmentation method based on knowledge distillation and countertraining, and aims to construct an efficient and highly robust three-dimensional voxel image segmentation method. The invention uses knowledge distillation to expand data, and can obtain a soft target aiming at the data without labels, thereby relieving the problem of less total data of the three-dimensional voxel image to a certain extent. The original label training is used, so that the final result is offset to a certain extent, and the accuracy deficiency of the teacher network can be made up. Knowledge transfer is performed by using knowledge distillation, and knowledge learned from a collection of a plurality of large models is transferred to a small model to realize model compression. Therefore, the method is beneficial to being put into practical production, and the robustness of the model is improved by using a mode of countertraining, so that the result image becomes smoother. Knowledge distillation and countermeasure training are combined, the judgment easiness of a discriminator in the countermeasure training can be improved, the robustness of a model is further improved, and the three-dimensional voxel image segmentation method model can be put into production and land more easily.
A three-dimensional voxel image segmentation method based on knowledge distillation and antagonistic training is divided into two big modules:
(1) training a plurality of teacher networks for integrated learning, wherein the plurality of teacher networks are used as prior knowledge to obtain soft targets for training student networks; the training teacher network is divided into two submodules, wherein the first submodule is a data preprocessing part and mainly used for carrying out random cutting, rotation and intensity offset on a three-dimensional voxel image to carry out data enhancement and then carrying out standardized processing on data by using a Z-Score (Z-Score); the second sub-module is used for carrying out supervision training on a teacher network, modifying a network structure by 3DUNet to serve as the teacher network, and then carrying out supervision by using image labels; saving a plurality of trained teacher models for subsequent student network training;
(2) the method comprises the following steps of training a student network in an antagonistic manner, carrying out supervision training by using a soft target and an image label of a teacher network, carrying out supervision by using a characteristic diagram in the middle of the teacher network, and adding a discriminator to construct GANs loss supervision training; the training student network is divided into four sub-modules, wherein the first sub-module is a data preprocessing part and mainly used for carrying out random cutting, rotation and intensity offset on a three-dimensional voxel image to carry out data enhancement and then carrying out standardized processing on data by using Z-Score; the second sub-module is transmitted to the bottom layer in a forward direction to carry out auxiliary supervision training; the third sub-module is used for forward propagation and finally output and jointly supervised training by using a soft target and an image label; the fourth submodule utilizes the GANs loss supervision training for constructing a discriminator; and 3DUNet is also modified as a student network structure, the number of layers of the student network structure is less than that of the teacher network structure, and compression is performed.
In order to achieve the purpose, the technical scheme adopted by the invention is as follows: the method comprises the following specific steps:
step (1) train multiple teacher networks
(1.1) cutting the original three-dimensional voxel image to a proper size by a random center, then randomly rotating, shifting the intensity, and finally performing pretreatment by Z-Score standardization;
the formula for Z-Score is as follows:
Figure BDA0002783410960000041
the Z-Score is zero-mean unit variance operation, so that training can be accelerated; x is an input image, X' is an output image, mu is the mean value of all images, and sigma is the variance of all images;
(1.2) training a teacher network, and training the teacher network end to end by using the modified 3DUNet network structure and using image labels; the 3DUNet structure mainly depends on three-dimensional convolution operation and three-dimensional maximum pooling operation, and a group of standardized operation and leakage correction unit activation functions are connected after the three-dimensional convolution operation is performed each time; the reason why batch normalization is not used is that the three-dimensional voxel data is large, and the batch sent into the network each time can only be small, so that batch normalization has little effect. After the image is subjected to convolution operation for two times, the resolution of the image is reduced to be half of the original resolution by using three-dimensional maximum pooling; the up-sampling uses a trilinear interpolation method, and is connected with the same convolution operation to form a coder decoder structure; the final activation function uses a sigmoid activation function to finally output a segmentation image; performing supervised training on a teacher network by using a soft dice loss function, wherein the condition of unbalanced segmentation labels usually exists on a three-dimensional voxel image; the training effect is not good if the softmax activation function and the multi-class cross entropy loss function are used, so the method uses the soft dice loss function to train the teacher network. Finally, a plurality of trained teacher networks are stored as prior knowledge to prepare for later training student networks;
the formula of the teacher network soft dice loss function is as follows:
Figure BDA0002783410960000042
teacher network Soft dice loss function, LtRepresenting the loss of the teacher network, K representing the class, t representing the segmentation result of the teacher network, and c representing the image label; ε is a constant close to 0 to avoid a denominator of 0.
Step (2) confrontation training student network, using the teacher network and the image label in the above and constructing a discriminator to supervise training;
(2.1) similarly to (1.1), the preprocessing part of the data is also performed, except that when the student network is trained, additional unlabelled data is used as an input image, because the soft loss of the unlabelled image is obtained by using the previously obtained trained teacher network, the unlabelled image is utilized, and the usable data set is expanded to a certain extent;
and (2.2) the teacher network and the student network are in a modified 3DUNet network structure, so that the intermediate results of the teacher network can be used for assisting in supervising the student network training. Because the student network is smaller than the teacher model and mainly embodied in that the number of layers is smaller than that of the teacher model, the student network can be mapped to a higher dimension through auxiliary convolution to be matched with the intermediate layer characteristic diagram of the teacher network, so that a loss function for auxiliary supervision can be obtained, and MSE loss function is used for auxiliary supervision;
the equation for MSE is as follows:
Figure BDA0002783410960000051
equation (3) is described as an aided supervised MSE loss function,
Figure BDA0002783410960000052
for the auxiliary loss of the student network, V is all voxels, as is the intermediate result of the student network, and at is the intermediate result of the teacher network;
and (2.3) propagating forward to the rearmost end, and training the student network by using the soft target of the teacher network and the image label in a joint supervision mode. The 3d onnet structure of the student network structure is very similar to that of the teacher network, the main difference being that the number of layers of the student network is less than that of the teacher network. And finally outputting the segmentation result of the student network through the encoder-decoder structure, and measuring the loss between the segmentation result of the student network and a soft target of a teacher network by using a soft dice loss function, and measuring the loss between the segmentation record of the student network and the image tag by using the soft dice loss function. The foregoing illustrates the use of unlabeled image data, where is how does unlabeled image data jointly supervise? The problem is solved by dynamically adjusting the weight of the joint loss function, joint supervision is used for data with image labels, and only soft target supervision is used for data without image labels;
the soft dice loss function for the student network is as follows:
Figure BDA0002783410960000061
in the formula (4), the first and second groups,
Figure BDA0002783410960000062
representing a loss function between the student network segmentation result and the teacher network soft target, s representing the segmentation result of the student network, and t representing the soft target of the teacher network;
Figure BDA0002783410960000063
in the formula (5), the first and second groups,
Figure BDA0002783410960000064
representing a loss function between the student network segmentation result and the image label, s representing the segmentation result of the student network, and c representing the image label;
Figure BDA0002783410960000065
in the formula (6), LsRepresents the total loss of the student network, as a weighted sum between several losses; lambda [ alpha ]1To represent
Figure BDA0002783410960000066
And
Figure BDA0002783410960000067
weight between, λ2To represent
Figure BDA0002783410960000068
A proportion weight;
(2.4) constructing a discriminator to construct the GANs loss, which has the regularization effect. The original image is spliced with the student network segmentation result or the teacher network soft target and input into the discriminator. The discriminator is a full convolution network, the characteristics are extracted through an average pooling layer in the middle, and finally two classification results are output to represent whether the image is true or false;
Figure BDA0002783410960000069
in the formula (7), the first and second groups,
Figure BDA00027834109600000610
for the discriminator to be lost, the image label from the student network needs to be set to 0, and the image label from the teacher network needs to be set to 1; e is a mathematical expectation; qtFor teacher networks, split labels, QsA split label for a student network; i is an original image; p is a radical oftRepresenting teacher distribution, psRepresenting a student distribution; d () represents the result output by the discriminator; i represents a splicing operation;
Figure BDA00027834109600000611
in the formula (8), the first and second groups,
Figure BDA00027834109600000612
for the loss of the GANs of the student network, the image label from the student network needs to be set to 1 for the student network;
Figure BDA0002783410960000071
loss L of final student networktotalIs a weighted sum of student network losses and GANs losses; lambda [ alpha ]ganTo represent
Figure BDA0002783410960000072
The weight of (c).
The invention has the beneficial effects that: the invention expands the available data using knowledge distillation and performs knowledge migration using knowledge distillation, enabling model compression. For the deep learning model of the three-dimensional voxel image, parameters are saved, and the model is easier to deploy. By combining knowledge distillation and antagonistic training, the regularization effect is more obvious, and the overfitting problem of the three-dimensional voxel image caused by the rare data volume is solved to a certain extent.
Drawings
Fig. 1 is an overall framework of the present invention.
FIG. 2 is a schematic diagram of a teacher model structure according to the present invention.
Fig. 3 is a schematic structural diagram of a student model of the invention.
FIG. 4 is a schematic diagram of the structure of the discriminator according to the present invention.
FIG. 5 is a flow chart of a teacher network training method of the present invention.
FIG. 6 is a flow chart of the student model and discriminator confrontation training method of the present invention.
Detailed Description
The following further describes a specific embodiment of the present invention with reference to the drawings and technical solutions.
The present invention will be described in further detail with reference to the attached drawings and detailed description, so that the conception and the technical scheme of the present invention can be more completely, accurately and deeply understood by those skilled in the art.
The invention can be used for the segmentation tasks of various three-dimensional voxel images, the overall framework of the invention is shown in figure 1 and is subdivided into each network model, the structure of a teacher model is shown in figure 2, the structure of a student model is shown in figure 3, and the structure of a discriminator is shown in figure 4.
The embodiment is applied to the tumor segmentation task of the three-dimensional nuclear magnetic resonance brain imaging graph, and the specific embodiment discussed is only used for illustrating the implementation mode of the invention and does not limit the scope of the invention.
The following describes embodiments of the present invention mainly in detail with respect to a tumor segmentation task of a three-dimensional nmr brain imaging graph, which is mainly divided into training of a teacher network (as shown in fig. 5) and training of a student network (as shown in fig. 6):
(1) training of the teacher network:
and performing data preprocessing according to a module I in the invention content, performing random center cutting, random rotation and intensity offset on the three-dimensional voxel image, and performing Z-Score processing and inputting into a trained teacher model. The input three-dimensional voxel image has the dimensions of 4x128x128x128, and the foremost 4 represents the mode images of 4 nuclear magnetic resonances, which are obtained by splicing. The image is transmitted forward, a segmentation graph is obtained through a coder decoder, soft dice loss is conducted on the segmentation graph and the image label, then optimization training is conducted through an Adam optimizer, and therefore end-to-end training is achieved. And optimizing until the model converges, training 3 teacher models with the same framework by using the same process, and storing.
(2) Training of a student network:
and performing data preprocessing according to a module II in the invention content, performing random center cutting, random rotation and intensity offset on the three-dimensional voxel image, performing Z-Score processing, inputting the processed three-dimensional voxel image into a trained student model, and simultaneously using the three-dimensional voxel data which is not labeled to expand the usable data amount. The input three-dimensional voxel image has the dimensions of 4x128x128x128, and the foremost 4 represents the mode images of 4 nuclear magnetic resonances, which are obtained by splicing. The confrontation training student model, specifically the alternate training student network and the discriminator, firstly randomly obtains a training teacher model for obtaining the teacher network intermediate result and the soft target.
When a student model is trained, an original three-dimensional voxel image is input into a student model and a teacher model, forward propagation is carried out, 3 losses can be obtained in the forward propagation process, soft dice losses are carried out on the segmentation result of the student model and a soft target of the teacher model at the tail end of a decoder to obtain loss1, soft dice losses are carried out on the segmentation result of the student model and an image label at the tail end of the decoder to obtain loss2, and MSE losses are carried out on the middle result of the student model and the middle result of the teacher model at the bottom layers of the student model and the teacher model to obtain loss 3. And then splicing the student network segmentation result and the original image, inputting the splicing result and the original image into a discriminator, carrying out forward propagation to obtain a binary classification result, carrying out GANS loss to obtain a loss4, and finally carrying out optimization training by using an Adam optimizer in loss weighted fusion. The parameters in the teacher model are fixed here because the teacher model is already trained and does not need to be trained again.
When the discriminator is trained, the spliced image and the soft target are sent into the discriminator to be transmitted forward, a binary classification result is obtained, the GANs loss is made, and an Adam optimizer is used for optimization training.

Claims (1)

1. A three-dimensional voxel image segmentation method based on knowledge distillation and antagonistic training is characterized by comprising the following steps:
step (1) train multiple teacher networks
(1.1) cutting the original three-dimensional voxel image at random center, then randomly rotating, shifting intensity, and finally carrying out pretreatment by Z-Score standardization;
the formula for Z-Score is as follows:
Figure FDA0002783410950000011
the Z-Score is zero-mean unit variance operation, so that training can be accelerated; x is an input image, X' is an output image, mu is the mean value of all images, and sigma is the variance of all images;
(1.2) training a teacher network, and training the teacher network end to end by using the modified 3DUNet network structure and using image labels; the 3DUNet structure mainly depends on three-dimensional convolution operation and three-dimensional maximum pooling operation, and a group of standardized operation and leakage correction unit activation functions are connected after the three-dimensional convolution operation is performed each time; after the image is subjected to convolution operation for two times, the resolution of the image is reduced to be half of the original resolution by using three-dimensional maximum pooling; the up-sampling uses a trilinear interpolation method, and is connected with the same convolution operation to form a coder decoder structure; the final activation function uses a sigmoid activation function to finally output a segmentation image; performing supervision training on the teacher network by using a soft dice loss function; finally, a plurality of trained teacher networks are stored as prior knowledge to prepare for later training student networks;
the formula of the teacher network soft dice loss function is as follows:
Figure FDA0002783410950000012
teacher network Soft dice loss function, LtRepresenting the loss of the teacher network, K representing the class, t representing the segmentation result of the teacher network, and c representing the image label; ε is a constant close to 0 to avoid a denominator of 0.
Step (2) confrontation training student network, using the teacher network and the image label in the above and constructing a discriminator to supervise training;
(2.1) similarly to (1.1), the preprocessing part of the data is also made, except that when the student network is trained, additional unlabelled data is used as an input image, because the soft loss of the unlabelled image is obtained by using the previously obtained teacher network with training elements, the unlabelled image is utilized, and the usable data set is enlarged;
(2.2) the teacher network and the student network are spread forward to a network bottom layer for auxiliary supervision, and as the structures of the teacher network and the student network are very similar and are modified 3DUNet network structures, the intermediate result of the teacher network is used for auxiliary supervision of student network training; because the student network is smaller than the teacher model and mainly embodied in that the number of layers is smaller than that of the teacher model, the student network is mapped to a higher dimension through auxiliary convolution to be matched with an intermediate layer characteristic diagram of the teacher network, so that an auxiliary supervision loss function is obtained, and MSE (mean square error) loss function is used for auxiliary supervision;
the equation for MSE is as follows:
Figure FDA0002783410950000021
equation (3) is described as an aided supervised MSE loss function,
Figure FDA0002783410950000022
for the auxiliary loss of the student network, V is all voxels, as is the intermediate result of the student network, and at is the intermediate result of the teacher network;
(2.3) propagating forward to the rearmost end, and using a teacher network soft target and an image label to jointly supervise and train a student network; finally, the segmentation result of the student network is output through the encoder-decoder structure, the loss between the segmentation result of the student network and a teacher network soft target is measured by using a soft dice loss function, and the loss between the segmentation record of the student network and the image label is measured by using the soft dice loss function; the method comprises the steps that joint supervision is used for data with image labels by dynamically adjusting the weight of a joint loss function, and only soft target supervision is used for data without the image labels;
the soft dice loss function for the student network is as follows:
Figure FDA0002783410950000031
in the formula (4), the first and second groups,
Figure FDA0002783410950000032
representing a loss function between the student network segmentation result and the teacher network soft target, s representing the segmentation result of the student network, and t representing the soft target of the teacher network;
Figure FDA0002783410950000033
in the formula (5), the first and second groups,
Figure FDA0002783410950000034
representing a loss function between the student network segmentation result and the image label, s representing the segmentation result of the student network, and c representing the image label;
Figure FDA0002783410950000035
in the formula (6), LsRepresents the total loss of the student network, as a weighted sum between several losses; lambda [ alpha ]1To represent
Figure FDA0002783410950000036
And
Figure FDA0002783410950000037
weight between, λ2To represent
Figure FDA0002783410950000038
A proportion weight;
(2.4) constructing a discriminator to construct the GANs loss, so that the regularization effect is achieved; splicing the original image with the student network segmentation result or the teacher network soft target and inputting the spliced original image into a discriminator; the discriminator is a full convolution network, the characteristics are extracted through an average pooling layer in the middle, and finally two classification results are output to represent whether the image is true or false;
Figure FDA0002783410950000039
in the formula (7), the first and second groups,
Figure FDA00027834109500000310
for the discriminator to be lost, the image label from the student network needs to be set to 0, and the image label from the teacher network needs to be set to 1; e is a mathematical expectation; qtFor teacher networks, split labels, QsA split label for a student network; i is an original image; p is a radical oftRepresenting teacher distribution, psRepresenting a student distribution; d () represents the result output by the discriminator; i represents a splicing operation;
Figure FDA00027834109500000311
in the formula (8), the first and second groups,
Figure FDA00027834109500000312
for the loss of the GANs of the student network, the image label from the student network needs to be set to 1 for the student network;
Figure FDA0002783410950000041
loss L of final student networktotalIs a weighted sum of student network losses and GANs losses; lambda [ alpha ]ganTo represent
Figure FDA0002783410950000042
The weight of (c).
CN202011289486.7A 2020-11-17 2020-11-17 Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining Pending CN112465111A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011289486.7A CN112465111A (en) 2020-11-17 2020-11-17 Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011289486.7A CN112465111A (en) 2020-11-17 2020-11-17 Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining

Publications (1)

Publication Number Publication Date
CN112465111A true CN112465111A (en) 2021-03-09

Family

ID=74837591

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011289486.7A Pending CN112465111A (en) 2020-11-17 2020-11-17 Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining

Country Status (1)

Country Link
CN (1) CN112465111A (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113112020A (en) * 2021-03-25 2021-07-13 厦门大学 Model network extraction and compression method based on generation network and knowledge distillation
CN113344896A (en) * 2021-06-24 2021-09-03 鹏城实验室 Breast CT image focus segmentation model training method and system
CN113379627A (en) * 2021-06-07 2021-09-10 北京百度网讯科技有限公司 Training method of image enhancement model and method for enhancing image
CN113421243A (en) * 2021-06-23 2021-09-21 深圳大学 Method and device for detecting type of fundus image based on knowledge distillation network
CN113449851A (en) * 2021-07-15 2021-09-28 北京字跳网络技术有限公司 Data processing method and device
CN113609965A (en) * 2021-08-03 2021-11-05 同盾科技有限公司 Training method and device of character recognition model, storage medium and electronic equipment
CN113793341A (en) * 2021-09-16 2021-12-14 湘潭大学 Automatic driving scene semantic segmentation method, electronic device and readable medium
CN113822339A (en) * 2021-08-27 2021-12-21 北京工业大学 Natural image classification method combining self-knowledge distillation and unsupervised method
CN113869512A (en) * 2021-10-09 2021-12-31 北京中科智眼科技有限公司 Supplementary label learning method based on self-supervision and self-distillation
CN113902761A (en) * 2021-11-02 2022-01-07 大连理工大学 Unsupervised segmentation method for lung disease focus based on knowledge distillation
CN113920540A (en) * 2021-11-04 2022-01-11 厦门市美亚柏科信息股份有限公司 Knowledge distillation-based pedestrian re-identification method, device, equipment and storage medium
CN114549901A (en) * 2022-02-24 2022-05-27 杭州电子科技大学 Multi-network joint auxiliary generation type knowledge distillation method
CN114943101A (en) * 2022-05-18 2022-08-26 广州大学 Privacy protection generative model construction method
CN115085805A (en) * 2022-06-09 2022-09-20 南京信息工程大学 Few-mode multi-core fiber optical performance monitoring method, system and device based on anti-distillation model and storage medium
CN115564024A (en) * 2022-10-11 2023-01-03 清华大学 Feature distillation method and device for generating network, electronic equipment and storage medium
CN116110022A (en) * 2022-12-10 2023-05-12 河南工业大学 Lightweight traffic sign detection method and system based on response knowledge distillation
CN116542321A (en) * 2023-07-06 2023-08-04 中科南京人工智能创新研究院 Image generation model compression and acceleration method and system based on diffusion model
WO2023207360A1 (en) * 2022-04-29 2023-11-02 北京字跳网络技术有限公司 Image segmentation method and apparatus, electronic device, and storage medium
CN116994309A (en) * 2023-05-06 2023-11-03 浙江大学 Face recognition model pruning method for fairness perception
CN113822339B (en) * 2021-08-27 2024-05-31 北京工业大学 Natural image classification method combining self-knowledge distillation and unsupervised method

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180268292A1 (en) * 2017-03-17 2018-09-20 Nec Laboratories America, Inc. Learning efficient object detection models with knowledge distillation
CN111462137A (en) * 2020-04-02 2020-07-28 中科人工智能创新技术研究院(青岛)有限公司 Point cloud scene segmentation method based on knowledge distillation and semantic fusion
WO2020186887A1 (en) * 2019-03-15 2020-09-24 深圳先进技术研究院 Target detection method, device and apparatus for continuous small sample images
CN111723812A (en) * 2020-06-05 2020-09-29 南强智视(厦门)科技有限公司 Real-time semantic segmentation method based on sequence knowledge distillation
CN111753878A (en) * 2020-05-20 2020-10-09 济南浪潮高新科技投资发展有限公司 Network model deployment method, equipment and medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180268292A1 (en) * 2017-03-17 2018-09-20 Nec Laboratories America, Inc. Learning efficient object detection models with knowledge distillation
WO2020186887A1 (en) * 2019-03-15 2020-09-24 深圳先进技术研究院 Target detection method, device and apparatus for continuous small sample images
CN111462137A (en) * 2020-04-02 2020-07-28 中科人工智能创新技术研究院(青岛)有限公司 Point cloud scene segmentation method based on knowledge distillation and semantic fusion
CN111753878A (en) * 2020-05-20 2020-10-09 济南浪潮高新科技投资发展有限公司 Network model deployment method, equipment and medium
CN111723812A (en) * 2020-06-05 2020-09-29 南强智视(厦门)科技有限公司 Real-time semantic segmentation method based on sequence knowledge distillation

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
EYTAN KATS 等: "Soft labeling by distilling anatomical knowledge for improved MS lesion segmentation", 《2019 IEEE 16TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING》, 11 July 2019 (2019-07-11), pages 1563 - 1566 *
YIFAN LIU 等: "Structured knowledge distillation for semantic segmentation", 《PROCEEDINGS OF THE IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》, 31 December 2019 (2019-12-31), pages 2604 - 2613 *
姜慧明: "基于生成对抗网络与知识蒸馏的人脸修复与表情识别", 《中国优秀硕士学位论文全文数据库 信息科技辑(月刊)计算机软件及计算机应用》, 15 August 2020 (2020-08-15), pages 138 - 499 *

Cited By (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113112020B (en) * 2021-03-25 2022-06-28 厦门大学 Model network extraction and compression method based on generation network and knowledge distillation
CN113112020A (en) * 2021-03-25 2021-07-13 厦门大学 Model network extraction and compression method based on generation network and knowledge distillation
CN113379627A (en) * 2021-06-07 2021-09-10 北京百度网讯科技有限公司 Training method of image enhancement model and method for enhancing image
CN113379627B (en) * 2021-06-07 2023-06-27 北京百度网讯科技有限公司 Training method of image enhancement model and method for enhancing image
CN113421243A (en) * 2021-06-23 2021-09-21 深圳大学 Method and device for detecting type of fundus image based on knowledge distillation network
CN113421243B (en) * 2021-06-23 2023-06-02 深圳大学 Method and device for detecting fundus image type based on knowledge distillation network
CN113344896A (en) * 2021-06-24 2021-09-03 鹏城实验室 Breast CT image focus segmentation model training method and system
CN113449851A (en) * 2021-07-15 2021-09-28 北京字跳网络技术有限公司 Data processing method and device
CN113609965B (en) * 2021-08-03 2024-02-13 同盾科技有限公司 Training method and device of character recognition model, storage medium and electronic equipment
CN113609965A (en) * 2021-08-03 2021-11-05 同盾科技有限公司 Training method and device of character recognition model, storage medium and electronic equipment
CN113822339A (en) * 2021-08-27 2021-12-21 北京工业大学 Natural image classification method combining self-knowledge distillation and unsupervised method
CN113822339B (en) * 2021-08-27 2024-05-31 北京工业大学 Natural image classification method combining self-knowledge distillation and unsupervised method
CN113793341B (en) * 2021-09-16 2024-02-06 湘潭大学 Automatic driving scene semantic segmentation method, electronic equipment and readable medium
CN113793341A (en) * 2021-09-16 2021-12-14 湘潭大学 Automatic driving scene semantic segmentation method, electronic device and readable medium
CN113869512A (en) * 2021-10-09 2021-12-31 北京中科智眼科技有限公司 Supplementary label learning method based on self-supervision and self-distillation
CN113869512B (en) * 2021-10-09 2024-05-21 北京中科智眼科技有限公司 Self-supervision and self-distillation-based supplementary tag learning method
CN113902761A (en) * 2021-11-02 2022-01-07 大连理工大学 Unsupervised segmentation method for lung disease focus based on knowledge distillation
CN113902761B (en) * 2021-11-02 2024-04-16 大连理工大学 Knowledge distillation-based unsupervised segmentation method for lung disease focus
CN113920540A (en) * 2021-11-04 2022-01-11 厦门市美亚柏科信息股份有限公司 Knowledge distillation-based pedestrian re-identification method, device, equipment and storage medium
CN114549901B (en) * 2022-02-24 2024-05-14 杭州电子科技大学 Multi-network combined auxiliary generation type knowledge distillation method
CN114549901A (en) * 2022-02-24 2022-05-27 杭州电子科技大学 Multi-network joint auxiliary generation type knowledge distillation method
WO2023207360A1 (en) * 2022-04-29 2023-11-02 北京字跳网络技术有限公司 Image segmentation method and apparatus, electronic device, and storage medium
CN114943101B (en) * 2022-05-18 2024-05-17 广州大学 Privacy protection generation model construction method
CN114943101A (en) * 2022-05-18 2022-08-26 广州大学 Privacy protection generative model construction method
CN115085805A (en) * 2022-06-09 2022-09-20 南京信息工程大学 Few-mode multi-core fiber optical performance monitoring method, system and device based on anti-distillation model and storage medium
CN115085805B (en) * 2022-06-09 2024-03-19 南京信息工程大学 Fiber optical performance monitoring method and system based on anti-distillation model
CN115564024B (en) * 2022-10-11 2023-09-15 清华大学 Characteristic distillation method, device, electronic equipment and storage medium for generating network
CN115564024A (en) * 2022-10-11 2023-01-03 清华大学 Feature distillation method and device for generating network, electronic equipment and storage medium
CN116110022A (en) * 2022-12-10 2023-05-12 河南工业大学 Lightweight traffic sign detection method and system based on response knowledge distillation
CN116110022B (en) * 2022-12-10 2023-09-05 河南工业大学 Lightweight traffic sign detection method and system based on response knowledge distillation
CN116994309B (en) * 2023-05-06 2024-04-09 浙江大学 Face recognition model pruning method for fairness perception
CN116994309A (en) * 2023-05-06 2023-11-03 浙江大学 Face recognition model pruning method for fairness perception
CN116542321B (en) * 2023-07-06 2023-09-01 中科南京人工智能创新研究院 Image generation model compression and acceleration method and system based on diffusion model
CN116542321A (en) * 2023-07-06 2023-08-04 中科南京人工智能创新研究院 Image generation model compression and acceleration method and system based on diffusion model

Similar Documents

Publication Publication Date Title
CN112465111A (en) Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining
CN106547880B (en) Multi-dimensional geographic scene identification method fusing geographic area knowledge
CN111950453B (en) Random shape text recognition method based on selective attention mechanism
CN113160234B (en) Unsupervised remote sensing image semantic segmentation method based on super-resolution and domain self-adaptation
CN112347970B (en) Remote sensing image ground object identification method based on graph convolution neural network
CN110363068B (en) High-resolution pedestrian image generation method based on multiscale circulation generation type countermeasure network
CN111652240B (en) CNN-based image local feature detection and description method
CN115272777B (en) Semi-supervised image analysis method for power transmission scene
CN112183448B (en) Method for dividing pod-removed soybean image based on three-level classification and multi-scale FCN
CN111833261A (en) Image super-resolution restoration method for generating countermeasure network based on attention
CN112529774A (en) Remote sensing simulation image generation method based on cycleGAN
CN114581552A (en) Gray level image colorizing method based on generation countermeasure network
CN116935043A (en) Typical object remote sensing image generation method based on multitasking countermeasure network
CN104036242A (en) Object recognition method based on convolutional restricted Boltzmann machine combining Centering Trick
CN112561782B (en) Method for improving reality degree of simulation picture of offshore scene
CN114049503A (en) Saliency region detection method based on non-end-to-end deep learning network
CN117437423A (en) Weak supervision medical image segmentation method and device based on SAM collaborative learning and cross-layer feature aggregation enhancement
CN113066074A (en) Visual saliency prediction method based on binocular parallax offset fusion
CN112330562A (en) Heterogeneous remote sensing image transformation method and system
Bai et al. Digital rock core images resolution enhancement with improved super resolution convolutional neural networks
Li et al. A new algorithm of vehicle license plate location based on convolutional neural network
CN113065407B (en) Financial bill seal erasing method based on attention mechanism and generation countermeasure network
CN114332491A (en) Saliency target detection algorithm based on feature reconstruction
CN114049567A (en) Self-adaptive soft label generation method and application in hyperspectral image classification
CN116343063B (en) Road network extraction method, system, equipment and computer readable storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information

Inventor after: Hou Yaqing

Inventor after: Zhang Qiang

Inventor after: Li Tianbo

Inventor after: Zhou Dongsheng

Inventor after: Wei Xiaopeng

Inventor before: Li Tianbo

Inventor before: Hou Yaqing

Inventor before: Zhou Dongsheng

Inventor before: Zhang Qiang

Inventor before: Wei Xiaopeng