CN113344896B - Breast CT image focus segmentation model training method and system - Google Patents
Breast CT image focus segmentation model training method and system Download PDFInfo
- Publication number
- CN113344896B CN113344896B CN202110705914.8A CN202110705914A CN113344896B CN 113344896 B CN113344896 B CN 113344896B CN 202110705914 A CN202110705914 A CN 202110705914A CN 113344896 B CN113344896 B CN 113344896B
- Authority
- CN
- China
- Prior art keywords
- image
- sample
- images
- training
- segmentation model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
- G06T7/0014—Biomedical image inspection using an image reference approach
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30096—Tumor; Lesion
Abstract
The invention discloses a semi-supervised chest CT image focus segmentation model training method and a system, wherein the method comprises the following steps: acquiring a sample CT image and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and an annotated-free image; according to the reliability of the sample CT image, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus; training a deep learning network according to the augmented CT image to obtain a focus segmentation model; and optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model. The invention provides a semi-supervised chest CT image lesion segmentation model training method, which retains lesion information, integrates the advantages of a fully supervised loss function and a common semi-supervised loss function, saves a large amount of labor cost and avoids the requirement of massive marking data.
Description
Technical Field
The invention relates to the technical field of lung CT image focus segmentation, in particular to a training method and a system of a breast CT image focus segmentation model.
Background
The focus segmentation of chest CT has important significance for diagnosing and treating new coronary pneumonia. Although the fully supervised algorithm can achieve a good segmentation effect, massive labeled data is required. However, lesion marking for breast CT requires a professional radiologist to do so, but such physicians are typically busy and have little off-hours. In addition, because the size and shape of the focus are different and the number of CT layers is large, which results in great labeling difficulty, a method for training a segmentation model according to a small amount of CT data for labeling the focus is urgently needed.
Thus, there is a need for improvements and enhancements in the art.
Disclosure of Invention
The invention provides a method and a system for training a breast CT image focus segmentation model, and aims to solve the problem that a method for training a segmentation model according to a small amount of CT data for marking focuses is lacked in the prior art.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows:
in a first aspect, the present invention provides a method for training a breast CT image lesion segmentation model, wherein the method includes:
acquiring a sample CT image and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and an annotated image;
according to the reliability of the sample CT image, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus;
training a deep learning network according to the augmented CT image to obtain a focus segmentation model;
and optimizing the self-learning strategy based on the teacher-student model to obtain the segmentation precision of the focus segmentation model.
In one implementation, the acquiring a sample CT image and calculating a confidence level of the sample CT image includes:
acquiring the sample CT image, and extracting semantic features of the sample CT image by using an encoder;
introducing noise features and disturbance factors into the semantic features to obtain the semantic features with noise and a corresponding probability map;
and calculating the average value of the probability map, and calculating the reliability of the CT image of each sample according to the average value of the probability map.
In one implementation, the acquiring a sample CT image and calculating the confidence level of the sample CT image further includes:
and obtaining a credibility mask according to the average value of the probability map and based on a threshold value.
In one implementation, the pairing CT images and generating an augmented CT image according to the confidence level of the sample CT image includes:
sequencing the sample CT images according to the credibility of the sample CT images;
matching the image with the highest reliability in the sample CT images with the image with the lowest reliability, matching the image with the second highest reliability in the sample CT images with the image with the second lowest reliability, and so on to complete the matching of the sample CT images;
and mixing the images containing the focuses by using the credibility mask to obtain the augmented CT image.
In one implementation, in the process of training the deep learning network according to the augmented CT image, the method for calculating the loss function includes:
L seg =L sup +βL sem
wherein, β is a dynamic parameter changing along with the training process, and the calculation mode is as follows:
wherein, b 1 And b 2 Two hyper-parameters are respectively set to be 0.1 and-5, and the value of beta is 0 to 0.1; l is sup Is a conventional loss function, L sem As a function of semi-supervised loss, e n Representing the number of steps of the current training, E max Representing the total number of steps of the training, the value being related to the data set, E max Taking the value of 200-300.
In one implementation, the conventional penalty function L sup The calculation method is as follows:
wherein L is ce Representing a cross-entropy function, F seg For the encoder, I is the sample CT image, I is the batch of sample CT images, M I N is a mask image of the sample CT images and indicates the number of sample CT images in the batch.
In one implementation, the conventional penalty function L sem The calculation method is as follows:
wherein K represents the logarithm of the image pair in the training batch;representing a blended image of the high confidence images in the k-th pair of images,a blended image mask representing a high confidence image in the kth pair of images,a blended image representing low confidence images in the kth pair of images,mixed image mask, L, representing low confidence images in the k-th pair of images ce Representing a cross entropy function.
In a second aspect, an embodiment of the present invention provides a training system for a semi-supervised breast CT image lesion segmentation model, where the system includes:
the uncertainty evaluation module is used for acquiring a sample CT image and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and an annotated-free image;
the image generation module is used for pairing the CT images according to the reliability of the sample CT image and generating an augmented CT image, wherein the CT image is an image only containing a focus;
the repairing network module is used for training a deep learning network according to the augmented CT image to obtain a focus segmentation model;
and the self-learning module is used for optimizing the segmentation precision of the focus segmentation model based on the self-learning strategy of the teacher-student model.
In a third aspect, an embodiment of the present invention further provides a terminal device, where the terminal device includes a memory, a processor, and a training program of a breast CT image lesion segmentation model stored in the memory and executable on the processor, and when the processor executes the training program of the breast CT image lesion segmentation model, the step of implementing the method for training a semi-supervised breast CT image lesion segmentation model according to any one of the above schemes is implemented.
In a fourth aspect, the present invention further provides a computer-readable storage medium, on which a training program of a breast CT image lesion segmentation model is stored, where the training program of the breast CT image lesion segmentation model is executed by a processor, and the steps of the method for training a semi-supervised breast CT image lesion segmentation model according to any one of the above schemes are implemented.
Has the advantages that: compared with the prior art, the invention provides a semi-supervised breast CT image lesion segmentation model training method. Firstly, a sample CT image is obtained, and the reliability of the sample CT image is calculated, wherein the sample CT image comprises an annotated image and an annotated image. And then according to the reliability of the sample CT image, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus. And then, training a deep learning network according to the augmented CT image to obtain a focus segmentation model. And finally, optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model. The invention provides a semi-supervised chest CT image lesion segmentation model training method, which retains lesion information, integrates the advantages of a fully supervised loss function and a common semi-supervised loss function, saves a large amount of labor cost and avoids the requirement of massive marking data.
Drawings
Fig. 1 is a flowchart of a method for training a lesion segmentation model of a semi-supervised chest CT image according to an embodiment of the present invention.
Fig. 2 is a schematic flowchart of a method for training a lesion segmentation model of a semi-supervised chest CT image according to an embodiment of the present invention.
Fig. 3 is a schematic diagram of obtaining reliability in a training method of a semi-supervised breast CT image lesion segmentation model according to an embodiment of the present invention.
Fig. 4 is a schematic block diagram of a training system of a lesion segmentation model of a semi-supervised chest CT image according to an embodiment of the present invention.
Fig. 5 is a schematic block diagram of an internal structure of a terminal device according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and effects of the present invention clearer and clearer, the present invention is further described in detail below with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
Research shows that the focus segmentation of chest CT has important significance for diagnosing and treating new coronary pneumonia. Although the fully supervised algorithm can achieve a good segmentation effect, massive labeled data is required. However, lesion marking for breast CT requires a professional radiologist to do so, but such physicians are typically busy and have little off-hours. In addition, because the size and shape of the focus are different and the number of CT layers is large, which results in great labeling difficulty, a method for training a segmentation model according to a small amount of CT data for labeling the focus is urgently needed.
In order to solve the above problem, the present embodiment provides a method for training a lesion segmentation model of a semi-supervised chest CT image. Firstly, a sample CT image is obtained, and the reliability of the sample CT image is calculated, wherein the sample CT image comprises an annotated image and an annotated image. And then according to the reliability of the sample CT images, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus. And then, training a deep learning network according to the augmented CT image to obtain a focus segmentation model. And finally, optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model. The invention provides a semi-supervised chest CT image focus segmentation model training method, which can train a focus segmentation model under the condition that the data volume labeled by a professional doctor is less, obtain a segmentation effect equivalent to that when a large amount of labeled data is used, retain focus information, integrate the advantages of a fully supervised loss function and a common semi-supervised loss function, save a large amount of labor cost and avoid the requirement of massive labeled data.
Exemplary method
As shown in fig. 1 and fig. 2, the method for training a semi-supervised breast CT image lesion segmentation model provided in this embodiment specifically includes the following steps:
s100, obtaining a sample CT image, and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and a non-annotated image.
In this embodiment, the sample CT image, that is, a batch of images in fig. 2, is obtained first, where the batch of images includes n unlabeled images and n labeled images. Semantic features of the sample CT image are then extracted using an encoder. And then introducing noise features and disturbance factors into the semantic features to obtain the semantic features with noise and a corresponding probability map. And finally, calculating the average value of the probability map, calculating the credibility of each Zhang Yangben CT image according to the probability map, and obtaining a credibility mask according to the average value of the probability map and a threshold value. Specifically, as shown in fig. 3, the present embodiment first uses the encoder F seg Providing semantic features f E R of input sample CT image I C×H×W Then introduce noise z i ∈R C×H×W Where i ∈ {1,2, …, T }, and T is a hyperparameter, which represents the number of noise perturbations added. Here, the noise is randomly collected from a uniform distribution U [ -0.3, +0.3]. After noise is introduced, noisy semantic features can be obtained
WhereinRepresenting the product of the elements. T different perturbations [ z ] 1 ,z 2 ,…,z T ]Will obtain T probability maps P 1 ,P 2 ,…,P T ]In which P is i =F seg (I;z t ). Next, the average of these probability maps is calculatedObtaining confidence mask M by threshold tau con :
Wherein the threshold tau is a dynamic parameter related to the number of training steps, and is calculated as follows,
wherein gamma is 1 、γ 2 And gamma 3 For the over-parameters, here set to 0.3, -5 and 0.5, respectively. And then calculating the credibility score s of each image to obtain the credibility of each sample CT image, wherein the calculation mode is as follows:
wherein, x represents the pixel in the sample CT image, | I | represents the number of pixels in the sample CT image I, the H function is an entropy calculation function, and the calculation formula is as follows: h (p) = -plog 2 p-(1-p)log 2 (1-p)。
And S200, matching the CT images according to the reliability of the sample CT image, and generating an augmented CT image, wherein the CT image is an image only containing a focus.
In this embodiment, the sample CT images are first sorted according to their reliability. And then, matching the image with the highest reliability in the sample CT images with the image with the lowest reliability, matching the image with the second highest reliability in the sample CT images with the image with the second lowest reliability, and so on to complete the matching of the sample CT images. And finally, mixing the images containing the focuses by using the credibility mask to obtain the augmented CT image. And forming an image pair by using the CT with low reliability and the CT with high reliability, and reserving the focus information of the image pair to generate a new mixed image.
In specific implementation, the reliability s of all images in each batch of CT sample images is calculated, then the sample CT images of the batch are sorted from small to large according to s, and the image with the highest reliability and the image with the lowest reliability form a pair. For example, the image with the highest confidence level s and the image with the lowest confidence level s are paired, and the image with the second highest confidence level s and the image with the second lowest confidence level s are paired. Followed by confidence mask M con Obtaining images of only mixed lesions
Where n denotes the two image indices in each pair of images.Indicating that an image of only the mixed lesion is obtained from image n. Alpha is alpha n Represents the blending weight of image n, here randomly acquired from a uniform distribution U (0,1). A blended image is then generated according to the following formula:
where m and n represent the subscripts of the two images in a pair, respectively, and if m is 1, then n is 2, and vice versa.Representing the generated blended image of image n.
Then, a mask of the mixed image is generated according to the following formula,
And step S300, training a deep learning network according to the augmented CT image to obtain a focus segmentation model.
In this embodiment, the deep learning network is trained according to the augmented data obtained above. For this purpose, we have designed a special loss function L seg ,
L seg =L sup +βL sem
Wherein, β is a dynamic parameter changing along with the training process, and the calculation mode is as follows:
wherein, b 1 And b 2 The values of the two hyper-parameters are respectively set to be 0.1 and-5, and the value of beta changes from 0 to 0.1 along with the increase of the training steps. L is sup Is a conventional loss function, L sem As a function of semi-supervised loss, e n Representing the number of steps of the current training, E max Representing the total number of steps of the training, the value being related to the data set, E max Taking the value of 200-300.L is a radical of an alcohol sup The calculation method of (c) is as follows:
wherein L is ce Representing a cross-entropy function, F seg For the encoder, I is the sample CT image, I is the batch of sample CT images, M I N is a mask image of the sample CT images and indicates the number of sample CT images in the batch.
L sem The calculation method of (c) is as follows:
where K represents the logarithm of the image pair in the training batch.Representing a blended image of the high confidence images in the k-th pair of images,a hybrid image mask representing the high confidence images in the kth pair of images,a blended image representing low confidence images in the kth pair of images,mixed image mask, L, representing low confidence images in the k-th pair of images ce Representing a cross entropy function.
And S400, optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model.
In the embodiment, in order to obtain a high-quality pseudo label during training, a Teacher-Student-based self-training process is introduced. The specific process is as follows:
inputting:
D l : labeled CT images;
D u : a CT image without annotation;
G l : an annotation mask for annotating the CT image;
G u : pseudo-label mask of the label-free CT image;
F t (·|θ t ): a Teacher model;
F s (·|θ s ): a Student model;
s1: initializing parameters, and enabling the parameter theta of the Teacher model t = parameter θ of Student model s Let step e =0;
S4: if the step number e is a multiple of K, executing S5-S10;
S11: S2-S10 max times are performed in sequence, unless there is no labeled CT data set D u If there is no more data in the data, the process is terminated.
Through the self-training process of the Teacher-Student, the segmentation effect of the focus segmentation model is improved.
To sum up, in this embodiment, a sample CT image is first obtained, and the reliability of the sample CT image is calculated, where the sample CT image includes an annotated image and an annotated image. And then according to the reliability of the sample CT image, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus. And then, training a deep learning network according to the augmented CT image to obtain a focus segmentation model. And finally, optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of the teacher-student model. The embodiment provides a training method of a breast CT image lesion segmentation model for semi-supervision, which retains lesion information, integrates the advantages of a fully-supervised loss function and a common semi-supervised loss function, saves a large amount of labor cost, and avoids the requirement of massive marking data.
Exemplary device
An embodiment of the present invention further provides a training system for a lesion segmentation model of a semi-supervised chest CT image, as shown in fig. 4, the system includes: uncertainty evaluation module 10, image generation module 20, remediation network module 30, self-learning module 40.
In specific implementation, the uncertainty evaluation module 10 is configured to acquire a sample CT image and calculate the reliability of the sample CT image, where the sample CT image includes an annotated image and an annotated-free image. Specifically, in the uncertainty evaluation module 10, the sample CT image is first acquired, and a semantic feature of the sample CT image is extracted using an encoder. And then introducing noise features and disturbance factors into the semantic features to obtain the semantic features with noise and a corresponding probability map. The mean of the probability maps is then calculated, and the confidence level of each Zhang Yangben CT image is calculated from the probability maps.
The image generation module 20 is configured to pair CT images according to the reliability of the sample CT image, and generate an augmented CT image, where the CT image is an image containing only a lesion. Specifically, the sample CT images are first sorted according to their confidence level. And then, matching the image with the highest reliability in the sample CT images with the image with the lowest reliability, matching the image with the second highest reliability in the sample CT images with the image with the second lowest reliability, and so on to complete the matching of the sample CT images. And mixing the images containing the focuses by using the credibility mask to obtain the augmented CT image.
The repairing network module 30 is configured to train a deep learning network according to the augmented CT image to obtain a focus segmentation model. The self-learning module 40 is used for optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model.
Based on the above embodiments, the present invention further provides a terminal device, and a schematic block diagram thereof may be as shown in fig. 5. The terminal equipment comprises a processor, a memory, a network interface, a display screen and a temperature sensor which are connected through a system bus. Wherein the processor of the terminal device is configured to provide computing and control capabilities. The memory of the terminal equipment comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The network interface of the terminal device is used for connecting and communicating with an external terminal through a network. The computer program is executed by a processor to implement a method for training a breast CT image lesion segmentation model. The display screen of the terminal equipment can be a liquid crystal display screen or an electronic ink display screen, and the temperature sensor of the terminal equipment is arranged in the terminal equipment in advance and used for detecting the operating temperature of the internal equipment.
It will be understood by those skilled in the art that the block diagram shown in fig. 5 is only a block diagram of a part of the structure related to the solution of the present invention, and does not constitute a limitation to the terminal device to which the solution of the present invention is applied, and a specific terminal device may include more or less components than those shown in the figure, or combine some components, or have a different arrangement of components.
In one embodiment, a terminal device is provided, where the terminal device includes a memory, a processor, and a program of a training method for a breast CT image lesion segmentation model stored in the memory and executable on the processor, and when the processor executes the program of the training method for the breast CT image lesion segmentation model, the following operation instructions are implemented:
acquiring a sample CT image and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and an annotated-free image;
according to the reliability of the sample CT image, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus;
training a deep learning network according to the augmented CT image to obtain a focus segmentation model;
and optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware related to instructions of a computer program, which can be stored in a non-volatile computer-readable storage medium, and when executed, the computer program can include the processes of the embodiments of the methods described above. Any reference to memory, storage, databases, or other media used in embodiments provided herein may include non-volatile and/or volatile memory. Non-volatile memory can include read-only memory (ROM), programmable ROM (PROM), electrically Programmable ROM (EPROM), electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double Data Rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous Link DRAM (SLDRAM), rambus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
In summary, the invention discloses a method and a system for training a lesion segmentation model of a semi-supervised chest CT image, wherein the method comprises the following steps: acquiring a sample CT image and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and an annotated-free image; according to the reliability of the sample CT image, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus; training a deep learning network according to the augmented CT image; and optimizing the segmentation precision based on a self-learning strategy of a teacher-student model to obtain a focus segmentation model. The invention provides a semi-supervised chest CT image lesion segmentation model training method, which retains lesion information, integrates the advantages of a full-supervised loss function and a common semi-supervised loss function, saves a large amount of labor cost and avoids the requirement of massive marking data.
Finally, it should be noted that: the above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.
Claims (7)
1. A method for training a semi-supervised breast CT image lesion segmentation model, which is characterized by comprising the following steps:
acquiring a sample CT image and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and an annotated-free image;
according to the reliability of the sample CT image, pairing the CT images and generating an augmented CT image, wherein the CT image is an image only containing a focus;
training a deep learning network according to the augmented CT image to obtain a focus segmentation model;
optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model;
the acquiring a sample CT image and calculating the credibility of the sample CT image comprises the following steps:
acquiring the sample CT images, and extracting semantic features of each sample CT image by using an encoder;
introducing noise features and disturbance factors into the semantic features to obtain the semantic features with noise and a corresponding probability map;
calculating the average value of the probability map, and calculating the reliability of each sample CT image according to the average value of the probability map;
obtaining a credibility mask according to the average value of the probability map and based on a threshold value;
the pairing of the CT images and the generation of the augmented CT image according to the credibility of the sample CT image comprises the following steps:
sequencing the sample CT images according to the credibility of the sample CT images;
matching the image with the highest reliability in the sample CT images with the image with the lowest reliability, matching the image with the second highest reliability in the sample CT images with the image with the second lowest reliability, and so on to complete the matching of the sample CT images;
mixing the images containing the focus by using the credibility mask to obtain the augmented CT image;
wherein the image containing the lesion is:
where n denotes the subscripts of the two images in each pair, M con Is a confidence mask;is a confidence mask obtained according to the image n;representing an image from which only mixed lesions are obtained from image n; alpha is alpha n Representing a blending weight for image n;
wherein, the threshold τ is a dynamic parameter related to the training step number, and the calculation method is as follows:
wherein gamma is 1 、γ 2 And gamma 3 The parameters are respectively set to be 0.3, -5 and 0.5;
P avg for T different perturbations [ z ] 1 ,z 2 ,…,z T ]Obtain T probability maps [ P 1 ,P 2 ,…,P T ]Average value of (d); p T =F seg (I;z t ),F seg Is an encoder;
an augmented CT image is generated according to the following formula:
2. The method for training the breast CT image lesion segmentation model according to claim 1, wherein in the process of training the deep learning network according to the augmented CT image, the method for calculating the loss function comprises the following steps:
L seg =L sup +βL sem
wherein, β is a dynamic parameter changing along with the training process, and the calculation mode is as follows:
wherein, b 1 And b 2 Two hyper-parameters are respectively set to be 0.1 and-5, and the value of beta is 0 to 0.1; l is a radical of an alcohol sup Is a conventional loss function, L sem As a function of semi-supervised loss, e n Representing the number of steps of the current training, E max Representing the total number of steps of the training, the value being related to the data set, E max Taking the value of 200-300.
3. The method as claimed in claim 2, wherein the conventional loss function L is a function of a lesion segmentation model of the chest CT image sup The calculation method is as follows:
wherein L is ce Representing a cross entropy function, F seg For the encoder, I is the sample CT image, I is the batch of sample CT images, M I N is a mask image of the sample CT images and indicates the number of sample CT images in the batch.
4. The method as claimed in claim 3, wherein the conventional loss function L is a function of a lesion segmentation model of the chest CT image sem The calculation method is as follows:
wherein K represents the logarithm of the image pair in the training batch;a blended image representing a high confidence image in the kth pair of images,a hybrid image mask representing the high confidence images in the kth pair of images,a blended image representing low confidence images in the kth pair of images,mixed image mask, L, representing low confidence images in the kth pair of images ce Representing a cross entropy function.
5. A system for training a lesion segmentation model of a semi-supervised chest CT image, the system comprising:
the uncertainty evaluation module is used for acquiring a sample CT image and calculating the reliability of the sample CT image, wherein the sample CT image comprises an annotated image and an annotated-free image;
the image generation module is used for pairing the CT images according to the reliability of the sample CT image and generating an augmented CT image, wherein the CT image is an image only containing a focus;
the repairing network module is used for training a deep learning network according to the augmented CT image to obtain a focus segmentation model;
the self-learning module is used for optimizing the segmentation precision of the focus segmentation model based on a self-learning strategy of a teacher-student model;
the uncertainty evaluation module comprises:
acquiring the sample CT images, and extracting semantic features of each sample CT image by using an encoder;
introducing noise features and disturbance factors into the semantic features to obtain the semantic features with noise and a corresponding probability map;
calculating the average value of the probability map, and calculating the reliability of each sample CT image according to the average value of the probability map;
obtaining a credibility mask according to the average value of the probability map and based on a threshold value;
the image generation module comprises:
sequencing the sample CT images according to the credibility of the sample CT images;
matching the image with the highest reliability in the sample CT images with the image with the lowest reliability, matching the image with the second highest reliability in the sample CT images with the image with the second lowest reliability, and so on to complete the matching of the sample CT images;
mixing the images containing the focus by using the credibility mask to obtain the augmented CT image;
wherein the image containing the lesion is:
where n denotes the subscripts of the two images in each pair, M con Is a confidence mask;is a confidence mask obtained from the image n;an image representing that only mixed lesions are obtained according to the image n; alpha is alpha n Representing a blending weight of image n;
wherein, the threshold τ is a dynamic parameter related to the training step number, and the calculation method is as follows:
wherein gamma is 1 、γ 2 And gamma 3 The parameters are respectively set to be 0.3, -5 and 0.5;
P avg for T different perturbations z 1 ,z 2 ,…,z t ]Obtain T probability maps [ P 1 ,P 2 ,…,P T ]Average value of (d); p T =F seg (I;z t ),F seg Is an encoder;
an augmented CT image is generated according to the following formula:
6. A terminal device, characterized in that the terminal device comprises a memory, a processor and a training program of a breast CT image lesion segmentation model stored in the memory and capable of running on the processor, and when the processor executes the training program of the breast CT image lesion segmentation model, the steps of the training method of the breast CT image lesion segmentation model as claimed in any one of claims 1 to 4 are implemented.
7. A computer-readable storage medium, on which a training program of a breast CT image lesion segmentation model is stored, and when the training program of the breast CT image lesion segmentation model is executed by a processor, the steps of the training method of the breast CT image lesion segmentation model according to any one of claims 1 to 4 are implemented.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110705914.8A CN113344896B (en) | 2021-06-24 | 2021-06-24 | Breast CT image focus segmentation model training method and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110705914.8A CN113344896B (en) | 2021-06-24 | 2021-06-24 | Breast CT image focus segmentation model training method and system |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113344896A CN113344896A (en) | 2021-09-03 |
CN113344896B true CN113344896B (en) | 2023-01-17 |
Family
ID=77478477
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110705914.8A Active CN113344896B (en) | 2021-06-24 | 2021-06-24 | Breast CT image focus segmentation model training method and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113344896B (en) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113763413B (en) * | 2021-09-30 | 2023-11-24 | 深圳大学 | Training method of image segmentation model, image segmentation method and storage medium |
CN114140639A (en) * | 2021-11-04 | 2022-03-04 | 杭州医派智能科技有限公司 | Deep learning-based renal blood vessel extreme urine pole classification method in image, computer equipment and computer readable storage medium |
CN114066871A (en) * | 2021-11-19 | 2022-02-18 | 江苏科技大学 | Method for training new coronary pneumonia focus region segmentation model |
CN114926471B (en) * | 2022-05-24 | 2023-03-28 | 北京医准智能科技有限公司 | Image segmentation method and device, electronic equipment and storage medium |
CN115147426B (en) * | 2022-09-06 | 2022-11-29 | 北京大学 | Model training and image segmentation method and system based on semi-supervised learning |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020143309A1 (en) * | 2019-01-09 | 2020-07-16 | 平安科技(深圳)有限公司 | Segmentation model training method, oct image segmentation method and apparatus, device and medium |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10430946B1 (en) * | 2019-03-14 | 2019-10-01 | Inception Institute of Artificial Intelligence, Ltd. | Medical image segmentation and severity grading using neural network architectures with semi-supervised learning techniques |
CN112150478B (en) * | 2020-08-31 | 2021-06-22 | 温州医科大学 | Method and system for constructing semi-supervised image segmentation framework |
CN112036335B (en) * | 2020-09-03 | 2023-12-26 | 南京农业大学 | Inverse convolution guided semi-supervised plant leaf disease identification and segmentation method |
CN112465111A (en) * | 2020-11-17 | 2021-03-09 | 大连理工大学 | Three-dimensional voxel image segmentation method based on knowledge distillation and countertraining |
CN112560964A (en) * | 2020-12-18 | 2021-03-26 | 深圳赛安特技术服务有限公司 | Method and system for training Chinese herbal medicine pest and disease identification model based on semi-supervised learning |
-
2021
- 2021-06-24 CN CN202110705914.8A patent/CN113344896B/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2020143309A1 (en) * | 2019-01-09 | 2020-07-16 | 平安科技(深圳)有限公司 | Segmentation model training method, oct image segmentation method and apparatus, device and medium |
Also Published As
Publication number | Publication date |
---|---|
CN113344896A (en) | 2021-09-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113344896B (en) | Breast CT image focus segmentation model training method and system | |
Zhang et al. | Learn from all: Erasing attention consistency for noisy label facial expression recognition | |
WO2020168934A1 (en) | Medical image segmentation method, apparatus, computer device, and storage medium | |
US11842487B2 (en) | Detection model training method and apparatus, computer device and storage medium | |
WO2020215984A1 (en) | Medical image detection method based on deep learning, and related device | |
CN111932561A (en) | Real-time enteroscopy image segmentation method and device based on integrated knowledge distillation | |
US11922654B2 (en) | Mammographic image processing method and apparatus, system and medium | |
CN113129309B (en) | Medical image semi-supervised segmentation system based on object context consistency constraint | |
CN110693486B (en) | Electrocardiogram abnormity labeling method and device | |
CN110797101B (en) | Medical data processing method, medical data processing device, readable storage medium and computer equipment | |
CN112232407B (en) | Neural network model training method and device for pathological image samples | |
CN110660055B (en) | Disease data prediction method and device, readable storage medium and electronic equipment | |
CN111275118B (en) | Chest film multi-label classification method based on self-correction type label generation network | |
CN115294086A (en) | Medical image segmentation method, segmentation model training method, medium, and electronic device | |
CN110570425B (en) | Pulmonary nodule analysis method and device based on deep reinforcement learning algorithm | |
CN114067233B (en) | Cross-mode matching method and system | |
CN109785940B (en) | Method for typesetting medical image film | |
Hu et al. | Unsupervised computed tomography and cone-beam computed tomography image registration using a dual attention network | |
CN116402838B (en) | Semi-supervised image segmentation method and system for intracranial hemorrhage | |
CN113435469A (en) | Kidney tumor enhanced CT image automatic identification system based on deep learning and training method thereof | |
Cai et al. | 3D Medical Image Segmentation with Sparse Annotation via Cross-Teaching Between 3D and 2D Networks | |
CN114972291A (en) | Medical image structured automatic labeling method and system based on hybrid enhanced intelligence | |
CN111178174B (en) | Urine formed component image identification method based on deep convolutional neural network | |
Ferber et al. | In-context learning enables multimodal large language models to classify cancer pathology images | |
CN110766652B (en) | Network training method, device, segmentation method, computer equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |