WO2024031219A1 - Image segmentation model training method, image segmentation method, and apparatus - Google Patents

Image segmentation model training method, image segmentation method, and apparatus Download PDF

Info

Publication number
WO2024031219A1
WO2024031219A1 PCT/CN2022/110779 CN2022110779W WO2024031219A1 WO 2024031219 A1 WO2024031219 A1 WO 2024031219A1 CN 2022110779 W CN2022110779 W CN 2022110779W WO 2024031219 A1 WO2024031219 A1 WO 2024031219A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
trained
segmentation
medical image
target object
Prior art date
Application number
PCT/CN2022/110779
Other languages
French (fr)
Chinese (zh)
Inventor
廖湘湘
许灿强
陈必超
李美
黎宇翔
张勇
Original Assignee
深圳华大生命科学研究院
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳华大生命科学研究院 filed Critical 深圳华大生命科学研究院
Priority to PCT/CN2022/110779 priority Critical patent/WO2024031219A1/en
Priority to PCT/CN2022/136698 priority patent/WO2024001051A1/en
Publication of WO2024031219A1 publication Critical patent/WO2024031219A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis

Definitions

  • This application belongs to the field of artificial intelligence technology, and specifically relates to a training method of an image segmentation model, an image segmentation method, device and equipment.
  • cell segmentation is the basic prerequisite for identifying and counting cell images.
  • segmentation of target objects in medical images can be implemented using deep learning-based segmentation algorithms.
  • current segmentation algorithms based on deep learning cannot handle the boundaries of target objects well, resulting in inaccurate segmentation results.
  • embodiments of the present application provide a training method for an image segmentation model, an image segmentation method, apparatus and equipment, so as to improve the segmentation accuracy of target objects.
  • a method for training an image segmentation model comprising:
  • the model includes an encoder module and a decoder module.
  • the encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
  • the pyramid squeeze attention module includes:
  • the slicing unit is used to divide the input features into N groups, perform convolution operations with different convolution kernel sizes on each group, generate N feature maps, merge the N feature maps, and generate a merged feature map.
  • N is a positive integer
  • An attention unit is used to perform an attention operation on the merged feature map and generate a feature map after the attention operation
  • the activation unit uses the softmax function to activate the feature map after the attention operation, performs a dot multiplication operation with the merged feature map, and outputs the final feature map.
  • the obtaining of the medical image to be trained and the segmentation result of the medical image to be trained includes:
  • the classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background is determined as a segmentation result of the medical image to be trained.
  • the image segmentation model is trained according to the medical image to be trained and the segmentation results of the medical image to be trained until a preset condition is reached, and a trained image segmentation model is obtained, including :
  • the loss value is calculated according to the predicted classification result of each pixel in the medical image to be trained and the classification label of each pixel in the medical image to be trained belonging to the target object, the target object outline or the image background.
  • the loss value is used to train the image segmentation model
  • the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the target object, and the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the image background. loss weight.
  • An image segmentation method includes:
  • the medical image to be segmented is input into the image segmentation model, and the initial segmentation result of each pixel in the medical image to be segmented is obtained.
  • the initial segmentation result includes belonging to the target object, the outline of the target object, or the image background;
  • the image segmentation model is based on It is trained by the above training method of image segmentation model;
  • the pixel points that the initial segmentation result belongs to the target object or the outline of the target object are determined as pixel points that belong to the target object, and the segmentation result of the medical image to be segmented is output.
  • the method further includes:
  • a watershed algorithm is used to filter target objects with an area smaller than a threshold in the segmentation results of the medical image to be segmented, and/or to repair incomplete target objects.
  • An image segmentation model training device includes:
  • An acquisition unit configured to acquire a medical image to be trained and a segmentation result of the medical image to be trained
  • a training unit configured to train an image segmentation model based on the medical image to be trained and the segmentation results of the medical image to be trained until a preset condition is reached, and a trained image segmentation model is obtained;
  • the model includes an encoder module and a decoder module.
  • the encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
  • An image segmentation device includes:
  • An input unit is used to input the medical image to be segmented into the image segmentation model, and obtain the initial segmentation result of each pixel in the medical image to be segmented, where the initial segmentation result includes the target object, the target object outline, or the image background;
  • the image segmentation model is trained according to the above image segmentation model training method;
  • a determination unit configured to determine the pixel points of the initial segmentation result belonging to the target object or the outline of the target object as pixel points belonging to the target object, and output the segmentation result of the medical image to be segmented.
  • a training device for an image segmentation model including: a memory, a processor, and a computer program stored in the memory and executable on the processor.
  • the processor executes the computer program, it implements the above Training method for image segmentation model.
  • An image segmentation device including: a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, the above image segmentation is implemented. method.
  • a computer-readable storage medium Instructions are stored in the computer-readable storage medium.
  • the terminal device causes the terminal device to execute the training method of the image segmentation model as mentioned above, or execute as The above image segmentation method.
  • the image segmentation model generated by training in the embodiment of this application consists of an encoder module and a decoder module.
  • the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units.
  • the efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, the pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module, making the image segmentation model pay more attention to target objects with salient features, thus improving the segmentation accuracy of the image segmentation model.
  • Figure 1 is a schematic diagram of an application scenario provided by an embodiment of this application.
  • Figure 2 is a flow chart of the training method of the image segmentation model provided by the embodiment of the present application.
  • Figure 3 is a schematic structural diagram of the image segmentation model in the embodiment of the present application.
  • Figure 4 is a schematic structural diagram of an efficient pyramid segmentation attention unit in an embodiment of the present application.
  • Figure 5 is a schematic structural diagram of the image segmentation model in the embodiment of the present application.
  • Figure 6 is a schematic structural diagram of the pyramid squeeze attention module in the embodiment of the present application.
  • Figure 7 is a flow chart of the image segmentation method provided in the embodiment of the present application.
  • Figure 8 is a diagram showing the segmentation effect of the image segmentation model of the embodiment of the present application on a mouse brain image
  • Figure 9 is a diagram of the segmentation effect of the image segmentation model on monkey brain images according to the embodiment of the present application.
  • Figure 10 is a diagram showing the segmentation effect of the image segmentation model according to the embodiment of the present application on a salamander brain image
  • Figure 11 shows the segmentation results of various algorithms on mouse brain images
  • Figure 12 is a schematic structural diagram of a training device for an image segmentation model provided by an embodiment of the present application.
  • Figure 13 is a schematic structural diagram of an image segmentation device provided by an embodiment of the present application.
  • segmenting target objects in medical images is one of the important fields in medical image processing, such as segmenting cells in medical images, segmenting organs and tissues in medical images, etc.
  • current segmentation algorithms based on deep learning cannot handle the boundaries of target objects well when segmenting target objects, resulting in inaccurate segmentation results.
  • cell segmentation is the basic prerequisite for identifying and counting cell images. Due to the complexity of cell segmentation images, uneven illumination of microscope images, and grayscale changes of the target object itself, there are still some problems in the segmented images, such as cell adhesion and overlap.
  • the image segmentation model generated by training consists of an encoder module and a decoder module.
  • the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units.
  • the efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, a pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module and make the image segmentation model pay more attention to target objects with salient features.
  • the target objects in medical images are segmented, which improves the accuracy of image segmentation.
  • Figure 1 is a schematic framework diagram of an exemplary application scenario provided by an embodiment of the present application.
  • the method can be applied to terminal devices or servers.
  • the image segmentation model can be trained based on the medical images to be trained and the segmentation results of the medical images to be trained.
  • the segmentation results of the medical image to be trained can be considered as label data, which can include classification labels indicating that each pixel in the medical image to be trained belongs to the target object, the outline of the target object, or the image background.
  • the prediction and classification results of each pixel in the medical image to be trained can be obtained.
  • the prediction classification results of each pixel in the medical image to be trained and the segmentation results of the medical image to be trained are used to calculate the loss value, and the image segmentation model is adjusted using the loss value.
  • the next medical image to be trained and the segmentation result of the medical image to be trained will be used to continue to adjust the image segmentation model until the training is completed, and the trained image segmentation model is obtained.
  • FIG. 1 is only an example in which the embodiments of the present application can be implemented.
  • the scope of application of the embodiments of this application is not limited by any aspect of this framework.
  • the image segmentation model training method may include steps S201-S202:
  • the medical images to be trained are images used to train the image segmentation model. They can be images under a microscope or images scanned by medical equipment, such as X-ray images, CT images, etc.
  • the type of medical images to be trained can be selected based on the target objects that the image segmentation model needs to segment. For example, if the target object is a cell, then the medical image to be trained is an image under a microscope; if the target object is an organ or tissue, then the medical image to be trained is an image scanned by a medical device.
  • the embodiment of this application does not limit the type of medical images to be trained.
  • the segmentation results of the medical images to be trained can be considered as the label data for training the image segmentation model.
  • the segmentation results of the medical image to be trained may include classification labels indicating that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background.
  • the segmentation results of the medical images to be trained can be manually annotated. However, in order to improve the efficiency of obtaining label data, the segmentation results of the medical images to be trained can also be obtained through a preset image segmentation algorithm.
  • the specific implementation of S201 to obtain the medical image to be trained and the segmentation results of the medical image to be trained may include A1-A3:
  • A1 Obtain the medical images to be trained.
  • A2 Use the preset image segmentation algorithm to extract the target object, target object outline and image background in the medical image to be trained.
  • the preset image segmentation algorithm can be used to segment the medical image to be trained, and it is obtained that each pixel in the medical image to be trained belongs to the target object, the outline of the target object, or the image background.
  • the medical image to be trained is an image under a microscope including cells
  • the deepcell and cellprofiler algorithms can be used to process the medical image to be trained, and extract the cells, cell outlines, and image background in the medical image to be trained.
  • A3 Determine the classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background as the segmentation result of the medical image to be trained.
  • the classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background can be determined as the segmentation result of the medical image to be trained, that is, the label data of the training image segmentation model is obtained.
  • segmentation results corresponding to a large number of medical images to be trained can be quickly obtained.
  • the image prediction segmentation result By inputting the medical image to be trained into the image segmentation model, the image prediction segmentation result can be obtained.
  • the image prediction segmentation result is compared with the segmentation result of the medical image to be trained, and the loss value is calculated.
  • the image segmentation model can be adjusted according to the loss value, specifically adjusting the image. Model parameters for the segmentation model.
  • the image segmentation model training ends, and the trained image segmentation model is obtained.
  • the medical image to be trained is input into the image segmentation model again, and the image segmentation model is adjusted according to the loss value until the preset conditions are reached.
  • the preset condition is reaching the preset training times.
  • the preset condition is that the loss value reaches a preset loss threshold. The preset training times and preset loss threshold can be set according to the actual situation, and there are no restrictions here.
  • S202 trains the image segmentation model based on the medical image to be trained and the segmentation results of the medical image to be trained until the preset conditions are reached.
  • the specific implementation of obtaining the trained image segmentation model may include B1- B3:
  • B1 Input the medical image to be trained into the image segmentation model and obtain the prediction classification results of each pixel in the medical image to be trained.
  • the image prediction segmentation result can be obtained, which specifically can be the prediction classification result of each pixel in the medical image to be trained.
  • Predicted classification results include belonging to the target object, target object outline, or image background.
  • B2 Calculate the loss value based on the predicted classification results of each pixel in the medical image to be trained and the classification label of each pixel in the medical image to be trained belonging to the target object, target object outline, or image background, and train the image segmentation model based on the loss value.
  • the medical image to be trained is input into the image segmentation model again, and the loss value is calculated based on the predicted classification results of each pixel in the medical image to be trained and the classification label of each pixel in the medical image to be trained belonging to the target object, the outline of the target object, or the image background.
  • the image segmentation model is trained until the preset conditions are reached, and the trained image segmentation model is obtained.
  • the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the target object
  • the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the image background.
  • the loss weight of the target object contour should be greater than the loss weight of the target object and the image background.
  • the weight loss of the target object and the image background are 0.2
  • the loss weight of the target object outline is 0.6.
  • the loss function for the target object and image background can be Cross Entropy
  • the loss function for the contour of the target object can be focal loss.
  • the structure of the image segmentation model is improved.
  • Figure 3 shows a schematic structural diagram of the image segmentation model in the embodiment of the present application.
  • the image segmentation model includes an encoder module and a decoder module.
  • the encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units that are connected in sequence; the efficient pyramid segmentation attention unit includes a sequentially connected The second convolutional layer, the pyramid squeeze attention module, and the third convolutional layer.
  • the Unet network structure is improved. Based on the Resnet network structure, the last fc (Full Connection, full connection) layer is removed, and each layer in the Resnet network structure is replaced with the Pyramid Squeezing Attention Module (PSA Module). A 3x3 convolution of the residual block.
  • the improved residual block is named Epsablock (ie, efficient pyramid segmentation attention unit). See Figure 4, which shows the schematic structural diagram of Epsablock.
  • the efficient pyramid segmentation attention unit (Epsablock) includes a second convolution layer, a pyramid squeeze attention module (PSA Module) and a third convolution layer that are connected in sequence. . Among them, the second convolution layer and the third convolution layer can be 1x1 convolution (Conv 1x1).
  • Epsablock replaces the original residual block as an improved resnet.
  • the improved resnet includes the first convolutional layer, the pooling layer, and multiple efficient pyramid segmentation attention units (Epsablock) connected in sequence.
  • the improved resnet is used as the encoder of the Unet network structure, and the decoder is not modified.
  • the improved Unet network structure is the image segmentation model of this application.
  • the pyramid squeeze attention module includes:
  • the slicing unit is used to divide the input features into N groups, perform convolution operations with different convolution kernel sizes on each group, generate N feature maps, merge the N feature maps, and generate a merged feature map; N is a positive integer;
  • the attention unit is used to perform attention operations on the merged feature maps and generate feature maps after attention operations
  • the activation unit uses the softmax function to activate the feature map after the attention operation, performs a dot multiplication operation with the merged feature map, and outputs the final feature map.
  • the slicing unit first divides the input features into N groups using 1x1 convolution, for example, N is 4. Convolution operations with different convolution kernel sizes are performed on each group, and the convolution kernel sizes of each group increase sequentially. For example, the convolution kernel sizes are 3, 5, 7, and 9 in order. After convolutions of different sizes, the generated N feature maps are merged to generate a merged feature map.
  • the output of the segmentation unit is passed through the attention unit to obtain the channel attention value, that is, the feature map after the attention operation is obtained.
  • the purpose of this is to obtain the attention weights of feature maps of different scales. In this way, contextual information of different scales is integrated and better pixel-level attention is generated.
  • the activation unit performs softmax normalization on the feature map after the attention operation, performs dot product operation (channel-wise product) with the merged feature map, and outputs the final feature map.
  • the PSA module allows the image segmentation model to pay more attention to targets with salient features.
  • the image segmentation model generated by training in the embodiment of this application consists of an encoder module and a decoder module.
  • the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units.
  • the efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, the pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module, making the image segmentation model pay more attention to target objects with salient features, thus improving the segmentation accuracy of the image segmentation model.
  • FIG. 7 is a flow chart of an image segmentation method provided by an embodiment of the present application. As shown in Figure 7, the method includes S701-S702:
  • S701 Input the medical image to be segmented into the image segmentation model, and obtain the initial segmentation results of each pixel in the medical image to be segmented.
  • the initial segmentation results include belonging to the target object, the outline of the target object, or the image background; wherein, the image segmentation model is based on the above
  • the image segmentation model is trained using the training method.
  • the probability value that each pixel in the medical image to be segmented belongs to the target object, the outline of the target object, or the image background can be obtained.
  • Each pixel in the medical image to be segmented is determined based on whether the probability value is greater than the threshold.
  • the initial segmentation result is to determine whether each pixel in the medical image to be segmented belongs to the target object, the target object outline or the image background.
  • the medical image to be segmented can be expanded into a three-channel (such as RGB three-channel) image, and the image segmentation model is input to obtain the initial segmentation result of each pixel in each channel image, that is, each pixel in each channel image
  • the points belong to the target object, the outline of the target object or the image background, and then the initial segmentation results of each pixel point in each channel image are combined to determine the initial segmentation result of each pixel point in the medical image to be segmented.
  • the image segmentation model is trained according to the training method of the image segmentation model provided in the above embodiments of the present application. For relevant description, please refer to the above embodiments and will not be described again here.
  • S702 Determine pixel points that the initial segmentation result belongs to the target object or the outline of the target object as pixel points belonging to the target object, and output the segmentation result of the medical image to be segmented.
  • Fusion of pixel points belonging to the target object or the contour of the target object to achieve segmentation of the target object that is, determining the pixel points belonging to the target object or the contour of the target object as pixel points belonging to the target object, performing image segmentation, and outputting the medical image to be segmented Segmentation results.
  • the watershed algorithm can also be used to filter the target objects whose area is smaller than the threshold in the segmentation results of the medical image to be segmented, and/or to repair the incomplete target objects.
  • the target object By filtering target objects with an area smaller than the threshold and/or repairing incomplete target objects, the target object is corrected, the post-processing of the segmentation results of the medical image to be segmented is completed, and the accuracy of image segmentation is improved.
  • the target object is cells
  • algorithms such as cell filtering and correction are used to filter very small cells and correct incomplete cells to make cell segmentation more accurate.
  • the image segmentation model used in the embodiments of this application adopts a pyramid squeeze attention module in the encoder module, which improves the feature extraction capability of the encoder module and enables the image segmentation model to pay more attention to target objects with salient features, thereby improving improve the segmentation accuracy of the image segmentation model.
  • Embodiments of the present application are applied to cell segmentation in spatiotemporal omics, as shown in Figure 8 , which demonstrates the segmentation effect of the image segmentation model of the embodiment of the present application on mouse brain images.
  • a part of the mouse brain image above has been intercepted for display.
  • the lower right image can be achieved using the image segmentation model proposed in the embodiment of this application. Separate independent cells from adherent cells and provide more reliable boundaries.
  • Figures 9 and 10 are respectively the segmentation results of the image segmentation model provided by the embodiment of the present application on monkey brain images and salamander brain images. Parts of the original images are cut out for display. It can be seen from the results that , the image segmentation model provided by the embodiments of this application can also process other images, indicating that the image segmentation model has generalization properties.
  • the left picture is a part of the mouse brain image.
  • the middle picture is the segmentation result of the mouse brain image using the original Unet network.
  • the right picture is the mouse brain image using the watershed algorithm.
  • the segmentation results of the Unet network and the image segmentation model proposed in the embodiment of the present application are similar, but in The Unet network's ability to process boundaries at adhesion cells is not as good as the image segmentation model proposed in the embodiments of this application.
  • the watershed algorithm has poor ability to handle boundaries and over-segments cells, forcing cells to be divided into multiple cells.
  • Table 1 is a comparison of the quantitative indicators of the Unet network and the image segmentation model proposed in the embodiment of the present application on five mouse brain images.
  • Quantitative indicators include dice coefficient, IoU (Intersection over Union), precision (precision rate) and recall (recall rate).
  • the segmentation effect of the image segmentation model proposed in the embodiment of this application is better than the Unet network and the watershed algorithm.
  • the embodiment of the present application also provides a training device of the image segmentation model.
  • the training device of the image segmentation model will be described below with reference to the accompanying drawings.
  • Figure 12 is a schematic structural diagram of an image segmentation model training device provided by an embodiment of the present application.
  • the training device of the image segmentation model includes:
  • the acquisition unit 1201 is used to acquire the medical image to be trained and the segmentation result of the medical image to be trained;
  • the training unit 1202 is configured to train the image segmentation model according to the medical image to be trained and the segmentation results of the medical image to be trained until the preset conditions are reached, and the trained image segmentation model is obtained;
  • the model includes an encoder module and a decoder module.
  • the encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
  • the pyramid squeeze attention module includes:
  • the slicing unit is used to divide the input features into N groups, perform convolution operations with different convolution kernel sizes on each group, generate N feature maps, merge the N feature maps, and generate a merged feature map.
  • N is a positive integer
  • An attention unit is used to perform an attention operation on the merged feature map and generate a feature map after the attention operation
  • the activation unit uses the softmax function to activate the feature map after the attention operation, performs a dot multiplication operation with the merged feature map, and outputs the final feature map.
  • the acquisition unit includes:
  • Acquisition subunit used to obtain medical images to be trained
  • An extraction subunit is used to extract the target object, target object contour and image background in the medical image to be trained using a preset image segmentation algorithm
  • Determining subunit configured to determine the classification label that each pixel of the medical image to be trained belongs to the target object, the target object outline, or the image background as a segmentation result of the medical image to be trained.
  • the training unit includes:
  • An input subunit used to input the medical image to be trained into an image segmentation model to obtain the prediction classification results of each pixel in the medical image to be trained;
  • a training subunit configured to calculate based on the predicted classification results of each pixel in the medical image to be trained and the classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background. Loss value, the image segmentation model is trained according to the loss value;
  • Loop subunit used to return to the input subunit and the training subunit to execute the input of the medical image to be trained into the image segmentation model, obtain the prediction classification results of each pixel in the medical image to be trained, and subsequent steps, until Reach the preset conditions and obtain the trained image segmentation model;
  • the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the target object, and the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the image background. loss weight.
  • the embodiment of the present application also provides an image segmentation device.
  • the image segmentation device will be described below with reference to the accompanying drawings.
  • Figure 13 is a schematic structural diagram of an image segmentation device provided by an embodiment of the present application. As shown in Figure 13, the image segmentation device includes:
  • the input unit 1301 is used to input the medical image to be segmented into the image segmentation model, and obtain the initial segmentation result of each pixel point in the medical image to be segmented, where the initial segmentation result includes the target object, the outline of the target object, or the image background;
  • the image segmentation model is trained according to the above image segmentation model training method;
  • the determination unit 1302 is configured to determine the pixel points of the initial segmentation result belonging to the target object or the outline of the target object as pixel points belonging to the target object, and output the segmentation result of the medical image to be segmented.
  • the device further includes:
  • a correction unit configured to use a watershed algorithm to filter target objects with an area smaller than a threshold in the segmentation results of the medical image to be segmented, and/or to repair incomplete target objects.
  • embodiments of the present application also provide an image segmentation model training device, including: a memory, a processor, and a computer program stored in the memory and executable on the processor.
  • the processor executes When the computer program is used, the above-mentioned training method of the image segmentation model is implemented.
  • An embodiment of the present application also provides an image segmentation device, including: a memory, a processor, and a computer program stored on the memory and executable on the processor.
  • an image segmentation device including: a memory, a processor, and a computer program stored on the memory and executable on the processor.
  • the processor executes the computer program , implement the image segmentation method as mentioned above.
  • Embodiments of the present application also provide a computer-readable storage medium. Instructions are stored in the computer-readable storage medium. When the instructions are run on a terminal device, the terminal device executes the image segmentation model as described above. training method, or perform the image segmentation method as above.
  • the image segmentation model generated by training in the embodiment of this application consists of an encoder module and a decoder module.
  • the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units.
  • the efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, the pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module, making the image segmentation model pay more attention to target objects with salient features, thus improving the segmentation accuracy of the image segmentation model.
  • At least one (item) refers to one or more, and “plurality” refers to two or more.
  • “And/or” is used to describe the relationship between associated objects, indicating that there can be three relationships. For example, “A and/or B” can mean: only A exists, only B exists, and A and B exist simultaneously. , where A and B can be singular or plural. The character “/” generally indicates that the related objects are in an "or” relationship. “At least one of the following” or similar expressions thereof refers to any combination of these items, including any combination of a single item (items) or a plurality of items (items).
  • At least one of a, b or c can mean: a, b, c, "a and b", “a and c", “b and c", or "a and b and c” ”, where a, b, c can be single or multiple.
  • RAM random access memory
  • ROM read-only memory
  • electrically programmable ROM electrically erasable programmable ROM
  • registers hard disks, removable disks, CD-ROMs, or anywhere in the field of technology. any other known form of storage media.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

Provided in the present application are an image segmentation model training method, an image segmentation method, an apparatus, and a device. The training method comprises: acquiring medical images for training and segmentation results of said medical images; and according to said medical images and the segmentation results of said medical images, training an image segmentation model until a preset condition is met, and obtaining the trained image segmentation model. The model comprises an encoder module and a decoder module; the encoder module comprises, connected in sequence, a first convolutional layer, a pooling layer and a plurality of efficient pyramid split attention units; each efficient pyramid split attention unit comprises, connected in sequence, a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer. By using the pyramid squeeze attention module in the encoder module, the feature extraction capability of the encoder module is improved, so that the image segmentation model can pay more attention to a target object having a saliency feature, thus improving the segmentation accuracy of the image segmentation model.

Description

一种图像分割模型的训练方法、图像分割方法及装置An image segmentation model training method, image segmentation method and device 技术领域Technical field
本申请属于人工智能技术领域,具体涉及一种图像分割模型的训练方法、一种图像分割方法、装置及设备。This application belongs to the field of artificial intelligence technology, and specifically relates to a training method of an image segmentation model, an image segmentation method, device and equipment.
背景技术Background technique
目前,对医学图像中目标对象进行分割是医学图像处理中的重要领域之一,例如,细胞分割是对细胞图像进行识别和计数的基本前提。Currently, segmenting target objects in medical images is one of the important fields in medical image processing. For example, cell segmentation is the basic prerequisite for identifying and counting cell images.
在相关技术中,对医学图像中目标对象进行分割可以基于深度学习的分割算法实现。但是,目前基于深度学习的分割算法不能很好处理目标对象的边界,导致分割结果不准确。In related technologies, segmentation of target objects in medical images can be implemented using deep learning-based segmentation algorithms. However, current segmentation algorithms based on deep learning cannot handle the boundaries of target objects well, resulting in inaccurate segmentation results.
发明内容Contents of the invention
有鉴于此,本申请实施例提供一种图像分割模型的训练方法、一种图像分割方法、装置及设备,以提高目标对象的分割准确性。In view of this, embodiments of the present application provide a training method for an image segmentation model, an image segmentation method, apparatus and equipment, so as to improve the segmentation accuracy of target objects.
为解决上述问题,本申请实施例提供的技术方案如下:In order to solve the above problems, the technical solutions provided by the embodiments of this application are as follows:
一种图像分割模型的训练方法,所述方法包括:A method for training an image segmentation model, the method comprising:
获取待训练医学图像以及所述待训练医学图像的分割结果;Obtaining a medical image to be trained and a segmentation result of the medical image to be trained;
根据所述待训练医学图像以及所述待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型;Train the image segmentation model according to the medical image to be trained and the segmentation results of the medical image to be trained until the preset conditions are reached, and a trained image segmentation model is obtained;
所述模型包括编码器模块以及解码器模块,所述编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元;所述高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。The model includes an encoder module and a decoder module. The encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
在一种可能的实现方式中,所述金字塔挤压注意力模块包括:In a possible implementation, the pyramid squeeze attention module includes:
切片单元,用于将输入特征划分为N组,对每一组进行不同卷积核大小的卷积操作,生成N个特征图,对所述N个特征图进行合并,生成合并后的特征图;N为正整数;The slicing unit is used to divide the input features into N groups, perform convolution operations with different convolution kernel sizes on each group, generate N feature maps, merge the N feature maps, and generate a merged feature map. ;N is a positive integer;
注意力单元,用于将所述合并后的特征图进行注意力操作,生成注意力操作后的特征图;An attention unit is used to perform an attention operation on the merged feature map and generate a feature map after the attention operation;
激活单元,使用softmax函数激活所述注意力操作后的特征图,与所述合并后的特征图进行点乘操作,输出最终特征图。The activation unit uses the softmax function to activate the feature map after the attention operation, performs a dot multiplication operation with the merged feature map, and outputs the final feature map.
在一种可能的实现方式中,所述获取待训练医学图像以及所述待训练医学图像的分割结果,包括:In a possible implementation, the obtaining of the medical image to be trained and the segmentation result of the medical image to be trained includes:
获取待训练医学图像;Obtain medical images to be trained;
采用预设图像分割算法提取所述待训练医学图像中的目标对象、目标对象轮廓以及图像背景;Using a preset image segmentation algorithm to extract the target object, target object outline and image background in the medical image to be trained;
将所述待训练医学图像各个像素属于所述目标对象、所述目标对象轮廓或者所述图像背景的分类标签确定为所述待训练医学图像的分割结果。The classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background is determined as a segmentation result of the medical image to be trained.
在一种可能的实现方式中,所述根据所述待训练医学图像以及所述待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型,包括:In a possible implementation, the image segmentation model is trained according to the medical image to be trained and the segmentation results of the medical image to be trained until a preset condition is reached, and a trained image segmentation model is obtained, including :
将所述待训练医学图像输入图像分割模型,获得所述待训练医学图像中各个像素的预测分类结果;Input the medical image to be trained into an image segmentation model to obtain the prediction classification results of each pixel in the medical image to be trained;
根据所述待训练医学图像中各个像素的预测分类结果以及所述待训练医学图像各个像素属于所述目标对象、所述目标对象轮廓或者所述图像背景的分类标签,计算损失值,根据所述损失值对图像分割模型进行训练;The loss value is calculated according to the predicted classification result of each pixel in the medical image to be trained and the classification label of each pixel in the medical image to be trained belonging to the target object, the target object outline or the image background. The loss value is used to train the image segmentation model;
重复执行将所述待训练医学图像输入图像分割模型,获得所述待训练医学图像中各个像素的预测分类结果以及后续步骤,直到达到预设条件,获得训练完成的图像分割模型;Repeatedly executing the input of the medical image to be trained into the image segmentation model to obtain the prediction classification results of each pixel in the medical image to be trained and subsequent steps until the preset conditions are reached and the image segmentation model that has been trained is obtained;
其中,在计算损失值时,分类标签为所述目标对象轮廓的损失权重大于分类标签为所述目标对象的损失权重,分类标签为所述目标对象轮廓的损失权重大于分类标签为所述图像背景的损失权重。Wherein, when calculating the loss value, the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the target object, and the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the image background. loss weight.
一种图像分割方法,所述方法包括:An image segmentation method, the method includes:
将待分割医学图像输入图像分割模型,获得所述待分割医学图像中各个像素点的初始分割结果,所述初始分割结果包括属于目标对象、目标对象轮廓或者图像背景;所述图像分割模型是根据上述的图像分割模型的训练方法训 练得到的;The medical image to be segmented is input into the image segmentation model, and the initial segmentation result of each pixel in the medical image to be segmented is obtained. The initial segmentation result includes belonging to the target object, the outline of the target object, or the image background; the image segmentation model is based on It is trained by the above training method of image segmentation model;
将所述初始分割结果属于目标对象或者目标对象轮廓的像素点确定为属于目标对象的像素点,输出所述待分割医学图像的分割结果。The pixel points that the initial segmentation result belongs to the target object or the outline of the target object are determined as pixel points that belong to the target object, and the segmentation result of the medical image to be segmented is output.
在一种可能的实现方式中,所述方法还包括:In a possible implementation, the method further includes:
利用分水岭算法过滤所述待分割医学图像的分割结果中面积小于阈值的目标对象,和/或,对不完整的目标对象进行修复。A watershed algorithm is used to filter target objects with an area smaller than a threshold in the segmentation results of the medical image to be segmented, and/or to repair incomplete target objects.
一种图像分割模型的训练装置,所述装置包括:An image segmentation model training device, the device includes:
获取单元,用于获取待训练医学图像以及所述待训练医学图像的分割结果;An acquisition unit, configured to acquire a medical image to be trained and a segmentation result of the medical image to be trained;
训练单元,用于根据所述待训练医学图像以及所述待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型;A training unit configured to train an image segmentation model based on the medical image to be trained and the segmentation results of the medical image to be trained until a preset condition is reached, and a trained image segmentation model is obtained;
所述模型包括编码器模块以及解码器模块,所述编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元;所述高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。The model includes an encoder module and a decoder module. The encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
一种图像分割装置,所述装置包括:An image segmentation device, the device includes:
输入单元,用于将待分割医学图像输入图像分割模型,获得所述待分割医学图像中各个像素点的初始分割结果,所述初始分割结果包括属于目标对象、目标对象轮廓或者图像背景;所述图像分割模型是根据上述的图像分割模型的训练方法训练得到的;An input unit is used to input the medical image to be segmented into the image segmentation model, and obtain the initial segmentation result of each pixel in the medical image to be segmented, where the initial segmentation result includes the target object, the target object outline, or the image background; The image segmentation model is trained according to the above image segmentation model training method;
确定单元,用于将所述初始分割结果属于目标对象或者目标对象轮廓的像素点确定为属于目标对象的像素点,输出所述待分割医学图像的分割结果。A determination unit configured to determine the pixel points of the initial segmentation result belonging to the target object or the outline of the target object as pixel points belonging to the target object, and output the segmentation result of the medical image to be segmented.
一种图像分割模型的训练设备,包括:存储器,处理器,及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如上述的图像分割模型的训练方法。A training device for an image segmentation model, including: a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, it implements the above Training method for image segmentation model.
一种图像分割设备,包括:存储器,处理器,及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如上述的图像分割方法。An image segmentation device, including: a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, the above image segmentation is implemented. method.
一种计算机可读存储介质,所述计算机可读存储介质中存储有指令,当所述指令在终端设备上运行时,使得所述终端设备执行如上述的图像分割模型的训练方法,或者执行如上述的图像分割方法。A computer-readable storage medium. Instructions are stored in the computer-readable storage medium. When the instructions are run on a terminal device, the terminal device causes the terminal device to execute the training method of the image segmentation model as mentioned above, or execute as The above image segmentation method.
由此可见,本申请实施例具有如下有益效果:It can be seen that the embodiments of the present application have the following beneficial effects:
本申请实施例训练生成的图像分割模型由编码器模块和解码器模块组成。其中,编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元。高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。即在编码器模块中采用了金字塔挤压注意力模块,提高了编码器模块的特征提取能力,使图像分割模型更加关注具有显著性特征的目标对象,从而提高了图像分割模型的分割准确性。The image segmentation model generated by training in the embodiment of this application consists of an encoder module and a decoder module. Among them, the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units. The efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, the pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module, making the image segmentation model pay more attention to target objects with salient features, thus improving the segmentation accuracy of the image segmentation model.
附图说明Description of drawings
图1为本申请实施例提供的应用场景示意图;Figure 1 is a schematic diagram of an application scenario provided by an embodiment of this application;
图2为本申请实施例提供的图像分割模型的训练方法的流程图;Figure 2 is a flow chart of the training method of the image segmentation model provided by the embodiment of the present application;
图3为本申请实施例中图像分割模型的结构示意图;Figure 3 is a schematic structural diagram of the image segmentation model in the embodiment of the present application;
图4为本申请实施例中高效金字塔切分注意力单元的结构示意图;Figure 4 is a schematic structural diagram of an efficient pyramid segmentation attention unit in an embodiment of the present application;
图5为本申请实施例中图像分割模型的结构示意图;Figure 5 is a schematic structural diagram of the image segmentation model in the embodiment of the present application;
图6为本申请实施例中金字塔挤压注意力模块的结构示意图;Figure 6 is a schematic structural diagram of the pyramid squeeze attention module in the embodiment of the present application;
图7为本申请实施例中提供的图像分割方法的流程图;Figure 7 is a flow chart of the image segmentation method provided in the embodiment of the present application;
图8为本申请实施例的图像分割模型在鼠脑图像上的分割效果图;Figure 8 is a diagram showing the segmentation effect of the image segmentation model of the embodiment of the present application on a mouse brain image;
图9为本申请实施例的图像分割模型在猴脑图像上的分割效果图;Figure 9 is a diagram of the segmentation effect of the image segmentation model on monkey brain images according to the embodiment of the present application;
图10为本申请实施例的图像分割模型在蝾螈脑图像上的分割效果图;Figure 10 is a diagram showing the segmentation effect of the image segmentation model according to the embodiment of the present application on a salamander brain image;
图11为多种算法在鼠脑图像上的分割结果图;Figure 11 shows the segmentation results of various algorithms on mouse brain images;
图12为本申请实施例提供的图像分割模型的训练装置的结构示意图;Figure 12 is a schematic structural diagram of a training device for an image segmentation model provided by an embodiment of the present application;
图13为本申请实施例提供的图像分割装置的结构示意图。Figure 13 is a schematic structural diagram of an image segmentation device provided by an embodiment of the present application.
具体实施方式Detailed ways
为使本申请的上述目的、特征和优点能够更加明显易懂,下面结合附图和具体实施方式对本申请实施例作进一步详细的说明。In order to make the above objects, features and advantages of the present application more obvious and understandable, the embodiments of the present application will be further described in detail below in conjunction with the accompanying drawings and specific implementation modes.
为便于理解本申请实施例提供的技术方案,下面将先对本申请实施例涉及的背景技术进行说明。In order to facilitate understanding of the technical solutions provided by the embodiments of the present application, the background technology involved in the embodiments of the present application will be described below.
目前,对医学图像中目标对象进行分割是医学图像处理中的重要领域之一,例如,对医学图像中的细胞进行分割、对医学图像中的器官、组织进行分割等。但是,目前基于深度学习的分割算法在进行目标对象分割时,不能很好处理目标对象的边界,导致分割结果不准确。Currently, segmenting target objects in medical images is one of the important fields in medical image processing, such as segmenting cells in medical images, segmenting organs and tissues in medical images, etc. However, current segmentation algorithms based on deep learning cannot handle the boundaries of target objects well when segmenting target objects, resulting in inaccurate segmentation results.
以细胞分割为例进行说明,细胞分割是对细胞图像进行识别和计数的基本前提。由于细胞分割图像的复杂性以及显微镜图像光照不均匀,目标物体本身灰度变化等问题,分割后的图像中还存在一些问题,存在细胞黏连、重叠等现象。Taking cell segmentation as an example to illustrate, cell segmentation is the basic prerequisite for identifying and counting cell images. Due to the complexity of cell segmentation images, uneven illumination of microscope images, and grayscale changes of the target object itself, there are still some problems in the segmented images, such as cell adhesion and overlap.
基于此,本申请实施例提供了一种图像分割模型的训练方法、图像分割方法、装置及设备,训练生成的图像分割模型由编码器模块和解码器模块组成。其中,编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元。高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。即在编码器模块中采用了金字塔挤压注意力模块,提高了编码器模块的特征提取能力,使图像分割模型更加关注具有显著性特征的目标对象。基于深度学习注意力机制实现对医学图像中目标对象的分割,提高了图像分割的准确性。Based on this, embodiments of the present application provide an image segmentation model training method, image segmentation method, device and equipment. The image segmentation model generated by training consists of an encoder module and a decoder module. Among them, the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units. The efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, a pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module and make the image segmentation model pay more attention to target objects with salient features. Based on the deep learning attention mechanism, the target objects in medical images are segmented, which improves the accuracy of image segmentation.
为了便于理解本申请实施例提供的一种图像分割模型的训练方法、图像分割方法,下面结合图1所示的场景示例进行说明。参见图1,该图为本申请实施例提供的示例性应用场景的框架示意图。作为一种可选示例,该方法可应用于终端设备或服务器。In order to facilitate understanding of an image segmentation model training method and image segmentation method provided by embodiments of the present application, description is given below with reference to the scene example shown in Figure 1 . Refer to Figure 1, which is a schematic framework diagram of an exemplary application scenario provided by an embodiment of the present application. As an optional example, the method can be applied to terminal devices or servers.
在实际应用中,可以根据待训练医学图像以及待训练医学图像的分割结果,对图像分割模型进行训练。在一种可能的实现方式中,待训练医学图像的分割结果可以认为是标签数据,可以包括待训练医学图像中各个像素属于目标对象、目标对象轮廓或者图像背景的分类标签。将待训练医学图像输入图像分割模型,可以获得待训练医学图像中各个像素的预测分类结果。利用待训练医 学图像中各个像素的预测分类结果以及待训练医学图像的分割结果,计算损失值,利用损失值调整图像分割模型。再将利用下一条待训练医学图像以及待训练医学图像的分割结果继续调整图像分割模型,直到训练完成,获得训练完成的图像分割模型。In practical applications, the image segmentation model can be trained based on the medical images to be trained and the segmentation results of the medical images to be trained. In one possible implementation, the segmentation results of the medical image to be trained can be considered as label data, which can include classification labels indicating that each pixel in the medical image to be trained belongs to the target object, the outline of the target object, or the image background. By inputting the medical image to be trained into the image segmentation model, the prediction and classification results of each pixel in the medical image to be trained can be obtained. The prediction classification results of each pixel in the medical image to be trained and the segmentation results of the medical image to be trained are used to calculate the loss value, and the image segmentation model is adjusted using the loss value. Then, the next medical image to be trained and the segmentation result of the medical image to be trained will be used to continue to adjust the image segmentation model until the training is completed, and the trained image segmentation model is obtained.
本领域技术人员可以理解,图1所示的框架示意图仅是本申请的实施方式可以在其中得以实现的一个示例。本申请实施方式的适用范围不受到该框架任何方面的限制。Those skilled in the art can understand that the schematic framework diagram shown in FIG. 1 is only an example in which the embodiments of the present application can be implemented. The scope of application of the embodiments of this application is not limited by any aspect of this framework.
基于上述说明,下面将结合附图对本申请实施例提供的图像分割模型的训练方法以及图像分割方法进行详细说明。Based on the above description, the training method of the image segmentation model and the image segmentation method provided by the embodiments of the present application will be described in detail below with reference to the accompanying drawings.
参见图2,该图为本申请实施例提供的一种图像分割模型的训练方法的流程图,如图2所示,该图像分割模型的训练方法可以包括步骤S201-S202:Refer to Figure 2, which is a flow chart of an image segmentation model training method provided by an embodiment of the present application. As shown in Figure 2, the image segmentation model training method may include steps S201-S202:
S201:获取待训练医学图像以及待训练医学图像的分割结果。S201: Obtain the medical image to be trained and the segmentation results of the medical image to be trained.
待训练医学图像是用于训练图像分割模型的图像,可以为显微镜下的图像、通过医疗设备扫描得到的图像,如X光图像、CT图像等。待训练医学图像的类型可以根据图像分割模型需要分割的目标对象进行选取。例如,目标对象为细胞,则待训练医学图像为显微镜下的图像,目标对象为某器官或组织,则待训练医学图像为通过医疗设备扫描得到的图像。本申请实施例对待训练医学图像的类型不进行限定。The medical images to be trained are images used to train the image segmentation model. They can be images under a microscope or images scanned by medical equipment, such as X-ray images, CT images, etc. The type of medical images to be trained can be selected based on the target objects that the image segmentation model needs to segment. For example, if the target object is a cell, then the medical image to be trained is an image under a microscope; if the target object is an organ or tissue, then the medical image to be trained is an image scanned by a medical device. The embodiment of this application does not limit the type of medical images to be trained.
待训练医学图像的分割结果可以认为是训练图像分割模型的标签数据。待训练医学图像的分割结果可以包括待训练医学图像各个像素属于目标对象、目标对象轮廓或者图像背景的分类标签。待训练医学图像的分割结果可以通过人工标注,但是,为了提高获取标签数据的效率,也可以通过预设图像分割算法实现待训练医学图像的分割结果的获取。The segmentation results of the medical images to be trained can be considered as the label data for training the image segmentation model. The segmentation results of the medical image to be trained may include classification labels indicating that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background. The segmentation results of the medical images to be trained can be manually annotated. However, in order to improve the efficiency of obtaining label data, the segmentation results of the medical images to be trained can also be obtained through a preset image segmentation algorithm.
则在一种可能的实现方式中,S201获取待训练医学图像以及待训练医学图像的分割结果的具体实现可以包括A1-A3:In one possible implementation, the specific implementation of S201 to obtain the medical image to be trained and the segmentation results of the medical image to be trained may include A1-A3:
A1:获取待训练医学图像。A1: Obtain the medical images to be trained.
A2:采用预设图像分割算法提取待训练医学图像中的目标对象、目标对象轮廓以及图像背景。A2: Use the preset image segmentation algorithm to extract the target object, target object outline and image background in the medical image to be trained.
在获取待训练医学图像之后,可以利用预设图像分割算法对待训练医 学图像进行分割,得到待训练医学图像中每个像素属于目标对象、目标对象轮廓或是图像背景。例如,如果待训练医学图像是包括细胞的显微镜下的图像,则可以使用deepcell和cellprofiler算法处理待训练医学图像,提取待训练医学图像中的细胞、细胞的轮廓以及图像背景。After obtaining the medical image to be trained, the preset image segmentation algorithm can be used to segment the medical image to be trained, and it is obtained that each pixel in the medical image to be trained belongs to the target object, the outline of the target object, or the image background. For example, if the medical image to be trained is an image under a microscope including cells, the deepcell and cellprofiler algorithms can be used to process the medical image to be trained, and extract the cells, cell outlines, and image background in the medical image to be trained.
A3:将待训练医学图像各个像素属于目标对象、目标对象轮廓或者图像背景的分类标签确定为待训练医学图像的分割结果。A3: Determine the classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background as the segmentation result of the medical image to be trained.
在对待训练医学图像进行分割后,可以将将待训练医学图像各个像素属于目标对象、目标对象轮廓或者图像背景的分类标签确定为待训练医学图像的分割结果,即得到训练图像分割模型的标签数据。After segmenting the medical image to be trained, the classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background can be determined as the segmentation result of the medical image to be trained, that is, the label data of the training image segmentation model is obtained. .
通过预设图像分割算法确定待训练医学图像的分割结果,可以快速获取大量待训练医学图像对应的分割结果。By determining the segmentation results of the medical images to be trained by preset image segmentation algorithms, segmentation results corresponding to a large number of medical images to be trained can be quickly obtained.
S202:根据待训练医学图像以及待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型。S202: Train the image segmentation model based on the medical image to be trained and the segmentation results of the medical image to be trained until the preset conditions are reached, and the trained image segmentation model is obtained.
将待训练医学图像输入图像分割模型,可以得到图像预测分割结果,将图像预测分割结果与待训练医学图像的分割结果进行比较,计算损失值,根据损失值可以调整图像分割模型,具体为调整图像分割模型的模型参数。By inputting the medical image to be trained into the image segmentation model, the image prediction segmentation result can be obtained. The image prediction segmentation result is compared with the segmentation result of the medical image to be trained, and the loss value is calculated. The image segmentation model can be adjusted according to the loss value, specifically adjusting the image. Model parameters for the segmentation model.
在调整图像分割模型之后,需要确定是否达到预设条件。当达到预设条件时,图像分割模型训练结束,获取训练完成的图像分割模型。当未达到预设条件时,再次待训练医学图像输入图像分割模型,根据损失值调整图像分割模型,直至达到预设条件。作为一种可选示例,预设条件为达到预设训练次数。作为另一种可选示例,预设条件为损失值达到预设损失阈值。预设训练次数和预设损失阈值可根据实际情况进行设定,这里不进行限制。After adjusting the image segmentation model, it is necessary to determine whether the preset conditions are met. When the preset conditions are reached, the image segmentation model training ends, and the trained image segmentation model is obtained. When the preset conditions are not reached, the medical image to be trained is input into the image segmentation model again, and the image segmentation model is adjusted according to the loss value until the preset conditions are reached. As an optional example, the preset condition is reaching the preset training times. As another optional example, the preset condition is that the loss value reaches a preset loss threshold. The preset training times and preset loss threshold can be set according to the actual situation, and there are no restrictions here.
在一种可能的实现方式中,S202根据待训练医学图像以及待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型的具体实现可以包括B1-B3:In a possible implementation, S202 trains the image segmentation model based on the medical image to be trained and the segmentation results of the medical image to be trained until the preset conditions are reached. The specific implementation of obtaining the trained image segmentation model may include B1- B3:
B1:将待训练医学图像输入图像分割模型,获得待训练医学图像中各个像素的预测分类结果。B1: Input the medical image to be trained into the image segmentation model and obtain the prediction classification results of each pixel in the medical image to be trained.
将待训练医学图像输入图像分割模型,可以得到图像预测分割结果,具体可以为待训练医学图像中各个像素的预测分类结果。预测分类结果包括 属于目标对象、目标对象轮廓或者图像背景。By inputting the medical image to be trained into the image segmentation model, the image prediction segmentation result can be obtained, which specifically can be the prediction classification result of each pixel in the medical image to be trained. Predicted classification results include belonging to the target object, target object outline, or image background.
B2:根据待训练医学图像中各个像素的预测分类结果以及待训练医学图像各个像素属于目标对象、目标对象轮廓或者图像背景的分类标签,计算损失值,根据损失值对图像分割模型进行训练。B2: Calculate the loss value based on the predicted classification results of each pixel in the medical image to be trained and the classification label of each pixel in the medical image to be trained belonging to the target object, target object outline, or image background, and train the image segmentation model based on the loss value.
在待训练医学图像中逐个像素比较预测分类结果,与该像素属于目标对象、目标对象轮廓或者图像背景的分类标签,利用待训练医学图像中各个像素的比较结果,计算损失值,根据损失值对图像分割模型进行训练。即根据损失值对对图像分割模型的模型参数进行调整。Compare the predicted classification results pixel by pixel in the medical image to be trained, with the classification label that the pixel belongs to the target object, target object outline, or image background. Use the comparison results of each pixel in the medical image to be trained to calculate the loss value, and calculate the loss value based on the loss value. Image segmentation model is trained. That is, the model parameters of the image segmentation model are adjusted according to the loss value.
B3:重复执行将待训练医学图像输入图像分割模型,获得待训练医学图像中各个像素的预测分类结果以及后续步骤,直到达到预设条件,获得训练完成的图像分割模型。B3: Repeat the input of the medical image to be trained into the image segmentation model to obtain the prediction classification results of each pixel in the medical image to be trained and subsequent steps until the preset conditions are reached and the trained image segmentation model is obtained.
再次将待训练医学图像输入图像分割模型,根据待训练医学图像中各个像素的预测分类结果以及待训练医学图像各个像素属于目标对象、目标对象轮廓或者图像背景的分类标签,计算损失值,根据损失值对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型。The medical image to be trained is input into the image segmentation model again, and the loss value is calculated based on the predicted classification results of each pixel in the medical image to be trained and the classification label of each pixel in the medical image to be trained belonging to the target object, the outline of the target object, or the image background. The image segmentation model is trained until the preset conditions are reached, and the trained image segmentation model is obtained.
其中,在计算损失值时,分类标签为目标对象轮廓的损失权重大于分类标签为目标对象的损失权重,分类标签为目标对象轮廓的损失权重大于分类标签为图像背景的损失权重。Among them, when calculating the loss value, the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the target object, and the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the image background.
由于目标对象轮廓在图像分割中作用较为重要,在计算损失值时,目标对象轮廓的损失权重应大于目标对象以及图像背景的损失权重。例如,目标对象的权重损失和图像背景的损失权重为0.2,目标对象轮廓的损失权重为0.6。在实际应用中,针对目标对象、图像背景的损失函数可以为Cross Entropy(交叉熵)、针对目标对象轮廓的损失函数可以为focal loss。Since the contour of the target object plays an important role in image segmentation, when calculating the loss value, the loss weight of the target object contour should be greater than the loss weight of the target object and the image background. For example, the weight loss of the target object and the image background are 0.2, and the loss weight of the target object outline is 0.6. In practical applications, the loss function for the target object and image background can be Cross Entropy, and the loss function for the contour of the target object can be focal loss.
在本申请实施例中,对图像分割模型的结构进行了改进,参见图3所示,示出了本申请实施例中图像分割模型的结构示意图。In the embodiment of the present application, the structure of the image segmentation model is improved. Refer to Figure 3, which shows a schematic structural diagram of the image segmentation model in the embodiment of the present application.
图像分割模型包括编码器模块以及解码器模块,编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元;高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。The image segmentation model includes an encoder module and a decoder module. The encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units that are connected in sequence; the efficient pyramid segmentation attention unit includes a sequentially connected The second convolutional layer, the pyramid squeeze attention module, and the third convolutional layer.
在本申请实施例中,对Unet网络结构进行了改进,基于Resnet网络结构,去掉最后的fc(Full Connection,全连接)层,用金字塔挤压注意力模块(PSA Module)替换Resnet网络结构中每个残差模块(residual block)的3x3卷积。改进后residual block命名为Epsablock(即高效金字塔切分注意力单元)。参见图4所示,示出了Epsablock的结构示意图,高效金字塔切分注意力单元(Epsablock)包括依次相连的第二卷积层、金字塔挤压注意力模块(PSA Module)以及第三卷积层。其中,第二卷积层以及第三卷积层可以为1x1卷积(Conv 1x1)。In the embodiment of this application, the Unet network structure is improved. Based on the Resnet network structure, the last fc (Full Connection, full connection) layer is removed, and each layer in the Resnet network structure is replaced with the Pyramid Squeezing Attention Module (PSA Module). A 3x3 convolution of the residual block. The improved residual block is named Epsablock (ie, efficient pyramid segmentation attention unit). See Figure 4, which shows the schematic structural diagram of Epsablock. The efficient pyramid segmentation attention unit (Epsablock) includes a second convolution layer, a pyramid squeeze attention module (PSA Module) and a third convolution layer that are connected in sequence. . Among them, the second convolution layer and the third convolution layer can be 1x1 convolution (Conv 1x1).
参见图5所示,示出了在实际应用中,图像分割模型的结构示意图。将Epsablock替换原来的residual block作为改进的resnet。改进的resnet包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元(Epsablock)。将改进的resnet作为Unet网络结构的编码器,解码器不做修改,改进后的Unet网络结构即为本申请的图像分割模型。Refer to Figure 5, which shows a schematic structural diagram of the image segmentation model in practical applications. Epsablock replaces the original residual block as an improved resnet. The improved resnet includes the first convolutional layer, the pooling layer, and multiple efficient pyramid segmentation attention units (Epsablock) connected in sequence. The improved resnet is used as the encoder of the Unet network structure, and the decoder is not modified. The improved Unet network structure is the image segmentation model of this application.
在一种可能的实现方式中,金字塔挤压注意力模块包括:In a possible implementation, the pyramid squeeze attention module includes:
切片单元,用于将输入特征划分为N组,对每一组进行不同卷积核大小的卷积操作,生成N个特征图,对N个特征图进行合并,生成合并后的特征图;N为正整数;The slicing unit is used to divide the input features into N groups, perform convolution operations with different convolution kernel sizes on each group, generate N feature maps, merge the N feature maps, and generate a merged feature map; N is a positive integer;
注意力单元,用于将合并后的特征图进行注意力操作,生成注意力操作后的特征图;The attention unit is used to perform attention operations on the merged feature maps and generate feature maps after attention operations;
激活单元,使用softmax函数激活注意力操作后的特征图,与合并后的特征图进行点乘操作,输出最终特征图。The activation unit uses the softmax function to activate the feature map after the attention operation, performs a dot multiplication operation with the merged feature map, and outputs the final feature map.
参见图6所示,示出了金字塔挤压注意力模块(PSA module)的原来示意图。切片单元先将输入特征使用1x1卷积划分为N组,例如,N为4。对每一组进行不同卷积核大小的卷积操作,每组的卷积核大小依次增大,例如卷积核大小依次为3、5、7、9。在经过不同大小的卷积后,将生成的N个特征图进行合并,生成合并后的特征图See Figure 6, which shows the original schematic diagram of the pyramid squeeze attention module (PSA module). The slicing unit first divides the input features into N groups using 1x1 convolution, for example, N is 4. Convolution operations with different convolution kernel sizes are performed on each group, and the convolution kernel sizes of each group increase sequentially. For example, the convolution kernel sizes are 3, 5, 7, and 9 in order. After convolutions of different sizes, the generated N feature maps are merged to generate a merged feature map.
将切分单元的输出通过注意力单元获得通道注意力值,即得到注意力操作后的特征图。这样做的目的是获得不同尺度特征图的注意力权值,通过这样的做法,融合了不同尺度的上下文信息,并产生了更好的像素级注意力。最 后激活单元将注意力操作后的特征图,进行softmax归一化,与合并后的特征图进行点乘操作(channel-wise product),输出最终特征图。The output of the segmentation unit is passed through the attention unit to obtain the channel attention value, that is, the feature map after the attention operation is obtained. The purpose of this is to obtain the attention weights of feature maps of different scales. In this way, contextual information of different scales is integrated and better pixel-level attention is generated. Finally, the activation unit performs softmax normalization on the feature map after the attention operation, performs dot product operation (channel-wise product) with the merged feature map, and outputs the final feature map.
PSA module可以让图像分割模型更加关注具有显著性特征的目标。The PSA module allows the image segmentation model to pay more attention to targets with salient features.
本申请实施例训练生成的图像分割模型由编码器模块和解码器模块组成。其中,编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元。高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。即在编码器模块中采用了金字塔挤压注意力模块,提高了编码器模块的特征提取能力,使图像分割模型更加关注具有显著性特征的目标对象,从而提高了图像分割模型的分割准确性。The image segmentation model generated by training in the embodiment of this application consists of an encoder module and a decoder module. Among them, the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units. The efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, the pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module, making the image segmentation model pay more attention to target objects with salient features, thus improving the segmentation accuracy of the image segmentation model.
基于上述提供的图像分割模型的训练方法的实施例,本申请实施例还提供了一种图像分割方法。参见图7,图7为本申请实施例提供的一种图像分割方法的流程图。如图7所示,该方法包括S701-S702:Based on the embodiments of the image segmentation model training method provided above, embodiments of the present application also provide an image segmentation method. Referring to Figure 7, Figure 7 is a flow chart of an image segmentation method provided by an embodiment of the present application. As shown in Figure 7, the method includes S701-S702:
S701:将待分割医学图像输入图像分割模型,获得待分割医学图像中各个像素点的初始分割结果,初始分割结果包括属于目标对象、目标对象轮廓或者图像背景;其中,图像分割模型是根据上述的图像分割模型的训练方法训练得到的。S701: Input the medical image to be segmented into the image segmentation model, and obtain the initial segmentation results of each pixel in the medical image to be segmented. The initial segmentation results include belonging to the target object, the outline of the target object, or the image background; wherein, the image segmentation model is based on the above The image segmentation model is trained using the training method.
将待分割医学图像输入上述图像分割模型,可以得到待分割医学图像中各个像素点属于目标对象、目标对象轮廓或者图像背景的概率值,根据概率值是否大于阈值确定待分割医学图像中各个像素点的初始分割结果,即确定待分割医学图像中各个像素点属于目标对象、目标对象轮廓或者图像背景。在实际应用中,可以将待分割医学图像扩展为三通道(如RGB三通道)图像,输入图像分割模型,得到每一通道图像中各个像素点的初始分割结果,即每一通道图像中各个像素点属于目标对象、目标对象轮廓或者图像背景,再综合各通道图像中各个像素点的初始分割结果,确定待分割医学图像中各个像素点的初始分割结果。By inputting the medical image to be segmented into the above image segmentation model, the probability value that each pixel in the medical image to be segmented belongs to the target object, the outline of the target object, or the image background can be obtained. Each pixel in the medical image to be segmented is determined based on whether the probability value is greater than the threshold. The initial segmentation result is to determine whether each pixel in the medical image to be segmented belongs to the target object, the target object outline or the image background. In practical applications, the medical image to be segmented can be expanded into a three-channel (such as RGB three-channel) image, and the image segmentation model is input to obtain the initial segmentation result of each pixel in each channel image, that is, each pixel in each channel image The points belong to the target object, the outline of the target object or the image background, and then the initial segmentation results of each pixel point in each channel image are combined to determine the initial segmentation result of each pixel point in the medical image to be segmented.
图像分割模型根据本申请上述实施例提供的图像分割模型的训练方法训练得到的,相关说明可以参见上述实施例,在此不再赘述。The image segmentation model is trained according to the training method of the image segmentation model provided in the above embodiments of the present application. For relevant description, please refer to the above embodiments and will not be described again here.
S702:将初始分割结果属于目标对象或者目标对象轮廓的像素点确定为属于目标对象的像素点,输出待分割医学图像的分割结果。S702: Determine pixel points that the initial segmentation result belongs to the target object or the outline of the target object as pixel points belonging to the target object, and output the segmentation result of the medical image to be segmented.
将属于目标对象或者目标对象轮廓的像素点进行融合,实现目标对象的分割,即将属于目标对象或者目标对象轮廓的像素点确定为属于目标对象的像素点,进行图像分割,输出待分割医学图像的分割结果。Fusion of pixel points belonging to the target object or the contour of the target object to achieve segmentation of the target object, that is, determining the pixel points belonging to the target object or the contour of the target object as pixel points belonging to the target object, performing image segmentation, and outputting the medical image to be segmented Segmentation results.
在本申请实施例中,还可以利用分水岭算法过滤待分割医学图像的分割结果中面积小于阈值的目标对象,和/或,对不完整的目标对象进行修复。In the embodiment of the present application, the watershed algorithm can also be used to filter the target objects whose area is smaller than the threshold in the segmentation results of the medical image to be segmented, and/or to repair the incomplete target objects.
通过过滤面积小于阈值的目标对象和/或修复不完整的目标对象,实现对目标对象的修正,完成待分割医学图像的分割结果的后处理,提高图像分割的准确性。例如,目标对象为细胞时,通过细胞过滤和修正等算法,过滤非常小的细胞,以及修正不完整的细胞,使细胞分割更为准确。By filtering target objects with an area smaller than the threshold and/or repairing incomplete target objects, the target object is corrected, the post-processing of the segmentation results of the medical image to be segmented is completed, and the accuracy of image segmentation is improved. For example, when the target object is cells, algorithms such as cell filtering and correction are used to filter very small cells and correct incomplete cells to make cell segmentation more accurate.
本申请实施例使用的图像分割模型,在编码器模块中采用了金字塔挤压注意力模块,提高了编码器模块的特征提取能力,使图像分割模型更加关注具有显著性特征的目标对象,从而提高了图像分割模型的分割准确性。The image segmentation model used in the embodiments of this application adopts a pyramid squeeze attention module in the encoder module, which improves the feature extraction capability of the encoder module and enables the image segmentation model to pay more attention to target objects with salient features, thereby improving improve the segmentation accuracy of the image segmentation model.
将本申请实施例应用于时空组学中的细胞分割,参见图8所示,展示了本申请实施例的图像分割模型在鼠脑图像上的分割效果。截取了上图鼠脑图像的一部分作为展示,从左下图中可以看出,鼠脑图像中存细胞黏连情况,部分细胞边界严重模糊,右下图利用本申请实施例提出的图像分割模型可以从黏连细胞中分割出独立细胞,并提供比较可靠的边界。Embodiments of the present application are applied to cell segmentation in spatiotemporal omics, as shown in Figure 8 , which demonstrates the segmentation effect of the image segmentation model of the embodiment of the present application on mouse brain images. A part of the mouse brain image above has been intercepted for display. As can be seen from the lower left image, there are cell adhesions in the mouse brain image, and some cell boundaries are seriously blurred. The lower right image can be achieved using the image segmentation model proposed in the embodiment of this application. Separate independent cells from adherent cells and provide more reliable boundaries.
参见图9、图10所示,分别是本申请实施例提供的图像分割模型在猴脑图像和蝾螈脑图像上的分割结果,从原图中截取了部分作为展示,从结果上可以看出,本申请实施例提供的图像分割模型同样能处理其他图像,说明图像分割模型具有泛化性。Refer to Figures 9 and 10, which are respectively the segmentation results of the image segmentation model provided by the embodiment of the present application on monkey brain images and salamander brain images. Parts of the original images are cut out for display. It can be seen from the results that , the image segmentation model provided by the embodiments of this application can also process other images, indicating that the image segmentation model has generalization properties.
参见图11是多种算法在鼠脑图像上的分割结果,左图为鼠脑图像的一部分,中图是使用原始Unet网络对鼠脑图像的分割结果,右图是使用分水岭算法对鼠脑图像的分割结果,和本申请实施例提出的图像分割模型的分割结果(见图8)对比,在清晰细胞的分割上,Unet网络和本申请实施例提出的图像分割模型的分割结果相近,但在黏连细胞处Unet网络处理边界的能力不如本申请实施例提出的图像分割模型。而分水岭算法处理边界的能力较差,对细胞过度分割,强制将细胞分割多个细胞。See Figure 11 for the segmentation results of multiple algorithms on mouse brain images. The left picture is a part of the mouse brain image. The middle picture is the segmentation result of the mouse brain image using the original Unet network. The right picture is the mouse brain image using the watershed algorithm. Compared with the segmentation results of the image segmentation model proposed in the embodiment of the present application (see Figure 8), in terms of segmentation of clear cells, the segmentation results of the Unet network and the image segmentation model proposed in the embodiment of the present application are similar, but in The Unet network's ability to process boundaries at adhesion cells is not as good as the image segmentation model proposed in the embodiments of this application. The watershed algorithm has poor ability to handle boundaries and over-segments cells, forcing cells to be divided into multiple cells.
进一步对比Unet网络和本申请实施例提出的图像分割模型,表1是Unet 网络和本申请实施例提出的图像分割模型在5张鼠脑图像上的量化指标对比。量化指标包括dice系数、IoU(交并比)、precision(精确率)以及recall(召回率)。To further compare the Unet network and the image segmentation model proposed in the embodiment of the present application, Table 1 is a comparison of the quantitative indicators of the Unet network and the image segmentation model proposed in the embodiment of the present application on five mouse brain images. Quantitative indicators include dice coefficient, IoU (Intersection over Union), precision (precision rate) and recall (recall rate).
从结果可以看出,本申请实施例提出的图像分割模型具有更好的分割性能。It can be seen from the results that the image segmentation model proposed in the embodiment of this application has better segmentation performance.
表1Table 1
Figure PCTCN2022110779-appb-000001
Figure PCTCN2022110779-appb-000001
综上所述本申请实施例提出的图像分割模型的分割效果优于Unet网络和分水岭算法。In summary, the segmentation effect of the image segmentation model proposed in the embodiment of this application is better than the Unet network and the watershed algorithm.
基于上述方法实施例提供的一种图像分割模型的训练方法,本申请实施例还提供了一种图像分割模型的训练装置,下面将结合附图对图像分割模型的训练装置进行说明。Based on the training method of the image segmentation model provided by the above method embodiment, the embodiment of the present application also provides a training device of the image segmentation model. The training device of the image segmentation model will be described below with reference to the accompanying drawings.
参见图12,图12为本申请实施例提供的一种图像分割模型的训练装置的结构示意图。如图12所示,该图像分割模型的训练装置包括:Referring to Figure 12, Figure 12 is a schematic structural diagram of an image segmentation model training device provided by an embodiment of the present application. As shown in Figure 12, the training device of the image segmentation model includes:
获取单元1201,用于获取待训练医学图像以及所述待训练医学图像的分割结果;The acquisition unit 1201 is used to acquire the medical image to be trained and the segmentation result of the medical image to be trained;
训练单元1202,用于根据所述待训练医学图像以及所述待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型;The training unit 1202 is configured to train the image segmentation model according to the medical image to be trained and the segmentation results of the medical image to be trained until the preset conditions are reached, and the trained image segmentation model is obtained;
所述模型包括编码器模块以及解码器模块,所述编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元;所述高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。The model includes an encoder module and a decoder module. The encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
在一种可能的实现方式中,所述金字塔挤压注意力模块包括:In a possible implementation, the pyramid squeeze attention module includes:
切片单元,用于将输入特征划分为N组,对每一组进行不同卷积核大小的卷积操作,生成N个特征图,对所述N个特征图进行合并,生成合并后的特征图;N为正整数;The slicing unit is used to divide the input features into N groups, perform convolution operations with different convolution kernel sizes on each group, generate N feature maps, merge the N feature maps, and generate a merged feature map. ;N is a positive integer;
注意力单元,用于将所述合并后的特征图进行注意力操作,生成注意力操作后的特征图;An attention unit is used to perform an attention operation on the merged feature map and generate a feature map after the attention operation;
激活单元,使用softmax函数激活所述注意力操作后的特征图,与所述合并后的特征图进行点乘操作,输出最终特征图。The activation unit uses the softmax function to activate the feature map after the attention operation, performs a dot multiplication operation with the merged feature map, and outputs the final feature map.
在一种可能的实现方式中,所述获取单元,包括:In a possible implementation, the acquisition unit includes:
获取子单元,用于获取待训练医学图像;Acquisition subunit, used to obtain medical images to be trained;
提取子单元,用于采用预设图像分割算法提取所述待训练医学图像中的目标对象、目标对象轮廓以及图像背景;An extraction subunit is used to extract the target object, target object contour and image background in the medical image to be trained using a preset image segmentation algorithm;
确定子单元,用于将所述待训练医学图像各个像素属于所述目标对象、所述目标对象轮廓或者所述图像背景的分类标签确定为所述待训练医学图像的分割结果。Determining subunit, configured to determine the classification label that each pixel of the medical image to be trained belongs to the target object, the target object outline, or the image background as a segmentation result of the medical image to be trained.
在一种可能的实现方式中,所述训练单元,包括:In a possible implementation, the training unit includes:
输入子单元,用于将所述待训练医学图像输入图像分割模型,获得所述待训练医学图像中各个像素的预测分类结果;An input subunit, used to input the medical image to be trained into an image segmentation model to obtain the prediction classification results of each pixel in the medical image to be trained;
训练子单元,用于根据所述待训练医学图像中各个像素的预测分类结果以及所述待训练医学图像各个像素属于所述目标对象、所述目标对象轮廓或者所述图像背景的分类标签,计算损失值,根据所述损失值对图像分割模型进行训练;A training subunit, configured to calculate based on the predicted classification results of each pixel in the medical image to be trained and the classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background. Loss value, the image segmentation model is trained according to the loss value;
循环子单元,用于返回所述输入子单元以及所述训练子单元执行将所述待训练医学图像输入图像分割模型,获得所述待训练医学图像中各个像素的预测分类结果以及后续步骤,直到达到预设条件,获得训练完成的图像分割模型;Loop subunit, used to return to the input subunit and the training subunit to execute the input of the medical image to be trained into the image segmentation model, obtain the prediction classification results of each pixel in the medical image to be trained, and subsequent steps, until Reach the preset conditions and obtain the trained image segmentation model;
其中,在计算损失值时,分类标签为所述目标对象轮廓的损失权重大于分类标签为所述目标对象的损失权重,分类标签为所述目标对象轮廓的损失权重大于分类标签为所述图像背景的损失权重。Wherein, when calculating the loss value, the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the target object, and the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the image background. loss weight.
需要说明的是,本申请实施例提供的图像分割模型的训练装置的技术详情可参见上述图像分割模型的训练方法的相关实施例,这里不再赘述。It should be noted that the technical details of the image segmentation model training device provided by the embodiments of the present application can be found in the relevant embodiments of the above image segmentation model training method, and will not be described again here.
基于上述方法实施例提供的一种图像分割方法,本申请实施例还提供了一种图像分割装置,下面将结合附图对图像分割装置进行说明。Based on the image segmentation method provided by the above method embodiment, the embodiment of the present application also provides an image segmentation device. The image segmentation device will be described below with reference to the accompanying drawings.
参见图13,图13为本申请实施例提供的一种图像分割装置的结构示意图。如图13所示,该图像分割装置包括:Referring to Figure 13, Figure 13 is a schematic structural diagram of an image segmentation device provided by an embodiment of the present application. As shown in Figure 13, the image segmentation device includes:
输入单元1301,用于将待分割医学图像输入图像分割模型,获得所述待分割医学图像中各个像素点的初始分割结果,所述初始分割结果包括属于目标对象、目标对象轮廓或者图像背景;所述图像分割模型是根据上述的图像分割模型的训练方法训练得到的;The input unit 1301 is used to input the medical image to be segmented into the image segmentation model, and obtain the initial segmentation result of each pixel point in the medical image to be segmented, where the initial segmentation result includes the target object, the outline of the target object, or the image background; The image segmentation model is trained according to the above image segmentation model training method;
确定单元1302,用于将所述初始分割结果属于目标对象或者目标对象轮廓的像素点确定为属于目标对象的像素点,输出所述待分割医学图像的分割结果。The determination unit 1302 is configured to determine the pixel points of the initial segmentation result belonging to the target object or the outline of the target object as pixel points belonging to the target object, and output the segmentation result of the medical image to be segmented.
在一种可能的实现方式中,所述装置还包括:In a possible implementation, the device further includes:
修正单元,用于利用分水岭算法过滤所述待分割医学图像的分割结果中面积小于阈值的目标对象,和/或,对不完整的目标对象进行修复。A correction unit configured to use a watershed algorithm to filter target objects with an area smaller than a threshold in the segmentation results of the medical image to be segmented, and/or to repair incomplete target objects.
需要说明的是,本申请实施例提供的图像分割装置的技术详情可参见上述图像分割方法的相关实施例,这里不再赘述。It should be noted that the technical details of the image segmentation device provided by the embodiments of the present application can be found in the relevant embodiments of the above image segmentation method, and will not be described again here.
另外,本申请实施例还提供了一种图像分割模型的训练设备,包括:存储器,处理器,及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如上述的图像分割模型的训练方法。In addition, embodiments of the present application also provide an image segmentation model training device, including: a memory, a processor, and a computer program stored in the memory and executable on the processor. The processor executes When the computer program is used, the above-mentioned training method of the image segmentation model is implemented.
本申请实施例还提供了一种图像分割设备,包括:存储器,处理器,及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如上述的图像分割方法。An embodiment of the present application also provides an image segmentation device, including: a memory, a processor, and a computer program stored on the memory and executable on the processor. When the processor executes the computer program , implement the image segmentation method as mentioned above.
本申请实施例还提供了一种计算机可读存储介质,所述计算机可读存储介质中存储有指令,当所述指令在终端设备上运行时,使得所述终端设备执行如上述的图像分割模型的训练方法,或者执行如上述的图像分割方法。Embodiments of the present application also provide a computer-readable storage medium. Instructions are stored in the computer-readable storage medium. When the instructions are run on a terminal device, the terminal device executes the image segmentation model as described above. training method, or perform the image segmentation method as above.
本申请实施例训练生成的图像分割模型由编码器模块和解码器模块组成。其中,编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元。高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。即在编码器模块中采用了金字塔挤压注意力模块,提高了编码器模块的特征提取能力,使图像分割模型更加关注具有显著性特征的目标对象,从而提高了图像分割模型的分割准确性。The image segmentation model generated by training in the embodiment of this application consists of an encoder module and a decoder module. Among them, the encoder module includes a sequentially connected first convolution layer, pooling layer, and multiple efficient pyramid segmentation attention units. The efficient pyramid segmentation attention unit includes a second convolutional layer, a pyramid squeeze attention module and a third convolutional layer that are connected in sequence. That is, the pyramid squeeze attention module is used in the encoder module to improve the feature extraction capability of the encoder module, making the image segmentation model pay more attention to target objects with salient features, thus improving the segmentation accuracy of the image segmentation model.
需要说明的是,本说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是与其他实施例的不同之处,各个实施例之间相同相似部分互相参见即可。对于实施例公开的系统或装置而言,由于其与实施例公开的方法相对应,所以描述的比较简单,相关之处参见方法部分说明即可。It should be noted that each embodiment in this specification is described in a progressive manner, and each embodiment focuses on its differences from other embodiments. The same and similar parts between the various embodiments can be referred to each other. As for the system or device disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple. For relevant details, please refer to the description in the method section.
应当理解,在本申请中,“至少一个(项)”是指一个或者多个,“多个”是指两个或两个以上。“和/或”,用于描述关联对象的关联关系,表示可以存在三种关系,例如,“A和/或B”可以表示:只存在A,只存在B以及同时存在A和B三种情况,其中A,B可以是单数或者复数。字符“/”一般表示前后关联对象是一种“或”的关系。“以下至少一项(个)”或其类似表达,是指这些项中的任意组合,包括单项(个)或复数项(个)的任意组合。例如,a,b或c中的至少一项(个),可以表示:a,b,c,“a和b”,“a和c”,“b和c”,或“a和b和c”,其中a,b,c可以是单个,也可以是多个。It should be understood that in this application, "at least one (item)" refers to one or more, and "plurality" refers to two or more. "And/or" is used to describe the relationship between associated objects, indicating that there can be three relationships. For example, "A and/or B" can mean: only A exists, only B exists, and A and B exist simultaneously. , where A and B can be singular or plural. The character "/" generally indicates that the related objects are in an "or" relationship. “At least one of the following” or similar expressions thereof refers to any combination of these items, including any combination of a single item (items) or a plurality of items (items). For example, at least one of a, b or c can mean: a, b, c, "a and b", "a and c", "b and c", or "a and b and c" ”, where a, b, c can be single or multiple.
还需要说明的是,在本文中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的 其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should also be noted that in this article, relational terms such as first and second are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply that these entities or operations There is no such actual relationship or sequence between them. Furthermore, the terms "comprises," "comprises," or any other variations thereof are intended to cover a non-exclusive inclusion such that a process, method, article, or apparatus that includes a list of elements includes not only those elements, but also those not expressly listed other elements, or elements inherent to the process, method, article or equipment. Without further limitation, an element defined by the statement "comprises a..." does not exclude the presence of additional identical elements in a process, method, article, or apparatus that includes the stated element.
结合本文中所公开的实施例描述的方法或算法的步骤可以直接用硬件、处理器执行的软件模块,或者二者的结合来实施。软件模块可以置于随机存储器(RAM)、内存、只读存储器(ROM)、电可编程ROM、电可擦除可编程ROM、寄存器、硬盘、可移动磁盘、CD-ROM、或技术领域内所公知的任意其它形式的存储介质中。The steps of the methods or algorithms described in conjunction with the embodiments disclosed herein may be implemented directly in hardware, in software modules executed by a processor, or in a combination of both. Software modules may be located in random access memory (RAM), memory, read-only memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disks, removable disks, CD-ROMs, or anywhere in the field of technology. any other known form of storage media.
对所公开的实施例的上述说明,使本领域专业技术人员能够实现或使用本申请。对这些实施例的多种修改对本领域的专业技术人员来说将是显而易见的,本文中所定义的一般原理可以在不脱离本申请的精神或范围的情况下,在其它实施例中实现。因此,本申请将不会被限制于本文所示的这些实施例,而是要符合与本文所公开的原理和新颖特点相一致的最宽的范围。The above description of the disclosed embodiments enables those skilled in the art to implement or use the present application. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be practiced in other embodiments without departing from the spirit or scope of the application. Therefore, the present application is not to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (11)

  1. 一种图像分割模型的训练方法,其特征在于,所述方法包括:A method for training an image segmentation model, characterized in that the method includes:
    获取待训练医学图像以及所述待训练医学图像的分割结果;Obtaining a medical image to be trained and a segmentation result of the medical image to be trained;
    根据所述待训练医学图像以及所述待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型;Train the image segmentation model according to the medical image to be trained and the segmentation results of the medical image to be trained until the preset conditions are reached, and a trained image segmentation model is obtained;
    所述模型包括编码器模块以及解码器模块,所述编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元;所述高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模块以及第三卷积层。The model includes an encoder module and a decoder module. The encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
  2. 根据权利要求1所述的方法,其特征在于,所述金字塔挤压注意力模块包括:The method according to claim 1, characterized in that the pyramid squeeze attention module includes:
    切片单元,用于将输入特征划分为N组,对每一组进行不同卷积核大小的卷积操作,生成N个特征图,对所述N个特征图进行合并,生成合并后的特征图;N为正整数;The slicing unit is used to divide the input features into N groups, perform convolution operations with different convolution kernel sizes on each group, generate N feature maps, merge the N feature maps, and generate a merged feature map. ;N is a positive integer;
    注意力单元,用于将所述合并后的特征图进行注意力操作,生成注意力操作后的特征图;An attention unit is used to perform an attention operation on the merged feature map and generate a feature map after the attention operation;
    激活单元,使用softmax函数激活所述注意力操作后的特征图,与所述合并后的特征图进行点乘操作,输出最终特征图。The activation unit uses the softmax function to activate the feature map after the attention operation, performs a dot multiplication operation with the merged feature map, and outputs the final feature map.
  3. 根据权利要求1所述的方法,其特征在于,所述获取待训练医学图像以及所述待训练医学图像的分割结果,包括:The method according to claim 1, characterized in that said obtaining the medical image to be trained and the segmentation result of the medical image to be trained includes:
    获取待训练医学图像;Obtain medical images to be trained;
    采用预设图像分割算法提取所述待训练医学图像中的目标对象、目标对象轮廓以及图像背景;Using a preset image segmentation algorithm to extract the target object, target object outline and image background in the medical image to be trained;
    将所述待训练医学图像各个像素属于所述目标对象、所述目标对象轮廓或者所述图像背景的分类标签确定为所述待训练医学图像的分割结果。The classification label that each pixel of the medical image to be trained belongs to the target object, the outline of the target object, or the image background is determined as a segmentation result of the medical image to be trained.
  4. 根据权利要求3所述的方法,其特征在于,所述根据所述待训练医学图像以及所述待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型,包括:The method according to claim 3, characterized in that the image segmentation model is trained according to the medical image to be trained and the segmentation result of the medical image to be trained until a preset condition is reached and a trained image is obtained. Segmentation models, including:
    将所述待训练医学图像输入图像分割模型,获得所述待训练医学图像中 各个像素的预测分类结果;Input the medical image to be trained into an image segmentation model to obtain the prediction classification results of each pixel in the medical image to be trained;
    根据所述待训练医学图像中各个像素的预测分类结果以及所述待训练医学图像各个像素属于所述目标对象、所述目标对象轮廓或者所述图像背景的分类标签,计算损失值,根据所述损失值对图像分割模型进行训练;The loss value is calculated according to the predicted classification result of each pixel in the medical image to be trained and the classification label of each pixel in the medical image to be trained belonging to the target object, the target object outline or the image background. The loss value is used to train the image segmentation model;
    重复执行将所述待训练医学图像输入图像分割模型,获得所述待训练医学图像中各个像素的预测分类结果以及后续步骤,直到达到预设条件,获得训练完成的图像分割模型;Repeatedly executing the input of the medical image to be trained into the image segmentation model to obtain the prediction classification results of each pixel in the medical image to be trained and subsequent steps until the preset conditions are reached and the image segmentation model that has been trained is obtained;
    其中,在计算损失值时,分类标签为所述目标对象轮廓的损失权重大于分类标签为所述目标对象的损失权重,分类标签为所述目标对象轮廓的损失权重大于分类标签为所述图像背景的损失权重。Wherein, when calculating the loss value, the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the target object, and the loss weight when the classification label is the outline of the target object is greater than the loss weight when the classification label is the image background. loss weight.
  5. 一种图像分割方法,其特征在于,所述方法包括:An image segmentation method, characterized in that the method includes:
    将待分割医学图像输入图像分割模型,获得所述待分割医学图像中各个像素点的初始分割结果,所述初始分割结果包括属于目标对象、目标对象轮廓或者图像背景;所述图像分割模型是根据权利要求1-4任一项所述的图像分割模型的训练方法训练得到的;The medical image to be segmented is input into the image segmentation model, and the initial segmentation result of each pixel in the medical image to be segmented is obtained. The initial segmentation result includes belonging to the target object, the outline of the target object, or the image background; the image segmentation model is based on Obtained by training by the training method of the image segmentation model described in any one of claims 1-4;
    将所述初始分割结果属于目标对象或者目标对象轮廓的像素点确定为属于目标对象的像素点,输出所述待分割医学图像的分割结果。The pixel points that the initial segmentation result belongs to the target object or the outline of the target object are determined as pixel points that belong to the target object, and the segmentation result of the medical image to be segmented is output.
  6. 根据权利要求5所述的方法,其特征在于,所述方法还包括:The method of claim 5, further comprising:
    利用分水岭算法过滤所述待分割医学图像的分割结果中面积小于阈值的目标对象,和/或,对不完整的目标对象进行修复。A watershed algorithm is used to filter target objects with an area smaller than a threshold in the segmentation results of the medical image to be segmented, and/or to repair incomplete target objects.
  7. 一种图像分割模型的训练装置,其特征在于,所述装置包括:An image segmentation model training device, characterized in that the device includes:
    获取单元,用于获取待训练医学图像以及所述待训练医学图像的分割结果;An acquisition unit, configured to acquire a medical image to be trained and a segmentation result of the medical image to be trained;
    训练单元,用于根据所述待训练医学图像以及所述待训练医学图像的分割结果对图像分割模型进行训练,直到达到预设条件,获得训练完成的图像分割模型;A training unit configured to train an image segmentation model based on the medical image to be trained and the segmentation results of the medical image to be trained until a preset condition is reached, and a trained image segmentation model is obtained;
    所述模型包括编码器模块以及解码器模块,所述编码器模块包括依次相连的第一卷积层、池化层、多个高效金字塔切分注意力单元;所述高效金字塔切分注意力单元包括依次相连的第二卷积层、金字塔挤压注意力模 块以及第三卷积层。The model includes an encoder module and a decoder module. The encoder module includes a first convolution layer, a pooling layer, and multiple efficient pyramid segmentation attention units connected in sequence; the efficient pyramid segmentation attention unit It includes the second convolutional layer, the pyramid squeeze attention module and the third convolutional layer connected in sequence.
  8. 一种图像分割装置,其特征在于,所述装置包括:An image segmentation device, characterized in that the device includes:
    输入单元,用于将待分割医学图像输入图像分割模型,获得所述待分割医学图像中各个像素点的初始分割结果,所述初始分割结果包括属于目标对象、目标对象轮廓或者图像背景;所述图像分割模型是根据权利要求1-4任一项所述的图像分割模型的训练方法训练得到的;An input unit is used to input the medical image to be segmented into the image segmentation model, and obtain the initial segmentation result of each pixel in the medical image to be segmented, where the initial segmentation result includes the target object, the target object outline, or the image background; The image segmentation model is trained according to the training method of the image segmentation model described in any one of claims 1-4;
    确定单元,用于将所述初始分割结果属于目标对象或者目标对象轮廓的像素点确定为属于目标对象的像素点,输出所述待分割医学图像的分割结果。A determination unit configured to determine the pixel points of the initial segmentation result belonging to the target object or the outline of the target object as pixel points belonging to the target object, and output the segmentation result of the medical image to be segmented.
  9. 一种图像分割模型的训练设备,其特征在于,包括:存储器,处理器,及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如权利要求1-4任一项所述的图像分割模型的训练方法。An image segmentation model training device, characterized in that it includes: a memory, a processor, and a computer program stored on the memory and executable on the processor. When the processor executes the computer program , implement the training method of the image segmentation model as described in any one of claims 1-4.
  10. 一种图像分割设备,其特征在于,包括:存储器,处理器,及存储在所述存储器上并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时,实现如权利要求5-6任一项所述的图像分割方法。An image segmentation device, characterized in that it includes: a memory, a processor, and a computer program stored in the memory and executable on the processor. When the processor executes the computer program, it implements the following: The image segmentation method according to any one of claims 5-6.
  11. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质中存储有指令,当所述指令在终端设备上运行时,使得所述终端设备执行如权利要求1-4任一项所述的图像分割模型的训练方法,或者执行如权利要求5-6任一项所述的图像分割方法。A computer-readable storage medium, characterized in that instructions are stored in the computer-readable storage medium. When the instructions are run on a terminal device, the terminal device is caused to execute any one of claims 1-4. The training method of the image segmentation model, or the image segmentation method as described in any one of claims 5-6.
PCT/CN2022/110779 2022-06-29 2022-08-08 Image segmentation model training method, image segmentation method, and apparatus WO2024031219A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2022/110779 WO2024031219A1 (en) 2022-08-08 2022-08-08 Image segmentation model training method, image segmentation method, and apparatus
PCT/CN2022/136698 WO2024001051A1 (en) 2022-06-29 2022-12-05 Spatial omics single cell data acquisition method and apparatus, and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/110779 WO2024031219A1 (en) 2022-08-08 2022-08-08 Image segmentation model training method, image segmentation method, and apparatus

Publications (1)

Publication Number Publication Date
WO2024031219A1 true WO2024031219A1 (en) 2024-02-15

Family

ID=89850156

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/110779 WO2024031219A1 (en) 2022-06-29 2022-08-08 Image segmentation model training method, image segmentation method, and apparatus

Country Status (1)

Country Link
WO (1) WO2024031219A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112070772A (en) * 2020-08-27 2020-12-11 闽江学院 Blood leukocyte image segmentation method based on UNet + + and ResNet
CN113421276A (en) * 2021-07-02 2021-09-21 深圳大学 Image processing method, device and storage medium
CN114663655A (en) * 2022-03-01 2022-06-24 新疆大学 Image segmentation model training method, image semantic segmentation device and related equipment

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112070772A (en) * 2020-08-27 2020-12-11 闽江学院 Blood leukocyte image segmentation method based on UNet + + and ResNet
CN113421276A (en) * 2021-07-02 2021-09-21 深圳大学 Image processing method, device and storage medium
CN114663655A (en) * 2022-03-01 2022-06-24 新疆大学 Image segmentation model training method, image semantic segmentation device and related equipment

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ZHANG HU, ZU KEKE, LU JIAN, ZOU YURU, MENG DEYU: "EPSANet: An efficient pyramid squeeze attention block on convolutional neural network", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, ARXIV.ORG, ITHACA, 30 May 2021 (2021-05-30), Ithaca, XP093138036, Retrieved from the Internet <URL:https://arxiv.org/abs/2105.14447> DOI: 10.48550/arXiv.2105.14447 *

Similar Documents

Publication Publication Date Title
Shin et al. Deep vessel segmentation by learning graphical connectivity
CN111325739B (en) Method and device for detecting lung focus and training method of image detection model
CN109685060B (en) Image processing method and device
CN108335303B (en) Multi-scale palm skeleton segmentation method applied to palm X-ray film
CN110599528A (en) Unsupervised three-dimensional medical image registration method and system based on neural network
CN109191424B (en) Breast mass detection and classification system and computer-readable storage medium
CN111369550B (en) Image registration and defect detection method, model, training method, device and equipment
TW201833867A (en) Method and device for training neural network model for image processing and computer readable storage medium
CN108961180B (en) Infrared image enhancement method and system
CN113728335A (en) Method and system for classification and visualization of 3D images
WO2021136368A1 (en) Method and apparatus for automatically detecting pectoralis major region in molybdenum target image
CN105321155A (en) Ring artifact elimination method for CBCT image
CN111091575B (en) Medical image segmentation method based on reinforcement learning method
CN103026379A (en) Method for estimating image noise level
CN111967464B (en) Weak supervision target positioning method based on deep learning
CN112365514A (en) Semantic segmentation method based on improved PSPNet
CN110009628A (en) A kind of automatic testing method for polymorphic target in continuous two dimensional image
CN106023205A (en) Medical image segmentation method based on simplified PSO (Particle Swarm Optimization) and 2D maximum entropy threshold
Asheghi et al. A comprehensive review on content-aware image retargeting: From classical to state-of-the-art methods
JP2023515367A (en) Out-of-distribution detection of input instances to model
WO2021058867A1 (en) Image analysis in pathology
CN112381811A (en) Method, device and equipment for realizing medical image data labeling
WO2024031219A1 (en) Image segmentation model training method, image segmentation method, and apparatus
CN113034528A (en) Target area and organ-at-risk delineation contour accuracy testing method based on image omics
CN117437423A (en) Weak supervision medical image segmentation method and device based on SAM collaborative learning and cross-layer feature aggregation enhancement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22954199

Country of ref document: EP

Kind code of ref document: A1