CN116433586A - Mammary gland ultrasonic tomography image segmentation model establishment method and segmentation method - Google Patents

Mammary gland ultrasonic tomography image segmentation model establishment method and segmentation method Download PDF

Info

Publication number
CN116433586A
CN116433586A CN202310149152.7A CN202310149152A CN116433586A CN 116433586 A CN116433586 A CN 116433586A CN 202310149152 A CN202310149152 A CN 202310149152A CN 116433586 A CN116433586 A CN 116433586A
Authority
CN
China
Prior art keywords
tomographic image
breast ultrasound
ultrasound tomographic
image segmentation
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310149152.7A
Other languages
Chinese (zh)
Inventor
丁明跃
蔡梦媛
周亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huazhong University of Science and Technology
Original Assignee
Huazhong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huazhong University of Science and Technology filed Critical Huazhong University of Science and Technology
Priority to CN202310149152.7A priority Critical patent/CN116433586A/en
Publication of CN116433586A publication Critical patent/CN116433586A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30068Mammography; Breast

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Radiology & Medical Imaging (AREA)
  • Databases & Information Systems (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Ultra Sonic Daignosis Equipment (AREA)

Abstract

本发明公开了一种乳腺超声断层图像分割模型建立方法及分割方法,属于医学图像分割领域,包括:构建由已标注出病灶区域的乳腺超声断层图像构成的训练集;基于UNet模型构建初始的乳腺超声断层图像分割模型;以Ltotal=LE+αLvntrast+βLuncer为训练损失函数,利用训练集训练所构建的初始模型,完成乳腺超声断层图像分割模型的建立;其中,Ltotal表示总体损失;LE表示乳腺超声断层图像分割模型的分割误差;Lcontrast表示解码结构输出中间结果与标注结果之间的对比损失函数,Luncer为解码结构输出中间结果与标注结果之间的不确定性损失函数,α、β表示权重系数;优选地,模型在UNet中插入了CrossFormer模块和多尺度注意力模块。本发明能够有效提高提高乳腺超声断层图像的分割精度。

Figure 202310149152

The invention discloses a method for establishing a breast ultrasound tomographic image segmentation model and a segmentation method, belonging to the field of medical image segmentation, including: constructing a training set composed of breast ultrasound tomographic images with marked lesion areas; constructing an initial breast mammary gland based on a UNet model Ultrasonic tomographic image segmentation model; take L total = L E + α L vntrast + β L uncer as the training loss function, use the initial model constructed by training set training, and complete the establishment of the breast ultrasonic tomographic image segmentation model; wherein, L total represents the overall loss ; L E represents the segmentation error of the breast ultrasound tomographic image segmentation model; L contrast represents the contrast loss function between the output intermediate result of the decoding structure and the labeling result, and L uncer is the uncertainty loss between the output intermediate result of the decoding structure and the labeling result function, α and β represent weight coefficients; preferably, the model inserts a CrossFormer module and a multi-scale attention module in UNet. The invention can effectively improve the segmentation precision of the breast ultrasound tomographic image.

Figure 202310149152

Description

一种乳腺超声断层图像分割模型建立方法及分割方法A method for establishing a breast ultrasound tomographic image segmentation model and a segmentation method

技术领域technical field

本发明属于医学图像分割领域,更具体地,涉及一种乳腺超声断层图像分割模型建立方法及分割方法。The invention belongs to the field of medical image segmentation, and more specifically relates to a method for establishing a segmentation model of a breast ultrasound tomographic image and a segmentation method.

背景技术Background technique

乳腺癌是女性中仅次于肺癌的第二大致命癌症,而定期的乳腺筛查能够有效降低乳腺癌死亡率。超声断层成像技术相比传统超声具有高灵敏度和标准化操作的优点,其诊断结果不受医生经验的影响,在乳腺肿瘤的筛查方面具有广阔的应用前景。超声断层成像技术可以分为两种模态,分别是反射模态和透射模态,其中反射模态图像相比传统超声图像分辨率更高。透射图像可以提供病灶的功能信息,从而进一步辅助医生的诊断,因此,在针对乳腺成像时,往往使用透射模态。超声断层成像系统采用步进式扫描获得乳腺组织不同断层的图像,进而重建出乳腺组织的三维图像,使得病灶显示更为直观。Breast cancer is the second most deadly cancer in women after lung cancer, and regular breast screening can effectively reduce breast cancer mortality. Ultrasound tomography has the advantages of high sensitivity and standardized operation compared with traditional ultrasound, and its diagnostic results are not affected by the doctor's experience, so it has broad application prospects in the screening of breast tumors. Ultrasound tomography can be divided into two modes, reflection mode and transmission mode, in which reflection mode images have higher resolution than traditional ultrasound images. Transmission images can provide functional information of lesions, thereby further assisting doctors in diagnosis. Therefore, transmission modalities are often used for breast imaging. The ultrasonic tomography system uses step-by-step scanning to obtain images of different sections of breast tissue, and then reconstructs a three-dimensional image of breast tissue, making the display of lesions more intuitive.

超声断层成像系统使用的探头频率通常为2-3MHz,相比传统超声探头使用的大于10MHz频率低很多,导致图像分辨力较低,因此辨认乳腺超声断层图像中的病灶是一件具有挑战性且耗时的任务。目前计算机辅助诊断在医学中应用广泛,主要用于辅助医生进行诊断以及制定后续治疗方案,提高医生诊断的特异性和敏感性。在计算机辅助乳腺的诊断中,自动医学图像分割是提高诊断效率和准确率最关键的一个步骤。通过将病灶从超声断层图像中分割出来,有利于医生确定病灶的体积大小和病变程度,重建直观的三维形状,便于医生做出正确的诊断,准确地制定治疗方案。The frequency of the probe used in the ultrasound tomography system is usually 2-3MHz, which is much lower than the frequency of more than 10MHz used by the traditional ultrasound probe, resulting in low image resolution. Therefore, it is challenging to identify lesions in the breast ultrasound tomography image. time consuming task. At present, computer-aided diagnosis is widely used in medicine. It is mainly used to assist doctors in diagnosis and formulate follow-up treatment plans, and improve the specificity and sensitivity of doctors' diagnosis. In computer-aided breast diagnosis, automatic medical image segmentation is the most critical step to improve the efficiency and accuracy of diagnosis. By segmenting the lesion from the ultrasonic tomographic image, it is beneficial for the doctor to determine the size and extent of the lesion, reconstruct the intuitive three-dimensional shape, and facilitate the doctor to make a correct diagnosis and formulate a treatment plan accurately.

UNet模型在医学图像分割应用中具有良好的效果,因此,也广泛应用于乳腺超声断层图像分割中。UNet是一种典型的编码-解码结构,其结构如图1所示,左边卷积网络部分为编码结构,负责完成特征提取,主要由卷积层和下采样层构成,可以看到特征图尺寸不断减小;右边为解码结构,对应的是上采样过程,通过与不同卷积层的信息进行长跳跃连接(concat方式),恢复到和原图接近的大小。为了进一步提高乳腺超声断层图像的分割效果,研究人员对传统UNet模型进行改进,提出了一些改进的模型,例如Attention-UNet模型、UNet++模型、ResUNet模型。The UNet model has good results in medical image segmentation applications, so it is also widely used in breast ultrasound tomographic image segmentation. UNet is a typical encoding-decoding structure. Its structure is shown in Figure 1. The convolutional network part on the left is the encoding structure, which is responsible for feature extraction. It is mainly composed of convolutional layers and downsampling layers. You can see the size of the feature map It keeps decreasing; on the right is the decoding structure, corresponding to the upsampling process, through long jump connections (concat method) with information of different convolutional layers, to restore the size close to the original image. In order to further improve the segmentation effect of breast ultrasound tomographic images, researchers improved the traditional UNet model and proposed some improved models, such as Attention-UNet model, UNet++ model, and ResUNet model.

以上模型在训练完成后,能够有效分割出乳腺超声断层图像中的病灶区域,但是,不同于其他医学图像,乳腺超声断层图像的背景复杂且病灶与其他组织的对比度较低,而上述模型在解码阶段,上采样操作会损失大量信息,导致在实际应用中仍然面临着误分割的问题。此外,由于乳腺超声断层图像中,病灶往往较小,现有的模型也不能很好地将小病灶区域精确地分割出来。总体而言,现有的乳腺超声断层图像分割方法的分割精度有待进一步提高。After the above model is trained, it can effectively segment the lesion area in the breast ultrasound tomographic image. However, unlike other medical images, the breast ultrasound tomographic image has a complex background and a low contrast between the lesion and other tissues. stage, the upsampling operation will lose a lot of information, resulting in the problem of mis-segmentation still faced in practical applications. In addition, since the lesions are often small in breast ultrasound tomographic images, the existing models cannot accurately segment small lesions. Overall, the segmentation accuracy of existing breast ultrasound tomographic image segmentation methods needs to be further improved.

发明内容Contents of the invention

针对现有技术的缺陷和改进需求,本发明提供了一种乳腺超声断层图像分割模型建立方法及分割方法,其目的在于,提高乳腺超声断层图像的分割精度。Aiming at the defects and improvement needs of the prior art, the present invention provides a method for establishing a breast ultrasound tomographic image segmentation model and a segmentation method, the purpose of which is to improve the segmentation accuracy of the breast ultrasound tomographic image.

为实现上述目的,按照本发明的一个方面,提供了一种乳腺超声断层图像分割模型建立方法,包括:In order to achieve the above object, according to one aspect of the present invention, a method for establishing a breast ultrasound tomographic image segmentation model is provided, including:

构建训练集;训练集中,每一条训练数据为已标注出病灶区域的乳腺超声断层图像;Construct a training set; in the training set, each piece of training data is a breast ultrasound tomographic image with a marked lesion area;

基于UNet模型构建初始的乳腺超声断层图像分割模型,用于从乳腺超声断层图像中分割出病灶区域;Construct an initial breast ultrasound tomographic image segmentation model based on the UNet model, which is used to segment the lesion area from the breast ultrasound tomographic image;

以Ltotal=LE+αLcontrast+βLuncer为训练损失函数,利用训练集训练初始的乳腺超声断层图像分割模型,完成乳腺超声断层图像分割模型的建立;Taking L total = L E + α L contrast + β L uncer as the training loss function, using the training set to train the initial breast ultrasound tomographic image segmentation model, and completing the establishment of the breast ultrasound tomographic image segmentation model;

其中,Ltotal表示总体损失;LE表示乳腺超声断层图像分割模型的分割误差;Lcontrast表示乳腺超声断层图像分割模型中解码结构输出的中间结果与标注结果之间的对比损失函数,α表示其权重系数;Luncer为不确定性损失函数,用于表示乳腺超声断层图像分割模型中解码结构输出的中间结果与标注结果间的差异,β表示其权重系数;α≥0,β≥0,且α、β不同时为0。Among them, L total represents the overall loss; L E represents the segmentation error of the breast ultrasound tomographic image segmentation model; L contrast represents the contrast loss function between the intermediate results output by the decoding structure and the labeling results in the breast ultrasound tomographic image segmentation model, and α represents its Weight coefficient; Luncer is an uncertainty loss function, which is used to represent the difference between the intermediate result of the decoding structure output and the labeling result in the breast ultrasound tomographic image segmentation model, and β represents its weight coefficient; α≥0, β≥0, and α and β are not 0 at the same time.

进一步地,further,

Figure BDA0004090170100000031
Figure BDA0004090170100000031

Figure BDA0004090170100000032
Figure BDA0004090170100000032

Figure BDA0004090170100000033
Figure BDA0004090170100000033

其中,J表示模型的中间输出的数量;LCE()表示交叉熵损失;l表示标注结果;pj表示乳腺超声断层图像分割模型中解码结构输出的中间结果,

Figure BDA0004090170100000034
表示中间结果pj的第i个通道;DKL表示KL散度,C表示类别数,/>
Figure BDA0004090170100000035
表示分配给第j个中间结果中的像素p的权重。Among them, J represents the number of intermediate outputs of the model; L CE () represents the cross-entropy loss; l represents the labeling result; pj represents the intermediate result of the decoding structure output in the breast ultrasound tomographic image segmentation model,
Figure BDA0004090170100000034
Represents the i-th channel of the intermediate result p j ; D KL represents the KL divergence, C represents the number of categories, />
Figure BDA0004090170100000035
denotes the weight assigned to pixel p in the jth intermediate result.

进一步地,further,

Figure BDA0004090170100000036
Figure BDA0004090170100000036

其中,L表示像素之间的相似度;so表示中间结果中的特征,sl表示标注结果中的特征;m和n表示特征类别;

Figure BDA0004090170100000037
表示同一类别的特征,/>
Figure BDA0004090170100000038
表示不同类别的特征;τ表示温度系数。Among them, L represents the similarity between pixels; s o represents the feature in the intermediate result, s l represents the feature in the labeling result; m and n represent the feature category;
Figure BDA0004090170100000037
Represents features of the same class, />
Figure BDA0004090170100000038
Indicates the characteristics of different categories; τ indicates the temperature coefficient.

进一步地,乳腺超声断层图像分割模型还包括:插入在UNet模型中编码结构的最后一层与解码结构之间的CrossFormer模块。Further, the breast ultrasound tomographic image segmentation model further includes: a CrossFormer module inserted between the last layer of the encoding structure and the decoding structure in the UNet model.

进一步地,乳腺超声断层图像分割模型还包括:插入在UNet模型中编码结构与解码结构间长跳跃连接中的多尺度注意力模块;Further, the breast ultrasound tomographic image segmentation model also includes: a multi-scale attention module inserted in the long skip connection between the encoding structure and the decoding structure in the UNet model;

多尺度注意力模块包括:空洞空间卷积池化金字塔和特征增强模块;The multi-scale attention module includes: hole space convolution pooling pyramid and feature enhancement module;

空洞卷积池化金字塔用于对编码结构中对应层输出的特征进行多尺度特征提取,得到多尺度特征;The hollow convolution pooling pyramid is used to extract multi-scale features from the features output by the corresponding layers in the encoding structure to obtain multi-scale features;

特征增强模块,以多尺度特征和解码结构中对应层输出的解码特征为输入,用于增强与任务相关的特征并抑制与特征不相关的特征。The feature enhancement module, which takes multi-scale features and decoded features output by corresponding layers in the decoding structure as input, is used to enhance task-related features and suppress feature-irrelevant features.

进一步地,特征增强模块包括第一卷积层、第二卷积层、第三卷积层、ReLU激活层、Sigmoid激活层、像素相加层以及像素相乘层;Further, the feature enhancement module includes a first convolutional layer, a second convolutional layer, a third convolutional layer, a ReLU activation layer, a Sigmoid activation layer, a pixel addition layer, and a pixel multiplication layer;

第一卷积层用于输入多尺度特征,并进行卷积操作;The first convolutional layer is used to input multi-scale features and perform convolution operations;

第二卷积层用于输入解码特征,并进行卷积操作;The second convolutional layer is used to input decoding features and perform convolution operations;

像素相加层用于对第一卷积层和第二卷积层输出的结果进行逐像素相加;The pixel addition layer is used to add the results output by the first convolution layer and the second convolution layer pixel by pixel;

ReLU激活层用于对像素相加层输出的结果进行激活处理;The ReLU activation layer is used to activate the output of the pixel addition layer;

第三卷积层用于对ReLU激活层输出的结果进行卷积操作;The third convolutional layer is used to perform convolution operations on the results output by the ReLU activation layer;

Sigmoid激活层用于对第三卷积层输出的结果进行激活处理;The Sigmoid activation layer is used to activate the output of the third convolutional layer;

像素相乘层用于对Sigmoid激活层输出的结果以及多尺度特征进行逐像素相乘。The pixel multiplication layer is used to multiply the results output by the Sigmoid activation layer and the multi-scale features pixel by pixel.

进一步地,构建训练集,包括:Further, build a training set, including:

获得由乳腺超声断层图像构成的原始数据集,并标注其中的每一张乳腺超声断层图像中的病灶区域,得到对应的病灶区域掩码图像;Obtaining an original data set composed of breast ultrasound tomographic images, and marking the lesion area in each of the breast ultrasound tomographic images, and obtaining a corresponding lesion area mask image;

将连续的n张乳腺超声断层图像在通道维度上拼接,由拼接后的乳腺超声断层图像和对应的病灶区域掩码图像构成训练集;Stitching n consecutive breast ultrasound tomographic images in the channel dimension, and forming a training set from the spliced breast ultrasound tomographic images and corresponding lesion area mask images;

其中,n为大于1的正整数。Wherein, n is a positive integer greater than 1.

进一步地,n=3;Further, n=3;

并且,将连续的3张乳腺超声断层图像在通道维度上拼接,包括:In addition, three consecutive breast ultrasound tomographic images are spliced in the channel dimension, including:

从每张乳腺超声断层图像中提取一个通道的信息,得到三个通道的信息;Extract the information of one channel from each breast ultrasound tomographic image to obtain the information of three channels;

将三个通道的信息在通道维度上拼接。The information of the three channels is spliced in the channel dimension.

按照本发明的另一个方面,提供了一种乳腺超声断层图像分割方法,包括:According to another aspect of the present invention, there is provided a breast ultrasound tomographic image segmentation method, comprising:

将待分割的乳腺超声断层图像输入至乳腺超声断层图像分割模型,得到病灶区域分割结果;Input the breast ultrasound tomographic image to be segmented into the breast ultrasound tomographic image segmentation model to obtain the lesion area segmentation result;

其中,乳腺超声断层图像分割模型由本发明提供的乳腺超声断层图像分割模型建立方法所建立。Wherein, the breast ultrasound tomographic image segmentation model is established by the breast ultrasound tomographic image segmentation model establishment method provided by the present invention.

按照本发明的又一个方面,提供了一种计算机可读存储介质,包括存储的计算机程序;计算机程序被处理器执行时,控制计算机可读存储介质所在设备执行本发明提供的上述乳腺超声断层图像分割模型建立方法,和/或,本发明提供的上述乳腺超声断层图像分割方法。According to another aspect of the present invention, a computer-readable storage medium is provided, including a stored computer program; when the computer program is executed by a processor, it controls the device where the computer-readable storage medium is located to execute the above-mentioned breast ultrasound tomographic image provided by the present invention A method for establishing a segmentation model, and/or the above-mentioned method for segmenting breast ultrasound tomographic images provided by the present invention.

总体而言,通过本发明所构思的以上技术方案,能够取得以下有益效果:Generally speaking, through the above technical solutions conceived by the present invention, the following beneficial effects can be obtained:

(1)本发明在对所建立的乳腺超声断层图像分割模型进行训练时,在训练损失函数中引入了解码结构输出的中间结果与标注结果间对比损失函数和不确定性损失函数中的至少一项;对比损失函数的引入,使模型具备区分属于不同类别(即病灶区域或背景)像素的能力,更准确地对边界附近的像素进行分类,从而缓解乳腺超声断层图像对比度低对于分割精度的影响;不确定性损失函数的引入,使得解码结构输出的中间结果与标注结果之间的差异最小化,使得网络能够在早期学习到更多具有辨别性和可靠的知识,并有效减少解码阶段上采样操作导致的细节信息的损失,使得模型在对比度低的情况下,也能准确分割出病灶区域。总体而言,本发明通过对模型训练损失函数进行改进,能够有效提高乳腺超声断层图像的分割精度。(1) The present invention introduces at least one of the comparison loss function and the uncertainty loss function between the intermediate results output by the decoding structure and the labeling results in the training loss function when the established breast ultrasound tomographic image segmentation model is trained item; the introduction of the contrast loss function enables the model to have the ability to distinguish pixels belonging to different categories (ie, lesion area or background), and more accurately classify pixels near the boundary, thereby alleviating the impact of low contrast of breast ultrasound tomographic images on segmentation accuracy ; The introduction of the uncertainty loss function minimizes the difference between the intermediate results output by the decoding structure and the labeling results, enabling the network to learn more discriminative and reliable knowledge in the early stage, and effectively reduce the upsampling in the decoding stage The loss of detail information caused by the operation enables the model to accurately segment the lesion area even in the case of low contrast. Generally speaking, the present invention can effectively improve the segmentation accuracy of breast ultrasound tomographic images by improving the model training loss function.

(2)模型输出与标签之间的差异通常使用交叉熵损失和Dice损失进行衡量,但在乳腺超声断层图像分割中,仅基于交叉熵损失构建不确定性损失函数,对于上采样导致的细节信息损失的缓解效果有限;在本发明的优选方案中,所设计的不确定性损失函数,在交叉熵损失的基础上,引入了基于KL散度的正则项,可以最大程度上减少解码阶段上采样操作导致的细节信息的损失,进一步提高模型的分割精度。(2) The difference between the model output and the label is usually measured using cross-entropy loss and Dice loss, but in breast ultrasound tomographic image segmentation, the uncertainty loss function is only constructed based on cross-entropy loss, and the detailed information caused by upsampling The mitigation effect of the loss is limited; in the preferred solution of the present invention, the designed uncertainty loss function, on the basis of the cross-entropy loss, introduces a regular term based on KL divergence, which can minimize the upsampling in the decoding stage The loss of detail information caused by the operation further improves the segmentation accuracy of the model.

(3)在本发明的优选方案中,所设计的对比损失函数使得中间输出与标注结果中,属于相同类别的像素相似性最大化,并且属于不同类别的相似性最小化,能够进一步提高模型区分不同类别像素的能力。(3) In the preferred solution of the present invention, the designed comparison loss function maximizes the similarity of pixels belonging to the same category in the intermediate output and labeling results, and minimizes the similarity of pixels belonging to different categories, which can further improve model differentiation. Capabilities for different classes of pixels.

(4)乳腺超声断层图像分割任务属于密集预测任务,局部特征和全局特征对于该类任务都非常重要;在本发明的优选方案中,在UNet模型中编码结构的最后一层与解码结构之间插入了CrossFormer模块,在编码结构中的卷积层和池化层提取丰富的局部信息的基础上,由CrossFormer模块提取全局信息,并通过编码全局信息和局部信息之间的依赖关系,能使得小病灶的分割更加精确。(4) breast ultrasound tomographic image segmentation task belongs to intensive prediction task, and local feature and global feature are all very important for this type of task; The CrossFormer module is inserted. Based on the rich local information extracted by the convolutional layer and the pooling layer in the encoding structure, the CrossFormer module extracts the global information, and by encoding the dependency between the global information and the local information, it can make small The segmentation of lesions is more precise.

(5)在本发明的优选方案中,在UNet模型的长跳跃连接中插入了多尺度注意力模块,该模块具体由空洞空间卷积池化金字塔和特征增强模块构成,其中的空洞空间卷积池化金字塔使得网络在增大感受野和捕捉多尺度信息的同时尽可能地保留局部细节信息,特征增强模块则对空洞空间卷积池化金字塔提取的多尺度特征做进一步的处理,增强与任务相关的特征并抑制与特征不相关的特征,最终使得模型对于小病灶也具有较高的分割精度。(5) In the preferred solution of the present invention, a multi-scale attention module is inserted into the long skip connection of the UNet model, which is specifically composed of an atrous spatial convolution pooling pyramid and a feature enhancement module, in which the atrous spatial convolution The pooling pyramid enables the network to retain local detail information as much as possible while increasing the receptive field and capturing multi-scale information. The feature enhancement module further processes the multi-scale features extracted by the hollow space convolution pooling pyramid, enhancing and task The relevant features and suppress the features that are not related to the features, and finally make the model have higher segmentation accuracy for small lesions.

(6)在超声断层成像过程中,对于同一个对象会生成多张图像;在本发明的优选方案中,在构建用于模型训练的数据集时,将连续的多张图像在通道维度上拼接后作为模型输入,能够有效利用超声断层图像的层间信息,进一步提高分割精度。在进一步优选的方案中,具体选取连续的三张图像分别提取一个通道拼接后作为模型输入,一方面可以避免拼接后的图像包含过多的背景区域特征,另一方面,拼接后的图像维度与原始图像维度一致,简化了后续分割过程中模型输入的处理。(6) In the process of ultrasonic tomography, multiple images will be generated for the same object; in the preferred solution of the present invention, when constructing a data set for model training, multiple continuous images are spliced on the channel dimension Finally, as a model input, the interlayer information of the ultrasonic tomographic image can be effectively used to further improve the segmentation accuracy. In a further preferred solution, three consecutive images are specifically selected to extract a channel and then spliced as the model input. On the one hand, it can avoid the spliced image from including too many background region features; on the other hand, the spliced image dimension and The dimensionality of the original image is consistent, which simplifies the processing of model input in the subsequent segmentation process.

附图说明Description of drawings

图1为现有的UNet模型结构示意图;Figure 1 is a schematic diagram of the existing UNet model structure;

图2为本发明实施例提供的乳腺超声断层图像及其对应的掩码图像;其中,(a)为乳腺超声断层图像,(b)为掩码图像;Fig. 2 is the ultrasound tomographic image of the breast and its corresponding mask image provided by the embodiment of the present invention; wherein, (a) is the ultrasound tomogram of the breast, and (b) is the mask image;

图3为本发明实施例提供的乳腺超声断层图像分割模型示意图;FIG. 3 is a schematic diagram of a breast ultrasound tomographic image segmentation model provided by an embodiment of the present invention;

图4为现有的CrossFormer模块示意图;Figure 4 is a schematic diagram of the existing CrossFormer module;

图5为本发明实施例提供的多尺度注意力模块示意图;FIG. 5 is a schematic diagram of a multi-scale attention module provided by an embodiment of the present invention;

图6为发明实施例提供的分割方法与其他基于UNet模型的分割方法的分割结果;其中,(a)为乳腺超声断层图像的病灶区域标签,(b)为UNet模型的分割结果,(c)为Attention-UNet模型分割结果,(d)为UNet++模型分割结果,(e)为ResUNet模型分割结果,(f)为TransUNet模型分割结果,(g)本实施例的分割结果。Fig. 6 is the segmentation result of the segmentation method provided by the embodiment of the invention and other segmentation methods based on the UNet model; wherein, (a) is the lesion area label of the breast ultrasound tomographic image, (b) is the segmentation result of the UNet model, (c) (d) is the segmentation result of the UNet++ model, (e) is the segmentation result of the ResUNet model, (f) is the segmentation result of the TransUNet model, and (g) is the segmentation result of this embodiment.

具体实施方式Detailed ways

为了使本发明的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。此外,下面所描述的本发明各个实施方式中所涉及到的技术特征只要彼此之间未构成冲突就可以相互组合。In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention. In addition, the technical features involved in the various embodiments of the present invention described below can be combined with each other as long as they do not constitute a conflict with each other.

在本发明中,本发明及附图中的术语“第一”、“第二”等(如果存在)是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。In the present invention, the terms "first", "second" and the like (if any) in the present invention and drawings are used to distinguish similar objects, and are not necessarily used to describe a specific order or sequence.

为了提高乳腺超声断层图像分割的精度,本发明提供了一种乳腺超声断层图像分割模型建立方法及分割方法,其整体思路在于:通过对模型训练过程的损失函数进行优化,使模型具备区分不同类别像素的能力,和/或,在早期学习到更多具有辨别性和可靠的知识,并减少由于上采样操作导致的细节信息的损失,使模型能够从对比度较低的乳腺超声断层图像中准确分割出病灶区域。在此基础上,进一步对模型的结构进行改进,使模型更好地利用图像中全局信息和局部的细节信息,进一步提高分割精度。In order to improve the accuracy of breast ultrasound tomographic image segmentation, the present invention provides a method for establishing a breast ultrasound tomographic image segmentation model and a segmentation method. pixels, and/or, learn more discriminative and reliable knowledge at an early stage, and reduce the loss of detail information due to upsampling operations, enabling the model to accurately segment from low-contrast breast ultrasound tomographic images out of the lesion area. On this basis, the structure of the model is further improved, so that the model can better utilize the global information and local detail information in the image, and further improve the segmentation accuracy.

以下为实施例。The following are examples.

实施例1:Example 1:

一种乳腺超声断层图像分割模型建立方法,包括:A method for establishing a breast ultrasound tomographic image segmentation model, comprising:

构建训练集;训练集中,每一条训练数据为已标注出病灶区域的乳腺超声断层图像;图2所示为训练数据的一个示例,图2中的(a)为采集到的乳腺超声断层图像,(b)为标注出病灶区域的掩码图像。Build a training set; in the training set, each piece of training data is a breast ultrasound tomographic image that has marked the lesion area; Figure 2 shows an example of training data, and (a) in Figure 2 is the breast ultrasound tomographic image collected, (b) is the mask image of the marked lesion area.

可选地,本实施例中,构建训练集的具体包括如下步骤:Optionally, in this embodiment, constructing the training set specifically includes the following steps:

(1)使用2-3MHz的超声断层成像系统采集乳腺超声断层图像,共采集到372张图像,并将其按7:1:2的比例划分为训练集、验证集和测试集;(1) Use the ultrasound tomography system of 2-3MHz to collect breast ultrasound tomography images, collect 372 images in total, and divide them into training set, verification set and test set according to the ratio of 7:1:2;

(2)对采集到的372张图像采用专业、同样的方式标注出病灶区域,得到对应的掩码图像,作为模型训练时的标签;(2) Mark the lesion area in a professional and the same way on the collected 372 images, and obtain the corresponding mask image as the label during model training;

(3)对训练集中的乳腺超声断层图像和对应的标签通过图像翻转、旋转、缩放等操作进行数据增强,扩充训练集至原始训练集的8倍,共计2088张;(3) Carry out data enhancement to the mammary gland ultrasound tomographic images and corresponding labels in the training set through operations such as image flipping, rotation, and zooming, and expand the training set to 8 times the original training set, with a total of 2088 pieces;

通过数据增强,能够提高模型的鲁棒性以及避免过拟合;Through data enhancement, the robustness of the model can be improved and overfitting can be avoided;

(4)将数据增强后的训练图像尺寸统一归一化到512×512,以适应模型对于输入图像尺寸的要求,之后将连续的三张图像在通道维度上进行拼接,具体拼接方式为:从每张图像中提取一个通道的信息,然后将提取的三个通道的信息在通道维度上拼接,得到的三通道图像作为分割模型的输入,即输入图像尺寸为512×512×3;(4) Normalize the size of the training image after data enhancement to 512×512 to meet the requirements of the model for the input image size, and then stitch the three consecutive images in the channel dimension. The specific stitching method is: from The information of one channel is extracted from each image, and then the information of the extracted three channels is spliced in the channel dimension, and the obtained three-channel image is used as the input of the segmentation model, that is, the input image size is 512×512×3;

在超声断层成像过程中,对于同一个对象会生成多张(通常为30张)图像,本实施例将连续的多张图像拼接后作为模型输入,能够有效利用层间信息;乳腺超声断层图像中的病灶往往较小,不同层的图像之间往往具有一定的层间距,此外,原始的乳腺超声断层图像为三通道图像,每个通道的信号相同,本实施例具体从连续的三张超声图像中分别提取一个通道后再拼接为三通道图像作为模型输入,在利用层间信息的基础上,既避免了避免拼接后的图像包含过多的背景区域特征,又能够保持拼接后的图像维度与原始图像维度一致,使得实际分割过程中,原始的乳腺超声断层图像的通道维度与模型输入要求一致,简化了后续分割过程中模型输入的处理。In the ultrasonic tomographic imaging process, multiple (usually 30) images will be generated for the same object. In this embodiment, the continuous multiple images are spliced and used as model input, which can effectively use the interlayer information; in the breast ultrasonic tomographic image The lesions are often small, and there is often a certain layer spacing between images of different layers. In addition, the original breast ultrasound tomographic image is a three-channel image, and the signals of each channel are the same. This embodiment specifically uses three consecutive ultrasound images One channel is extracted respectively and then spliced into a three-channel image as the model input. On the basis of using inter-layer information, it not only avoids avoiding the spliced image from including too many background region features, but also keeps the spliced image dimension and The dimensionality of the original image is consistent, so that in the actual segmentation process, the channel dimension of the original breast ultrasound tomographic image is consistent with the model input requirements, which simplifies the processing of the model input in the subsequent segmentation process.

需要说明的是,以上训练集的构建方式,仅为本发明优选的实施方式,不应理解为对本发明的唯一限定,其他由标注了病灶区域的乳腺超声断层图像的数据集均可用于本发明;容易理解的是,上述步骤中,图像数量、图像尺寸等参数仅为示例性描述,不应理解为对本发明的唯一限定,在实际应用中可按需调整。It should be noted that the construction method of the above training set is only a preferred embodiment of the present invention, and should not be understood as the only limitation of the present invention. Other data sets of breast ultrasound tomographic images marked with lesion areas can be used in the present invention It is easy to understand that in the above steps, parameters such as the number of images and image size are only exemplary descriptions, and should not be understood as the only limitation to the present invention, and can be adjusted as needed in practical applications.

本实施例进一步包括:基于UNet模型构建初始的乳腺超声断层图像分割模型,用于从乳腺超声断层图像中分割出病灶区域。This embodiment further includes: constructing an initial breast ultrasound tomographic image segmentation model based on the UNet model, for segmenting lesion areas from the breast ultrasound tomographic image.

考虑到原始的UNet模型及基于UNet模型改进得到的模型,由于在解码阶段的上采样过程中会损失细节信息,无法较好地分割出乳腺超声断层图像中的小病灶,本实施例在UNet模型的基础上改进得到了一种新的乳腺超声断层图像分割模型,所选用的UNet模型具体为图1所示的模型,相关的改进包括:在UNet模型中编码结构的最后一层与解码结构之间插入CrossFormer模块,以及,在UNet模型的长跳跃连接中插入多尺度注意力模块(MFEModule)。Considering the original UNet model and the improved model based on the UNet model, since the detailed information will be lost in the up-sampling process of the decoding stage, the small lesions in the breast ultrasound tomographic image cannot be well segmented. In this embodiment, the UNet model A new breast ultrasound tomographic image segmentation model was obtained based on the improvement of the model. The selected UNet model is specifically the model shown in Figure 1. The related improvements include: the last layer of the encoding structure and the decoding structure in the UNet model Insert the CrossFormer module between them, and insert the multi-scale attention module (MFEModule) in the long skip connection of the UNet model.

改进后的分割模型结构如图3所示,可分为十二层:第一层由两个卷积层构成,第二层至第四层均是由一个2×2池化层和两个卷积层组成的模块,卷积层由一个3×3卷积、一个批量归一化模块以及一个ReLU激活函数所构成;第一层的输出特征图大小为512×512×64(512表示特征图的宽高为512,64表示特征图的通道数),第二层至第四层输出特征图的通道数是上一层输出特征图的两倍,宽高是上一层的1/2,第四层的输出特征图大小为64×64×512,第五层由一个池化层和一个卷积层构成,第五层输出特征图的通道数与第四层相同,但宽高仍然减半,即第五层输出特征图大小为32×32×512。The structure of the improved segmentation model is shown in Figure 3, which can be divided into twelve layers: the first layer is composed of two convolutional layers, and the second to fourth layers are composed of a 2×2 pooling layer and two A module composed of a convolutional layer. The convolutional layer consists of a 3×3 convolution, a batch normalization module, and a ReLU activation function; the output feature map size of the first layer is 512×512×64 (512 represents the feature The width and height of the map is 512, 64 represents the number of channels of the feature map), the number of channels of the output feature map of the second layer to the fourth layer is twice that of the output feature map of the previous layer, and the width and height are 1/2 of the previous layer , the size of the output feature map of the fourth layer is 64×64×512, the fifth layer consists of a pooling layer and a convolutional layer, the number of channels of the output feature map of the fifth layer is the same as that of the fourth layer, but the width and height are still Halved, that is, the output feature map size of the fifth layer is 32×32×512.

参阅图3,乳腺超声断层图像分割模型中,第六层为插入的CrossFormer模块;CrossFormer模块由两个连续的CrossFormerBlock构成,如图4所示。其中一个CrossFormerBlock包含一个跨尺度嵌入层(CEL)、两个层归一化(Layer Norm)、一个相对位置编码层(RPB)、一个短距离注意力层(SDA)和一个多层感知机层(MLP)。另一个CrossFormerBlock包含两个层归一化(Layer Norm)、一个相对位置编码层(RPB)、一个长距离注意力层(LDA)和一个多层感知机层(MLP)。第六层的输出特征图尺寸与第五层的相同,即仍为32×32×512。第七层由一个卷积层和一个上采样层构成,第七层输出特征图的通道数与第六层相同,宽高是第六层的2倍,即第七层输出特征图大小为64×64×512。Referring to Fig. 3, in the breast ultrasound tomographic image segmentation model, the sixth layer is the inserted CrossFormer module; the CrossFormer module is composed of two continuous CrossFormerBlocks, as shown in Fig. 4 . One of the CrossFormerBlocks consists of a cross-scale embedding layer (CEL), two layers of normalization (Layer Norm), a relative position encoding layer (RPB), a short distance attention layer (SDA) and a multi-layer perceptron layer ( MLP). Another CrossFormerBlock contains two layers of normalization (Layer Norm), a relative position encoding layer (RPB), a long-distance attention layer (LDA) and a multi-layer perceptron layer (MLP). The output feature map size of the sixth layer is the same as that of the fifth layer, that is, still 32×32×512. The seventh layer is composed of a convolutional layer and an upsampling layer. The number of channels of the output feature map of the seventh layer is the same as that of the sixth layer, and the width and height are twice that of the sixth layer, that is, the output feature map size of the seventh layer is 64 ×64×512.

本实施例在UNet模型中编码结构的最后一层与解码结构之间插入了CrossFormer模块,在编码结构中的卷积层和池化层提取丰富的局部信息的基础上,由CrossFormer模块提取全局信息,并通过编码全局信息和局部信息之间的依赖关系,能使得小病灶的分割更加精确。In this embodiment, a CrossFormer module is inserted between the last layer of the encoding structure and the decoding structure in the UNet model. On the basis of extracting rich local information from the convolutional layer and pooling layer in the encoding structure, the CrossFormer module extracts global information. , and by encoding the dependency between global information and local information, it can make the segmentation of small lesions more accurate.

参阅图3,乳腺超声断层图像分割模型中,第八层至第十层均由两个卷积层和一个上采样层构成,每一层输出特征图的通道数变为输入特征图通道数的1/2,宽高变为输入特征图宽高的两倍。第十层输出特征图的宽高恢复到输入图像的大小,为512×512。第十一层由两个卷积层构成,第十二层由一个1×1卷积层构成,输出通道数为种类数,输出特征图大小为512×512×2。Referring to Figure 3, in the breast ultrasound tomographic image segmentation model, the eighth to tenth layers are composed of two convolutional layers and an upsampling layer, and the number of channels of the output feature map of each layer becomes the number of channels of the input feature map 1/2, the width and height become twice the width and height of the input feature map. The width and height of the output feature map of the tenth layer are restored to the size of the input image, which is 512×512. The eleventh layer is composed of two convolutional layers, the twelfth layer is composed of a 1×1 convolutional layer, the number of output channels is the number of categories, and the output feature map size is 512×512×2.

以CrossFormer模块为区分,左侧结构对应编码结构,右侧结构对应解码结构。The CrossFormer module is used as a distinction, the structure on the left corresponds to the encoding structure, and the structure on the right corresponds to the decoding structure.

参阅图3,乳腺超声断层图像分割模型中,编码结构中的第一、二、三、四层输出经过插入的多尺度注意力模块(MFEModule)与解码结构中的第十、九、八、七层输出使用一个长跳跃连接拼接起来。第一层与第十层输出特征图的大小均为512×512×64,第二层与第九层输出特征图的大小均为256×256×128,第三层与第八层输出特征图的大小均为128×128×256,第四层与第七层输出特征图的大小均为64×64×512。即:第一层和第十层的输出特征图拼接起来为512×512×128后再进行卷积操作,第二层和第九层的输出特征图拼接起来为256×256×256后再进行卷积操作,第三层和第八层的输出特征图拼接起来为128×128×512后再进行卷积操作,第四层和第七层的输出特征图拼接起来为64×64×1024后再进行卷积操作。Referring to Figure 3, in the breast ultrasound tomographic image segmentation model, the output of the first, second, third, and fourth layers in the encoding structure is inserted into the multi-scale attention module (MFEModule) and the tenth, ninth, eighth, and seventh layers in the decoding structure Layer outputs are concatenated using a long skip connection. The size of the output feature map of the first layer and the tenth layer is 512×512×64, the size of the output feature map of the second layer and the ninth layer is 256×256×128, and the output feature map of the third layer and the eighth layer The size of the output feature maps of the fourth layer and the seventh layer are both 64×64×512. That is: the output feature maps of the first layer and the tenth layer are concatenated to 512×512×128 and then the convolution operation is performed, and the output feature maps of the second layer and the ninth layer are concatenated to be 256×256×256 and then performed Convolution operation, the output feature maps of the third layer and the eighth layer are concatenated into 128×128×512 and then the convolution operation is performed, and the output feature maps of the fourth and seventh layers are concatenated into 64×64×1024 Then perform the convolution operation.

本实施例中,多尺度注意力模块(MFEModule)的结构如图5所示,包含两个分支,其中一个分支是一个空洞空间卷积池化金字塔(ASPPModule),用于对编码结构中对应层输出的特征进行多尺度特征提取,得到多尺度特征;ASPPModule包含一个1×1的卷积、三个膨胀系数分别为6、12和18的3×3空洞空间卷积和一个1×1的池化层,ASPP的结构使得网络在增大感受野和捕捉多尺度信息的同时尽可能地保留局部细节信息;另一个分支是一个特征增强模块,包含三个1×1的卷积(即第一卷积层、第二卷积层和第三卷积层)、一个ReLU激活层、一个Sigmoid激活层、一个像素相加层以及像素相乘层,如图5所示:In this embodiment, the structure of the multi-scale attention module (MFEModule) is shown in Figure 5, which includes two branches, one of which is an atrous space convolution pooling pyramid (ASPPModule), which is used for the corresponding layer in the coding structure The output features are multi-scale feature extracted to obtain multi-scale features; ASPPModule contains a 1×1 convolution, three 3×3 hole space convolutions with expansion coefficients of 6, 12 and 18, and a 1×1 pool The structure of ASPP enables the network to preserve local details as much as possible while increasing the receptive field and capturing multi-scale information; the other branch is a feature enhancement module, which contains three 1×1 convolutions (that is, the first Convolutional layer, second convolutional layer and third convolutional layer), a ReLU activation layer, a Sigmoid activation layer, a pixel addition layer and a pixel multiplication layer, as shown in Figure 5:

第一卷积层用于输入多尺度特征,并进行卷积操作;The first convolutional layer is used to input multi-scale features and perform convolution operations;

第二卷积层用于输入解码特征,并进行卷积操作;The second convolutional layer is used to input decoding features and perform convolution operations;

像素相加层用于对第一卷积层和第二卷积层输出的结果进行逐像素相加;The pixel addition layer is used to add the results output by the first convolution layer and the second convolution layer pixel by pixel;

ReLU激活层用于对像素相加层输出的结果进行激活处理;The ReLU activation layer is used to activate the output of the pixel addition layer;

第三卷积层用于对ReLU激活层输出的结果进行卷积操作;The third convolutional layer is used to perform convolution operations on the results output by the ReLU activation layer;

Sigmoid激活层用于对第三卷积层输出的结果进行激活处理;The Sigmoid activation layer is used to activate the output of the third convolutional layer;

像素相乘层用于对Sigmoid激活层输出的结果以及多尺度特征进行逐像素相乘;相乘后的结果作为下一层的输入。The pixel multiplication layer is used to multiply the results output by the Sigmoid activation layer and the multi-scale features pixel by pixel; the multiplied result is used as the input of the next layer.

通过该特征增强模块对ASPP提取的多尺度特征做进一步的处理,能够增强与任务相关的特征并抑制与特征不相关的特征;需要说明的是,此处关于特征增强模块的结构描述,为优选方案,不应理解为对本发明的唯一限定,其他基于注意力机制实现的能够增强与任务相关的特征并抑制与特征不相关的特征的模块,也可用于本发明。Through the feature enhancement module to further process the multi-scale features extracted by ASPP, the features related to the task can be enhanced and the features not related to the feature can be suppressed; it should be noted that the structural description of the feature enhancement module here is the best The solution should not be construed as the only limitation to the present invention, and other modules based on the attention mechanism that can enhance the features related to the task and suppress the features that are not related to the feature can also be used in the present invention.

总体而言,本实施例通过在UNet模型的长跳跃连接中插入上述特征增强模块,能够使得模型对于小病灶也具有较高的分割精度。In general, this embodiment can make the model have higher segmentation accuracy for small lesions by inserting the above feature enhancement module into the long skip connection of the UNet model.

需要说明的是,上述模型仅为本发明优选的模型,不应理解为对本发明的唯一限定;在本发明其他的实施例中,在分割精度满足应用需求的情况下,可以仅在UNet模型中相应位置插入CrossFormer模块,也可以仅插入多尺度注意力模块,也可以直接使用现有的UNet模型或者Attention-UNet模型、UNet++模型、ResUNet模型等改进模型。It should be noted that the above model is only the preferred model of the present invention, and should not be understood as the only limitation of the present invention; in other embodiments of the present invention, when the segmentation accuracy meets the application requirements, it can only be used in the UNet model Insert the CrossFormer module at the corresponding position, or only insert the multi-scale attention module, or directly use the existing UNet model or Attention-UNet model, UNet++ model, ResUNet model and other improved models.

容易理解的是,当选用的UNet模型结构发生变化时,改进后的乳腺超声断层图像分割模型的层数、每一层中的参数也可能相应发生变化,以上关于模型参数的说明,仅为一种示例性的描述,不应理解为对本发明的唯一限定。It is easy to understand that when the structure of the selected UNet model changes, the number of layers of the improved breast ultrasound tomographic image segmentation model and the parameters in each layer may also change accordingly. The above description of the model parameters is only a summary This exemplary description should not be construed as the only limitation of the present invention.

上述初始的乳腺超声断层图像分割模型需要经过训练之后,才能完成实际的分割任务;模型输出与标签之间的差异通常使用交叉熵损失和Dice损失进行衡量,但是,由于乳腺超声断层图像中病灶区域和背景之间的对比度较低,而UNet模型及其改进模型中,在解码阶段,上采样操作会损失大量信息,现有的训练方法训练所得模型仍会面临误分割的问题,针对该问题,本实施例在模型的训练损失函数进中引入了解码结构输出的中间结果(即图3中的O2,O3,O4)与标注结果间的对比损失函数和不确定性损失函数,改进后的损失函数为:The above initial breast ultrasound tomographic image segmentation model needs to be trained before it can complete the actual segmentation task; the difference between the model output and the label is usually measured by cross-entropy loss and Dice loss. The contrast between the background and the background is low, and in the UNet model and its improved model, in the decoding stage, the upsampling operation will lose a lot of information, and the model trained by the existing training method will still face the problem of mis-segmentation. To solve this problem, This embodiment introduces the comparison loss function and uncertainty loss function between the intermediate results output by the decoding structure (ie O 2 , O 3 , O 4 in Figure 3 ) and the marked results in the training loss function of the model, improving The final loss function is:

Ltotal=LE+αLcontrast+βLuncer L total =L E +αL contrast +βL uncer

其中,Ltotal表示总体损失;LE表示乳腺超声断层图像分割模型的分割误差;Lcontrast表示乳腺超声断层图像分割模型中解码结构输出的中间结果与标注结果之间的对比损失函数,α表示其权重系数;Luncer为不确定性损失函数,用于表示乳腺超声断层图像分割模型中解码结构输出的中间结果与标注结果间的差异,β表示其权重系数;α≥0,β≥0,且α、β不同时为0;α、β的具体取值可根据实际的分割效果灵活调整,可选地,本实施例中,α、β分别设置为0.2和0.3。Among them, L total represents the overall loss; L E represents the segmentation error of the breast ultrasound tomographic image segmentation model; L contrast represents the contrast loss function between the intermediate results output by the decoding structure and the labeling results in the breast ultrasound tomographic image segmentation model, and α represents its Weight coefficient; Luncer is an uncertainty loss function, which is used to represent the difference between the intermediate result of the decoding structure output and the labeling result in the breast ultrasound tomographic image segmentation model, and β represents its weight coefficient; α≥0, β≥0, and α and β are different from 0 at the same time; the specific values of α and β can be flexibly adjusted according to the actual segmentation effect. Optionally, in this embodiment, α and β are set to 0.2 and 0.3 respectively.

本实施例所设计的上述损失函数中,通过引入不确定性损失函数,使得解码结构输出的中间结果与标注结果之间的差异最小化,使得网络能够在早期学习到更多具有辨别性和可靠的知识,并有效减少解码阶段上采样操作导致的细节信息的损失,使得模型在对比度低的情况下,也能准确分割出病灶区域。实验发现,仅基于交叉熵损失构建不确定性损失函数,对于上采样导致的细节信息损失的缓解效果有限,为了最大程度上减少解码阶段上采样操作导致的细节信息的损失,本实施例在交叉熵损失的基础上,引入了基于KL散度的正则项,最终,不确定性损失函数的表达式具体为:In the above loss function designed in this embodiment, by introducing the uncertainty loss function, the difference between the intermediate result output by the decoding structure and the labeling result is minimized, so that the network can learn more discriminative and reliable knowledge, and effectively reduce the loss of detail information caused by the upsampling operation in the decoding stage, so that the model can accurately segment the lesion area even in the case of low contrast. Experiments have found that constructing an uncertainty loss function based only on cross-entropy loss has a limited mitigation effect on the loss of detail information caused by upsampling. In order to minimize the loss of detail information caused by upsampling operations in the decoding stage, this embodiment uses cross On the basis of entropy loss, a regular term based on KL divergence is introduced. Finally, the expression of the uncertainty loss function is specifically:

Figure BDA0004090170100000131
Figure BDA0004090170100000131

Figure BDA0004090170100000132
Figure BDA0004090170100000132

Figure BDA0004090170100000133
Figure BDA0004090170100000133

其中,J表示模型的中间输出的数量;LCE()表示交叉熵损失;l表示标注结果;pj表示乳腺超声断层图像分割模型中解码结构输出的中间结果,

Figure BDA0004090170100000141
表示中间结果pj的第i个通道;DKL表示KL散度,C表示类别数,/>
Figure BDA0004090170100000142
表示分配给第j个中间结果中的像素p的权重。Among them, J represents the number of intermediate outputs of the model; L CE () represents the cross-entropy loss; l represents the labeling result; pj represents the intermediate result of the decoding structure output in the breast ultrasound tomographic image segmentation model,
Figure BDA0004090170100000141
Represents the i-th channel of the intermediate result p j ; D KL represents the KL divergence, C represents the number of categories, />
Figure BDA0004090170100000142
denotes the weight assigned to pixel p in the jth intermediate result.

对比损失函数的引入,使模型具备区分属于不同类别(即病灶区域或背景)像素的能力,更准确地对边界附近的像素进行分类,从而缓解乳腺超声断层图像对比度低对于分割精度的影响;本实施例中,对比损失函数的表达式为具体:The introduction of the contrast loss function enables the model to have the ability to distinguish pixels belonging to different categories (ie, lesion area or background), and more accurately classify pixels near the boundary, thereby alleviating the impact of low contrast of breast ultrasound tomographic images on segmentation accuracy; In the embodiment, the expression of the comparison loss function is specific:

Figure BDA0004090170100000143
Figure BDA0004090170100000143

其中,L表示像素之间的相似度;so表示中间结果中的特征,sl表示标注结果中的特征;m和n表示特征类别,

Figure BDA0004090170100000144
表示标注结果中的第m类特征,/>
Figure BDA0004090170100000145
和/>
Figure BDA0004090170100000146
分别表示中间结果so中的第m、n类特征;/>
Figure BDA0004090170100000147
表示同一类别的特征,/>
Figure BDA0004090170100000148
表示不同类别的特征;τ表示温度系数,可选地,本实施例中,其值为2;Among them, L represents the similarity between pixels; s o represents the feature in the intermediate result, s l represents the feature in the labeling result; m and n represent the feature category,
Figure BDA0004090170100000144
Indicates the mth class feature in the labeling result, />
Figure BDA0004090170100000145
and />
Figure BDA0004090170100000146
Respectively represent the m and nth class features in the intermediate result s o ; />
Figure BDA0004090170100000147
Represents features of the same class, />
Figure BDA0004090170100000148
Represents the characteristics of different categories; τ represents the temperature coefficient, optionally, in this embodiment, its value is 2;

可选地,可选地,本实施例中,具体采用余弦相似度进行衡量,相应的表达式为Optionally, optionally, in this embodiment, cosine similarity is specifically used for measurement, and the corresponding expression is

Figure BDA0004090170100000149
Figure BDA0004090170100000149

应当说明的是,余弦相似度仅为可选的特征相似度衡量方式,不应理解为对本发明的唯一限定,其他衡量方式,如皮尔逊相关系数、马氏距离、欧式距离等,也可以用于本发明。It should be noted that the cosine similarity is only an optional feature similarity measurement method, and should not be understood as the only limitation to the present invention. Other measurement methods, such as Pearson correlation coefficient, Mahalanobis distance, Euclidean distance, etc., can also be used in the present invention.

基于该对比损失函数,使得中间输出与标注结果中,属于相同类别的像素相似性最大化,并且属于不同类别的相似性最小化,能够进一步提高模型区分不同类别像素的能力。Based on the comparison loss function, the similarity of pixels belonging to the same category in the intermediate output and labeling results is maximized, and the similarity of pixels belonging to different categories is minimized, which can further improve the ability of the model to distinguish pixels of different categories.

可选地,在上述损失函数中,乳腺超声断层图像分割模型的分割误差LE,即病灶区域分割结果O1和标签之间的误差,仍使用交叉熵损失函数LCE和Dice损失函数LDice表示。Optionally, in the above loss function, the segmentation error LE of the breast ultrasound tomographic image segmentation model, that is, the error between the lesion region segmentation result O 1 and the label, still uses the cross-entropy loss function L CE and the Dice loss function L Dice express.

最终,总体损失函数表达式如下:Finally, the overall loss function expression is as follows:

Figure BDA0004090170100000151
Figure BDA0004090170100000151

LCE=-[ylogl+(1-y)log(1-l)]L CE =-[ylogl+(1-y)log(1-l)]

Figure BDA0004090170100000152
Figure BDA0004090170100000152

其中,y表示乳腺超声断层图像分割模型输出的分割结果,l表示标注结果;Wherein, y represents the segmentation result output by the breast ultrasound tomographic image segmentation model, and l represents the labeling result;

基于以上分析,本实施例在数据集和初始的乳腺超声断层图像分割模型构建之后,进一步包括:Based on the above analysis, after the data set and the initial breast ultrasound tomographic image segmentation model are constructed, this embodiment further includes:

以Ltotal=LE+αLcontrast+βLuncer为训练损失函数,利用训练集训练初始的乳腺超声断层图像分割模型,完成乳腺超声断层图像分割模型的建立。Taking L total = L E + αL contrast + βL uncer as the training loss function, the initial breast ultrasound tomographic image segmentation model is trained using the training set, and the establishment of the breast ultrasound tomographic image segmentation model is completed.

需要说明的是,在本发明其他的一些实施例中,超参数,α、β中的一项可以为0。It should be noted that, in some other embodiments of the present invention, one of the hyperparameters α and β may be 0.

容易理解的是,在模型训练完成之后,为了确保分割结果满足要求,进一步会利用测试集和验证集对训练后的模型进行测试和验证,测试过程中,可使用Dice系数和IoU作为评价指标对分割结果进行评价。It is easy to understand that after the model training is completed, in order to ensure that the segmentation results meet the requirements, the trained model will be further tested and verified using the test set and verification set. During the test, the Dice coefficient and IoU can be used as evaluation indicators for Segmentation results are evaluated.

实施例2:Example 2:

一种乳腺超声断层图像分割方法,包括:A breast ultrasound tomographic image segmentation method, comprising:

将待分割的乳腺超声断层图像输入至乳腺超声断层图像分割模型,得到病灶区域分割结果;Input the breast ultrasound tomographic image to be segmented into the breast ultrasound tomographic image segmentation model to obtain the lesion area segmentation result;

其中,乳腺超声断层图像分割模型由本发明提供的乳腺超声断层图像分割模型建立方法所建立。Wherein, the breast ultrasound tomographic image segmentation model is established by the breast ultrasound tomographic image segmentation model establishment method provided by the present invention.

实施例3:Example 3:

一种计算机可读存储介质,包括存储的计算机程序;计算机程序被处理器执行时,控制计算机可读存储介质所在设备执行本发明提供的上述乳腺超声断层图像分割模型建立方法,和/或,本发明提供的上述乳腺超声断层图像分割方法。A computer-readable storage medium, including a stored computer program; when the computer program is executed by a processor, it controls the device where the computer-readable storage medium is located to execute the method for establishing a breast ultrasound tomographic image segmentation model provided by the present invention, and/or, the present invention The above-mentioned breast ultrasound tomographic image segmentation method provided by the invention.

以下结合不同模型的分割效果对本发明所能取得的有益效果做进一步的分析说明。The beneficial effects obtained by the present invention will be further analyzed and described below in combination with the segmentation effects of different models.

对于图3所示的乳腺超声断层图像分割模型,使用Pytorch作为深度学习框架,实验环境Nvidia GeForce RTX 2080Ti(11GB)GPU,采用上述实施例1提供的UNet网络结构(即图3所示的乳腺超声断层图像分割模型)作为分割模型,模型中的C0=2,C1=64,C2=128,C3=256,C4=512。输入的图像尺寸为512×512×3。采用随机梯度下降算法(weight decay=0.0001,momentum=0.9)作为优化器,初始学习率设置为0.01,每经过100个epoch学习率降低至原始值的0.1倍,共训练300个epoch,batch size设置为4。使用的数据集共327张,按照7:1:2的比例划分数据集为训练集、验证集和测试集,将训练集扩充至原始的8倍。损失函数使用交叉熵函数、Dice损失函数、对比损失函数及不确定损失函数,Dice系数和交并比(IoU)作为评价指标。For the breast ultrasound tomographic image segmentation model shown in Figure 3, use Pytorch as the deep learning framework, the experimental environment Nvidia GeForce RTX 2080Ti (11GB) GPU, adopt the UNet network structure provided by the above-mentioned embodiment 1 (ie the breast ultrasound shown in Figure 3 Tomographic image segmentation model) as the segmentation model, C0=2, C1=64, C2=128, C3=256, C4=512 in the model. The input image size is 512×512×3. The stochastic gradient descent algorithm (weight decay=0.0001, momentum=0.9) is used as the optimizer, the initial learning rate is set to 0.01, and the learning rate is reduced to 0.1 times the original value after every 100 epochs, and a total of 300 epochs are trained, and the batch size is set for 4. A total of 327 data sets were used. The data set was divided into training set, verification set and test set according to the ratio of 7:1:2, and the training set was expanded to 8 times of the original. The loss function uses cross-entropy function, Dice loss function, contrast loss function and uncertainty loss function, Dice coefficient and intersection-over-union ratio (IoU) as evaluation indicators.

以现有的UNet模型、Attention-UNet模型、UNet++模型以及ResUNet模型、TransUNet模型作为的对比模型,对乳腺超声断层图像使用不同模型进行分割,其结果如图6所示。图6中,(a)为多张乳腺超声断层图像的病灶区域标签,(b)为UNet模型的分割结果,(c)为Attention-UNet模型分割结果,(d)为UNet++模型分割结果,(e)为ResUNet模型分割结果,(f)为TransUNet模型分割结果(g)本实施例的分割结果。根据附图6所示结果可以看出,基于本发明所建立的乳腺超声断层图像分割模型分割出来的边界更接近真实的边界,错误分割的情况也相对更少。为便于描述,以下将上述实施例1所建立的模型简记为“MFE-DSCrossUNet”,各模型分割结果的Dice系数和IoU系数如表1所示。Using the existing UNet model, Attention-UNet model, UNet++ model, ResUNet model, and TransUNet model as comparison models, different models are used to segment breast ultrasound tomographic images, and the results are shown in Figure 6. In Figure 6, (a) is the lesion area label of multiple breast ultrasound tomographic images, (b) is the segmentation result of the UNet model, (c) is the segmentation result of the Attention-UNet model, (d) is the segmentation result of the UNet++ model, ( e) is the segmentation result of the ResUNet model, (f) is the segmentation result of the TransUNet model (g) the segmentation result of this embodiment. According to the results shown in Fig. 6, it can be seen that the boundary segmented by the breast ultrasound tomographic image segmentation model based on the present invention is closer to the real boundary, and the situation of wrong segmentation is relatively less. For the convenience of description, the model established in the above-mentioned embodiment 1 is abbreviated as "MFE-DSCrossUNet" below, and the Dice coefficient and IoU coefficient of each model segmentation result are shown in Table 1.

表1不同模型分割结果的Dice系数和IoU系数Table 1 Dice coefficient and IoU coefficient of different model segmentation results

Figure BDA0004090170100000171
Figure BDA0004090170100000171

根据表1所示评价指标可以看出,本发明分割的结果Dice系数达到了0.8883,IoU系数达到了0.7990。与UNet的分割结果相比,Dice系数提高了3.68%,IoU系数提高了5.76%。充分证明了本发明中所建立的乳腺超声断层图像分割模型的有效性和优越性。According to the evaluation indicators shown in Table 1, it can be seen that the Dice coefficient of the segmentation result of the present invention reaches 0.8883, and the IoU coefficient reaches 0.7990. Compared with the segmentation results of UNet, the Dice coefficient is improved by 3.68%, and the IoU coefficient is improved by 5.76%. It fully proves the validity and superiority of the breast ultrasound tomographic image segmentation model established in the present invention.

本领域的技术人员容易理解,以上所述仅为本发明的较佳实施例而已,并不用以限制本发明,凡在本发明的精神和原则之内所作的任何修改、等同替换和改进等,均应包含在本发明的保护范围之内。It is easy for those skilled in the art to understand that the above descriptions are only preferred embodiments of the present invention, and are not intended to limit the present invention. Any modifications, equivalent replacements and improvements made within the spirit and principles of the present invention, All should be included within the protection scope of the present invention.

Claims (10)

1.一种乳腺超声断层图像分割模型建立方法,其特征在于,包括:1. A breast ultrasound tomographic image segmentation model building method, is characterized in that, comprising: 构建训练集;所述训练集中,每一条训练数据为已标注出病灶区域的乳腺超声断层图像;Constructing a training set; in the training set, each piece of training data is a breast ultrasound tomographic image that has marked a lesion area; 基于UNet模型构建初始的乳腺超声断层图像分割模型,用于从乳腺超声断层图像中分割出病灶区域;Construct an initial breast ultrasound tomographic image segmentation model based on the UNet model, which is used to segment the lesion area from the breast ultrasound tomographic image; 以Ltotal=LE+αLcontrast+βLuncer为训练损失函数,利用所述训练集训练初始的乳腺超声断层图像分割模型,完成所述乳腺超声断层图像分割模型的建立;Taking L total = LE +αL contrast + βLuncer as the training loss function, using the training set to train the initial breast ultrasound tomographic image segmentation model, and completing the establishment of the breast ultrasound tomographic image segmentation model; 其中,Ltotal表示总体损失;LE表示乳腺超声断层图像分割模型的分割误差;Lcontrast表示乳腺超声断层图像分割模型中解码结构输出的中间结果与标注结果之间的对比损失函数,α表示其权重系数;Luncer为不确定性损失函数,用于表示乳腺超声断层图像分割模型中解码结构输出的中间结果与标注结果间的差异,β表示其权重系数;α≥0,β≥0,且α、β不同时为0。Among them, L total represents the overall loss; L E represents the segmentation error of the breast ultrasound tomographic image segmentation model; L contrast represents the contrast loss function between the intermediate results output by the decoding structure and the labeling results in the breast ultrasound tomographic image segmentation model, and α represents its Weight coefficient; Luncer is an uncertainty loss function, which is used to represent the difference between the intermediate result of the decoding structure output and the labeling result in the breast ultrasound tomographic image segmentation model, and β represents its weight coefficient; α≥0, β≥0, and α and β are not 0 at the same time. 2.如权利要求1所述的乳腺超声断层图像分割模型建立方法,其特征在于,2. the breast ultrasound tomographic image segmentation model establishment method as claimed in claim 1, is characterized in that,
Figure FDA0004090170090000011
Figure FDA0004090170090000011
Figure FDA0004090170090000012
Figure FDA0004090170090000012
Figure FDA0004090170090000013
Figure FDA0004090170090000013
其中,J表示模型的中间输出的数量;LCE()表示交叉熵损失;l表示标注结果;pj表示乳腺超声断层图像分割模型中解码结构输出的中间结果,
Figure FDA0004090170090000014
表示中间结果pj的第i个通道;DKL表示KL散度,C表示类别数,/>
Figure FDA0004090170090000015
表示分配给第j个中间结果中的像素p的权重。
Among them, J represents the number of intermediate outputs of the model; L CE () represents the cross-entropy loss; l represents the labeling result; pj represents the intermediate result of the decoding structure output in the breast ultrasound tomographic image segmentation model,
Figure FDA0004090170090000014
Represents the i-th channel of the intermediate result p j ; D KL represents the KL divergence, C represents the number of categories, />
Figure FDA0004090170090000015
denotes the weight assigned to pixel p in the jth intermediate result.
3.如权利要求1或2所述的乳腺超声断层图像分割模型建立方法,其特征在于,3. the breast ultrasound tomographic image segmentation model building method as claimed in claim 1 or 2, is characterized in that,
Figure FDA0004090170090000021
Figure FDA0004090170090000021
其中,L表示像素之间的相似度;so表示中间结果中的特征,sl表示标注结果中的特征;m和n表示特征类别,
Figure FDA0004090170090000022
表示同一类别的特征,/>
Figure FDA0004090170090000023
表示不同类别的特征;τ表示温度系数。
Among them, L represents the similarity between pixels; s o represents the feature in the intermediate result, s l represents the feature in the labeling result; m and n represent the feature category,
Figure FDA0004090170090000022
Represents features of the same class, />
Figure FDA0004090170090000023
Indicates the characteristics of different categories; τ indicates the temperature coefficient.
4.如权利要求1所述的乳腺超声断层图像分割模型建立方法,其特征在于,所述乳腺超声断层图像分割模型还包括:插入在UNet模型中编码结构的最后一层与解码结构之间的CrossFormer模块。4. the breast ultrasound tomographic image segmentation model establishment method as claimed in claim 1, is characterized in that, described breast ultrasound tomographic image segmentation model also comprises: insert between the last layer of encoding structure and decoding structure in UNet model CrossFormer module. 5.如权利要求1或4所述的乳腺超声断层图像分割模型建立方法,其特征在于,所述乳腺超声断层图像分割模型还包括:插入在UNet模型中编码结构与解码结构间长跳跃连接中的多尺度注意力模块;5. the breast ultrasound tomographic image segmentation model establishment method as claimed in claim 1 or 4, is characterized in that, described breast ultrasound tomographic image segmentation model also comprises: insert in the long skip connection between encoding structure and decoding structure in UNet model The multi-scale attention module of ; 所述多尺度注意力模块包括:空洞空间卷积池化金字塔和特征增强模块;The multi-scale attention module includes: a hollow space convolution pooling pyramid and a feature enhancement module; 所述空洞卷积池化金字塔用于对编码结构中对应层输出的特征进行多尺度特征提取,得到多尺度特征;The hollow convolution pooling pyramid is used to perform multi-scale feature extraction on the features output by the corresponding layer in the coding structure to obtain multi-scale features; 所述特征增强模块,以所述多尺度特征和解码结构中对应层输出的解码特征为输入,用于增强与任务相关的特征并抑制与特征不相关的特征。The feature enhancement module takes the multi-scale features and the decoded features output by the corresponding layer in the decoding structure as input, and is used to enhance the features related to the task and suppress the features not related to the feature. 6.如权利要求5所述的乳腺超声断层图像分割模型建立方法,其特征在于,所述特征增强模块包括第一卷积层、第二卷积层、第三卷积层、ReLU激活层、Sigmoid激活层、像素相加层以及像素相乘层;6. the breast ultrasound tomographic image segmentation model establishment method as claimed in claim 5, is characterized in that, described feature enhancement module comprises the first convolutional layer, the second convolutional layer, the 3rd convolutional layer, ReLU activation layer, Sigmoid activation layer, pixel addition layer and pixel multiplication layer; 所述第一卷积层用于输入所述多尺度特征,并进行卷积操作;The first convolutional layer is used to input the multi-scale features and perform a convolution operation; 所述第二卷积层用于输入所述解码特征,并进行卷积操作;The second convolutional layer is used to input the decoding feature and perform a convolution operation; 所述像素相加层用于对所述第一卷积层和所述第二卷积层输出的结果进行逐像素相加;The pixel addition layer is used to add the results output by the first convolution layer and the second convolution layer pixel by pixel; 所述ReLU激活层用于对所述像素相加层输出的结果进行激活处理;The ReLU activation layer is used to activate the result output by the pixel addition layer; 所述第三卷积层用于对所述ReLU激活层输出的结果进行卷积操作;The third convolutional layer is used to perform a convolution operation on the result output by the ReLU activation layer; 所述Sigmoid激活层用于对所述第三卷积层输出的结果进行激活处理;The Sigmoid activation layer is used to activate the result output by the third convolutional layer; 所述像素相乘层用于对所述Sigmoid激活层输出的结果以及所述多尺度特征进行逐像素相乘。The pixel multiplication layer is used to multiply the result output by the sigmoid activation layer and the multi-scale features pixel by pixel. 7.如权利要求1所述的乳腺超声断层图像分割模型建立方法,其特征在于,构建训练集,包括:7. the breast ultrasound tomographic image segmentation model establishment method as claimed in claim 1, is characterized in that, constructs training set, comprises: 获得由乳腺超声断层图像构成的原始数据集,并标注其中的每一张乳腺超声断层图像中的病灶区域,得到对应的病灶区域掩码图像;Obtaining an original data set composed of breast ultrasound tomographic images, and marking the lesion area in each of the breast ultrasound tomographic images, and obtaining a corresponding lesion area mask image; 将连续的n张乳腺超声断层图像在通道维度上拼接,由拼接后的乳腺超声断层图像和对应的病灶区域掩码图像构成所述训练集;Stitching n consecutive breast ultrasound tomographic images in the channel dimension, and forming the training set from the spliced breast ultrasound tomographic images and corresponding lesion area mask images; 其中,n为大于1的正整数。Wherein, n is a positive integer greater than 1. 8.如权利要求7所述的乳腺超声断层图像分割模型建立方法,其特征在于,n=3;8. the breast ultrasound tomographic image segmentation model establishment method as claimed in claim 7, is characterized in that, n=3; 并且,将连续的3张乳腺超声断层图像在通道维度上拼接,包括:In addition, three consecutive breast ultrasound tomographic images are spliced in the channel dimension, including: 从每张乳腺超声断层图像中提取一个通道的信息,得到三个通道的信息;Extract the information of one channel from each breast ultrasound tomographic image to obtain the information of three channels; 将三个通道的信息在通道维度上拼接。The information of the three channels is spliced in the channel dimension. 9.一种乳腺超声断层图像分割方法,其特征在于,包括:9. A breast ultrasound tomographic image segmentation method, characterized in that, comprising: 将待分割的乳腺超声断层图像输入至乳腺超声断层图像分割模型,得到病灶区域分割结果;Input the breast ultrasound tomographic image to be segmented into the breast ultrasound tomographic image segmentation model to obtain the lesion area segmentation result; 其中,所述乳腺超声断层图像分割模型由权利要求1~8任一项所述的乳腺超声断层图像分割模型建立方法所建立。Wherein, the breast ultrasound tomographic image segmentation model is established by the method for establishing a breast ultrasound tomographic image segmentation model according to any one of claims 1-8. 10.一种计算机可读存储介质,其特征在于,包括存储的计算机程序;所述计算机程序被处理器执行时,控制所述计算机可读存储介质所在设备执行权利要求1~8任一项所述的乳腺超声断层图像分割模型建立方法,和/或,权利要求9所述的乳腺超声断层图像分割方法。10. A computer-readable storage medium, characterized by comprising a stored computer program; when the computer program is executed by a processor, it controls the device where the computer-readable storage medium is located to execute any one of claims 1-8. The method for establishing a breast ultrasound tomographic image segmentation model described above, and/or the breast ultrasound tomographic image segmentation method described in claim 9.
CN202310149152.7A 2023-02-22 2023-02-22 Mammary gland ultrasonic tomography image segmentation model establishment method and segmentation method Pending CN116433586A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310149152.7A CN116433586A (en) 2023-02-22 2023-02-22 Mammary gland ultrasonic tomography image segmentation model establishment method and segmentation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202310149152.7A CN116433586A (en) 2023-02-22 2023-02-22 Mammary gland ultrasonic tomography image segmentation model establishment method and segmentation method

Publications (1)

Publication Number Publication Date
CN116433586A true CN116433586A (en) 2023-07-14

Family

ID=87080422

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310149152.7A Pending CN116433586A (en) 2023-02-22 2023-02-22 Mammary gland ultrasonic tomography image segmentation model establishment method and segmentation method

Country Status (1)

Country Link
CN (1) CN116433586A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117422927A (en) * 2023-11-09 2024-01-19 什维新智医疗科技(上海)有限公司 Mammary gland ultrasonic image classification method, system, electronic equipment and medium
CN118196102A (en) * 2024-05-17 2024-06-14 华侨大学 Method and device for detecting breast ultrasonic tumor lesion area based on double-network shadow removal
CN118261925A (en) * 2024-04-17 2024-06-28 徐州医科大学 A breast ultrasound image segmentation method with large receptive field attention enhancement
CN119027946A (en) * 2024-08-05 2024-11-26 华平祥晟(上海)医疗科技有限公司 A video auxiliary analysis system and method based on artificial intelligence
CN119379720A (en) * 2024-12-30 2025-01-28 中国科学院深圳先进技术研究院 A breast ultrasound image segmentation model construction method, segmentation method and device
CN119792086A (en) * 2025-03-12 2025-04-11 广东工业大学 AI-assisted intelligent processing equipment and method for Chinese herbal medicine

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117422927A (en) * 2023-11-09 2024-01-19 什维新智医疗科技(上海)有限公司 Mammary gland ultrasonic image classification method, system, electronic equipment and medium
CN118261925A (en) * 2024-04-17 2024-06-28 徐州医科大学 A breast ultrasound image segmentation method with large receptive field attention enhancement
CN118196102A (en) * 2024-05-17 2024-06-14 华侨大学 Method and device for detecting breast ultrasonic tumor lesion area based on double-network shadow removal
CN119027946A (en) * 2024-08-05 2024-11-26 华平祥晟(上海)医疗科技有限公司 A video auxiliary analysis system and method based on artificial intelligence
CN119379720A (en) * 2024-12-30 2025-01-28 中国科学院深圳先进技术研究院 A breast ultrasound image segmentation model construction method, segmentation method and device
CN119379720B (en) * 2024-12-30 2025-04-08 中国科学院深圳先进技术研究院 Mammary gland ultrasonic image segmentation model construction method, segmentation method and device
CN119792086A (en) * 2025-03-12 2025-04-11 广东工业大学 AI-assisted intelligent processing equipment and method for Chinese herbal medicine

Similar Documents

Publication Publication Date Title
CN116433586A (en) Mammary gland ultrasonic tomography image segmentation model establishment method and segmentation method
Oghli et al. Automatic fetal biometry prediction using a novel deep convolutional network architecture
CN109523521B (en) Pulmonary nodule classification and lesion location method and system based on multi-slice CT images
Shuvo et al. CNL-UNet: A novel lightweight deep learning architecture for multimodal biomedical image segmentation with false output suppression
Sugimori Classification of computed tomography images in different slice positions using deep learning
CN110599499B (en) MRI image heart structure segmentation method based on multipath convolutional neural network
Misra et al. Deep learning‐based multimodal fusion network for segmentation and classification of breast cancers using B‐mode and elastography ultrasound images
CN110310289A (en) Lung tissue image segmentation method based on deep learning
CN114663445B (en) A 3D cardiac image segmentation method based on multi-scale edge perception
CN116645380A (en) Automatic segmentation method of tumor area in CT images of esophageal cancer based on two-stage progressive information fusion
Du et al. Segmentation and visualization of left atrium through a unified deep learning framework
CN115409859A (en) Coronary artery blood vessel image segmentation method and device, storage medium and terminal
CN117635625A (en) Pancreatic tumor segmentation method based on automatic data enhancement strategy and multi-attention-assisted UNet
CN118898773A (en) Medical image segmentation method and device based on multi-scale feature fusion
CN118037791A (en) Construction method and application of multi-mode three-dimensional medical image segmentation registration model
Ozcan et al. Enhanced-TransUNet for ultrasound segmentation of thyroid nodules
Wen et al. Segmenting medical MRI via recurrent decoding cell
CN115471512A (en) Medical image segmentation method based on self-supervision contrast learning
Zhang et al. MCBTNet: Multi-Feature Fusion CNN and Bi-Level Routing Attention Transformer-based Medical Image Segmentation Network
CN118037615A (en) A tumor segmentation-guided magnetic resonance image translation method, system, device and medium
CN118072017A (en) Lung CT image lung nodule full-automatic segmentation method and system based on deep learning
CN117409202A (en) Methods, systems and equipment for multi-organ segmentation in medical images integrating frequency domain sensing
CN117198511A (en) Deep learning-based diagnosis method for children's posterior fossa tumor
CN114155215B (en) A method and system for nasopharyngeal cancer identification and tumor segmentation based on MR images
CN115526898A (en) Medical image segmentation method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination