WO2022007957A1 - Network architecture for automatically processing images, program carrier, and workstation - Google Patents

Network architecture for automatically processing images, program carrier, and workstation Download PDF

Info

Publication number
WO2022007957A1
WO2022007957A1 PCT/CN2021/105538 CN2021105538W WO2022007957A1 WO 2022007957 A1 WO2022007957 A1 WO 2022007957A1 CN 2021105538 W CN2021105538 W CN 2021105538W WO 2022007957 A1 WO2022007957 A1 WO 2022007957A1
Authority
WO
WIPO (PCT)
Prior art keywords
images
path
image
network architecture
encoding
Prior art date
Application number
PCT/CN2021/105538
Other languages
French (fr)
Chinese (zh)
Inventor
王少彬
陈颀
陈宇
Original Assignee
北京医智影科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京医智影科技有限公司 filed Critical 北京医智影科技有限公司
Publication of WO2022007957A1 publication Critical patent/WO2022007957A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20092Interactive image processing based on input by user
    • G06T2207/20104Interactive definition of region of interest [ROI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion

Definitions

  • the invention relates to a network architecture for automatic processing of images, a corresponding computer-readable program carrier and a corresponding workstation.
  • image processing such as image segmentation to identify specific objects in the image or to automatically outline specific objects in the image.
  • IMRT Intensity-modulated radiation therapy
  • CTV Clinical target volume
  • OAR organ-at-risk
  • the object of the present invention is to provide an improved network architecture for the automatic processing of images, a corresponding computer-readable program carrier and a corresponding workstation.
  • a network architecture for automatic processing of images comprising: an input module for inputting an image to be processed; an encoding path configured to use a dual The path network performs feature extraction on the input image; a decoding path is configured to establish a connection with the encoding path; a central module is configured for transition from the encoding path to the decoding path to refine high-dimensional image features; and an output a module configured to output an image processing result from a decoding path; wherein the decoding path is configured to perform a decoding operation on the corresponding encoding result of the encoding path and the image feature extraction result of the central module using the decoding path of the Unet architecture.
  • the dual-path network comprises a series of concatenated micro-blocks, the micro-blocks being correspondingly embedded in the decoders of the decoding paths.
  • the encoding path includes 5 encoders
  • the decoding path includes 4 decoders
  • each decoder uses a corresponding microblock in the encoding path.
  • the central module is configured to implement: Conv(3*3)+BN+ReLu; and the decoder is configured to implement: Microblock+ Among them, Conv(3*3) is a 3*3 convolution operation, BN is a batch normalization operation, ReLU is a linear rectification function, Represents bilinear upsampling.
  • the input module receives an image or images
  • the output module outputs the image or images
  • the output module is configured to implement: +Conv(3*3)+BN+ReLU+ in, represents the sigmoid activation function.
  • the image is a computed tomography image
  • the multiple images are multiple adjacent computed tomography images
  • the network architecture is configured to analyze the images Perform automatic segmentation processing.
  • the input module receives an image or images
  • the output module outputs a classification result
  • the output module is configured to implement: +Conv(3*3)+BN+ReLU+ in, Represents a fully connected layer.
  • a loss function is configured at the end point of the network architecture, and the loss function is a cross-entropy loss function or a weight adaptive loss function:
  • FIG y is label, which value 0 or 1; I is the input image; N is the total number of segmented object; p and y have the shape of I * N; P n is figured out every The proportion of the size of the target area of each category to the size of the overall image volume.
  • the formula is used as the loss function when the network architecture is configured to output multiple segmented images.
  • the network architecture is configured for automatic segmentation or classification of images for radiotherapy of tumors, especially images for radiotherapy of cervical cancer.
  • a method for processing an image using the network architecture comprising the steps of: inputting the image through an input module; processing the image through an encoding path, a central module, and a decoding path; and The image processing results are output through the output module.
  • a computer readable program carrier storing program instructions for implementing the method when executed by a processor.
  • a workstation configured to include the computer-readable program carrier.
  • the workstation is configured as a doctor's workstation for automatic segmentation or classification of medical images.
  • Figure 1 shows a simplified block diagram of a network architecture for automatic segmentation or classification of images according to an exemplary embodiment of the present invention.
  • Figure 2 schematically shows a simplified block diagram of a network architecture according to an exemplary embodiment of the present invention.
  • FIG. 3 schematically shows a simplified block diagram of a network architecture for automatic image segmentation according to another exemplary embodiment of the present invention.
  • FIG. 4 schematically shows a simplified block diagram of a network architecture for automatic segmentation of images according to yet another exemplary embodiment of the present invention.
  • FIG. 5 schematically shows a simplified block diagram for identifying which types of regions of interest are in an image, according to an exemplary embodiment of the present invention.
  • Figure 1 shows a simplified block diagram of a network architecture for automatic segmentation or classification of images according to an exemplary embodiment of the present invention.
  • the network architecture 1 may mainly include an input module 11 , an encoding path 12 , a central module 13 , a decoding path 14 and an output module 15 , wherein the illustrated input module 11 is used to receive images, such as CT images, and the encoding path 12 is configured to use a dual-path network (DPN) for feature extraction on the input image, the decoding path 14 is associated with the encoding path 12 to decode the encoding result of the encoding path 12 accordingly, and the central module 13 is used to extract from the encoding path 12.
  • the transition from encoding path 12 to decoding path 14 is used to refine high-dimensional image features, and output module 15 is used to output output results that can be used for automatic contouring.
  • the decoding path 14 is configured to decode using the decoding path of the Unet architecture.
  • the dual-path network also includes the residual network (ResNet) path and the dense convolutional network (DenseNet) path.
  • ResNet supports the reuse of features
  • DenseNet supports the exploration of new features, so the dual-path network combines the advantages of both.
  • the dual path network consists of a series of micro-blocks connected in series. Microblocks are at the heart of DPN.
  • microblocks are embedded in the decoder part to replace the standard convolution operation.
  • Figure 2 schematically shows a simplified block diagram of a network architecture 1 according to an exemplary embodiment of the present invention.
  • the single CT image 16 input to the encoding path 12 usually has 512*512 pixels.
  • 1*512*512 represents a CT image with 512*512 pixels.
  • the CT image 16 undergoes a series of encoder processes in the encoding path 12 to extract features of the image.
  • the final encoded result of the encoding path 12 is input to the central module 13 , and then processed by the central module 13 and then output to the decoding path 14 , so as to realize the transition from the encoding path 12 to the decoding path 14 .
  • the decoding path 14 is also associated with the encoding path 12 as schematically shown by arrow 17 .
  • the encoding path 12 encodes a series of images 16 through 5 encoders/microblocks 18, while the decoding path 14 includes 4 decoders 19, each decoder 19 Each uses a corresponding microblock/encoder 18 in the encoding path 12 .
  • exemplary operations/operations within some modules are also schematically represented by symbols.
  • the darker arrow represents Conv(3*3)+BN+ReLu, where Conv(3*3) is a 3*3 convolution operation, BN (Batch Normalization) is a batch normalization operation, ReLU (Rectified Linear Unit) is a linear rectification function.
  • Lighter colored arrows indicate microblocks. Represents a connection relationship. represents bilinear upsampling, represents the sigmoid activation function.
  • the central module 13 may be configured as: Conv(3*3)+BN+ReLu.
  • the output module 15 After being processed by the decoding path 14, it is output to the output module 15, and the output module 15 also outputs an automatically segmented image 20 of 512*512.
  • the output module 15 may be configured to: +Conv(3*3)+BN+ReLU+ Among them, the Sigmoid activation function It is used to output the probability value that each pixel belongs to the region of interest (ROI, region of interest).
  • the decoder 19 may be configured as: microblock+
  • the present exemplary embodiment can be used for collaborative tumor radiation therapy, especially radiation therapy with complex tumor location organs, such as cervical cancer radiation therapy planning, and more specifically, for automatic contouring of clinical target volumes and organs at risk , to generate intermediate results usable by oncology radiation therapists.
  • FIG. 3 schematically shows a simplified block diagram of a network architecture for automatic image segmentation according to another exemplary embodiment of the present invention.
  • the main difference between FIG. 3 and FIG. 2 is that the input module 11 inputs three adjacent CT images 16 before and after, so the neural network obtains the image information before and after, which is a quasi-or-like 3D image. In this way, the segmentation results of the front and rear layers are smoother, avoiding the sudden change of the segmentation results.
  • the input CT images are not necessarily three images, but can be any suitable number of multiple images.
  • the output is also an automatically segmented image 20 .
  • Others are similar to FIG. 2, and are not repeated here for the sake of clarity.
  • FIG. 4 schematically shows a simplified block diagram of a network architecture for automatic segmentation of images according to yet another exemplary embodiment of the present invention.
  • the output is a multi-channel output.
  • the multi-channel output indicates that the network can not only do 2-class image segmentation, but also has the function of multi-class classification.
  • 8 kinds of ROI segmentation results can be output at the same time.
  • the 8 ROIs include 7 organs at risk and 1 clinical target area, which are: 1. Bladder, 2. Bone marrow, 3. Left femoral head, 4. Right femoral head, 5. Rectum, 6. Small intestine, 7. Spinal cord 8. Clinical target area of cervical cancer.
  • multi-channel output for other tumors is also supported, and of course other numbers of channels.
  • Multi-channel output means outputting multiple segmented images, each segmented image corresponds to a region of interest.
  • FIG. 3 Others are similar to FIG. 3 , and are not repeated here for the sake of clarity.
  • Figure 5 schematically shows a simplified block diagram for identifying which types of regions of interest are in an image, according to an exemplary embodiment of the present invention.
  • the main difference between FIG. 5 and FIG. 4 is that the last output part of the output module 15 is replaced with a fully connected layer (Full Connected Layer, FC), so that the segmentation network becomes a classification network.
  • FC Full Connected Layer
  • the network can judge which types of ROIs are included in the input CT image, but does not judge the positions of various ROIs, that is, does not do specific segmentation.
  • Sigmoid activation function shown in Figure 2-4 Replaced with a fully connected layer Other identical parts will not be repeated.
  • a loss function is usually provided at the end of the network, with the predicted value of the network and the prepared gold standard as input, the loss (bias or error) is calculated through the loss function, and then reversed
  • the corresponding, eg, all layers in the network are propagated to update the corresponding weight values, eg, all weight values, in the network, so that the network prediction value is getting closer and closer to the gold standard. For example, when the loss drops to an acceptable level, such as a predetermined level, the optimal weight value is also updated, and the network is finally determined, and the training can be ended at this time.
  • the loss function not only affects the training process of the network, but also directly affects the final performance of the network.
  • the loss function is a cross-entropy loss function (1) or a weight adaptive loss function (2):
  • p is the output of the sigmoid activation function
  • y is the ground truth map, which takes the value 0 or 1
  • I is the input image
  • N is the total number of segmentation targets. Both p and y have the shape I*N
  • P n is the ratio of the size of the target area of each category, such as the size of the organ at risk, to the size of the overall image volume.
  • formula (2) is used as the loss function.
  • the multi-channel segmentation network needs to segment multiple regions of interest, such as multiple organs at risk and clinical target areas. If the volume sizes of multiple regions of interest are very different, it is difficult to obtain a better segmentation effect using the cross-entropy loss function.
  • the weight adaptive loss function can achieve very good segmentation results.
  • Equation (1) can be used as the loss function.
  • CT images mainly takes CT images as an example, but obviously the images can also be X-ray imaging images, magnetic resonance imaging (MRI) images, single photon emission computed tomography (SPECT) images, positron emission type images. Computed tomography (PET) images, superimaging images, etc.
  • MRI magnetic resonance imaging
  • SPECT single photon emission computed tomography
  • PET positron emission type images
  • Computed tomography (PET) images superimaging images, etc.
  • the present invention provides a neural network architecture for automatic image processing, such as automatic segmentation or classification, which can be used in the medical field, especially for assisting in the analysis of CT images. Segmentation or classification.
  • the neural network architecture of the present invention is suitable for assisting the operations involved in the automatic delineation of clinical target volumes and organs at risk for radiotherapy of tumors (especially cervical cancer), and then physicians can automatically delineate the results based on the processing results, such as Follow-up operations, such as audits, can be modified if necessary for subsequent treatment purposes.
  • the network architecture of the present invention is not directly used for treatment purposes when automatically processing CT images, but generates intermediate results, and the main purpose is to improve the working efficiency of doctors.
  • the present invention also provides a computer-readable program carrier, which stores program instructions, and when the program instructions are executed by the processor, is used to implement the functions of the above-mentioned network architecture, to Perform image processing.
  • the computer-readable program carrier may be constructed as a flash memory or the like.
  • the invention further provides a workstation configured to include the computer-readable program carrier.
  • the workstation can be configured as a doctor's workstation for use by a doctor.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biophysics (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Image Processing (AREA)

Abstract

Disclosed is a network architecture (1) for automatically processing images, the architecture comprising: an input module (11) for inputting an image to be processed; an encoding path (12), which is configured to use a dual-path network to perform feature extraction on an inputted image; a decoding path (14), which is configured to establish a connection with the encoding path (12); a central module (13), which is configured as transition from the encoding path (12) to the decoding path (14) so as to refine high-dimensional image features; and an output module (15), which is configured to output an image processing result from the decoding path (14). The decoding path (14) is configured to use a decoding path of a Unet architecture to decode a corresponding encoding result of the encoding path (12) as well as the image feature extraction result of the central module (13). Also disclosed are a corresponding method, a corresponding program carrier, and a corresponding workstation. According to the present invention, better image segmentation and classification can be performed.

Description

用于对图像进行自动处理的网络架构、程序载体和工作站Network architecture, program carrier and workstation for automatic processing of images 技术领域technical field
本发明涉及一种用于对图像进行自动处理的网络架构、一种相应的计算机可读程序载体以及一种相应的工作站。The invention relates to a network architecture for automatic processing of images, a corresponding computer-readable program carrier and a corresponding workstation.
背景技术Background technique
有许多场合需要对图像进行处理,例如对图像进行分割,以识别出图像中的特定目标或对图像中的特定目标进行自动轮廓勾画。There are many occasions where image processing is required, such as image segmentation to identify specific objects in the image or to automatically outline specific objects in the image.
随着现代医学的发展,越来越多的疾病需要借助医学图像进行诊断和治疗,例如肿瘤放射治疗。With the development of modern medicine, more and more diseases need to be diagnosed and treated with the help of medical images, such as tumor radiation therapy.
宫颈癌目前已经是全球15至44岁女性中第二常见的癌症。调强放射疗法(IMRT,Intensity-modulated radiation therapy)已成为宫颈癌治疗的首选放射疗法。IMRT的有效性取决于临床靶区(CTV,Clinical target volume)和危及器官(OAR,organ-at-risk)轮廓勾画的准确性。临床靶区和危及器官的轮廓目前是由放射治疗肿瘤医生通过费力、繁琐的手动勾画完成。轮廓勾画非常耗时,非常依赖放射治疗肿瘤医生的经验。尽管有标准指南可用,但放射治疗肿瘤医生的经验差异仍然是制定放射治疗计划的主要挑战之一。因此,如果可以在合理的时间内自动分割解剖结构,则可以大大减轻放射治疗肿瘤医生的手动工作量。Cervical cancer is now the second most common cancer in women aged 15 to 44 worldwide. Intensity-modulated radiation therapy (IMRT) has become the radiation therapy of choice for cervical cancer treatment. The effectiveness of IMRT depends on the accuracy of clinical target volume (CTV, Clinical target volume) and organ-at-risk (OAR, organ-at-risk) delineation. Outlines of clinical target volumes and organs at risk are currently performed by radiation therapy oncologists through laborious and tedious manual delineation. Contouring is time-consuming and relies heavily on the experience of the radiation oncologist. Despite the availability of standard guidelines, differences in the experience of radiation oncologists remain one of the main challenges in planning radiation therapy. Therefore, if the anatomy can be automatically segmented in a reasonable time, the manual workload of radiation therapy oncologists can be greatly reduced.
传统的自动分割方法源自统计模型或基于图集的模型。然而,这两种方法均具有诸多限制。特别是,宫颈癌所处位置的器官较为复杂,CT图像的边界也不太清晰。因此,当前的自动分割方法在CTV的轮廓勾画方面表现较差,各种复杂的组织和器官可能与CTV的边界混淆,而且在CT图像中可能无法检测到CTV中的肿瘤组织或亚临床疾病的潜在扩散。Traditional automatic segmentation methods are derived from statistical models or atlas-based models. However, both methods have limitations. In particular, the organs where cervical cancer is located are complex, and the boundaries of CT images are not very clear. Therefore, current automatic segmentation methods perform poorly in CTV contouring, various complex tissues and organs may be confused with CTV boundaries, and tumor tissue or subclinical disease in CTV may not be detected in CT images. potential spread.
因此,尤其是对于宫颈癌之类目前仍难以良好地进行自动轮廓勾画的癌症,迫切需要进行改进。Therefore, there is an urgent need for improvement, especially for cancers such as cervical cancer, for which automatic contouring is still difficult to perform well.
发明内容SUMMARY OF THE INVENTION
本发明的目的是通过一种改进的用于对图像进行自动处理的网络架构、一种相应的计算机可读程序载体以及一种相应的工作站。The object of the present invention is to provide an improved network architecture for the automatic processing of images, a corresponding computer-readable program carrier and a corresponding workstation.
根据本发明的第一个方面,提供了一种用于对图像进行自动处理的网络架构,所述网络架构包括:用于输入待处理的图像的输入模块;编码路径,其被配置成使用双路径网络对输入的图像进行特征提取;解码路径,其被配置成与编码路径建立连接;中央模块,其被配置成用于从编码路径到解码路径的过渡,以提炼高维图像特征;以及输出模块,其被配置成从解码路径输出图像处理结果;其中,所述解码路径被配置成使用Unet架构的解码路径对编码路径的相应编码结果以及中央模块的图像特征提炼结果进行解码操作。According to a first aspect of the present invention, there is provided a network architecture for automatic processing of images, the network architecture comprising: an input module for inputting an image to be processed; an encoding path configured to use a dual The path network performs feature extraction on the input image; a decoding path is configured to establish a connection with the encoding path; a central module is configured for transition from the encoding path to the decoding path to refine high-dimensional image features; and an output a module configured to output an image processing result from a decoding path; wherein the decoding path is configured to perform a decoding operation on the corresponding encoding result of the encoding path and the image feature extraction result of the central module using the decoding path of the Unet architecture.
根据本发明的一个可选实施例,所述双路径网络包括一系列串联的微块,所述微块被相应地嵌入解码路径的解码器中。According to an optional embodiment of the invention, the dual-path network comprises a series of concatenated micro-blocks, the micro-blocks being correspondingly embedded in the decoders of the decoding paths.
根据本发明的一个可选实施例,所述编码路径包括5个编码器,所述解码路径包括4个解码器,每个解码器分别使用所述编码路径中的相应的微块。According to an optional embodiment of the present invention, the encoding path includes 5 encoders, the decoding path includes 4 decoders, and each decoder uses a corresponding microblock in the encoding path.
根据本发明的一个可选实施例,所述中央模块被配置成实施:Conv(3*3)+BN+ReLu;以及所述解码器被配置成实施:微块+
Figure PCTCN2021105538-appb-000001
其中,Conv(3*3)为3*3卷积运算,BN为批量标准化运算,ReLU为线性整流函数,
Figure PCTCN2021105538-appb-000002
表示双线性上采样。
According to an optional embodiment of the present invention, the central module is configured to implement: Conv(3*3)+BN+ReLu; and the decoder is configured to implement: Microblock+
Figure PCTCN2021105538-appb-000001
Among them, Conv(3*3) is a 3*3 convolution operation, BN is a batch normalization operation, ReLU is a linear rectification function,
Figure PCTCN2021105538-appb-000002
Represents bilinear upsampling.
根据本发明的一个可选实施例,所述输入模块接收一张图像或多张图像,所述输出模块输出一张图像或多张图像;以及所述输出模块被配置成实施:
Figure PCTCN2021105538-appb-000003
+Conv(3*3)+BN+ReLU+
Figure PCTCN2021105538-appb-000004
其中,
Figure PCTCN2021105538-appb-000005
表示Sigmoid激活函数。
According to an optional embodiment of the present invention, the input module receives an image or images, the output module outputs the image or images; and the output module is configured to implement:
Figure PCTCN2021105538-appb-000003
+Conv(3*3)+BN+ReLU+
Figure PCTCN2021105538-appb-000004
in,
Figure PCTCN2021105538-appb-000005
represents the sigmoid activation function.
根据本发明的一个可选实施例,所述图像为计算机断层扫描图像,所述多张图像为前后相邻的多张计算机断层扫描图像;和/或所述网络架构被配置成用于对图像进行自动分割处理。According to an optional embodiment of the present invention, the image is a computed tomography image, and the multiple images are multiple adjacent computed tomography images; and/or the network architecture is configured to analyze the images Perform automatic segmentation processing.
根据本发明的一个可选实施例,所述输入模块接收一张图像或多张图像,所述输出模块输出分类结果;以及所述输出模块被配置成实施:
Figure PCTCN2021105538-appb-000006
+Conv(3*3)+BN+ReLU+
Figure PCTCN2021105538-appb-000007
其中,
Figure PCTCN2021105538-appb-000008
表示全连接层。
According to an optional embodiment of the present invention, the input module receives an image or images, the output module outputs a classification result; and the output module is configured to implement:
Figure PCTCN2021105538-appb-000006
+Conv(3*3)+BN+ReLU+
Figure PCTCN2021105538-appb-000007
in,
Figure PCTCN2021105538-appb-000008
Represents a fully connected layer.
根据本发明的一个可选实施例,在所述网络架构的终点配置损失函数, 所述损失函数为交叉熵损失函数或权重自适应损失函数:According to an optional embodiment of the present invention, a loss function is configured at the end point of the network architecture, and the loss function is a cross-entropy loss function or a weight adaptive loss function:
Figure PCTCN2021105538-appb-000009
Figure PCTCN2021105538-appb-000009
Figure PCTCN2021105538-appb-000010
Figure PCTCN2021105538-appb-000010
其中,in,
Figure PCTCN2021105538-appb-000011
Figure PCTCN2021105538-appb-000011
Figure PCTCN2021105538-appb-000012
Figure PCTCN2021105538-appb-000012
p是Sigmoid激活函数输出;y是标签图,其取值0或1;I为输入图像;N是总的分割目标的数量;p和y均具有形状I*N;P n是统计出来的每个类别的目标区域的体积大小占整体图像体积的大小的比例。 p is a Sigmoid activation function output; FIG y is label, which value 0 or 1; I is the input image; N is the total number of segmented object; p and y have the shape of I * N; P n is figured out every The proportion of the size of the target area of each category to the size of the overall image volume.
根据本发明的一个可选实施例,当网络架构被配置成输出多张分割图像时,使用公式作为损失函数。According to an optional embodiment of the present invention, the formula is used as the loss function when the network architecture is configured to output multiple segmented images.
根据本发明的一个可选实施例,所述网络架构被配置成用于对肿瘤放射治疗用图像、尤其是宫颈癌放射治疗用图像进行自动分割或分类。According to an optional embodiment of the present invention, the network architecture is configured for automatic segmentation or classification of images for radiotherapy of tumors, especially images for radiotherapy of cervical cancer.
根据本发明的第二个方面,提供了一种使用所述网络架构对图像进行处理的方法,包括以下步骤:通过输入模块输入图像;通过编码路径、中央模块和解码路径对图像进行处理;以及通过输出模块输出图像处理结果。According to a second aspect of the present invention, there is provided a method for processing an image using the network architecture, comprising the steps of: inputting the image through an input module; processing the image through an encoding path, a central module, and a decoding path; and The image processing results are output through the output module.
根据本发明的第三个方面,提供了一种计算机可读程序载体,其存储有程序指令,当所述程序指令被处理器执行时用于实施所述方法。According to a third aspect of the present invention, there is provided a computer readable program carrier storing program instructions for implementing the method when executed by a processor.
根据本发明的第四个方面,提供了一种工作站,所述工作站被配置成包括所述计算机可读程序载体。According to a fourth aspect of the present invention, there is provided a workstation configured to include the computer-readable program carrier.
根据本发明的一个可选实施例,所述工作站被配置成医生工作站,用于对医疗图像进行自动分割或分类。According to an optional embodiment of the present invention, the workstation is configured as a doctor's workstation for automatic segmentation or classification of medical images.
根据本发明,不仅可以提高图像处理的性能,而且还能使网络更快地建立。According to the present invention, not only the performance of image processing can be improved, but also the network can be established faster.
附图说明Description of drawings
下面,通过参看附图更详细地描述本发明,可以更好地理解本发明的原理、特点和优点。附图包括:The principles, features and advantages of the present invention may be better understood by describing the present invention in more detail below with reference to the accompanying drawings. The accompanying drawings include:
图1示出了根据本发明的一个示例性实施例的用于对图像进行自动分割或分类的网络架构的简化框图。Figure 1 shows a simplified block diagram of a network architecture for automatic segmentation or classification of images according to an exemplary embodiment of the present invention.
图2示意性地示出了根据本发明的一个示例性实施例的网络架构的简化框图。Figure 2 schematically shows a simplified block diagram of a network architecture according to an exemplary embodiment of the present invention.
图3示意性地示出了根据本发明的另一个示例性实施例的用于对图像进行自动分割的网络架构的简化框图。FIG. 3 schematically shows a simplified block diagram of a network architecture for automatic image segmentation according to another exemplary embodiment of the present invention.
图4示意性地示出了根据本发明的又一个示例性实施例的用于对图像进行自动分割的网络架构的简化框图。FIG. 4 schematically shows a simplified block diagram of a network architecture for automatic segmentation of images according to yet another exemplary embodiment of the present invention.
图5示意性地示出了根据本发明的一个示例性实施例的用于识别图像中有哪几种感兴趣区域的简化框图。FIG. 5 schematically shows a simplified block diagram for identifying which types of regions of interest are in an image, according to an exemplary embodiment of the present invention.
具体实施方式detailed description
为了使本发明所要解决的技术问题、技术方案以及有益的技术效果更加清楚明白,以下将结合附图以及多个示例性实施例对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅用于解释本发明,而不是用于限制本发明的保护范围。In order to make the technical problems, technical solutions and beneficial technical effects to be solved by the present invention clearer, the present invention will be further described in detail below with reference to the accompanying drawings and multiple exemplary embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention, rather than to limit the protection scope of the present invention.
图1示出了根据本发明的一个示例性实施例的用于对图像进行自动分割或分类的网络架构的简化框图。Figure 1 shows a simplified block diagram of a network architecture for automatic segmentation or classification of images according to an exemplary embodiment of the present invention.
如图1所示,网络架构1主要可包括输入模块11、编码路径12、中央模块13、解码路径14和输出模块15,其中,所示输入模块11用于接收图像、例如CT图像,编码路径12被配置成使用双路径网络(DPN),以对输入图像进行特征提取,所述解码路径14与编码路径12关联,以相应地对编码路径12的编码结果进行解码,中央模块13用于从编码路径12到解码路径14的过渡,以用于提炼高维图像特征,输出模块15用于输出可用于自 动轮廓勾画的输出结果,所述解码路径14被配置使用Unet架构的解码路径进行解码。As shown in FIG. 1 , the network architecture 1 may mainly include an input module 11 , an encoding path 12 , a central module 13 , a decoding path 14 and an output module 15 , wherein the illustrated input module 11 is used to receive images, such as CT images, and the encoding path 12 is configured to use a dual-path network (DPN) for feature extraction on the input image, the decoding path 14 is associated with the encoding path 12 to decode the encoding result of the encoding path 12 accordingly, and the central module 13 is used to extract from the encoding path 12. The transition from encoding path 12 to decoding path 14 is used to refine high-dimensional image features, and output module 15 is used to output output results that can be used for automatic contouring. The decoding path 14 is configured to decode using the decoding path of the Unet architecture.
双路径网络又包括残差网络(ResNet)路径和密集卷积网络(DenseNet)路径,ResNet支持特征的重复使用,而DenseNet支持新特征的探索,因此双路径网络结合了这两者的优点。The dual-path network also includes the residual network (ResNet) path and the dense convolutional network (DenseNet) path. ResNet supports the reuse of features, while DenseNet supports the exploration of new features, so the dual-path network combines the advantages of both.
双路径网络包括一系列串联的微块(Micro-block)。微块是DPN的核心。通过使用双路径网络作为本发明的网络架构的编码器部分,可实现更好的特征提取能力。The dual path network consists of a series of micro-blocks connected in series. Microblocks are at the heart of DPN. By using a dual-path network as the encoder part of the network architecture of the present invention, better feature extraction capabilities can be achieved.
根据本发明的一个示例性实施例,为了使解码器部分具有恢复抽象特征的相同性能,将微块嵌入到解码器部分中以替换标准卷积操作。According to an exemplary embodiment of the present invention, in order to give the decoder part the same performance of recovering abstract features, microblocks are embedded in the decoder part to replace the standard convolution operation.
图2示意性地示出了根据本发明的一个示例性实施例的网络架构1的简化框图。Figure 2 schematically shows a simplified block diagram of a network architecture 1 according to an exemplary embodiment of the present invention.
如图2所示,向编码路径12输入的单张CT图像16,通常像素为512*512,在此以1*512*512表示1张像素为512*512的CT图像。CT图像16在编码路径12中经过一系列编码器处理而提取图像的特征。编码路径12最后编码后的结果输入到中央模块13,然后经过中央模块13的处理后输出给解码路径14,以实现从编码路径12到解码路径14的过渡。还可以从图2中看出,解码路径14还如箭头17示意性所示地与编码路径12关联。As shown in FIG. 2 , the single CT image 16 input to the encoding path 12 usually has 512*512 pixels. Here, 1*512*512 represents a CT image with 512*512 pixels. The CT image 16 undergoes a series of encoder processes in the encoding path 12 to extract features of the image. The final encoded result of the encoding path 12 is input to the central module 13 , and then processed by the central module 13 and then output to the decoding path 14 , so as to realize the transition from the encoding path 12 to the decoding path 14 . It can also be seen from FIG. 2 that the decoding path 14 is also associated with the encoding path 12 as schematically shown by arrow 17 .
在图2所示的示例性实施例中,编码路径12通过5个编码器/微块18对图像16进行一系列编码,而在解码路径14中包括4个解码器19,每个解码器19均使用了编码路径12中的一个相应的微块/编码器18。In the exemplary embodiment shown in Figure 2, the encoding path 12 encodes a series of images 16 through 5 encoders/microblocks 18, while the decoding path 14 includes 4 decoders 19, each decoder 19 Each uses a corresponding microblock/encoder 18 in the encoding path 12 .
在图2中,还用符号示意性地表示一些模块内的示例性运算/操作。具体地讲,颜色较深的箭头表示Conv(3*3)+BN+ReLu,其中,Conv(3*3)为3*3卷积运算,BN(批量标准化,Batch Normalization)为批量标准化运算,ReLU(Rectified Linear Unit)为线性整流函数。颜色较浅的箭头表示微块。
Figure PCTCN2021105538-appb-000013
表示连接关系。
Figure PCTCN2021105538-appb-000014
表示双线性上采样,
Figure PCTCN2021105538-appb-000015
表示Sigmoid激活函数。
In Figure 2, exemplary operations/operations within some modules are also schematically represented by symbols. Specifically, the darker arrow represents Conv(3*3)+BN+ReLu, where Conv(3*3) is a 3*3 convolution operation, BN (Batch Normalization) is a batch normalization operation, ReLU (Rectified Linear Unit) is a linear rectification function. Lighter colored arrows indicate microblocks.
Figure PCTCN2021105538-appb-000013
Represents a connection relationship.
Figure PCTCN2021105538-appb-000014
represents bilinear upsampling,
Figure PCTCN2021105538-appb-000015
represents the sigmoid activation function.
根据本发明的一个示例性实施例,如图2所示,中央模块13可被配置成:Conv(3*3)+BN+ReLu。According to an exemplary embodiment of the present invention, as shown in FIG. 2 , the central module 13 may be configured as: Conv(3*3)+BN+ReLu.
在图2所示的示例性实施例中,经过解码路径14处理后输出到输出模块15,输出模块15也输出一张自动分割后的512*512的图像20。In the exemplary embodiment shown in FIG. 2 , after being processed by the decoding path 14, it is output to the output module 15, and the output module 15 also outputs an automatically segmented image 20 of 512*512.
根据本发明的一个示例性实施例,如图2所示,输出模块15可被配置成:
Figure PCTCN2021105538-appb-000016
+Conv(3*3)+BN+ReLU+
Figure PCTCN2021105538-appb-000017
其中,Sigmoid激活函数
Figure PCTCN2021105538-appb-000018
用于输出每个像素属于感兴趣区域(ROI,region of interest)的概率值。
According to an exemplary embodiment of the present invention, as shown in FIG. 2 , the output module 15 may be configured to:
Figure PCTCN2021105538-appb-000016
+Conv(3*3)+BN+ReLU+
Figure PCTCN2021105538-appb-000017
Among them, the Sigmoid activation function
Figure PCTCN2021105538-appb-000018
It is used to output the probability value that each pixel belongs to the region of interest (ROI, region of interest).
根据本发明的一个示例性实施例,如图2所示,解码器19可被配置成:微块+
Figure PCTCN2021105538-appb-000019
According to an exemplary embodiment of the present invention, as shown in FIG. 2 , the decoder 19 may be configured as: microblock+
Figure PCTCN2021105538-appb-000019
本示例性实施例可用于协作肿瘤放射治疗、尤其是肿瘤位置器官较为复杂的放射治疗、例如宫颈癌的放射治疗的计划制定,更特别地是用于对临床靶区和危及器官进行自动轮廓勾画,以生成肿瘤放射治疗医生可用的中间结果。The present exemplary embodiment can be used for collaborative tumor radiation therapy, especially radiation therapy with complex tumor location organs, such as cervical cancer radiation therapy planning, and more specifically, for automatic contouring of clinical target volumes and organs at risk , to generate intermediate results usable by oncology radiation therapists.
图3示意性地示出了根据本发明的另一个示例性实施例的用于对图像进行自动分割的网络架构的简化框图。FIG. 3 schematically shows a simplified block diagram of a network architecture for automatic image segmentation according to another exemplary embodiment of the present invention.
图3与图2的主要不同之处在于,通过输入模块11输入的是前后相邻的三张CT图像16,因此神经网络获得了前后的图像信息,是一种准或类3D图像。这样,前后层的分割结果更加流畅,避免了分割结果的突变。The main difference between FIG. 3 and FIG. 2 is that the input module 11 inputs three adjacent CT images 16 before and after, so the neural network obtains the image information before and after, which is a quasi-or-like 3D image. In this way, the segmentation results of the front and rear layers are smoother, avoiding the sudden change of the segmentation results.
当然,本领域的技术人员可以理解,输入的CT图像也不一定是三张图像,而是可以为任何合适数量的多张图像。Of course, those skilled in the art can understand that the input CT images are not necessarily three images, but can be any suitable number of multiple images.
输出的也是一张自动分割后的图像20。其它类似于图2,在此为了清楚起见,不再赘述。The output is also an automatically segmented image 20 . Others are similar to FIG. 2, and are not repeated here for the sake of clarity.
图4示意性地示出了根据本发明的又一个示例性实施例的用于对图像进行自动分割的网络架构的简化框图。FIG. 4 schematically shows a simplified block diagram of a network architecture for automatic segmentation of images according to yet another exemplary embodiment of the present invention.
图4与图3的主要不同之处在于,输出是多通道的输出。多通道的输出表示该网络不仅能做2分类的图像分割,还具有多类别分类的功能。这里以宫颈癌为例,可将8种ROI分割结果同时输出。8个ROI包括7个危及器官和1个临床靶区,具体分别为:1.膀胱、2.骨髓、3.左股骨头、4.右股骨头、5.直肠、6.小肠、7.脊髓、8.宫颈癌临床靶区。然而,也支持其它肿瘤的多通道输出,当然也支持其它数量的通道数。The main difference between Figure 4 and Figure 3 is that the output is a multi-channel output. The multi-channel output indicates that the network can not only do 2-class image segmentation, but also has the function of multi-class classification. Taking cervical cancer as an example here, 8 kinds of ROI segmentation results can be output at the same time. The 8 ROIs include 7 organs at risk and 1 clinical target area, which are: 1. Bladder, 2. Bone marrow, 3. Left femoral head, 4. Right femoral head, 5. Rectum, 6. Small intestine, 7. Spinal cord 8. Clinical target area of cervical cancer. However, multi-channel output for other tumors is also supported, and of course other numbers of channels.
多通道输出表示输出多张分割后的图像,每张分割后的图像对应一个感兴趣区域。Multi-channel output means outputting multiple segmented images, each segmented image corresponds to a region of interest.
其它类似于图3,在此为了清楚起见,也不再赘述。Others are similar to FIG. 3 , and are not repeated here for the sake of clarity.
图5示意性地示出了根据本发明的一个示例性实施例的用于识别图像 中有哪几种感兴趣区域的简化框图。Figure 5 schematically shows a simplified block diagram for identifying which types of regions of interest are in an image, according to an exemplary embodiment of the present invention.
图5与图4的主要不同之处在于,将输出模块15的最后的输出部分替换为全连接层(Full Connected Layer,FC),从而使分割网络变成了分类网络。该网络能够判断输入的CT图像中包含有哪几种感兴趣区域ROI,但是并不对各种ROI的位置进行判断,即不做具体的分割。具体地讲,在此是将图2-4中所示的Sigmoid激活函数
Figure PCTCN2021105538-appb-000020
替换为全连接层
Figure PCTCN2021105538-appb-000021
其它相同部分也不再赘述。
The main difference between FIG. 5 and FIG. 4 is that the last output part of the output module 15 is replaced with a fully connected layer (Full Connected Layer, FC), so that the segmentation network becomes a classification network. The network can judge which types of ROIs are included in the input CT image, but does not judge the positions of various ROIs, that is, does not do specific segmentation. Specifically, here is the Sigmoid activation function shown in Figure 2-4
Figure PCTCN2021105538-appb-000020
Replaced with a fully connected layer
Figure PCTCN2021105538-appb-000021
Other identical parts will not be repeated.
为了更好地评价、训练和更新网络,通常在网络的终点处提供一个损失函数,以网络的预测值和准备好的金标准作为输入,通过损失函数计算出损失(偏差或误差),然后反向传播到网络中的相应、例如所有层,以更新网络中的相应权重值、例如所有权重值,使得网络预测值越来越接近金标准。例如当损失下降到可接受的程度、例如预定程度时,也更新出最优的权重值,网络才最终确定,此时可结束训练。In order to better evaluate, train and update the network, a loss function is usually provided at the end of the network, with the predicted value of the network and the prepared gold standard as input, the loss (bias or error) is calculated through the loss function, and then reversed The corresponding, eg, all layers in the network are propagated to update the corresponding weight values, eg, all weight values, in the network, so that the network prediction value is getting closer and closer to the gold standard. For example, when the loss drops to an acceptable level, such as a predetermined level, the optimal weight value is also updated, and the network is finally determined, and the training can be ended at this time.
因此,损失函数不仅影响网络的训练过程,而且还直接影响网络的最终性能。Therefore, the loss function not only affects the training process of the network, but also directly affects the final performance of the network.
根据本发明的一个示例性实施例,所述损失函数为交叉熵损失函数(1)或权重自适应损失函数(2):According to an exemplary embodiment of the present invention, the loss function is a cross-entropy loss function (1) or a weight adaptive loss function (2):
Figure PCTCN2021105538-appb-000022
Figure PCTCN2021105538-appb-000022
Figure PCTCN2021105538-appb-000023
Figure PCTCN2021105538-appb-000023
其中,in,
Figure PCTCN2021105538-appb-000024
Figure PCTCN2021105538-appb-000024
Figure PCTCN2021105538-appb-000025
Figure PCTCN2021105538-appb-000025
p是Sigmoid激活函数输出;y是标签图(Ground Truth Map),其取值0或1;I为输入图像;N是总的分割目标的数量。p和y均具有形状I*N;P n是统计出来的每个类别的目标区域、例如危及器官的体积大小占整体图像体积的大小的比例。 p is the output of the sigmoid activation function; y is the ground truth map, which takes the value 0 or 1; I is the input image; N is the total number of segmentation targets. Both p and y have the shape I*N; P n is the ratio of the size of the target area of each category, such as the size of the organ at risk, to the size of the overall image volume.
优选地,对于图4中示例性示出的多通道分割网络,使用公式(2)作为损失函数。多通道分割网络需要分割多个感兴趣区域、例如多个危及器官和临床靶区,如果多个感兴趣区域的体积大小差异特别大,使用交叉熵损失函数难以获得比较好的分割效果,而采用权重自适应损失函数可以获得非常好的分割效果。Preferably, for the multi-channel segmentation network exemplarily shown in Figure 4, formula (2) is used as the loss function. The multi-channel segmentation network needs to segment multiple regions of interest, such as multiple organs at risk and clinical target areas. If the volume sizes of multiple regions of interest are very different, it is difficult to obtain a better segmentation effect using the cross-entropy loss function. The weight adaptive loss function can achieve very good segmentation results.
对于其它分割网络,可以使用公式(1)作为损失函数。For other segmentation networks, Equation (1) can be used as the loss function.
可以理解,上面主要以CT图像为例进行描述,但显然所述图像也可以是X射线成像图像、磁共振成像(MRI)图像、单光子发射计算机断层显像(SPECT)图像、正电子发射型计算机断层显像(PET)图像、超生成像图像等。It can be understood that the above description mainly takes CT images as an example, but obviously the images can also be X-ray imaging images, magnetic resonance imaging (MRI) images, single photon emission computed tomography (SPECT) images, positron emission type images. Computed tomography (PET) images, superimaging images, etc.
本领域的技术人员可以理解,总体而言,本发明提供了一种用于对图像进行自动处理,例如自动分割或分类的神经网络架构,其可以用于医学领域,尤其用于协助对CT图像进行分割或分类。特别地,本发明的神经网络架构适用于协助对肿瘤(尤其是宫颈癌)的放射治疗的临床靶区和危及器官进行自动轮廓勾画所涉及的操作,然后医生可以基于处理结果、例如自动勾画结果进行后续操作、例如审核,必要时可以进行修改,以用于后续治疗目的。Those skilled in the art can understand that, in general, the present invention provides a neural network architecture for automatic image processing, such as automatic segmentation or classification, which can be used in the medical field, especially for assisting in the analysis of CT images. Segmentation or classification. In particular, the neural network architecture of the present invention is suitable for assisting the operations involved in the automatic delineation of clinical target volumes and organs at risk for radiotherapy of tumors (especially cervical cancer), and then physicians can automatically delineate the results based on the processing results, such as Follow-up operations, such as audits, can be modified if necessary for subsequent treatment purposes.
因此,本发明的网络架构在对CT图像进行自动处理时并不是直接用于治疗目的,而是生成中间结果,主要目的是提高医生的工作效率。Therefore, the network architecture of the present invention is not directly used for treatment purposes when automatically processing CT images, but generates intermediate results, and the main purpose is to improve the working efficiency of doctors.
而且,本领域的技术人员还可以理解,本发明还提供了一种计算机可读程序载体,其存储有程序指令,当所述程序指令被处理器执行时用于实施上述网络架构的功能,以进行图像处理。例如,该计算机可读程序载体可被构造成闪存存储器等。Moreover, those skilled in the art can also understand that the present invention also provides a computer-readable program carrier, which stores program instructions, and when the program instructions are executed by the processor, is used to implement the functions of the above-mentioned network architecture, to Perform image processing. For example, the computer-readable program carrier may be constructed as a flash memory or the like.
本发明进一步提供了一种工作站,所述工作站被构造成包括所述计算机可读程序载体。所述工作站尤其可被构造成医生工作站,以供医生使用。The invention further provides a workstation configured to include the computer-readable program carrier. In particular, the workstation can be configured as a doctor's workstation for use by a doctor.
尽管这里详细描述了本发明的特定实施方式,但它们仅仅是为了解释 的目的而给出的,而不应认为它们对本发明的范围构成限制。在不脱离本发明精神和范围的前提下,各种替换、变更和改造可被构想出来。Although specific embodiments of the invention have been described in detail herein, they are presented for purposes of explanation only, and should not be considered as limiting the scope of the invention. Various substitutions, alterations and modifications may be devised without departing from the spirit and scope of the present invention.

Claims (14)

  1. 一种用于对图像进行自动处理的网络架构(1),所述网络架构(1)包括:A network architecture (1) for automatically processing images, the network architecture (1) comprising:
    用于输入待处理的图像的输入模块(11);an input module (11) for inputting an image to be processed;
    编码路径(12),其被配置成使用双路径网络对输入的图像进行特征提取;an encoding path (12) configured to perform feature extraction on the input image using a two-path network;
    解码路径(14),其被配置成与编码路径(12)建立连接;a decoding path (14) configured to establish a connection with the encoding path (12);
    中央模块(13),其被配置成用于从编码路径(12)到解码路径(14)的过渡,以提炼高维图像特征;以及a central module (13) configured for transition from an encoding path (12) to a decoding path (14) to refine high-dimensional image features; and
    输出模块(15),其被配置成从解码路径(14)输出图像处理结果;an output module (15) configured to output image processing results from the decoding path (14);
    其中,所述解码路径(14)被配置成使用Unet架构的解码路径对编码路径(12)的相应编码结果以及中央模块(13)的图像特征提炼结果进行解码操作。The decoding path (14) is configured to perform a decoding operation on the corresponding encoding result of the encoding path (12) and the image feature extraction result of the central module (13) using the decoding path of the Unet architecture.
  2. 根据权利要求1所述的网络架构(1),其中,The network architecture (1) according to claim 1, wherein,
    所述双路径网络包括一系列串联的微块(18),所述微块(18)被相应地嵌入解码路径(14)的解码器中。The dual path network comprises a series of concatenated microblocks (18) that are correspondingly embedded in the decoder of the decoding path (14).
  3. 根据权利要求1或2所述的网络架构(1),其中,The network architecture (1) according to claim 1 or 2, wherein,
    所述编码路径(12)包括5个编码器,所述解码路径(14)包括4个解码器(19),每个解码器(19)分别使用所述编码路径(12)中的相应的微块。The encoding path (12) includes 5 encoders, and the decoding path (14) includes 4 decoders (19), each decoder (19) using a corresponding microcomputer in the encoding path (12), respectively. piece.
  4. 根据权利要求3所述的网络架构(1),其中,The network architecture (1) according to claim 3, wherein,
    所述中央模块(13)被配置成实施:Conv(3*3)+BN+ReLu;以及The central module (13) is configured to implement: Conv(3*3)+BN+ReLu; and
    所述解码器(19)被配置成实施:
    Figure PCTCN2021105538-appb-100001
    The decoder (19) is configured to implement:
    Figure PCTCN2021105538-appb-100001
    其中,Conv(3*3)为3*3卷积运算,BN为批量标准化运算,ReLU为线性整流函数,
    Figure PCTCN2021105538-appb-100002
    表示双线性上采样。
    Among them, Conv(3*3) is a 3*3 convolution operation, BN is a batch normalization operation, ReLU is a linear rectification function,
    Figure PCTCN2021105538-appb-100002
    Represents bilinear upsampling.
  5. 根据权利要求4所述的网络架构(1),其中,The network architecture (1) according to claim 4, wherein,
    所述输入模块(11)接收一张图像或多张图像,所述输出模块(15)输出一张图像或多张图像;以及The input module (11) receives an image or images, and the output module (15) outputs the image or images; and
    所述输出模块(15)被配置成实施:
    Figure PCTCN2021105538-appb-100003
    其中,
    Figure PCTCN2021105538-appb-100004
    表示Sigmoid激活函数。
    The output module (15) is configured to implement:
    Figure PCTCN2021105538-appb-100003
    in,
    Figure PCTCN2021105538-appb-100004
    represents the sigmoid activation function.
  6. 根据权利要求5所述的网络架构(1),其中,The network architecture (1) according to claim 5, wherein,
    所述图像为计算机断层扫描图像、X射线成像图像、磁共振成像图像、单光子发射计算机断层显像图像、正电子发射型计算机断层显像图像、超生成像图像,所述多张图像为前后相邻的图像;和/或The images are computed tomography images, X-ray imaging images, magnetic resonance imaging images, single photon emission computed tomography images, positron emission computed tomography images, and super-generated imaging images, and the multiple images are before and after images. adjacent images; and/or
    所述网络架构(1)被配置成用于对图像进行自动分割处理。The network architecture (1) is configured for automatic segmentation processing of images.
  7. 根据权利要求4所述的网络架构(1),其中,The network architecture (1) according to claim 4, wherein,
    所述输入模块(11)接收一张图像或多张图像,所述输出模块(15)输出分类结果;以及The input module (11) receives an image or images, and the output module (15) outputs a classification result; and
    所述输出模块(15)被配置成实施:
    Figure PCTCN2021105538-appb-100005
    其中,
    Figure PCTCN2021105538-appb-100006
    表示全连接层。
    The output module (15) is configured to implement:
    Figure PCTCN2021105538-appb-100005
    in,
    Figure PCTCN2021105538-appb-100006
    Represents a fully connected layer.
  8. 根据权利要求6所述的网络架构(1),其中,The network architecture (1) according to claim 6, wherein,
    在所述网络架构(1)的终点配置损失函数,所述损失函数为交叉熵损失函数(1)或权重自适应损失函数(2):A loss function is configured at the end point of the network architecture (1), and the loss function is a cross-entropy loss function (1) or a weight adaptive loss function (2):
    Figure PCTCN2021105538-appb-100007
    Figure PCTCN2021105538-appb-100007
    Figure PCTCN2021105538-appb-100008
    Figure PCTCN2021105538-appb-100008
    其中,in,
    Figure PCTCN2021105538-appb-100009
    Figure PCTCN2021105538-appb-100009
    Figure PCTCN2021105538-appb-100010
    Figure PCTCN2021105538-appb-100010
    p是Sigmoid激活函数输出;y是标签图,其取值0或1;I为输入图像;N是总的分割目标的数量;p和y均具有形状I*N;P n是统计出来的每个类别的目标区域的体积大小占整体图像体积的大小的比例。 p is a Sigmoid activation function output; FIG y is label, which value 0 or 1; I is the input image; N is the total number of segmented object; p and y have the shape of I * N; P n is figured out every The proportion of the size of the target area of each category to the size of the overall image volume.
  9. 根据权利要求8所述的网络架构(1),其中,The network architecture (1) according to claim 8, wherein,
    当网络架构(1)被配置成输出多张分割图像时,使用公式(2)作为损失函数。Equation (2) is used as the loss function when the network architecture (1) is configured to output multiple segmented images.
  10. 根据权利要求1-9中任一所述的网络架构(1),其中,The network architecture (1) according to any one of claims 1-9, wherein,
    所述网络架构(1)被配置成用于对肿瘤放射治疗用图像、尤其是宫颈癌放射治疗用图像进行自动分割或分类。The network architecture (1) is configured for automatic segmentation or classification of images for radiotherapy of tumors, in particular images for radiotherapy of cervical cancer.
  11. 一种使用权利要求1-10中任一所述的网络架构(1)对图像进行处理的方法,包括以下步骤:A method for processing an image using the network architecture (1) according to any one of claims 1-10, comprising the following steps:
    通过输入模块(11)输入图像;Input the image through the input module (11);
    通过编码路径(12)、中央模块(13)和解码路径(14)对图像进行处理;以及image processing through an encoding path (12), a central module (13) and a decoding path (14); and
    通过输出模块(15)输出图像处理结果。The image processing result is output through the output module (15).
  12. 一种计算机可读程序载体,其存储有程序指令,当所述程序指令被处理器执行时用于实施根据权利要求12所述的方法。A computer readable program carrier storing program instructions for implementing the method of claim 12 when executed by a processor.
  13. 一种工作站,所述工作站被配置成包括根据权利要求12所述的计算机可读程序载体。A workstation configured to include the computer-readable program carrier of claim 12.
  14. 根据权利要求13所述的工作站,其中,所述工作站被配置成医生工作站,用于对医疗图像进行自动分割或分类。14. The workstation of claim 13, wherein the workstation is configured as a doctor's workstation for automatic segmentation or classification of medical images.
PCT/CN2021/105538 2020-07-10 2021-07-09 Network architecture for automatically processing images, program carrier, and workstation WO2022007957A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010663227.X 2020-07-10
CN202010663227.XA CN111784682B (en) 2020-07-10 2020-07-10 Network architecture system, program carrier and workstation for automatic processing of images

Publications (1)

Publication Number Publication Date
WO2022007957A1 true WO2022007957A1 (en) 2022-01-13

Family

ID=72768151

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/105538 WO2022007957A1 (en) 2020-07-10 2021-07-09 Network architecture for automatically processing images, program carrier, and workstation

Country Status (2)

Country Link
CN (1) CN111784682B (en)
WO (1) WO2022007957A1 (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111784682B (en) * 2020-07-10 2024-05-28 北京医智影科技有限公司 Network architecture system, program carrier and workstation for automatic processing of images
CN112734762B (en) * 2020-12-31 2022-10-11 西华师范大学 Dual-path UNet network tumor segmentation method based on covariance self-attention mechanism
CN113378933A (en) * 2021-06-11 2021-09-10 合肥合滨智能机器人有限公司 Thyroid ultrasound image classification and segmentation network, training method, device and medium
CN117197156B (en) * 2022-10-21 2024-04-02 南华大学 Lesion segmentation method and system based on double decoders UNet and Transformer

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150238148A1 (en) * 2013-10-17 2015-08-27 Siemens Aktiengesellschaft Method and system for anatomical object detection using marginal space deep neural networks
CN109598727A (en) * 2018-11-28 2019-04-09 北京工业大学 A kind of CT image pulmonary parenchyma three-dimensional semantic segmentation method based on deep neural network
CN111242288A (en) * 2020-01-16 2020-06-05 浙江工业大学 Multi-scale parallel deep neural network model construction method for lesion image segmentation
CN111369574A (en) * 2020-03-11 2020-07-03 合肥凯碧尔高新技术有限公司 Thoracic cavity organ segmentation method and device
CN111784682A (en) * 2020-07-10 2020-10-16 北京医智影科技有限公司 Network architecture, program carrier and workstation for automatic processing of images

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10755391B2 (en) * 2018-05-15 2020-08-25 Adobe Inc. Digital image completion by learning generation and patch matching jointly
CN110059772B (en) * 2019-05-14 2021-04-30 温州大学 Remote sensing image semantic segmentation method based on multi-scale decoding network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150238148A1 (en) * 2013-10-17 2015-08-27 Siemens Aktiengesellschaft Method and system for anatomical object detection using marginal space deep neural networks
CN109598727A (en) * 2018-11-28 2019-04-09 北京工业大学 A kind of CT image pulmonary parenchyma three-dimensional semantic segmentation method based on deep neural network
CN111242288A (en) * 2020-01-16 2020-06-05 浙江工业大学 Multi-scale parallel deep neural network model construction method for lesion image segmentation
CN111369574A (en) * 2020-03-11 2020-07-03 合肥凯碧尔高新技术有限公司 Thoracic cavity organ segmentation method and device
CN111784682A (en) * 2020-07-10 2020-10-16 北京医智影科技有限公司 Network architecture, program carrier and workstation for automatic processing of images

Also Published As

Publication number Publication date
CN111784682A (en) 2020-10-16
CN111784682B (en) 2024-05-28

Similar Documents

Publication Publication Date Title
WO2022007957A1 (en) Network architecture for automatically processing images, program carrier, and workstation
Dong et al. Automatic multiorgan segmentation in thorax CT images using U‐net‐GAN
Fu et al. A review of deep learning based methods for medical image multi-organ segmentation
Dong et al. Synthetic MRI-aided multi-organ segmentation on male pelvic CT using cycle consistent deep attention network
Men et al. Deep deconvolutional neural network for target segmentation of nasopharyngeal cancer in planning computed tomography images
Liu et al. MRI-based treatment planning for liver stereotactic body radiotherapy: validation of a deep learning-based synthetic CT generation method
Liu et al. Development and validation of a deep learning algorithm for auto-delineation of clinical target volume and organs at risk in cervical cancer radiotherapy
Zhan et al. Multi-constraint generative adversarial network for dose prediction in radiotherapy
JP6018077B2 (en) System and method for automatically generating an initial radiation treatment plan
CN112770811A (en) Method and system for radiation therapy treatment planning using a deep learning engine
Shi et al. Deep learning empowered volume delineation of whole-body organs-at-risk for accelerated radiotherapy
US9098912B2 (en) Method, system and computer readable medium for automatic segmentation of a medical image
US11682485B2 (en) Methods and systems for adaptive radiotherapy treatment planning using deep learning engines
CN110960803B (en) Computer system for performing adaptive radiation therapy planning
US9355447B2 (en) System and method for gradient assisted non-connected automatic region (GANAR) analysis
US20230290480A1 (en) Systems and methods for clinical target contouring in radiotherapy
Yoganathan et al. Automatic segmentation of magnetic resonance images for high‐dose‐rate cervical cancer brachytherapy using deep learning
WO2015189303A1 (en) Optimization of parameters for segmenting an image
Sartor et al. Auto-segmentations by convolutional neural network in cervical and anorectal cancer with clinical structure sets as the ground truth
Vandewinckele et al. Segmentation of head-and-neck organs-at-risk in longitudinal CT scans combining deformable registrations and convolutional neural networks
Huang et al. 3D lightweight network for simultaneous registration and segmentation of organs-at-risk in CT images of head and neck cancer
Kawahara et al. Stepwise deep neural network (stepwise-net) for head and neck auto-segmentation on CT images
Xu et al. Bg-net: Boundary-guided network for lung segmentation on clinical ct images
Losnegård et al. Semi-automated segmentation of the sigmoid and descending colon for radiotherapy planning using the fast marching method
Chen et al. Evaluation of multi atlas-based approaches for the segmentation of the thyroid gland in IMRT head-and-neck CT images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21837836

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21837836

Country of ref document: EP

Kind code of ref document: A1