WO2023040164A1 - Method and apparatus for training pet/ct-based lung adenocarcinoma and squamous carcinoma diagnosis model - Google Patents

Method and apparatus for training pet/ct-based lung adenocarcinoma and squamous carcinoma diagnosis model Download PDF

Info

Publication number
WO2023040164A1
WO2023040164A1 PCT/CN2022/074386 CN2022074386W WO2023040164A1 WO 2023040164 A1 WO2023040164 A1 WO 2023040164A1 CN 2022074386 W CN2022074386 W CN 2022074386W WO 2023040164 A1 WO2023040164 A1 WO 2023040164A1
Authority
WO
WIPO (PCT)
Prior art keywords
pet
pathological
lung adenocarcinoma
cell carcinoma
squamous cell
Prior art date
Application number
PCT/CN2022/074386
Other languages
French (fr)
Chinese (zh)
Inventor
朱闻韬
金源
黄海亮
薛梦凡
Original Assignee
之江实验室
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 之江实验室 filed Critical 之江实验室
Priority to JP2022562000A priority Critical patent/JP2023544466A/en
Publication of WO2023040164A1 publication Critical patent/WO2023040164A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10081Computed x-ray tomography [CT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10104Positron emission tomography [PET]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30061Lung
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30096Tumor; Lesion

Definitions

  • the present invention relates to the fields of medical imaging and deep learning, in particular to a fully automatic intelligent diagnosis model training method and device for lung adenocarcinoma and squamous cell carcinoma based on PET/CT and pathological slices.
  • Positron emission tomography is a functional imaging device at the molecular level.
  • the radioactive tracer needs to be injected into the patient before scanning, and the tracer decays and annihilates in the patient's body, producing a pair of 511keV gamma photons with emission directions about 180° opposite, and the detector will collect these gamma photons to reach the position of the crystal and time information.
  • image reconstruction algorithms to reconstruct and post-process the acquired information, the metabolism and uptake of the reaction tracer in the patient can be obtained.
  • Pathological examination that is, a pathological and morphological method used to examine pathological changes in organs, tissues or cells of the body, is the examination method with the highest diagnostic accuracy among all examinations.
  • the pathologist takes out a small piece of tissue from the diseased part of the patient's body (according to different situations, forceps, excision, or puncture suction can be used) or surgically removes the specimen to make pathological sections, and observes the morphological changes of cells and tissues to determine the pathological changes.
  • biopsy biopsy
  • Pathological examination is a commonly used and relatively accurate method in the diagnosis of tumors, and is known as the "gold standard" for diagnosis.
  • Deep convolutional neural network is one of the common methods for building medical artificial intelligence models in recent years. It extracts feature information of different dimensions of images through layered convolution processing, and the extracted features are input into subsequent specific Networks perform specific tasks such as classification, segmentation, registration, detection, noise reduction, etc.
  • the advantage of this method is that it can automatically learn high-order features that are significant for specific tasks through samples, but it has certain requirements for the amount of data used for training.
  • the purpose of the present invention is to address the deficiencies in the prior art, to provide a PET/CT-based lung adenocarcinoma squamous cell carcinoma diagnostic model training method and device, the present invention is based on multi-objective learning (Multi-Task Learning) algorithm, in order to pass Pathological features are used to optimize the PET/CT-based deep learning network to improve the training efficiency and upper limit of accuracy of the PET/CT network.
  • Multi-Task Learning Multi-Task Learning
  • a training method based on PET/CT lung adenocarcinoma and squamous cell carcinoma diagnostic classification model specifically comprising:
  • An initial neural network model is constructed, with PET/CT images as input, predicted pathological features and lung adenocarcinoma squamous cell carcinoma diagnosis results as output, and the acquired data is used to train the initial neural network model, and a PET/CT based lung adenocarcinoma squamous cell carcinoma is obtained.
  • Cancer diagnosis classification model ;
  • the input of the neural network A is a pathological image
  • the output is the diagnosis result of lung adenocarcinoma and squamous cell carcinoma, which is obtained through the training of the obtained pathological image data. Since the pathological image has a "gold standard" diagnosis and classification effect for lung cancer, it can be trained
  • the network A makes it have extremely high classification accuracy for the input pathological image, and the diagnostic classification accuracy is required to be greater than or equal to 0.95 in the present invention
  • the pathological feature is the output of a feature extraction layer of the neural network A. Preferably, it is the input of the previous layer of the output layer of the neural network A.
  • the initial neural network model specifically includes the following features:
  • the network contains two inputs. After the pre-processing convolution layer, the input PET and CT images are normalized in size, superimposed along the channel dimension, and then enter the main convolution layer.
  • the network is a multi-task target network, which is mainly established based on the regularization framework.
  • the target function is:
  • n i represents the number of training samples
  • l(.,.) represents a loss function, such as cross-entropy loss or mean square error loss
  • U ⁇ R d ⁇ d is a square transformation matrix
  • a ⁇ R d ⁇ m contains the parameters of each task
  • d is the dimension of the parameter
  • 2 2,1 is Its L2 regularization matrix
  • a i represents the model parameters of the i-th task
  • I is the identity matrix
  • is the regularization parameter.
  • the first part of the objective function represents the experience loss of all tasks
  • the second part uses L2 regularization to ensure that the learned rows are sparse and the orthogonalization of the constraint matrix U, then formula (1) can also be expressed as:
  • tr(.) represents the trace of the matrix
  • D ⁇ 0 indicates that the D matrix is a positive semi-definite matrix.
  • the optimization of the multi-task target network is to solve the covariance matrix D, which decouples multiple task problems and promotes their parallel computing.
  • the main task of the initial neural network model is the diagnosis and classification of lung cancer
  • the auxiliary task is the fitting of pathological features.
  • the input image outputs high-dimensional features after passing through the main convolutional layer, and the features output diagnostic classification results after being processed by the fully connected layer of the main task, and output fitting pathological features after being processed by the convolutional layer of the auxiliary task.
  • the two outputs are compared with the real diagnostic results of the case and the pathological features output by network A and the loss is calculated.
  • the two losses jointly determine the parameter update of the initial neural network model.
  • the parameter quantity of the initial neural network model should be less than or equal to the parameter quantity of the neural network A, so that the initial neural network model can learn the diagnostic classification features output by the neural network A, and avoid overfitting.
  • S mask and S are the area of the image marked by the doctor's mark with lung cancer information and the whole area including the background, respectively.
  • the neural network A adopts a ResNet-50 structure.
  • the lung adenocarcinoma and squamous cell carcinoma diagnostic classification model adopts the DenseNet-121 structure, and the input is the feature of fusion of PET and CT images along the channel dimension.
  • the initial neural network model is trained using the acquired data, and the PET/CT lung adenocarcinoma and squamous cell carcinoma diagnostic classification model is obtained as follows:
  • Loss loss1*(1- ⁇ )+loss2* ⁇
  • loss1 and loss2 are the errors between the diagnostic results and pathological features of lung adenocarcinoma squamous cell carcinoma and the corresponding true value, respectively, and ⁇ is a hyperparameter.
  • loss1 uses a cross-entropy loss function
  • loss2 uses a mean square loss function
  • the present invention also provides a training device for a PET/CT lung adenocarcinoma and squamous cell carcinoma diagnostic classification model, including:
  • a data acquisition unit configured to acquire corresponding PET/CT images, pathological images, and diagnosis result data of lung adenocarcinoma and squamous cell carcinoma;
  • a pathological feature acquisition unit configured to input pathological images to a trained neural network A to obtain pathological features
  • the training unit is used to construct the initial neural network model, and the PET/CT image is used as input, the predicted pathological features and lung adenocarcinoma squamous cell carcinoma diagnosis results are output, and the data obtained by the data acquisition unit and the pathological feature acquisition unit are used to train the described
  • the initial neural network model was used to obtain a diagnostic classification model for lung adenocarcinoma and squamous cell carcinoma based on PET/CT.
  • a data preprocessing unit is also included, which is used to process the corresponding PET/CT images and pathological images into pictures of the same size.
  • the trained lung adenocarcinoma and squamous cell carcinoma diagnostic classification model can directly output the diagnostic classification results based on PET/CT images without the participation of pathological data. specifically:
  • a PET/CT diagnostic classification device for lung adenocarcinoma and squamous cell carcinoma comprising:
  • a data acquisition module configured to acquire PET/CT images to be diagnosed
  • the diagnosis and classification module of lung adenocarcinoma and squamous cell carcinoma is used to input the PET/CT image to be diagnosed into the PET/CT-based diagnostic classification model of lung adenocarcinoma and squamous cell carcinoma trained by any one of the above training methods to obtain the result of diagnosis and classification.
  • the present invention uses a multi-task learning method to assist in training a lung adenocarcinoma and squamous cell carcinoma diagnosis and classification model based on PET/CT images through a diagnostic and classification neural network based on pathological images.
  • This method aims to assist the training of the diagnostic classification model of lung adenocarcinoma and squamous cell carcinoma through pathological features, and improve the accuracy of lung cancer diagnostic classification based on PET/CT images.
  • pathological images are only used as prior knowledge in the training process, and do not need to be used as the input of the network in practical applications.
  • this method improves the accuracy of PET/CT images used in the diagnosis and classification of lung cancer, which is conducive to its further promotion and application as a means of early lung cancer diagnosis, and provides clinicians with patient diagnosis and follow-up treatment plans.
  • pathological images can further improve the interpretability of pathological slices and help pathologists further extract pathological features.
  • Fig. 1 is a flow chart of training the lung adenocarcinoma squamous cell carcinoma diagnostic model based on PET/CT;
  • Fig. 2 is a neural network structure diagram trained on a PET/CT-based lung adenocarcinoma squamous cell carcinoma diagnostic model.
  • the following examples illustrate how to specifically apply this method to introduce pathological information into the PET/CT-based lung cancer diagnosis and classification network.
  • a kind of PET/CT-based lung adenocarcinoma squamous cell carcinoma diagnostic model training method of the present invention is specifically as follows:
  • Step 1 Obtain the corresponding PET/CT images, pathological images and lung adenocarcinoma squamous cell carcinoma diagnostic result data, establish a single input and output classification convolutional neural network, and combine the pathological images corresponding to PET/CT images and lung adenocarcinoma squamous cell carcinoma
  • the cancer diagnosis results are imported into the classification convolutional neural network. Since the pathological image has the "gold standard" diagnostic classification effect for lung cancer, the classification convolutional neural network can be trained to have a very high classification accuracy for the input pathological image, and then saved Parameters to obtain the trained neural network A. And input the pathological image to a trained neural network A to obtain the pathological features; specifically include the following sub-steps:
  • S mask and S are the area of the image marked by the doctor's mark with lung cancer information and the whole area including the background, respectively.
  • the purpose of formula (3) is to ensure that the input pathological slice images contain certain lung cancer classification features. Due to the different areas of lung cancer information representation images on different original pathological images, in order to ensure the balance of the sample size distribution during the training process, the Overlap-tile strategy is used in this example to keep the number of slices cut out from each original image. unanimous.
  • S(x) and x represent the input and output of the activation function, respectively.
  • Step 2 Establish an initial neural network model with multi-input and multi-task target output, use the paired PET and CT images as input, and the corresponding diagnostic classification results and pathological features as output, and train the initial neural network model. Specifically include the following sub-steps:
  • a lung cancer slice of size 224*224 is cut out from the original PET/CT image based on the location of the lung cancer lesion, and the slice is superimposed along the channel layer and then input into the initial neural network model.
  • the target output of the network One is the diagnosis and classification result of lung cancer of the case
  • the target output two is the pathological features of the case saved in step two (also normalized using the Sigmoid function).
  • the errors are calculated separately (in this example, the target output 1 uses the cross-entropy loss function, and the target output 2 uses the mean square loss function), and the actual error of the network is:
  • Loss loss1*(1- ⁇ )+loss2* ⁇ (5)
  • loss1 and loss2 are the errors of target output 1 and 2 respectively
  • is a hyperparameter, that is, the ratio of the error of target output 2 to the total network error, ⁇ (0,1).
  • the network parameters are updated and adjusted by the actual error, that is, the network is jointly trained by two target tasks, and a PET/CT-based lung adenocarcinoma and squamous cell carcinoma diagnostic classification model is obtained.
  • the trained lung adenocarcinoma squamous cell carcinoma diagnostic classification model can directly output diagnostic classification results based on PET/CT images without the participation of pathological data.
  • the lung adenocarcinoma squamous cell carcinoma diagnostic classification model has higher accuracy for the lung cancer diagnostic classification of PET/CT images, while the pathological images are only used in the training process of the network. Application does not need to provide. Therefore, the trained PET/CT lung cancer diagnosis classification network has higher accuracy and better stability than the network trained solely by PET/CT images, which has clinical practical significance for the early diagnosis of lung cancer.
  • the training device for the PET/CT lung adenocarcinoma squamous cell carcinoma diagnostic classification model constructed based on the above training method includes:
  • a data acquisition unit configured to acquire corresponding PET/CT images, pathological images, and diagnosis result data of lung adenocarcinoma and squamous cell carcinoma;
  • a pathological feature acquisition unit configured to input pathological images to a trained neural network A to obtain pathological features
  • the training unit is used to construct the initial neural network model, and the PET/CT image is used as input, the predicted pathological features and lung adenocarcinoma squamous cell carcinoma diagnosis results are output, and the data obtained by the data acquisition unit and the pathological feature acquisition unit are used to train the described
  • the initial neural network model was used to obtain a diagnostic classification model for lung adenocarcinoma and squamous cell carcinoma based on PET/CT.
  • the trained lung adenocarcinoma and squamous cell carcinoma diagnostic classification model can directly output the diagnostic classification results based on PET/CT images without the participation of pathological data. specifically:
  • a PET/CT diagnostic classification device for lung adenocarcinoma and squamous cell carcinoma based on the lung adenocarcinoma and squamous cell carcinoma diagnostic classification model obtained through training includes:
  • a data acquisition module configured to acquire PET/CT images to be diagnosed
  • the diagnosis and classification module of lung adenocarcinoma and squamous cell carcinoma is used to input the PET/CT image to be diagnosed into the PET/CT-based diagnostic classification model of lung adenocarcinoma and squamous cell carcinoma trained by any one of the above training methods to obtain the result of diagnosis and classification.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Radiology & Medical Imaging (AREA)
  • Quality & Reliability (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Image Analysis (AREA)
  • Nuclear Medicine (AREA)

Abstract

Provided in the present invention are a method and apparatus for training a PET/CT-based lung adenocarcinoma and squamous carcinoma diagnosis model. In the present invention, a multi-task learning method is used, and a pathological feature is extracted by a neural network that is obtained by means of diagnosis classification training based on pathological images, thereby assisting in the training of a PET/CT-image-based diagnosis classification neural network. By means of the present invention, the precision of lung carcinoma diagnosis and classification based on a PET/CT image is improved, and pathological images are only used as a priori knowledge during a training process, and do not need to be used as inputs of a network during practical applications. In the method, the concept of multi-dimension fusion is used, such that the precision of using a PET/CT image for lung carcinoma diagnosis and classification is improved, thereby facilitating the further popularization and application of using the PET/CT image as a means for early lung carcinoma diagnosis, and helping a clinician to provide diagnosis for a patient and helping with a subsequent treatment scheme; in addition, by means of the solution of using pathological images as a priori knowledge assistance, the interpretability of a pathological section is further improved, thereby facilitating a pathologist in further extracting a pathological feature.

Description

一种基于PET/CT的肺腺癌鳞癌诊断模型训练方法及装置A PET/CT-based diagnostic model training method and device for lung adenocarcinoma and squamous cell carcinoma 技术领域technical field
本发明涉及医学影像领域和深度学习领域,尤其涉及一种基于PET/CT和病理切片的全自动肺腺癌鳞癌的智能诊断模型训练方法及装置。The present invention relates to the fields of medical imaging and deep learning, in particular to a fully automatic intelligent diagnosis model training method and device for lung adenocarcinoma and squamous cell carcinoma based on PET/CT and pathological slices.
背景技术Background technique
正电子发射断层扫描仪(positron emission tomography,PET)是一种在分子层面上的功能性成像设备。扫描前需要对患者注入放射性示踪剂,示踪剂在患者体内进行衰变进而发生湮灭,产生一对发射方向约180°相反的511keV伽马光子,检测器会采集这些伽马光子达到晶体的位置和时间信息。通过使用图像重构建算法对采集的信息进行重构建并进行后处理,即可获得反应示踪剂在患者体内代谢和摄取的情况。医生根据PET/CT的影像结果,结合各项临床指标综合分析患者的病情,从而确定治疗方案。Positron emission tomography (PET) is a functional imaging device at the molecular level. The radioactive tracer needs to be injected into the patient before scanning, and the tracer decays and annihilates in the patient's body, producing a pair of 511keV gamma photons with emission directions about 180° opposite, and the detector will collect these gamma photons to reach the position of the crystal and time information. By using image reconstruction algorithms to reconstruct and post-process the acquired information, the metabolism and uptake of the reaction tracer in the patient can be obtained. According to the imaging results of PET/CT, doctors comprehensively analyze the patient's condition in combination with various clinical indicators, so as to determine the treatment plan.
病理检查,即用以检查机体器官、组织或细胞中的病理改变的病理形态学方法,是所有检查之中诊断准确率最高的一种检查方法。病理科医生从患者身体的病变部位取出小块组织(根据不同情况可采用钳取、切除或穿刺吸取等方法)或通过手术切除标本制成病理切片,观察细胞和组织的形态结构变化,以确定病变性质,作出病理诊断,称为活体组织检查(biopsy),简称活体。病理检查是诊断肿瘤方法中常用且较为准确的方法,被称为诊断的“金标准”。Pathological examination, that is, a pathological and morphological method used to examine pathological changes in organs, tissues or cells of the body, is the examination method with the highest diagnostic accuracy among all examinations. The pathologist takes out a small piece of tissue from the diseased part of the patient's body (according to different situations, forceps, excision, or puncture suction can be used) or surgically removes the specimen to make pathological sections, and observes the morphological changes of cells and tissues to determine the pathological changes. The nature of the lesion and the pathological diagnosis are called biopsy (biopsy), referred to as living body. Pathological examination is a commonly used and relatively accurate method in the diagnosis of tumors, and is known as the "gold standard" for diagnosis.
深度卷积神经网络(convolutional neural network,CNN)是近年构建医学人工智能模型的常用方法之一,它通过分层的卷积处理提取图像的不同维度的特征信息,所提取的特征则输入后续特定网络进行特定任务,如分类、分割、配准、检测、降噪等。该方法的优势在于可通过样本自动学习对特定任务具有显著意义的高阶特征,但是对于用于训练的数据量具有一定的要求。Deep convolutional neural network (CNN) is one of the common methods for building medical artificial intelligence models in recent years. It extracts feature information of different dimensions of images through layered convolution processing, and the extracted features are input into subsequent specific Networks perform specific tasks such as classification, segmentation, registration, detection, noise reduction, etc. The advantage of this method is that it can automatically learn high-order features that are significant for specific tasks through samples, but it has certain requirements for the amount of data used for training.
关于现有的基于PET/CT的肺癌诊断分类模型,由于数据规模以及精度的限制,无论该模型是基于单一中心数据还是多中心数据训练,其诊断分类精度均未能达到实用的要求。而被视为癌症诊断“金标准”的病理切片,由于采样往往需要对患者进行侵入性甚至有创性检查,使其较少应用于早期诊断。因此,开发一种基于PET/CT的能以较高正确率对肺癌进行早期诊断分类的模型,能够在一定程度上提高医院对于早期肺癌的诊断率以及帮助临床医生进行后续的治疗。Regarding the existing PET/CT-based lung cancer diagnosis and classification model, due to the limitation of data scale and accuracy, no matter whether the model is trained on single-center data or multi-center data, its diagnosis and classification accuracy cannot meet the practical requirements. Pathological slides, which are regarded as the "gold standard" for cancer diagnosis, are rarely used in early diagnosis because sampling often requires invasive or even invasive examinations on patients. Therefore, the development of a PET/CT-based model that can classify early lung cancer with a high accuracy rate can improve the hospital's diagnosis rate of early lung cancer to a certain extent and help clinicians to carry out follow-up treatment.
发明内容Contents of the invention
本发明的目的在于针对现有技术的不足,提供一种基于PET/CT的肺腺癌鳞癌诊断模型训练方法及装置,本发明基于多目标学习(Multi-Task Learning)的算法,用以通过病理特征来对基于PET/CT的深度学习网络进行优化,以提高PET/CT网络的训练效率以及精度上限。而在临床应用过程中,只需要输入PET/CT图像即可获得肺腺癌鳞癌的诊断分类信息,无需病理信息的参与。The purpose of the present invention is to address the deficiencies in the prior art, to provide a PET/CT-based lung adenocarcinoma squamous cell carcinoma diagnostic model training method and device, the present invention is based on multi-objective learning (Multi-Task Learning) algorithm, in order to pass Pathological features are used to optimize the PET/CT-based deep learning network to improve the training efficiency and upper limit of accuracy of the PET/CT network. In the clinical application process, only PET/CT images are needed to obtain the diagnostic classification information of lung adenocarcinoma and squamous cell carcinoma, without the participation of pathological information.
本发明的目的是通过以下技术方案来实现的:The purpose of the present invention is achieved through the following technical solutions:
一种基于PET/CT肺腺癌鳞癌诊断分类模型的训练方法,具体包括:A training method based on PET/CT lung adenocarcinoma and squamous cell carcinoma diagnostic classification model, specifically comprising:
获取对应的PET/CT图像、病理图像和肺腺癌鳞癌诊断结果数据,将病理图像输入至一训练好的神经网络A获取病理特征;Obtain the corresponding PET/CT images, pathological images, and lung adenocarcinoma squamous cell carcinoma diagnostic result data, and input the pathological images into a trained neural network A to obtain pathological features;
构建初始神经网络模型,以PET/CT图像为输入,预测的病理特征和肺腺癌鳞癌诊断结果为输出,利用获取的数据训练所述初始神经网络模型,得到基于PET/CT肺腺癌鳞癌诊断分类模型;An initial neural network model is constructed, with PET/CT images as input, predicted pathological features and lung adenocarcinoma squamous cell carcinoma diagnosis results as output, and the acquired data is used to train the initial neural network model, and a PET/CT based lung adenocarcinoma squamous cell carcinoma is obtained. Cancer diagnosis classification model;
其中,所述神经网络A的输入为病理图像,输出为肺腺癌鳞癌诊断结果,通过获取的病理图像数据训练获得,由于病理图像对于肺癌具有“金标准”的诊断分类效果,因此可训练网络A使其对于输入的病理图像具有极高的分类精度,本发明中要求诊断分类精度大于等于0.95;所述病理特征为神经网络A的一特征提取层的输出。优选地,为神经网络A输出层的前一层输入。Wherein, the input of the neural network A is a pathological image, and the output is the diagnosis result of lung adenocarcinoma and squamous cell carcinoma, which is obtained through the training of the obtained pathological image data. Since the pathological image has a "gold standard" diagnosis and classification effect for lung cancer, it can be trained The network A makes it have extremely high classification accuracy for the input pathological image, and the diagnostic classification accuracy is required to be greater than or equal to 0.95 in the present invention; the pathological feature is the output of a feature extraction layer of the neural network A. Preferably, it is the input of the previous layer of the output layer of the neural network A.
优选地,所述初始神经网络模型具体包括以下特征:Preferably, the initial neural network model specifically includes the following features:
1.网络包含两个输入,分别输入的PET和CT图像在经过前处理卷积层后,尺寸归一化,并沿着通道维度叠加,进入之后的主卷积层。1. The network contains two inputs. After the pre-processing convolution layer, the input PET and CT images are normalized in size, superimposed along the channel dimension, and then enter the main convolution layer.
2.网络为多任务目标网络,主要基于正则化框架建立,目标函数为:2. The network is a multi-task target network, which is mainly established based on the regularization framework. The target function is:
Figure PCTCN2022074386-appb-000001
Figure PCTCN2022074386-appb-000001
s.t.U U T=I stU U T =I
其中m表示任务数,n i表示训练样本数,
Figure PCTCN2022074386-appb-000002
表示第j个训练样本、第i个任务的标签,l(.,.)代表了一种损失函数,例如交叉熵损失或均方差损失,b=(b 1,...b m) T是所有任务中的偏移补偿,U∈R d×d是平方变换矩阵,A∈R d×m包含了各个任务的参数,d为参数的维度,而||A|| 2 2,1则是其L2正则化矩阵,a i表示其中第i个任务的模型参数,I是单位矩阵而λ则是正则化参数。目标函数的第一部分代表了所有任务的经验损失,第二部分则通过L2正则化确保了解的行稀疏以及约束矩阵U的正交化,则公式(1)也可表示为:
Where m represents the number of tasks, n i represents the number of training samples,
Figure PCTCN2022074386-appb-000002
Indicates the label of the j-th training sample and the i-th task, l(.,.) represents a loss function, such as cross-entropy loss or mean square error loss, b=(b 1 ,...b m ) T is Offset compensation in all tasks, U∈R d×d is a square transformation matrix, A∈R d×m contains the parameters of each task, d is the dimension of the parameter, and ||A|| 2 2,1 is Its L2 regularization matrix, a i represents the model parameters of the i-th task, I is the identity matrix and λ is the regularization parameter. The first part of the objective function represents the experience loss of all tasks, and the second part uses L2 regularization to ensure that the learned rows are sparse and the orthogonalization of the constraint matrix U, then formula (1) can also be expressed as:
Figure PCTCN2022074386-appb-000003
Figure PCTCN2022074386-appb-000003
其中
Figure PCTCN2022074386-appb-000004
表示公式(1)中的总训练损失,tr(.)代表矩阵的迹,W i=Ua i代表了第i个任务的模型参数,而D≥0则表示D矩阵是半正定矩阵。而多任务目标网络的优化,即为求解协方差矩阵D,使得多个任务问题解耦,促进它们的并行计算。
in
Figure PCTCN2022074386-appb-000004
Indicates the total training loss in formula (1), tr(.) represents the trace of the matrix, W i =Ua i represents the model parameters of the i-th task, and D≥0 indicates that the D matrix is a positive semi-definite matrix. The optimization of the multi-task target network is to solve the covariance matrix D, which decouples multiple task problems and promotes their parallel computing.
3.初始神经网络模型的主任务为肺癌的诊断分类,辅助任务为病理特征的拟合。输入的图像在经过主卷积层后输出高维特征,特征在经过主任务的全连接层处理后输出诊断分类结果,经过辅助任务的卷积层处理后输出拟合病理特征。两个输出分别与病例的真实诊断结果以及网络A输出的病理特征比较并求损失,两个损失共同决定了初始神经网络模型的参数更新。3. The main task of the initial neural network model is the diagnosis and classification of lung cancer, and the auxiliary task is the fitting of pathological features. The input image outputs high-dimensional features after passing through the main convolutional layer, and the features output diagnostic classification results after being processed by the fully connected layer of the main task, and output fitting pathological features after being processed by the convolutional layer of the auxiliary task. The two outputs are compared with the real diagnostic results of the case and the pathological features output by network A and the loss is calculated. The two losses jointly determine the parameter update of the initial neural network model.
进一步地,初始神经网络模型的参数量应少于等于神经网络A的参数量,使初始神经网络模型能够学习到神经网络A输出的诊断分类特征,而避免过拟合。Further, the parameter quantity of the initial neural network model should be less than or equal to the parameter quantity of the neural network A, so that the initial neural network model can learn the diagnostic classification features output by the neural network A, and avoid overfitting.
进一步地,所述病理图像满足:Further, the pathological image satisfies:
Figure PCTCN2022074386-appb-000005
Figure PCTCN2022074386-appb-000005
其中S mask与S all分别为医生标注下有肺癌信息表征的图像面积以及包括背景的全面积。 Among them, S mask and S all are the area of the image marked by the doctor's mark with lung cancer information and the whole area including the background, respectively.
进一步地,所述神经网络A采用ResNet-50结构。Further, the neural network A adopts a ResNet-50 structure.
进一步地,所述肺腺癌鳞癌诊断分类模型采用DenseNet-121结构,输入为PET与CT图像在沿通道维度融合的特征。Further, the lung adenocarcinoma and squamous cell carcinoma diagnostic classification model adopts the DenseNet-121 structure, and the input is the feature of fusion of PET and CT images along the channel dimension.
进一步地,所述利用获取的数据训练所述初始神经网络模型,得到基于PET/CT肺腺癌鳞癌诊断分类模型具体为:Further, the initial neural network model is trained using the acquired data, and the PET/CT lung adenocarcinoma and squamous cell carcinoma diagnostic classification model is obtained as follows:
计算所述初始神经网络模型的输出与对应真值的误差,根据误差更新所述初始神经网络模型的参数,直至误差最小,获得基于PET/CT肺腺癌鳞癌诊断分类模型;所述损失表示为:Calculate the error between the output of the initial neural network model and the corresponding true value, update the parameters of the initial neural network model according to the error until the error is minimum, and obtain a diagnostic classification model based on PET/CT lung adenocarcinoma and squamous cell carcinoma; the loss represents for:
Loss=loss1*(1-φ)+loss2*φLoss=loss1*(1-φ)+loss2*φ
其中loss1与loss2分别为肺腺癌鳞癌诊断结果和病理特征与对应真值的误差,φ为超参数。Among them, loss1 and loss2 are the errors between the diagnostic results and pathological features of lung adenocarcinoma squamous cell carcinoma and the corresponding true value, respectively, and φ is a hyperparameter.
进一步地,所述loss1采用交叉熵损失函数,loss2采用均方损失函数。Further, the loss1 uses a cross-entropy loss function, and loss2 uses a mean square loss function.
基于相同的发明思路,本发明还提供了一种PET/CT肺腺癌鳞癌诊断分类模型的训练装置,包括:Based on the same inventive idea, the present invention also provides a training device for a PET/CT lung adenocarcinoma and squamous cell carcinoma diagnostic classification model, including:
数据获取单元,用于获取对应的PET/CT图像、病理图像和肺腺癌鳞癌诊断结果数据;A data acquisition unit, configured to acquire corresponding PET/CT images, pathological images, and diagnosis result data of lung adenocarcinoma and squamous cell carcinoma;
病理特征获取单元,用于将病理图像输入至一训练好的神经网络A获取病理特征;A pathological feature acquisition unit, configured to input pathological images to a trained neural network A to obtain pathological features;
训练单元,用于构建初始神经网络模型,并以PET/CT图像为输入,预测的病理特征和肺腺癌鳞癌诊断结果为输出,利用数据获取单元和病理特征获取单元获取的数据训练所述初始神经网络模型,得到基于PET/CT肺腺癌鳞癌诊断分类模型。The training unit is used to construct the initial neural network model, and the PET/CT image is used as input, the predicted pathological features and lung adenocarcinoma squamous cell carcinoma diagnosis results are output, and the data obtained by the data acquisition unit and the pathological feature acquisition unit are used to train the described The initial neural network model was used to obtain a diagnostic classification model for lung adenocarcinoma and squamous cell carcinoma based on PET/CT.
进一步地,还包括数据预处理单元,用于将对应的PET/CT图像及病理图像处理成大小一致的图片。Further, a data preprocessing unit is also included, which is used to process the corresponding PET/CT images and pathological images into pictures of the same size.
训练完成的肺腺癌鳞癌诊断分类模型可以基于PET/CT图像,直接输出诊断分类结果,而无需病理数据的参与。具体地:The trained lung adenocarcinoma and squamous cell carcinoma diagnostic classification model can directly output the diagnostic classification results based on PET/CT images without the participation of pathological data. specifically:
一种PET/CT肺腺癌鳞癌诊断分类装置,包括:A PET/CT diagnostic classification device for lung adenocarcinoma and squamous cell carcinoma, comprising:
数据获取模块,用于获取待诊断的PET/CT图像;A data acquisition module, configured to acquire PET/CT images to be diagnosed;
肺腺癌鳞癌诊断分类模块,用于将待诊断的PET/CT图像输入至上述任一项训练方法训练获得的基于PET/CT肺腺癌鳞癌诊断分类模型中,获得诊断分类结果。The diagnosis and classification module of lung adenocarcinoma and squamous cell carcinoma is used to input the PET/CT image to be diagnosed into the PET/CT-based diagnostic classification model of lung adenocarcinoma and squamous cell carcinoma trained by any one of the above training methods to obtain the result of diagnosis and classification.
本发明使用了多任务学习的方法,通过基于病理图像的诊断分类神经网络,来辅助训练基于PET/CT图像的肺腺癌鳞癌诊断分类模型。本方法旨在通过病理特征来辅助肺腺癌鳞癌诊断分类模型的训练,提高基于PET/CT图像的肺癌诊断分类精度。同时病理图像仅用作训练过程中的先验知识,在实际应用中并不需要用作网络的输入。此方法通过多尺度融合的理念,提高了PET/CT图像用作肺癌诊断分类的精度,有利于其作为早期肺癌诊断的手段得到进一步推广与应用,为临床医师对患者的诊断以及后续治疗方案提供帮助;与此同时,病理图像作为先验知识辅助的方案,也进一步提升了病理切片的可解释性,有助于病理科医师对于病理特征的进一步提取。The present invention uses a multi-task learning method to assist in training a lung adenocarcinoma and squamous cell carcinoma diagnosis and classification model based on PET/CT images through a diagnostic and classification neural network based on pathological images. This method aims to assist the training of the diagnostic classification model of lung adenocarcinoma and squamous cell carcinoma through pathological features, and improve the accuracy of lung cancer diagnostic classification based on PET/CT images. At the same time, pathological images are only used as prior knowledge in the training process, and do not need to be used as the input of the network in practical applications. Through the concept of multi-scale fusion, this method improves the accuracy of PET/CT images used in the diagnosis and classification of lung cancer, which is conducive to its further promotion and application as a means of early lung cancer diagnosis, and provides clinicians with patient diagnosis and follow-up treatment plans. At the same time, as a priori knowledge-assisted solution, pathological images can further improve the interpretability of pathological slices and help pathologists further extract pathological features.
附图说明Description of drawings
图1是基于PET/CT的肺腺癌鳞癌诊断模型训练流程图;Fig. 1 is a flow chart of training the lung adenocarcinoma squamous cell carcinoma diagnostic model based on PET/CT;
图2是基于PET/CT的肺腺癌鳞癌诊断模型训练的神经网络结构图。Fig. 2 is a neural network structure diagram trained on a PET/CT-based lung adenocarcinoma squamous cell carcinoma diagnostic model.
具体实施方式Detailed ways
以下结合实例,说明如何具体的应用本方法在基于PET/CT的肺癌诊断分类网络中引入病理信息。The following examples illustrate how to specifically apply this method to introduce pathological information into the PET/CT-based lung cancer diagnosis and classification network.
如图1-2所示,本发明的一种基于PET/CT的肺腺癌鳞癌诊断模型训练方法,具体如下:As shown in Figure 1-2, a kind of PET/CT-based lung adenocarcinoma squamous cell carcinoma diagnostic model training method of the present invention is specifically as follows:
步骤一:获取对应的PET/CT图像、病理图像和肺腺癌鳞癌诊断结果数据,建立单一输入及输出的分类卷积神经网络,把同PET/CT图像对应的病理图像以及肺腺癌鳞癌诊断结果导入 分类卷积神经网络中,由于病理图像对于肺癌具有“金标准”的诊断分类效果,因此可训练分类卷积神经网络使其对于输入的病理图像具有极高的分类精度,之后保存参数获得训练好的神经网络A。并将病理图像输入至一训练好的神经网络A获取病理特征;具体包括以下子步骤:Step 1: Obtain the corresponding PET/CT images, pathological images and lung adenocarcinoma squamous cell carcinoma diagnostic result data, establish a single input and output classification convolutional neural network, and combine the pathological images corresponding to PET/CT images and lung adenocarcinoma squamous cell carcinoma The cancer diagnosis results are imported into the classification convolutional neural network. Since the pathological image has the "gold standard" diagnostic classification effect for lung cancer, the classification convolutional neural network can be trained to have a very high classification accuracy for the input pathological image, and then saved Parameters to obtain the trained neural network A. And input the pathological image to a trained neural network A to obtain the pathological features; specifically include the following sub-steps:
(1.1)建立分类卷积神经网络,在本实例中分类卷积神经网络采用ResNet-50结构,具体结构如表1:(1.1) Establish a classification convolutional neural network. In this example, the classification convolutional neural network adopts the ResNet-50 structure. The specific structure is shown in Table 1:
表1:ResNet-50网络结构Table 1: ResNet-50 network structure
Figure PCTCN2022074386-appb-000006
Figure PCTCN2022074386-appb-000006
(1.2)建立用于训练的病理切片数据集,由于使用的原始病理图像均为全视野数字切片(Whole Slide Image),具有极高的分辨率。为满足神经网络的输入尺寸要求以及网络训练所需的计算资源,本实施例在预处理阶段将所有的原始病理图像切成尺寸为224*224的切片。切片需满足的条件为:(1.2) Establish a pathological slice data set for training. Since the original pathological images used are all whole-field digital slices (Whole Slide Image), they have extremely high resolution. In order to meet the input size requirements of the neural network and the computing resources required for network training, this embodiment cuts all the original pathological images into slices with a size of 224*224 in the preprocessing stage. The conditions to be met for slices are:
Figure PCTCN2022074386-appb-000007
Figure PCTCN2022074386-appb-000007
其中S mask与S all分别为医生标注下有肺癌信息表征的图像面积以及包括背景的全面积。公式(3)目的为保证输入的病理切片图像均包含一定的肺癌分类特征。由于不同原始病理图像上的肺癌信息表征图像面积不同,为保证训练过程中各样本量分布平衡,在本实例中使用重叠平铺策略(Overlap-tile strategy)使得各原始图像切出的切片数量保持一致。 Among them, S mask and S all are the area of the image marked by the doctor's mark with lung cancer information and the whole area including the background, respectively. The purpose of formula (3) is to ensure that the input pathological slice images contain certain lung cancer classification features. Due to the different areas of lung cancer information representation images on different original pathological images, in order to ensure the balance of the sample size distribution during the training process, the Overlap-tile strategy is used in this example to keep the number of slices cut out from each original image. unanimous.
(1.3)将样本分为训练集与验证集,其中验证集需保证包含所有病例的切片图像,使 用训练集训练步骤(1.1)中建立的分类卷积神经网络,训练好的网络需保证在测试集中拥有极高的准确度,大于等于0.95,本实例中要求训练好的神经网络A对病理切片的肺癌诊断分类精度为0.99。(1.3) Divide the sample into a training set and a verification set, wherein the verification set needs to ensure that the slice images of all cases are included, using the classification convolutional neural network established in the training set training step (1.1), the trained network needs to be guaranteed to be in the test The concentration has a very high accuracy, greater than or equal to 0.95. In this example, the trained neural network A is required to have a lung cancer diagnosis and classification accuracy of 0.99 for pathological slices.
(1.4)对于每一组病例,有对应的PET、CT以及病理图像,将步骤(1.3)中建立的病理图像验证集数据输入到步骤一中训练好的神经网络A中,获取其对应的诊断分类结果,同时提取神经网络A输出层的前一层输入,作为该病例的病理特征保存,保存的特征需提前归一化,在本实例中使用Sigmoid函数(公式4)进行归一化,每一例PET/CT图像对应一张病理切片所提取的特征。(1.4) For each group of cases, there are corresponding PET, CT and pathological images. Input the pathological image verification set data established in step (1.3) into the neural network A trained in step 1 to obtain its corresponding diagnosis At the same time, extract the input of the previous layer of the output layer of the neural network A, and save it as the pathological feature of the case. The saved features need to be normalized in advance. In this example, the Sigmoid function (formula 4) is used for normalization. A PET/CT image corresponds to the features extracted from a pathological slice.
Figure PCTCN2022074386-appb-000008
S(x)和x分别表示激活函数的输入与输出。
Figure PCTCN2022074386-appb-000008
S(x) and x represent the input and output of the activation function, respectively.
步骤二:建立多输入及多任务目标输出的初始神经网络模型,将配对的PET以及CT图像作为输入,对应的诊断分类结果以及病理特征作为输出,对初始神经网络模型进行训练。具体包括以下子步骤:Step 2: Establish an initial neural network model with multi-input and multi-task target output, use the paired PET and CT images as input, and the corresponding diagnostic classification results and pathological features as output, and train the initial neural network model. Specifically include the following sub-steps:
(2.1)建立基于PET/CT的肺癌诊断分类的初始神经网络模型,并包含有病理特征的拟合输出。在本实例中初始神经网络模型采用DenseNet-121结构,主结构如表2所示。(2.1) Establish an initial neural network model for lung cancer diagnostic classification based on PET/CT, and include the fitting output of pathological features. In this example, the initial neural network model adopts the DenseNet-121 structure, and the main structure is shown in Table 2.
表2:DenseNet-121网络结构Table 2: DenseNet-121 network structure
Figure PCTCN2022074386-appb-000009
Figure PCTCN2022074386-appb-000009
(2.2)在预处理阶段,基于肺癌的病灶位置从PET/CT原始图像中切出尺寸为224*224的肺 癌切片,将切片沿着通道层叠加后输入初始神经网络模型中,网络的目标输出一为该病例的肺癌诊断分类结果,目标输出二为步骤二保存好的该病例的病理特征(亦使用Sigmoid函数进行归一化)。对于两个目标输出与其对应真值分别求误差(本实例中目标输出一使用交叉熵损失函数,目标输出二使用均方损失函数),网络实际误差为:(2.2) In the preprocessing stage, a lung cancer slice of size 224*224 is cut out from the original PET/CT image based on the location of the lung cancer lesion, and the slice is superimposed along the channel layer and then input into the initial neural network model. The target output of the network One is the diagnosis and classification result of lung cancer of the case, and the target output two is the pathological features of the case saved in step two (also normalized using the Sigmoid function). For the two target outputs and their corresponding true values, the errors are calculated separately (in this example, the target output 1 uses the cross-entropy loss function, and the target output 2 uses the mean square loss function), and the actual error of the network is:
Loss=loss1*(1-φ)+loss2*φ   (5)Loss=loss1*(1-φ)+loss2*φ (5)
其中loss1与loss2分别是目标输出一与二的误差,φ为超参数,即目标输出二的误差占全网络误差的比值,φ∈(0,1)。通过实际误差来更新调整网络参数,也即通过两个目标任务共同训练网络,获得基于PET/CT肺腺癌鳞癌诊断分类模型。Among them, loss1 and loss2 are the errors of target output 1 and 2 respectively, and φ is a hyperparameter, that is, the ratio of the error of target output 2 to the total network error, φ∈(0,1). The network parameters are updated and adjusted by the actual error, that is, the network is jointly trained by two target tasks, and a PET/CT-based lung adenocarcinoma and squamous cell carcinoma diagnostic classification model is obtained.
训练获得的肺腺癌鳞癌诊断分类模型可以基于PET/CT图像,直接输出诊断分类结果,而无需病理数据的参与。The trained lung adenocarcinoma squamous cell carcinoma diagnostic classification model can directly output diagnostic classification results based on PET/CT images without the participation of pathological data.
得益于病理图像带来的先验知识,肺腺癌鳞癌诊断分类模型对于PET/CT图像的肺癌诊断分类具有更高的准确性,而病理图像仅仅作用于网络的训练过程中,在临床应用中并不需要提供。由此,训练得到的PET/CT肺癌诊断分类网络比单纯由PET/CT图像训练得到的网络精度更高,稳定性更好,对于肺癌的早期诊断具有临床实际意义。Thanks to the prior knowledge brought by the pathological images, the lung adenocarcinoma squamous cell carcinoma diagnostic classification model has higher accuracy for the lung cancer diagnostic classification of PET/CT images, while the pathological images are only used in the training process of the network. Application does not need to provide. Therefore, the trained PET/CT lung cancer diagnosis classification network has higher accuracy and better stability than the network trained solely by PET/CT images, which has clinical practical significance for the early diagnosis of lung cancer.
另外,作为一优选方案,基于上述训练方法构建的PET/CT肺腺癌鳞癌诊断分类模型的训练装置,包括:In addition, as a preferred solution, the training device for the PET/CT lung adenocarcinoma squamous cell carcinoma diagnostic classification model constructed based on the above training method includes:
数据获取单元,用于获取对应的PET/CT图像、病理图像和肺腺癌鳞癌诊断结果数据;A data acquisition unit, configured to acquire corresponding PET/CT images, pathological images, and diagnosis result data of lung adenocarcinoma and squamous cell carcinoma;
病理特征获取单元,用于将病理图像输入至一训练好的神经网络A获取病理特征;A pathological feature acquisition unit, configured to input pathological images to a trained neural network A to obtain pathological features;
训练单元,用于构建初始神经网络模型,并以PET/CT图像为输入,预测的病理特征和肺腺癌鳞癌诊断结果为输出,利用数据获取单元和病理特征获取单元获取的数据训练所述初始神经网络模型,得到基于PET/CT肺腺癌鳞癌诊断分类模型。The training unit is used to construct the initial neural network model, and the PET/CT image is used as input, the predicted pathological features and lung adenocarcinoma squamous cell carcinoma diagnosis results are output, and the data obtained by the data acquisition unit and the pathological feature acquisition unit are used to train the described The initial neural network model was used to obtain a diagnostic classification model for lung adenocarcinoma and squamous cell carcinoma based on PET/CT.
训练完成的肺腺癌鳞癌诊断分类模型可以基于PET/CT图像,直接输出诊断分类结果,而无需病理数据的参与。具体地:The trained lung adenocarcinoma and squamous cell carcinoma diagnostic classification model can directly output the diagnostic classification results based on PET/CT images without the participation of pathological data. specifically:
作为一优选方案,基于训练获得的肺腺癌鳞癌诊断分类模型的一种PET/CT肺腺癌鳞癌诊断分类装置,包括:As a preferred solution, a PET/CT diagnostic classification device for lung adenocarcinoma and squamous cell carcinoma based on the lung adenocarcinoma and squamous cell carcinoma diagnostic classification model obtained through training includes:
数据获取模块,用于获取待诊断的PET/CT图像;A data acquisition module, configured to acquire PET/CT images to be diagnosed;
肺腺癌鳞癌诊断分类模块,用于将待诊断的PET/CT图像输入至上述任一项训练方法训练获得的基于PET/CT肺腺癌鳞癌诊断分类模型中,获得诊断分类结果。The diagnosis and classification module of lung adenocarcinoma and squamous cell carcinoma is used to input the PET/CT image to be diagnosed into the PET/CT-based diagnostic classification model of lung adenocarcinoma and squamous cell carcinoma trained by any one of the above training methods to obtain the result of diagnosis and classification.
显然,上述实施例仅仅是为清楚地说明所作的举例,而并非对实施方式的限定。对于所属领域的普通技术人员来说,在上述说明的基础上还可以做出其他不同形式的变化或变动。 这里无需也无法把所有的实施方式予以穷举。而由此所引申出的显而易见的变化或变动仍处于本发明的保护范围。Apparently, the above-mentioned embodiments are only examples for clear description, rather than limiting the implementation. For those of ordinary skill in the art, other changes or changes in different forms can be made on the basis of the above description. It is not necessary and impossible to exhaustively list all implementation modes here. However, the obvious changes or variations derived therefrom still fall within the protection scope of the present invention.

Claims (8)

  1. 一种基于PET/CT肺腺癌鳞癌诊断分类模型的训练方法,其特征在于,具体包括:A kind of training method based on PET/CT lung adenocarcinoma squamous cell carcinoma diagnostic classification model, it is characterized in that, specifically comprises:
    获取对应的PET/CT图像、病理图像和临床肺腺癌鳞癌诊断结果数据,并将病理图像输入至一训练好的神经网络A获取病理特征;Obtain the corresponding PET/CT images, pathological images, and clinical lung adenocarcinoma and squamous cell carcinoma diagnosis data, and input the pathological images into a trained neural network A to obtain pathological features;
    构建初始神经网络模型,以PET/CT图像为输入,预测的病理特征和肺腺癌鳞癌诊断结果为输出,利用获取的PET/CT图像、病理特征和临床肺腺癌鳞癌诊断结果数据训练所述初始神经网络模型,得到基于PET/CT肺腺癌鳞癌诊断分类模型,具体为:Construct an initial neural network model, take PET/CT images as input, predict pathological features and diagnosis results of lung adenocarcinoma and squamous cell carcinoma as output, and use the acquired PET/CT images, pathological features and clinical diagnosis results of lung adenocarcinoma and squamous cell carcinoma for training The initial neural network model is obtained based on PET/CT lung adenocarcinoma squamous cell carcinoma diagnostic classification model, specifically:
    计算所述初始神经网络模型的输出与对应真值的误差,根据误差更新所述初始神经网络模型的参数,直至误差最小,获得基于PET/CT肺腺癌鳞癌诊断分类模型;损失表示为:Calculate the error between the output of the initial neural network model and the corresponding true value, update the parameters of the initial neural network model according to the error, until the error is minimum, and obtain a diagnostic classification model based on PET/CT lung adenocarcinoma and squamous cell carcinoma; the loss is expressed as:
    Loss=loss1*(1-φ)+loss2*φLoss=loss1*(1-φ)+loss2*φ
    其中loss1与loss2分别为肺腺癌鳞癌诊断结果和病理特征与对应真值的误差,φ为超参数,φ∈(0,1);Among them, loss1 and loss2 are the errors between the diagnostic results of lung adenocarcinoma and squamous cell carcinoma, pathological features and corresponding true values, φ is a hyperparameter, φ∈(0,1);
    其中,所述神经网络A的输入为病理图像,输出为肺腺癌鳞癌诊断结果,诊断分类精度大于等于0.95;所述病理特征为神经网络A的特征提取层的输出。Wherein, the input of the neural network A is a pathological image, and the output is the diagnosis result of lung adenocarcinoma and squamous cell carcinoma, and the diagnostic classification accuracy is greater than or equal to 0.95; the pathological feature is the output of the feature extraction layer of the neural network A.
  2. 根据权利要求1所述的训练方法,其特征在于,所述病理图像满足:The training method according to claim 1, wherein the pathological image satisfies:
    Figure PCTCN2022074386-appb-100001
    Figure PCTCN2022074386-appb-100001
    其中S mask与S all分别为医生标注下有肺癌信息表征的图像面积以及包括背景的全面积。 Among them, S mask and S all are the area of the image marked by the doctor's mark with lung cancer information and the whole area including the background, respectively.
  3. 根据权利要求1所述的训练方法,其特征在于,所述神经网络A采用ResNet-50结构。The training method according to claim 1, wherein the neural network A adopts a ResNet-50 structure.
  4. 根据权利要求1所述的训练方法,其特征在于,所述初始神经网络模型采用DenseNet-121结构,输入为PET与CT图像在沿通道维度融合的特征。The training method according to claim 1, wherein the initial neural network model adopts a DenseNet-121 structure, and the input is a feature of fusion of PET and CT images along the channel dimension.
  5. 根据权利要求1所述的训练方法,其特征在于,所述loss1采用交叉熵损失函数,loss2采用均方损失函数。The training method according to claim 1, wherein said loss1 adopts a cross-entropy loss function, and loss2 adopts a mean square loss function.
  6. 一种基于权利要求1-5任一项训练方法的PET/CT肺腺癌鳞癌诊断分类模型的训练装置,其特征在于,包括:A training device for a PET/CT lung adenocarcinoma squamous cell carcinoma diagnostic classification model based on any one of the training methods of claims 1-5, characterized in that it comprises:
    数据获取单元,用于获取对应的PET/CT图像、病理图像和肺腺癌鳞癌诊断结果数据;A data acquisition unit, configured to acquire corresponding PET/CT images, pathological images, and diagnosis result data of lung adenocarcinoma and squamous cell carcinoma;
    病理特征获取单元,用于将病理图像输入至一训练好的神经网络A获取病理特征;A pathological feature acquisition unit, configured to input pathological images to a trained neural network A to obtain pathological features;
    训练单元,用于构建初始神经网络模型,并以PET/CT图像为输入,预测的病理特征和肺腺癌鳞癌诊断结果为输出,利用数据获取单元和病理特征获取单元获取的数据训练所述初始神经网络模型,得到基于PET/CT肺腺癌鳞癌诊断分类模型。The training unit is used to construct the initial neural network model, and the PET/CT image is used as input, the predicted pathological features and lung adenocarcinoma squamous cell carcinoma diagnosis results are output, and the data obtained by the data acquisition unit and the pathological feature acquisition unit are used to train the described The initial neural network model was used to obtain a diagnostic classification model for lung adenocarcinoma and squamous cell carcinoma based on PET/CT.
  7. 根据权利要求6所述的训练装置,其特征在于,还包括数据预处理单元,用于将对应的 PET/CT图像及病理图像处理成大小一致的图片。The training device according to claim 6, further comprising a data preprocessing unit for processing the corresponding PET/CT images and pathological images into pictures of the same size.
  8. 一种PET/CT肺腺癌鳞癌诊断分类装置,其特征在于,包括:A PET/CT diagnostic classification device for lung adenocarcinoma and squamous cell carcinoma, characterized in that it comprises:
    数据获取模块,用于获取待诊断的PET/CT图像;A data acquisition module, configured to acquire PET/CT images to be diagnosed;
    肺腺癌鳞癌诊断分类模块,用于将待诊断的PET/CT图像输入至权利要求1-6任一项训练方法训练获得的基于PET/CT肺腺癌鳞癌诊断分类模型中,获得诊断分类结果。Lung adenocarcinoma and squamous cell carcinoma diagnosis and classification module, for inputting the PET/CT image to be diagnosed into the PET/CT lung adenocarcinoma and squamous cell carcinoma diagnosis and classification model obtained by training according to any one of claims 1-6, to obtain a diagnosis classification results.
PCT/CN2022/074386 2021-09-14 2022-01-27 Method and apparatus for training pet/ct-based lung adenocarcinoma and squamous carcinoma diagnosis model WO2023040164A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2022562000A JP2023544466A (en) 2021-09-14 2022-01-27 Training method and device for diagnostic model of lung adenocarcinoma and squamous cell carcinoma based on PET/CT

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111075570.3 2021-09-14
CN202111075570.3A CN113516210B (en) 2021-09-14 2021-09-14 Lung adenocarcinoma squamous carcinoma diagnosis model training method and device based on PET/CT

Publications (1)

Publication Number Publication Date
WO2023040164A1 true WO2023040164A1 (en) 2023-03-23

Family

ID=78063210

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/074386 WO2023040164A1 (en) 2021-09-14 2022-01-27 Method and apparatus for training pet/ct-based lung adenocarcinoma and squamous carcinoma diagnosis model

Country Status (3)

Country Link
JP (1) JP2023544466A (en)
CN (1) CN113516210B (en)
WO (1) WO2023040164A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116468690A (en) * 2023-04-17 2023-07-21 北京透彻未来科技有限公司 Subtype analysis system of invasive non-mucous lung adenocarcinoma based on deep learning
CN116825363A (en) * 2023-08-29 2023-09-29 济南市人民医院 Early lung adenocarcinoma pathological type prediction system based on fusion deep learning network
CN117831757A (en) * 2024-03-05 2024-04-05 之江实验室 Pathological CT multi-mode priori knowledge-guided lung cancer diagnosis method and system

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113516210B (en) * 2021-09-14 2021-12-17 之江实验室 Lung adenocarcinoma squamous carcinoma diagnosis model training method and device based on PET/CT
CN114360718B (en) * 2022-03-10 2022-07-12 之江实验室 Feature fitting-based PET/CT automatic lung cancer diagnosis and classification system and construction method
CN114897849A (en) * 2022-05-21 2022-08-12 四川大学 Newborn auricle morphological deformity diagnosis method based on convolutional neural network
CN116152246B (en) 2023-04-19 2023-07-25 之江实验室 Image recognition method, device, equipment and storage medium
CN116664550B (en) * 2023-07-10 2024-04-12 广州医科大学附属第一医院(广州呼吸中心) Intelligent identification method and device for lung cancer tissue immunohistochemical PD-L1 pathological section

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108805858A (en) * 2018-04-10 2018-11-13 燕山大学 Hepatopathy CT image computers assistant diagnosis system based on data mining and method
US20200184643A1 (en) * 2018-12-11 2020-06-11 New York University Classification and mutation prediction from histopathology images using deep learning
CN111584073A (en) * 2020-05-13 2020-08-25 山东大学 Artificial intelligence fusion multi-modal information-based diagnosis model for constructing multiple pathological types of benign and malignant pulmonary nodules
CN111932562A (en) * 2020-09-22 2020-11-13 平安科技(深圳)有限公司 Image identification method and device based on CT sequence, electronic equipment and medium
CN113516210A (en) * 2021-09-14 2021-10-19 之江实验室 Lung adenocarcinoma squamous carcinoma diagnosis model training method and device based on PET/CT

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112419452B (en) * 2020-12-24 2022-08-23 福州大学 Rapid merging system and method for PD-L1 digital pathological section images of stomach cancer

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108805858A (en) * 2018-04-10 2018-11-13 燕山大学 Hepatopathy CT image computers assistant diagnosis system based on data mining and method
US20200184643A1 (en) * 2018-12-11 2020-06-11 New York University Classification and mutation prediction from histopathology images using deep learning
CN111584073A (en) * 2020-05-13 2020-08-25 山东大学 Artificial intelligence fusion multi-modal information-based diagnosis model for constructing multiple pathological types of benign and malignant pulmonary nodules
CN111932562A (en) * 2020-09-22 2020-11-13 平安科技(深圳)有限公司 Image identification method and device based on CT sequence, electronic equipment and medium
CN113516210A (en) * 2021-09-14 2021-10-19 之江实验室 Lung adenocarcinoma squamous carcinoma diagnosis model training method and device based on PET/CT

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116468690A (en) * 2023-04-17 2023-07-21 北京透彻未来科技有限公司 Subtype analysis system of invasive non-mucous lung adenocarcinoma based on deep learning
CN116468690B (en) * 2023-04-17 2023-11-14 北京透彻未来科技有限公司 Subtype analysis system of invasive non-mucous lung adenocarcinoma based on deep learning
CN116825363A (en) * 2023-08-29 2023-09-29 济南市人民医院 Early lung adenocarcinoma pathological type prediction system based on fusion deep learning network
CN116825363B (en) * 2023-08-29 2023-12-12 济南市人民医院 Early lung adenocarcinoma pathological type prediction system based on fusion deep learning network
CN117831757A (en) * 2024-03-05 2024-04-05 之江实验室 Pathological CT multi-mode priori knowledge-guided lung cancer diagnosis method and system
CN117831757B (en) * 2024-03-05 2024-05-28 之江实验室 Pathological CT multi-mode priori knowledge-guided lung cancer diagnosis method and system

Also Published As

Publication number Publication date
CN113516210A (en) 2021-10-19
JP2023544466A (en) 2023-10-24
CN113516210B (en) 2021-12-17

Similar Documents

Publication Publication Date Title
WO2023040164A1 (en) Method and apparatus for training pet/ct-based lung adenocarcinoma and squamous carcinoma diagnosis model
Xue et al. An application of transfer learning and ensemble learning techniques for cervical histopathology image classification
Latif et al. Medical imaging using machine learning and deep learning algorithms: a review
CN106056595B (en) Based on the pernicious assistant diagnosis system of depth convolutional neural networks automatic identification Benign Thyroid Nodules
CN109670510B (en) Deep learning-based gastroscope biopsy pathological data screening system
Mamalakis et al. DenResCov-19: A deep transfer learning network for robust automatic classification of COVID-19, pneumonia, and tuberculosis from X-rays
Kadry et al. Automated segmentation of leukocyte from hematological images—a study using various CNN schemes
CN107563383A (en) A kind of medical image auxiliary diagnosis and semi-supervised sample generation system
CN112085113B (en) Severe tumor image recognition system and method
Zhang et al. Fda: Feature decomposition and aggregation for robust airway segmentation
CN112263217A (en) Non-melanoma skin cancer pathological image lesion area detection method based on improved convolutional neural network
CN113889261B (en) Pathological feature assistance-based PET/CT automatic lung cancer diagnosis classification model training method
Abu Haeyeh et al. Development and evaluation of a novel deep-learning-based framework for the classification of renal histopathology images
Li et al. Convolutional neural networks for classifying cervical cancer types using histological images
Li et al. Research about tongue image of traditional chinese medicine (TCM) based on artificial intelligence technology
CN113539476A (en) Stomach endoscopic biopsy Raman image auxiliary diagnosis method and system based on artificial intelligence
CN112508943A (en) Breast tumor identification method based on ultrasonic image
Barshooi IoT-based classification and differential diagnosis of COVID-19 from CT Images using DexiNed filter, image blending and weighted ensemble CNNs
Sun et al. Liver tumor segmentation and subsequent risk prediction based on Deeplabv3+
Chaudhry et al. Unitochest: A lung image dataset for segmentation of cancerous nodules on ct scans
Darwish et al. Enhancing cervical pre-cancerous classification using advanced vision transformer
Rohith et al. Convolutional neural networks for placenta cell classification
Chae et al. An Investigation of Transfer Learning Approaches to Overcome Limited Labeled Data in Medical Image Analysis
Yumeng et al. Pneumonia Detection in chest X-rays: A deep learning approach based on ensemble RetinaNet and Mask R-CNN
Kalaivani et al. A Deep Ensemble Model for Automated Multiclass Classification Using Dermoscopy Images

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 2022562000

Country of ref document: JP

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22868544

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE