WO2021082416A1 - 网络模型训练方法及装置、病灶区域确定方法及装置 - Google Patents
网络模型训练方法及装置、病灶区域确定方法及装置 Download PDFInfo
- Publication number
- WO2021082416A1 WO2021082416A1 PCT/CN2020/092570 CN2020092570W WO2021082416A1 WO 2021082416 A1 WO2021082416 A1 WO 2021082416A1 CN 2020092570 W CN2020092570 W CN 2020092570W WO 2021082416 A1 WO2021082416 A1 WO 2021082416A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- network model
- training data
- lesion area
- model
- medical image
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30061—Lung
Definitions
- the present disclosure relates to the field of image processing technology, in particular to a network model training method and device, a lesion area determination method and device, a computer-readable storage medium, and electronic equipment.
- the embodiments of the present application provide a network model training method and device, a lesion area determination method and device, a computer-readable storage medium, and electronic equipment.
- an embodiment of the present disclosure provides a network model training method.
- the network model training method includes: determining first training data based on a sample image, wherein the sample image includes a lesion area, and the first training data includes a labeled first The coordinate information of the lesion area and the information of the first lesion type; an initial network model is determined, and the initial network model is trained based on the sample image to generate a network model for determining the lesion area in the medical image.
- an embodiment of the present disclosure provides a method for determining a lesion area.
- the method for determining a lesion area includes: determining a medical image of a lesion area that needs to be determined; and inputting the medical image to a network for determining the lesion area in the medical image
- the model is used to determine the coordinate information of the lesion area of the medical image, wherein the network model used to determine the lesion area in the medical image can be obtained based on the network model training method mentioned in the above embodiment.
- an embodiment of the present disclosure provides a network model training device.
- the network model training device includes: a first training data determining module for determining first training data based on a sample image, wherein the sample image includes a lesion area ,
- the first training data includes the coordinate information of the marked first lesion area and the first lesion type information;
- the training module is used to determine the initial network model, and train the initial network model based on the sample image to generate the information used to determine the medical image Network model of the lesion area.
- an embodiment of the present disclosure provides an apparatus for determining a lesion area.
- the apparatus for determining a lesion area includes: an image determining module for determining a medical image of a lesion area that needs to be determined; a lesion area determining module for comparing the medical image Input to the network model used to determine the focus area in the medical image to determine the focus area coordinate information of the medical image, where the network model used to determine the focus area in the medical image can be based on the network model training mentioned in the above embodiment Method to obtain.
- the embodiments of the present disclosure provide a computer-readable storage medium, the storage medium stores a computer program, and the computer program is used to execute the network model training method mentioned in the above-mentioned embodiment, or to execute the above-mentioned embodiment The mentioned method of determining the lesion area.
- an embodiment of the present disclosure provides an electronic device, the electronic device includes a processor and a memory for storing executable instructions of the processor, wherein the processor is used for executing the network model mentioned in the above embodiment Training method, or execute the method for determining the lesion area mentioned in the above embodiment.
- the network model training method determines the first training data based on the sample image, then determines the initial network model, and trains the initial network model based on the sample image to generate a network model for determining the lesion area in the medical image
- the method achieves the purpose of training an initial network model using sample images to generate a network model for determining the lesion area in a medical image.
- the sample image is a medical image that includes the lesion area
- the first training data determined based on the sample image includes the labeled first lesion area coordinate information and the first lesion type information, therefore, the training generated based on the sample image is used to determine the medical image
- the network model of the lesion area can be used to assist the doctor in determining the lesion area in any medical image of the same type as the sample image.
- using the network model training method provided by the embodiments of the present disclosure to train and generate the network model for determining the lesion area in the medical image can assist the doctor in determining the lesion area in the medical image of the same type as the sample image (such as determining the lesion area).
- the method for determining the focus area provided by the embodiment of the present disclosure realizes the determination by inputting the medical image that needs to determine the focus area into the network model for determining the focus area in the medical image to determine the focus area coordinate information of the medical image.
- the purpose of the coordinate information of the lesion area in the medical image Since the method for determining the lesion area provided by the embodiment of the present disclosure is implemented based on the network model used to determine the lesion area in the medical image, compared with the existing solution, the embodiment of the present disclosure does not need to determine the medical image of the lesion area. Perform complex image enhancement, filter transformation and other processing operations, thereby avoiding the failure of predicting the coordinate information of the lesion area due to factors such as image quality. That is, the method for determining the lesion area provided by the embodiment of the present disclosure has the advantages of high stability and good robustness.
- FIG. 1 is a schematic diagram of a scenario to which the embodiments of the present disclosure are applicable.
- FIG. 2 is a schematic diagram of another scenario to which the embodiments of the present disclosure are applicable.
- Fig. 3 is a schematic flowchart of a network model training method provided by an exemplary embodiment of the present disclosure.
- FIG. 4 is a schematic flowchart of training an initial network model based on sample images to generate a network model for determining a lesion area in a medical image according to an exemplary embodiment of the present disclosure.
- FIG. 5 is a schematic flowchart of training an initial network model based on sample images to generate a network model for determining a lesion area in a medical image according to another exemplary embodiment of the present disclosure.
- FIG. 6 shows the method of adjusting the network parameters of the prediction model and the image feature extraction model based on the second lesion type information in the second training data and the first lesion type information in the first training data provided by an exemplary embodiment of the present disclosure.
- FIG. 7 is a schematic flowchart of training an initial network model based on sample images to generate a network model for determining a lesion area in a medical image according to another exemplary embodiment of the present disclosure.
- Fig. 8 is a schematic structural diagram of an initial network model provided by an exemplary embodiment of the present disclosure.
- FIG. 9 shows the first parameter adjustment operation performed on the initial network model based on the first training data and the second training data provided by an exemplary embodiment of the present disclosure to generate a network model for determining the lesion area in a medical image Schematic diagram of the process.
- FIG. 10 is a schematic diagram of a process of determining first training data based on sample images according to an exemplary embodiment of the present disclosure.
- FIG. 11 is a schematic flowchart of a method for determining a lesion area provided by an exemplary embodiment of the present disclosure.
- FIG. 12 is a schematic flowchart of a method for determining a lesion area according to another exemplary embodiment of the present disclosure.
- FIG. 13 is a schematic diagram of region division of a medical image provided by an exemplary embodiment of the present disclosure.
- FIG. 14 is a schematic diagram of a process of performing a region division operation on a medical image to generate multiple divided regions according to an exemplary embodiment of the present disclosure.
- FIG. 15 is a schematic flowchart of determining the positional relationship between the lesion area and multiple divided areas based on the coordinate information of the lesion area according to an exemplary embodiment of the present disclosure.
- FIG. 16 is a schematic diagram of positioning the lesion area of the medical image including the lung field area based on the area division shown in FIG. 13.
- Fig. 17 is a schematic structural diagram of a network model training device provided by an exemplary embodiment of the present disclosure.
- Fig. 18 is a schematic structural diagram of a training module provided by an exemplary embodiment of the present disclosure.
- FIG. 19 is a schematic structural diagram of a training module provided by another exemplary embodiment of the present disclosure.
- Fig. 20 is a schematic structural diagram of a first training subunit provided by an exemplary embodiment of the present disclosure.
- FIG. 21 is a schematic structural diagram of a training module provided by another exemplary embodiment of the present disclosure.
- Fig. 22 is a schematic structural diagram of a training unit provided by an exemplary embodiment of the present disclosure.
- FIG. 23 is a schematic structural diagram of a first training data determining module provided by an exemplary embodiment of the present disclosure.
- FIG. 24 is a schematic structural diagram of a device for determining a lesion area provided by an exemplary embodiment of the present disclosure.
- FIG. 25 is a schematic structural diagram of an apparatus for determining a focus area provided by another exemplary embodiment of the present disclosure.
- FIG. 26 is a schematic structural diagram of a divided area generating module provided by an exemplary embodiment of the present disclosure.
- Fig. 27 is a schematic structural diagram of a position relationship determining module provided by an exemplary embodiment of the present disclosure.
- FIG. 28 is a schematic structural diagram of an electronic device provided by an exemplary embodiment of the present disclosure.
- Medical imaging is an image that uses a certain medium (such as X-rays, electromagnetic fields, ultrasound, etc.) to interact with the human body or animal body to present information such as the structure and density of the internal tissues and organs of the human body or animal body.
- a certain medium such as X-rays, electromagnetic fields, ultrasound, etc.
- medical imaging is mainly divided into anatomical structure images describing physiological forms and functional images describing human or animal body functions or metabolic functions.
- medical imaging is an important tool for disease prevention and treatment.
- the anatomical structure images describing physiological morphology mainly include X-ray images, Computed Tomography (CT) images, and Magnetic Resonance Imaging (MRI).
- CT Computed Tomography
- MRI Magnetic Resonance Imaging
- the imaging principle based on X-ray images can be divided into computer radiography (CR) and digital radiography (DR).
- CR computer radiography
- DR digital radiography
- the anatomical structure images describing the physiological morphology can clearly show the morphology and pathological conditions of the tissues and organs, which helps to determine the location information of the lesions in the tissues and organs and the type of lesions, and then provide accurate disease treatment
- the program provides prerequisites.
- tuberculosis is a chronic infectious disease caused by Mycobacterium tuberculosis, which can invade many organs, and its harmfulness is self-evident.
- tuberculosis there are six types of tuberculosis of tuberculosis including primary tuberculosis, hematologically disseminated tuberculosis, secondary tuberculosis, tracheobronchial tuberculosis, tuberculous pleurisy and old tuberculosis.
- the basic idea of the present disclosure is to propose a network model training method and device, a lesion area determination method and device, a computer-readable storage medium, and electronic equipment.
- the network model training method determines the first training data based on the sample image, then determines the initial network model, and trains the initial network model based on the sample image to generate a network model for determining the lesion area in the medical image.
- the sample image trains the initial network model for the purpose of generating a network model for determining the lesion area in the medical image. Since the sample image is a medical image that includes the lesion area, the first training data determined based on the sample image includes the labeled first lesion area coordinate information and the first lesion type information, therefore, the training generated based on the sample image is used to determine the medical image
- the network model of the lesion area can be used to assist the doctor in determining the lesion area in any medical image of the same type as the sample image.
- the embodiments of the present disclosure can determine the lesion area in the medical image of the same type as the sample image (for example, determining that the lesion area corresponds to Therefore, compared with the prior art, the embodiments of the present disclosure can effectively improve the efficiency and accuracy of determining the lesion area.
- the method for determining the focus area realizes the determination of the focus area in the medical image by inputting the medical image for determining the focus area into the network model for determining the focus area in the medical image to determine the focus area coordinate information of the medical image
- the purpose of the coordinate information Since the method for determining the lesion area provided by the embodiment of the present disclosure is implemented based on the network model used to determine the lesion area in the medical image, compared with the existing solution, the embodiment of the present disclosure does not need to determine the medical image of the lesion area. Perform complex image enhancement, filter transformation and other processing operations, thereby avoiding the failure of predicting the coordinate information of the lesion area due to factors such as image quality. That is, the method for determining the lesion area provided by the embodiment of the present disclosure has the advantages of high stability and good robustness.
- FIG. 1 is a schematic diagram of a scenario to which the embodiments of the present disclosure are applicable.
- the applicable scenario of the embodiment of the present disclosure includes a server 1 and an image acquisition device 2, where there is a communication connection relationship between the server 1 and the image acquisition device 2.
- the image acquisition device 2 is used to acquire a medical image including the lesion area as a sample image
- the server 1 is used to determine the first training data based on the sample image collected by the image acquisition device 2, and then determine the initial network model, and based on the sample
- the initial network model is trained on the image to generate a network model for determining the lesion area in the medical image, wherein the first training data includes the first lesion area coordinate information and the first lesion type information. That is, this scenario implements a network model training method.
- the image acquisition device 2 is used to acquire medical images that need to determine the lesion area
- the server 1 is used to input the medical image collected by the image acquisition device 2 into a network model for determining the lesion area in the medical image, so as to determine the medical image Coordinate information of the lesion area. That is, this scenario implements a method for determining the lesion area.
- the network model used to determine the lesion area in the medical image may be the network model generated in the above scene and used to determine the lesion area in the medical image. Since the above-mentioned scene shown in FIG. 1 uses the server 1 to implement the network model training method and/or the focus area determination method, not only can the adaptability of the scene be improved, but also the calculation amount of the image acquisition device 2 can be effectively reduced.
- FIG. 2 is a schematic diagram of another scenario to which the embodiments of the present disclosure are applicable.
- the image processing device 3 is included in the scene, and the image processing device 3 includes an image acquisition module 31 and a calculation module 32.
- the image acquisition module 31 in the image processing device 3 is used to acquire a medical image including the lesion area as a sample image
- the calculation module 32 in the image processing device 3 is used to determine the first image based on the sample image collected by the image acquisition module 31 A training data, and then determine an initial network model, and train the initial network model based on the sample image to generate a network model for determining the lesion area in the medical image, wherein the first training data includes the coordinate information of the first lesion area labeled And the first lesion type information. That is, this scenario implements a network model training method.
- the image acquisition module 31 in the image processing device 3 is used to acquire medical images that need to determine the lesion area
- the calculation module 32 in the image processing device 3 is used to input the medical images acquired by the image acquisition module 31 into the medical image for determining the medical image.
- the network model of the lesion area to determine the coordinate information of the lesion area of the medical image. That is, this scenario implements a method for determining the lesion area.
- the network model used to determine the lesion area in the medical image may be the network model generated in the above scene and used to determine the lesion area in the medical image. Since the above scenario shown in FIG.
- the 2 uses the image processing device 3 to implement the network model training method and/or the lesion area determination method, there is no need to perform data transmission operations with related devices such as servers. Therefore, the above scenario can ensure the network model training method or the lesion area.
- the real-time nature of the area determination method is the real-time nature of the area determination method.
- the image acquisition device 2 and the image acquisition module 31 mentioned in the above scenario include, but are not limited to, image acquisition devices such as X-ray machines, CT scanners, and MRI equipment.
- the medical images collected by the image acquisition device 2 and the image acquisition module 31 mentioned in the above scenes include, but are not limited to, X-ray images, CT images, MRI images, etc., which can capture the internal tissues and organs of the human or animal body, A medical image in which information such as density is presented as an image.
- the network model training method and the lesion area determination method provided by the embodiments of the present disclosure are not limited to the above-mentioned applicable scenarios of medical images. As long as the application scenarios determined based on the feature area are involved, they belong to the implementation of the present disclosure. The scope of application of the example. For example, monitoring the definite scene of the region of interest in the image.
- Fig. 3 is a schematic flowchart of a network model training method provided by an exemplary embodiment of the present disclosure. As shown in FIG. 3, the network model training method provided by the embodiment of the present disclosure includes the following steps.
- Step 10 Determine the first training data based on the sample image, where the sample image includes the lesion area, and the first training data includes the labeled first lesion area coordinate information and the first lesion type information.
- the sample image is a medical image including a lesion area.
- Step 20 Determine an initial network model, and train the initial network model based on the sample image to generate a network model for determining the lesion area in the medical image.
- the initial network model is a convolutional neural network (Convolutional Neural Networks, CNN) model.
- CNN convolutional Neural Networks
- the model structure of the initial network model and the network model used to determine the lesion area in the medical image are the same, and the difference between the initial network model and the network model used to determine the lesion area in the medical image is model Differences in network parameters. That is, the network parameters in the initial network model are the initial network parameters, and then the sample images are used to train the initial network model. During the training process, the initial network parameters are adjusted to finally generate the network model used to determine the lesion area in the medical image. Network parameters. For example, the network parameters of the initial network model are continuously adjusted based on the gradient descent method to finally generate the network parameters in the network model used to determine the lesion area in the medical image.
- the first training data is first determined based on the sample image, and then the initial network model is determined, and the initial network model is trained based on the sample image to generate a network model for determining the lesion area in the medical image.
- the network model training method determines the first training data based on the sample image, then determines the initial network model, and trains the initial network model based on the sample image to generate a network model for determining the lesion area in the medical image
- the method achieves the purpose of training an initial network model using sample images to generate a network model for determining the lesion area in a medical image.
- the sample image is a medical image that includes the lesion area
- the first training data determined based on the sample image includes the labeled first lesion area coordinate information and the first lesion type information, therefore, the training generated based on the sample image is used to determine the medical image
- the network model of the lesion area can be used to assist the doctor in determining the lesion area in any medical image of the same type as the sample image.
- using the network model training method provided by the embodiments of the present disclosure to train and generate the network model for determining the lesion area in the medical image can assist the doctor in determining the lesion area in the medical image of the same type as the sample image (such as determining the lesion area).
- the medical images of the same type as the sample image mentioned above refer to that the tissues and organs included in the medical image are of the same type as the tissues and organs in the sample image.
- the sample image is a chest radiograph image including the lung field area of the human body
- the medical image is also a chest radiograph image including the lung field area of the human body.
- the sample image is a head image including a human brain region
- the medical image is also a head image including a human brain region.
- the sample image is a lung image including a pulmonary tuberculosis lesion area
- the first lesion type information includes primary pulmonary tuberculosis, blood disseminated pulmonary tuberculosis, secondary pulmonary tuberculosis, tracheobronchial tuberculosis, At least one of tuberculous pleurisy and old tuberculosis.
- the network model for determining the lesion area in the medical image determined by the above embodiment can be used to achieve the purpose of predicting the coordinate information of the lesion area of the medical image including the pulmonary tuberculosis lesion area.
- FIG. 4 is a schematic flowchart of training an initial network model based on sample images to generate a network model for determining a lesion area in a medical image according to an exemplary embodiment of the present disclosure.
- the embodiment shown in Fig. 4 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in Fig. 4 and the embodiment shown in Fig. 3, and the similarities are not repeated here. .
- training an initial network model based on a sample image to generate a network model for determining a lesion area in a medical image includes the following steps.
- Step 21 Input the sample image into the initial network model to determine second training data corresponding to the first training data, where the second training data includes the second lesion area coordinate information and the second lesion type information.
- the second training data refers to the training data corresponding to the sample image determined by the initial network model after the sample image is input to the initial network model (wherein, the training data includes the focus area coordinate information and the focus type information).
- Step 22 Perform a first parameter adjustment operation on the initial network model based on the first training data and the second training data to generate a network model for determining the lesion area in the medical image.
- the specific network parameters adjusted in the first parameter adjustment operation can be determined according to actual conditions, including but not limited to learning rate, image size, etc.
- first determine the first training data based on the sample image then determine the initial network model, input the sample image to the initial network model to determine the second training data corresponding to the first training data, and based on the first training
- the data and the second training data perform a first parameter adjustment operation on the initial network model to generate a network model for determining the lesion area in the medical image.
- the sample image is input to the initial network model to determine the second training data corresponding to the first training data, and then the initial network model is compared based on the first training data and the second training data.
- Perform the first parameter adjustment operation to generate a network model for determining the focus area in the medical image, and realize the training of the initial network model based on the sample image to generate the network model for determining the focus area in the medical image purpose. Since the first training data is pre-labeled and the second training data is determined based on the initial network model, the difference between the first training data and the second training data can characterize the prediction accuracy of the initial network model.
- the embodiments of the present disclosure after the first parameter adjustment operation is performed on the initial network model based on the first training data and the second training data, the error range between the first training data and the second training data can be effectively reduced. Therefore, the embodiments of the present disclosure can effectively improve the prediction accuracy of the finally generated network model for determining the lesion area in the medical image.
- FIG. 5 is a schematic flowchart of training an initial network model based on sample images to generate a network model for determining a lesion area in a medical image according to another exemplary embodiment of the present disclosure.
- the embodiment shown in Fig. 5 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in Fig. 5 and the embodiment shown in Fig. 4, and the similarities will not be repeated. .
- the initial network model includes a signal-connected image feature extraction model and a prediction model.
- the image feature extraction model is used to extract image feature information of the medical image
- the prediction model is used to predict information corresponding to the medical image. Training data.
- the initial network model is subjected to the first parameter adjustment operation based on the first training data and the second training data to generate a medical image for determining
- the steps of the network model of the lesion area in include the following steps.
- Step 221 Based on the second lesion type information in the second training data and the first lesion type information in the first training data, the network parameters of the prediction model and the network parameters of the image feature extraction model are adjusted to generate the information used to determine the medical image. Network model of the lesion area.
- first determine the first training data based on the sample image then determine the initial network model, and input the sample image into the initial network model to determine the second training data corresponding to the first training data, and then based on the second training
- the second lesion type information in the data and the first lesion type information in the first training data adjust the network parameters of the prediction model and the network parameters of the image feature extraction model to generate a network model for determining the lesion area in the medical image.
- the network model training method provided by the embodiments of the present disclosure adjusts the network parameters of the prediction model and the network parameters of the image feature extraction model based on the second lesion type information in the second training data and the first lesion type information in the first training data
- the method achieves the purpose of performing the first parameter adjustment operation on the initial network model based on the first training data and the second training data. Since the lesion type information can help determine the coordinate information of the lesion area, the embodiments of the present disclosure can further improve the accuracy of the determined coordinate information of the lesion area, thereby improving the positioning accuracy of the lesion area.
- FIG. 6 shows the method of adjusting the network parameters of the prediction model and the image feature extraction model based on the second lesion type information in the second training data and the first lesion type information in the first training data provided by an exemplary embodiment of the present disclosure.
- the embodiment shown in Fig. 6 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in Fig. 6 and the embodiment shown in Fig. 5, and the similarities will not be repeated. .
- the prediction model includes a coordinate information prediction sub-model and a type information prediction sub-model.
- the coordinate information prediction sub-model is used to predict the coordinate information of the lesion area
- the type information prediction sub-model is used to predict the type information of the lesion.
- the network parameters of the prediction model are adjusted based on the second lesion type information in the second training data and the first lesion type information in the first training data.
- the network parameter step of the image feature extraction model includes the following steps.
- Step 2211 Adjust the network parameters of the type information prediction sub-model in the prediction model based on the second focus type information and the first focus type information.
- the network of the type information prediction sub-model in the prediction model can be adjusted based on the second lesion type information and the pre-labeled first lesion type information Parameters to further improve the prediction accuracy of the type information prediction sub-model.
- Step 2212 Adjust the network parameters of the image feature extraction model based on the adjusted type information prediction sub-model.
- the sub-model adjusts the network parameter operation of the image feature extraction model, which can further improve the accuracy of the image feature information extracted by the image feature extraction model.
- Step 2213 Adjust the network parameters of the coordinate information prediction sub-model in the prediction model based on the adjusted image feature extraction model.
- the coordinate information prediction sub-model in the prediction model uses the image feature information predicted by the image feature extraction model as input data, it is based on the adjusted image
- the feature extraction model adjusts the network parameter operation of the coordinate information prediction sub-model in the prediction model, which can further improve the accuracy of the coordinate information of the lesion area determined by the coordinate information prediction sub-model.
- first determine the first training data based on the sample image then determine the initial network model, and input the sample image into the initial network model to determine the second training data corresponding to the first training data, and then based on the second lesion Type information and first lesion type information adjust the network parameters of the type information prediction sub-model in the prediction model, and adjust the network parameters of the image feature extraction model based on the adjusted type information prediction sub-model, and based on the adjusted image feature extraction model
- the network parameters of the coordinate information prediction sub-model in the prediction model are adjusted to generate a network model for determining the lesion area in the medical image.
- the network model training method adjusts the network parameters of the prediction sub-model based on the type information of the second lesion type information and the first lesion type information, and then adjusts the image based on the adjusted type information prediction sub-model.
- the network parameters of the feature extraction model, and based on the adjusted image feature extraction model, the coordinate information in the prediction model is adjusted to predict the network parameters of the sub-model, so that the second lesion type information in the second training data and the first training are realized
- the purpose of adjusting the network parameters of the prediction model and the network parameters of the image feature extraction model for the first lesion type information in the data Based on the analysis content of the foregoing embodiment, it can be known that the embodiment of the present disclosure can further improve the accuracy of the determined coordinate information of the lesion area.
- FIG. 7 is a schematic flowchart of training an initial network model based on sample images to generate a network model for determining a lesion area in a medical image according to another exemplary embodiment of the present disclosure.
- the embodiment shown in Fig. 7 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in Fig. 7 and the embodiment shown in Fig. 5, and the similarities will not be repeated. .
- the initial network model is adjusted for the first time based on the first training data and the second training data, so as to generate the parameters used to determine the medical image.
- the steps of the network model of the lesion area include the following steps.
- Step 2214 Adjust network parameters of the prediction model and network parameters of the image feature extraction model based on the second lesion type information in the second training data and the first lesion type information in the first training data, and based on the first lesion type information in the second training data
- the second focus area coordinate information and the first focus area coordinate information in the first training data adjust the network parameters of the image feature extraction model to generate a network model for determining the focus area in the medical image.
- first determine the first training data based on the sample image then determine the initial network model, and input the sample image into the initial network model to determine the second training data corresponding to the first training data, and then based on the second training
- the second lesion type information in the data and the first lesion type information in the first training data adjust the network parameters of the prediction model and the network parameters of the image feature extraction model, based on the second lesion area coordinate information in the second training data and the first
- the coordinate information of the first lesion area in the training data adjusts the network parameters of the image feature extraction model to generate a network model for determining the lesion area in the medical image.
- the network model training method adjusts the network parameters of the prediction model and the network parameters of the image feature extraction model based on the second lesion type information in the second training data and the first lesion type information in the first training data , And adjust the network parameters of the image feature extraction model based on the coordinate information of the second lesion area in the second training data and the coordinate information of the first lesion area in the first training data.
- the data is the purpose of the first parameter adjustment operation on the initial network model.
- the embodiment of the present disclosure adds the step of adjusting the network parameters of the image feature extraction model based on the coordinate information of the second lesion area and the coordinate information of the first lesion area. Therefore, the embodiment of the present disclosure can further Improve the accuracy of the determined coordinate information of the lesion area, thereby improving the positioning accuracy of the lesion area.
- Fig. 8 is a schematic structural diagram of an initial network model provided by an exemplary embodiment of the present disclosure.
- the image feature extraction model includes a ResNext-50 network model 41 and a panoramic feature pyramid network model 42
- the prediction model includes a prediction network model 43.
- the type information prediction sub-model is the classification prediction module 431
- the coordinate information prediction sub-model is the coordinate prediction module 432.
- G represents the number of grouped convolutions.
- MP represents the maximum pooling layer. “ ⁇ 3”, “ ⁇ 4”, “ ⁇ 6” and “ ⁇ 3” in the ResNext-50 network model 41 indicate that the module is stacked 3 times, 4 times, 6 times, and 3 times, respectively.
- the sample image is input to the ResNext-50 network model 41 and the panoramic feature pyramid network model 42 for image feature extraction operations to output three feature layers P3, P4 and P5, and then these three feature layers P3, P4 and P5 are input to the classification prediction module 431 and the coordinate prediction module 432, respectively.
- the sizes of the three feature layers P3, P4, and P5 are batch ⁇ 256 ⁇ 64 ⁇ 64, batch ⁇ 256 ⁇ 32 ⁇ 32, batch ⁇ 256 ⁇ 16, respectively. ⁇ 16.
- batch represents the batch size, that is, the sample size used to calculate the gradient.
- the feature layers P4 and P5 are up-sampled by 2 times and 4 times, respectively, and then merged with the feature layer P3 to generate a feature map with a size of batch ⁇ 768 ⁇ 64 ⁇ 64.
- a batch ⁇ 2n matrix is obtained, where n represents the number of categories that need to be predicted, and finally the softmax classifier is used to obtain the predicted probability for each category.
- the classification prediction module 431 will use the feature layers P3, P4, and P5 to inversely affect the network parameters of the panoramic feature pyramid network model 42, and then continue to indirectly affect the performance of the ResNext-50 network model 41 Network parameters. Since the input data of the coordinate prediction module 432 is determined based on the ResNext-50 network model 41 and the panoramic feature pyramid network model 42, the classification prediction module 431 will indirectly affect the network parameters of the coordinate prediction module 432, so as to rely on the classification prediction module 431
- the lesion type information improves the prediction accuracy of the coordinate prediction module 432.
- the embodiments of the present disclosure can not only reduce over-fitting, but also further improve the accuracy of the determined coordinate information of the lesion area.
- the loss function can be used to evaluate the difference between the predicted result output by the network model and the actual result.
- the loss function is a non-negative real-valued function, and the loss value of the loss function can characterize the prediction performance of the network model, that is, the smaller the loss value of the loss function, the better the prediction performance of the network model.
- the purpose of the continuous iterative training process mentioned in the above embodiment is to make the loss value of the loss function as small as possible, so as to optimize the prediction performance of the network model. Therefore, the loss function is of great significance for improving the prediction performance of the network model.
- the loss function in the type information prediction sub-model is determined based on the following calculation formula (1).
- ⁇ represents the network parameters of the type information prediction sub-model
- m represents the number of types
- h represents the prediction probability
- y represents the label of each image.
- the loss function in the type information prediction sub-model described in the above calculation formula (1) is a cross-entropy loss function. Since the cross entropy loss function includes logarithmic function information, compared with the mean square error loss function, when the training prediction result is close to the real result, the cross entropy loss function can still maintain a high gradient state, that is, the convergence speed of the network model Will not be adversely affected.
- the loss function in the type information prediction sub-model is not limited to the loss function described in the above calculation formula (1). As long as the loss function includes logarithmic function information generated based on the predicted probability, the above embodiments can be implemented. And the beneficial effects.
- the number of samples of each type is not exactly the same, and the number of different types may be quite different.
- the number of different types may have large differences, such as the loss function recorded in the calculation formula (1), it may appear that the type with a large number has a larger proportion in the loss function, and the type with a small number is The proportion of the loss function is relatively small, which leads to a situation where the training effect of a small number of types is not as good as the training effect of a large number of types.
- the cross-entropy loss function represented by the calculation formula (1) mentioned in the above-mentioned embodiment extends another embodiment of the present disclosure.
- the loss factor corresponding to each type is correspondingly set with a weight parameter w i .
- the weight parameter w i is determined according to the proportion of the corresponding type in the entire sample data set, and the value range is between 0 and 1.
- the embodiment of the present disclosure achieves the purpose of further equalizing the training effect by setting a corresponding weight parameter for each type of corresponding loss factor, thereby further improving the prediction accuracy of the network model.
- the loss function in the coordinate information prediction sub-model is determined based on the following calculation formula (2).
- N represents the number of matched preset frames
- x represents whether the matched frame belongs to type P
- l represents the predicted frame
- g represents the real frame
- c represents the selected target belongs to type P Confidence.
- type P may be any type, which is not limited in the embodiment of the present disclosure.
- the loss function in the coordinate information prediction sub-model mentioned in the embodiment of the present disclosure can be applied to any prediction unit in the coordinate information prediction sub-model.
- the loss function described in the calculation formula (2) is applied to the category prediction unit and the coordinate prediction unit of the coordinate prediction module 432, that is, the loss function and the coordinate prediction unit in the category prediction unit
- the loss functions of are all the above calculation formula (2).
- FIG. 9 shows the first parameter adjustment operation performed on the initial network model based on the first training data and the second training data provided by an exemplary embodiment of the present disclosure to generate a network model for determining the lesion area in a medical image Schematic diagram of the process.
- the embodiment shown in Fig. 9 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in Fig. 9 and the embodiment shown in Fig. 4, and the similarities will not be repeated. .
- the initial network model is adjusted for the first time based on the first training data and the second training data, so as to generate the parameters used to determine the medical image.
- the steps of the network model of the lesion area include the following steps.
- Step 222 Perform a first parameter adjustment operation on the initial network model based on the first training data and the second training data.
- Step 223 Determine third training data corresponding to the first training data based on the sample image and the initial network model after the first parameter adjustment operation, where the third training data includes the third lesion area coordinate information and the third lesion type information.
- Step 224 Perform a second parameter adjustment operation on the initial network model after the first parameter adjustment operation based on the first training data and the third training data, so as to generate a network model for determining the lesion area in the medical image.
- first determine the first training data based on the sample image then determine the initial network model, input the sample image to the initial network model to determine the second training data corresponding to the first training data, and based on the first training
- the data and the second training data perform the first parameter adjustment operation on the initial network model, and then based on the sample image and the initial network model after the first parameter adjustment operation, the third training data corresponding to the first training data is determined, and based on The first training data and the third training data perform a second parameter adjustment operation on the initial network model after the first parameter adjustment operation is performed, so as to generate a network model for determining the lesion area in the medical image.
- the number of parameter adjustments for the initial network model is not limited to the two mentioned in the embodiment of the present disclosure, but can also be three, four or more times, until the generated value is used to determine the value of the medical image.
- the prediction accuracy of the network model of the lesion area only needs to meet the preset requirements.
- the network model training method provided by the embodiments of the present disclosure achieves the purpose of performing multiple parameter adjustment operations on the initial network model. Therefore, compared with the embodiment shown in FIG. 4, the embodiment of the present disclosure can further improve the prediction accuracy of the finally generated network model for determining the lesion area in the medical image.
- FIG. 10 is a schematic diagram of a process of determining first training data based on sample images according to an exemplary embodiment of the present disclosure.
- the embodiment shown in Fig. 10 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in Fig. 10 and the embodiment shown in Fig. 3, and the similarities will not be repeated. .
- the step of determining the first training data based on the sample image includes the following steps.
- Step 11 Determine the sample image and marking rules including the lesion area.
- the marking rule is pre-determined based on humans (such as doctors).
- the marking rule is to mark the lesion area coordinate information and the lesion type information corresponding to the lesion area in the sample image.
- Step 12 Perform a labeling operation on the sample image based on the labeling rule to generate first training data.
- the network model training method provided by the embodiment of the present disclosure realizes the determination of the first training data based on the sample image by determining the sample image including the lesion area and the marking rule, and performing the marking operation on the sample image based on the marking rule to generate the first training data.
- the purpose of the training data Since the marking rule can be determined in advance based on the actual situation of the sample image, the embodiments of the present disclosure can effectively improve the flexibility of marking, thereby improving the adaptability and wide application of the trained network model for determining the lesion area in the medical image .
- FIG. 11 is a schematic flowchart of a method for determining a lesion area provided by an exemplary embodiment of the present disclosure. As shown in FIG. 11, the method for determining a lesion area provided by an embodiment of the present disclosure includes the following steps.
- Step 50 It is determined that the medical image of the lesion area needs to be determined.
- Step 60 Input the medical image to a network model for determining the lesion area in the medical image, so as to determine the coordinate information of the lesion area of the medical image.
- the network model used to determine the lesion area in the medical image mentioned in step 60 may be obtained based on the network model training method mentioned in any of the foregoing embodiments.
- the medical image that needs to determine the lesion area is first determined, and then the medical image is input to the network model for determining the lesion area in the medical image to determine the coordinate information of the lesion area of the medical image.
- the method for determining the focus area provided by the embodiment of the present disclosure realizes the determination by inputting the medical image that needs to determine the focus area into the network model for determining the focus area in the medical image to determine the focus area coordinate information of the medical image.
- the purpose of the coordinate information of the lesion area in the medical image Since the method for determining the lesion area provided by the embodiment of the present disclosure is implemented based on the network model used to determine the lesion area in the medical image, compared with the existing solution, the embodiment of the present disclosure does not need to determine the medical image of the lesion area. Perform complex image enhancement, filter transformation and other processing operations, thereby avoiding the failure of predicting the coordinate information of the lesion area due to factors such as image quality. That is, the method for determining the lesion area provided by the embodiment of the present disclosure has the advantages of high stability and good robustness.
- FIG. 12 is a schematic flowchart of a method for determining a lesion area according to another exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 12 of the present disclosure is extended on the basis of the embodiment shown in FIG. 11 of the present disclosure.
- the following focuses on the differences between the embodiment shown in FIG. 12 and the embodiment shown in FIG. 11, and the similarities are not repeated here. .
- the method for determining the focus area provided by the embodiment of the present disclosure, after the step of inputting the medical image into the network model for determining the focus area in the medical image to determine the focus area coordinate information of the medical image, It also includes the following steps.
- Step 70 Perform a region division operation on the medical image to generate multiple divided regions.
- Step 80 Determine the positional relationship between the lesion area and the multiple divided areas based on the coordinate information of the lesion area.
- the area division operation is to generate multiple divided areas, and determine the positional relationship between the focus area and the multiple divided areas based on the focus area coordinate information.
- the method for determining the lesion area realizes the determination of the lesion by performing an area division operation on a medical image to generate multiple divided areas, and then determining the positional relationship between the lesion area and the multiple divided areas based on the coordinate information of the lesion area.
- the purpose of the positional relationship between the area coordinate information and the multiple divided areas Since the positional relationship between the coordinate information of the lesion area and the multiple divided areas can better realize the lesion area positioning operation, the embodiments of the present disclosure can further assist the subsequent disease diagnosis operation.
- FIG. 13 is a schematic diagram of region division of a medical image provided by an exemplary embodiment of the present disclosure.
- the medical image provided by the embodiment of the present disclosure is a medical image including a lung field area.
- the medical image includes key points 1 to 16, and based on the corresponding relationship between the key points 1 to 16, a plurality of area division lines are generated, and the plurality of area division lines divide the lung field area into Multiple divided areas.
- the lung field area is divided into the upper field inner zone, the upper field middle zone, the upper field outer zone, the middle field inner zone, the middle field middle zone, the middle field outer zone, the lower field inner zone, the lower field middle zone, and the lower field. Wilderness area.
- the position relationship of the lesion area in the multiple divided areas may be determined based on the determined coordinate information of the lesion area, and then a structured report is generated for the doctor's reference. For example, nodules can be seen in the upper field zone of the right lung.
- FIG. 14 is a schematic diagram of a process of performing a region division operation on a medical image to generate multiple divided regions according to an exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 14 of the present disclosure is extended on the basis of the embodiment shown in FIG. 12 of the present disclosure. The following focuses on the differences between the embodiment shown in FIG. 14 and the embodiment shown in FIG. 12, and the similarities are not repeated here. .
- performing a region division operation on a medical image to generate a plurality of division region steps includes the following steps.
- Step 71 Input the medical image to the key point network model to determine the coordinate information set of multiple key points corresponding to the medical image, where the coordinate information set is used to perform a region division operation on the medical image.
- the key point network model is a convolutional neural network (Convolutional Neural Networks, CNN) model.
- CNN convolutional Neural Networks
- Step 72 Perform a region division operation on the medical image based on the coordinate information set to generate multiple divided regions.
- first determine the medical image that needs to determine the focus area and then input the medical image into the network model used to determine the focus area in the medical image to determine the focus area coordinate information of the medical image, and then input the medical image To the key point network model to determine the coordinate information collection of multiple key points corresponding to the medical image, and perform the region division operation on the medical image based on the coordinate information collection to generate multiple divided regions, and finally determine the focus area based on the focus area coordinate information The positional relationship with multiple divided areas.
- a medical image is input to a key point network model to determine the coordinate information set of multiple key points corresponding to the medical image, and then the medical image is divided into regions based on the coordinate information set.
- the method of generating multiple divided regions achieves the purpose of performing region division operations on medical images to generate multiple divided regions. Since the region division operation mentioned in the embodiment of the present disclosure is implemented based on the key point network model, compared with the existing solution, the embodiment of the present disclosure does not need to perform complicated image enhancement and filter transformation on medical images that need to be divided into regions. And other processing operations, thereby avoiding the failure of regional division due to image quality and other factors.
- the embodiment of the present disclosure converts the area division problem into the positioning problem of the key point coordinate information, the embodiment of the present disclosure can greatly simplify the division complexity of the area division operation.
- FIG. 15 is a schematic flowchart of determining the positional relationship between the lesion area and multiple divided areas based on the coordinate information of the lesion area according to an exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 15 of the present disclosure is extended on the basis of the embodiment shown in FIG. 12 of the present disclosure. The following focuses on the differences between the embodiment shown in FIG. 15 and the embodiment shown in FIG. 12, and the similarities are not repeated here. .
- the step of determining the position relationship between the focus area and the multiple divided areas based on the focus area coordinate information includes the following steps.
- Step 81 Determine the position information of the center of gravity of the lesion area based on the coordinate information of the lesion area.
- Step 82 Determine the positional relationship between the focus area and the multiple divided areas based on the center of gravity and the positional relationship between the multiple divided areas.
- the position information that defines the center of gravity of the lesion area is the position information of the lesion area.
- the medical image In the actual application process, first determine the medical image that needs to determine the focus area, and then input the medical image into the network model used to determine the focus area in the medical image to determine the focus area coordinate information of the medical image, and then perform the medical image
- the area division operation is to generate multiple divided areas, and finally the location information of the center of gravity of the focus area is determined based on the coordinate information of the focus area, and the location relationship between the focus area and the multiple divided areas is determined based on the location relationship between the center of gravity and the multiple divided areas.
- the method for determining the focus area is realized by determining the position information of the center of gravity of the focus area, and determining the position information of the focus area and the multiple divided areas based on the position relationship between the center of gravity of the focus area and the multiple divided areas
- the purpose of determining the positional relationship between the focus area of the medical image and the multiple divided areas is achieved. Due to the ever-changing shape and volume of the lesion area, the embodiments of the present disclosure effectively ensure the positioning accuracy of the lesion area by determining the relative position information of the lesion area based on the center of gravity of the lesion area.
- FIG. 16 is a schematic diagram of positioning the lesion area of the medical image including the lung field area based on the area division shown in FIG. 13.
- the left lung field area and the right lung field area in the medical image can be framed.
- the first connection line is generated based on the first direction key points 1 and 2
- the second connection line is generated based on the first direction key points 3 and 4
- the third connection line is generated based on the second direction key points 9 and 10.
- the two-direction key points 11 and 12 generate a fourth connecting line.
- the first connecting line to the fourth connecting line can jointly form the left lung field area Contour lines.
- the connecting lines formed based on the first direction key points 5, 6, 7, and 8 and the second direction key points 13, 14, 15 and 16 can form a contour line that defines the right lung field area.
- the specific connection method can refer to the above-mentioned connection method for the contour line of the left lung field area, which is not repeated in the embodiment of the present disclosure.
- the lesion area in the medical image includes the lesion area M and the lesion area N, wherein the lesion area M has a regular boundary, and the boundary of the lesion area M is presented in the medical image It is a rectangular frame, and the center of gravity of the focus area M is m; the focus area N has an irregular boundary, the boundary of the focus area N appears as an irregular polygon box in the medical image, and the center of gravity of the focus area N is n.
- the location information of the focus area can be determined based on the location information of the center of gravity of the focus area. For example, since the center of gravity m of the lesion area M is located in the middle zone of the left lung field area, it can be determined that the lesion area M is located in the middle field zone of the left lung field area. Exemplarily, in the process of assisting the doctor in the diagnosis, it can be described as "the left lung field area with a visible lesion area M in the middle field". For another example, since the center of gravity n of the lesion area N is located in the upper field mid zone of the right lung field area, it can be determined that the lesion area N is located in the upper field mid zone of the right lung field area. Exemplarily, in the process of assisting the doctor in the diagnosis, it can be described as "the area with visible lesions in the upper field of the right lung field area N".
- Fig. 17 is a schematic structural diagram of a network model training device provided by an exemplary embodiment of the present disclosure. As shown in FIG. 17, the network model training device provided by the embodiment of the present disclosure includes:
- the first training data determining module 100 is configured to determine first training data based on a sample image, where the sample image includes a lesion area, and the first training data includes labeled first lesion area coordinate information and first lesion type information;
- the training module 200 is used to determine an initial network model, and to train the initial network model based on the sample image to generate a network model for determining the lesion area in the medical image.
- Fig. 18 is a schematic structural diagram of a training module provided by an exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 18 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in FIG. 18 and the embodiment shown in FIG. 17, and the similarities are not repeated here. .
- the training module 200 includes:
- the second training data determining unit 210 is configured to input sample images into the initial network model to determine second training data corresponding to the first training data, where the second training data includes coordinate information of the second lesion area and the second lesion Type information;
- the training unit 220 is configured to perform a first parameter adjustment operation on the initial network model based on the first training data and the second training data to generate a network model for determining the focus area in the medical image.
- FIG. 19 is a schematic structural diagram of a training module provided by another exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 19 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in FIG. 19 and the embodiment shown in FIG. 18, and the similarities are not repeated here. .
- the training unit 220 includes:
- the first training subunit 2210 is configured to adjust the network parameters of the prediction model and the network parameters of the image feature extraction model based on the second lesion type information in the second training data and the first lesion type information in the first training data to generate A network model used to determine the lesion area in a medical image.
- Fig. 20 is a schematic structural diagram of a first training subunit provided by an exemplary embodiment of the present disclosure.
- the embodiment shown in Fig. 20 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in Fig. 20 and the embodiment shown in Fig. 19, and the similarities will not be repeated. .
- the first training subunit 2210 includes:
- the first network parameter adjustment subunit 22110 is configured to adjust the network parameters of the type information prediction sub-model in the prediction model based on the second lesion type information and the first lesion type information;
- the second network parameter adjustment subunit 22120 is configured to adjust the network parameters of the image feature extraction model based on the adjusted type information prediction submodel;
- the third network parameter adjustment subunit 22130 is configured to adjust the network parameters of the coordinate information prediction submodel in the prediction model based on the adjusted image feature extraction model.
- FIG. 21 is a schematic structural diagram of a training module provided by another exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 21 of the present disclosure is extended on the basis of the embodiment shown in FIG. 19 of the present disclosure.
- the following focuses on the differences between the embodiment shown in FIG. 21 and the embodiment shown in FIG. 19, and the similarities are not repeated here. .
- the first training subunit 2210 includes:
- the second training subunit 22140 is used to adjust the network parameters of the prediction model and the network parameters of the image feature extraction model based on the second lesion type information in the second training data and the first lesion type information in the first training data, based on the first Second, the coordinate information of the second lesion area in the training data and the coordinate information of the first lesion area in the first training data adjust the network parameters of the image feature extraction model to generate a network model for determining the lesion area in the medical image.
- Fig. 22 is a schematic structural diagram of a training unit provided by an exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 22 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in FIG. 22 and the embodiment shown in FIG. 18, and the similarities are not repeated here. .
- the training unit 220 includes:
- the first parameter adjustment subunit 2220 is configured to perform the first parameter adjustment operation on the initial network model based on the first training data and the second training data;
- the third training data determining subunit 2230 is configured to determine third training data corresponding to the first training data based on the sample image and the initial network model after the first parameter adjustment operation, where the third training data includes a third lesion Regional coordinate information and third lesion type information;
- the second parameter adjustment subunit 2240 is configured to perform a second parameter adjustment operation on the initial network model after the first parameter adjustment operation is performed based on the first training data and the third training data, so as to generate a second parameter adjustment operation for determining the medical image Network model of the lesion area.
- FIG. 23 is a schematic structural diagram of a first training data determining module provided by an exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 23 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in FIG. 23 and the embodiment shown in FIG. 17, and the similarities will not be repeated. .
- the first training data determining module 100 includes:
- the determining unit 110 is configured to determine the sample image and marking rules including the lesion area
- the first training data generating unit 120 is configured to perform a labeling operation on the sample image based on the labeling rule to generate the first training data.
- FIG. 24 is a schematic structural diagram of a device for determining a lesion area provided by an exemplary embodiment of the present disclosure. As shown in FIG. 24, the device for determining a lesion area provided by an embodiment of the present disclosure includes:
- the image determining module 500 is used to determine the medical image that needs to determine the lesion area
- the lesion area determination module 600 is used to input the medical image into a network model for determining the lesion area in the medical image, so as to determine the lesion area coordinate information of the medical image.
- FIG. 25 is a schematic structural diagram of an apparatus for determining a focus area provided by another exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 25 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in FIG. 25 and the embodiment shown in FIG. 24, and the similarities are not repeated here. .
- the device for determining a lesion area further includes:
- the divided region generating module 700 is configured to perform a region dividing operation on the medical image to generate multiple divided regions;
- the position relationship determination module 800 is configured to determine the position relationship between the lesion area and multiple divided areas based on the coordinate information of the lesion area.
- FIG. 26 is a schematic structural diagram of a divided area generating module provided by an exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 26 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in FIG. 26 and the embodiment shown in FIG. 25, and the similarities will not be repeated here. .
- the divided area generating module 700 includes:
- the coordinate information set determining unit 710 is configured to input the medical image into the key point network model to determine the coordinate information set of multiple key points corresponding to the medical image, where the coordinate information set is used to perform region division operations on the medical image;
- the region dividing unit 720 is configured to perform a region dividing operation on the medical image based on the coordinate information set to generate multiple divided regions.
- Fig. 27 is a schematic structural diagram of a position relationship determining module provided by an exemplary embodiment of the present disclosure.
- the embodiment shown in FIG. 27 of the present disclosure is extended. The following focuses on the differences between the embodiment shown in FIG. 27 and the embodiment shown in FIG. 25, and the similarities will not be repeated. .
- the position relationship determining module 800 includes:
- the center of gravity determination unit 810 is configured to determine the position information of the center of gravity of the lesion area based on the coordinate information of the lesion area;
- the position relationship determining unit 820 is configured to determine the position relationship between the lesion area and the multiple divided areas based on the position relationship between the center of gravity and the multiple divided areas.
- the image determining module 500, the lesion area determining module 600, the divided area generating module 700, and the position relationship determining module 800 in the lesion area determining device provided in FIGS. 24 to 27, and the divided area generating module 700 include
- the operations and functions of the coordinate information set determining unit 710 and the area dividing unit 720, as well as the center of gravity determining unit 810 and the position relationship determining unit 820 included in the position relationship determining module 800 can refer to the lesion area determination methods provided in FIGS. 11 to 15. In order to avoid repetition, I won't repeat it here.
- FIG. 28 is a schematic structural diagram of an electronic device provided by an exemplary embodiment of the present disclosure.
- the electronic device 90 includes one or more processors 901 and a memory 902.
- the processor 901 may be a central processing unit (CPU) or another form of processing unit with data processing capability and/or instruction execution capability, and may control other components in the electronic device 90 to perform desired functions.
- CPU central processing unit
- the processor 901 may be a central processing unit (CPU) or another form of processing unit with data processing capability and/or instruction execution capability, and may control other components in the electronic device 90 to perform desired functions.
- the memory 902 may include one or more computer program products, and the computer program products may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory.
- the volatile memory may include random access memory (RAM) and/or cache memory (cache), for example.
- the non-volatile memory may include, for example, read-only memory (ROM), hard disk, flash memory, and the like.
- One or more computer program instructions may be stored on the computer-readable storage medium, and the processor 11 may run the program instructions to implement the network model training method and focus area of the various embodiments of the application described above. Determine the method and/or other desired functions.
- Various contents such as medical images can also be stored in the computer-readable storage medium.
- the electronic device 90 may further include: an input device 903 and an output device 904, and these components are interconnected by a bus system and/or other forms of connection mechanisms (not shown).
- the input device 903 may include, for example, a keyboard, a mouse, and so on.
- the output device 904 can output various information to the outside, including information on the determined lesion area and so on.
- the output device 904 may include, for example, a display, a speaker, a printer, a communication network and a remote output device connected to it, and so on.
- the electronic device 90 may also include any other appropriate components.
- the embodiments of the present application may also be computer program products, which include computer program instructions that, when run by a processor, cause the processor to execute the above-mentioned description of the present specification according to the present application. Steps in the network model training method and the method for determining the lesion area of various embodiments.
- the computer program product may use any combination of one or more programming languages to write program codes for performing the operations of the embodiments of the present application.
- the programming languages include object-oriented programming languages, such as Java, C++, etc. , Also includes conventional procedural programming languages, such as "C" language or similar programming languages.
- the program code can be executed entirely on the user's computing device, partly on the user's device, executed as an independent software package, partly on the user's computing device and partly executed on the remote computing device, or entirely on the remote computing device or server Executed on.
- the embodiments of the present application may also be a computer-readable storage medium, on which computer program instructions are stored.
- the processor executes the above-mentioned descriptions of the present specification according to the various aspects of the present application.
- the computer-readable storage medium may adopt any combination of one or more readable media.
- the readable medium may be a readable signal medium or a readable storage medium.
- the readable storage medium may include, but is not limited to, an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or a combination of any of the above, for example. More specific examples (non-exhaustive list) of readable storage media include: electrical connections with one or more wires, portable disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable Type programmable read only memory (EPROM or flash memory), optical fiber, portable compact disk read only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
- each component or each step can be decomposed and/or recombined.
- decompositions and/or recombinations shall be regarded as equivalent solutions of this application.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- General Physics & Mathematics (AREA)
- Medical Informatics (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Public Health (AREA)
- Evolutionary Computation (AREA)
- Pathology (AREA)
- Databases & Information Systems (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
Claims (19)
- 一种网络模型训练方法,包括:基于样本图像确定第一训练数据,其中,所述样本图像包括病灶区域,所述第一训练数据包括标记的第一病灶区域坐标信息和第一病灶类型信息;确定初始网络模型,并基于所述样本图像训练所述初始网络模型,以生成用于确定医学图像中的病灶区域的网络模型。
- 根据权利要求1所述的方法,其中,所述基于所述样本图像训练所述初始网络模型,以生成用于确定医学图像中的病灶区域的网络模型,包括:将所述样本图像输入至所述初始网络模型,以确定与所述第一训练数据对应的第二训练数据,其中,所述第二训练数据包括第二病灶区域坐标信息和第二病灶类型信息;基于所述第一训练数据和所述第二训练数据对所述初始网络模型进行第一次参数调整操作,以生成用于确定医学图像中的病灶区域的网络模型。
- 根据权利要求2所述的方法,其中,所述初始网络模型包括信号连接的图像特征提取模型和预测模型,所述基于所述第一训练数据和所述第二训练数据对所述初始网络模型进行第一次参数调整操作,包括:基于所述第二训练数据中的所述第二病灶类型信息与所述第一训练数据中的所述第一病灶类型信息调整所述预测模型的网络参数和所述图像特征提取模型的网络参数。
- 根据权利要求3所述的方法,其中,所述预测模型包括坐标信息预测子模型和类型信息预测子模型,所述基于所述第二训练数据中的所述第二病灶类型信息与所述第一训练数据中的所述第一病灶类型信息调整所述预测模型的网络参数和所述图像特征提取模型的网络参数,包括:基于所述第二病灶类型信息和所述第一病灶类型信息调整所述预测模型中的所述类型信息预测子模型的网络参数;基于调整后的所述类型信息预测子模型调整所述图像特征提取模型的网络参数;基于调整后的所述图像特征提取模型调整所述预测模型中的所述坐标信息预测子模型的网络参数。
- 根据权利要求4所述的方法,其中,所述类型信息预测子模型中的损失函数包括基于预测概率生成的对数函数。
- 根据权利要求4所述的方法,其中,所述类型信息预测子模型中的损失函数和/或所述坐标信息预测子模型中的损失函数为交叉熵损失函数。
- 根据权利要求3所述的方法,其中,所述基于所述第一训练数据和所述第二训练数据对所述初始网络模型进行第一次参数调整操作,还包括:基于所述第二训练数据中的所述第二病灶区域坐标信息和所述第一训练数据中的所述第一病灶区域坐标信息调整所述图像特征提取模型的网络参数。
- 根据权利要求3至7任一所述的方法,其中,所述图像特征提取模型包括ResNext-50网络模型和全景特征金字塔网络模型。
- 根据权利要求2所述的方法,其中,所述基于所述第一训练数据和所述第二训练数据对所述初始网络模型进行第一次参数调整操作,以生成用于确定医学图像中的病灶区域的网络模型,包括:基于所述第一训练数据和所述第二训练数据对所述初始网络模型进行第一次参数调整操作;基于所述样本图像和进行所述第一次参数调整操作后的初始网络模型确定与所述第一训练数据对应的第三训练数据,其中,所述第三训练数据包括第三病灶区域坐标信息和第三病灶类型信息;基于所述第一训练数据和所述第三训练数据对进行所述第一次参数调整操作后的初始网络模型进行第二次参数调整操作,以生成用于确定医学图像中的病灶区域的网络模型。
- 根据权利要求1至7任一所述的方法,其中,所述基于样本图像确定第一训练数据,包括:确定包括病灶区域的样本图像和标记规则;基于所述标记规则对所述样本图像进行标记操作,以生成第一训练数据。
- 根据权利要求1至7任一所述的方法,其中,所述样本图像为包括肺结核病灶区域的肺部图像,所述第一病灶类型信息包括原发性肺结核、血行播散型肺结核、继发性肺结核、气管支气管肺结核、结核性胸膜炎、陈旧性肺结核中的至少一种。
- 一种病灶区域确定方法,包括:确定需要确定病灶区域的医学图像;将所述医学图像输入至用于确定医学图像中的病灶区域的网络模型,以确定所述医学图像的病灶区域坐标信息,其中,所述网络模型基于上述权利要求1至11任一所述的网络模型训练方法获得。
- 根据权利要求12所述的方法,其中,在所述将所述医学图像输入至用于确定医学图像中的病灶区域的网络模型,以确定所述医学图像的病灶区域坐标信息后,还包括:对所述医学图像进行区域划分操作,以生成多个划分区域;基于所述病灶区域坐标信息确定所述病灶区域与所述多个划分区域的位置关系。
- 根据权利要求13所述的方法,其中,所述对所述医学图像进行区域划分操作,以生成多个划分区域,包括:将所述医学图像输入至关键点网络模型,以确定所述医学图像对应的多个关键点的坐标信息集合,其中,所述坐标信息集合用于对所述医学图像进行区域划分操作;基于所述坐标信息集合对所述医学图像进行区域划分操作,以生成所述多个划分区域。
- 根据权利要求13或14所述的方法,其中,所述基于所述病灶区域坐标信息确定所述病灶区域与所述多个划分区域的位置关系,包括:基于所述病灶区域坐标信息确定所述病灶区域的重心的位置信息;基于所述重心和所述多个划分区域的位置关系确定所述病灶区域与所述多个划分区域的位置关系。
- 一种网络模型训练装置,包括:第一训练数据确定模块,用于基于样本图像确定第一训练数据,其中,所述样本图像包括病灶区域,所述第一训练数据包括标记的第一病灶区域坐标信息和第一病灶类型信息;训练模块,用于确定初始网络模型,并基于所述样本图像训练所述初始网络模型,以生成用于确定医学图像中的病灶区域的网络模型。
- 一种病灶区域确定装置,包括:图像确定模块,用于确定需要确定病灶区域的医学图像;病灶区域确定模块,用于将所述医学图像输入至用于确定医学图像中的病灶区域的网络模型,以确定所述医学图像的病灶区域坐标信息,其中,所述网络模型基于上述权利要求1至11任一所述的网络模型训练方法获得。
- 一种计算机可读存储介质,所述存储介质存储有计算机程序,所述计算机程序用于执行上述权利要求1至11任一所述的网络模型训练方法,或者执行上述权利要求12至15任一所述的病灶区域确定方法。
- 一种电子设备,所述电子设备包括:处理器;用于存储所述处理器可执行指令的存储器;所述处理器,用于执行上述权利要求1至11任一所述的网络模型训练方法,或者执行上述权利要求12至15任一所述的病灶区域确定方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201911049680.5A CN110827294A (zh) | 2019-10-31 | 2019-10-31 | 网络模型训练方法及装置、病灶区域确定方法及装置 |
CN201911049680.5 | 2019-10-31 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021082416A1 true WO2021082416A1 (zh) | 2021-05-06 |
Family
ID=69551516
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/092570 WO2021082416A1 (zh) | 2019-10-31 | 2020-05-27 | 网络模型训练方法及装置、病灶区域确定方法及装置 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110827294A (zh) |
WO (1) | WO2021082416A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113782221A (zh) * | 2021-09-16 | 2021-12-10 | 平安科技(深圳)有限公司 | 基于自训练学习的疾病预测装置、设备及存储介质 |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110827294A (zh) * | 2019-10-31 | 2020-02-21 | 北京推想科技有限公司 | 网络模型训练方法及装置、病灶区域确定方法及装置 |
CN111353975A (zh) * | 2020-02-24 | 2020-06-30 | 北京推想科技有限公司 | 网络模型训练方法及装置、病灶定位方法及装置 |
CN111383328B (zh) * | 2020-02-27 | 2022-05-20 | 西安交通大学 | 一种面向乳腺癌病灶的3d可视化方法及系统 |
CN111325739B (zh) * | 2020-02-28 | 2020-12-29 | 推想医疗科技股份有限公司 | 肺部病灶检测的方法及装置,和图像检测模型的训练方法 |
CN111445456B (zh) * | 2020-03-26 | 2023-06-27 | 推想医疗科技股份有限公司 | 分类模型、网络模型的训练方法及装置、识别方法及装置 |
CN111899848B (zh) * | 2020-08-05 | 2023-07-07 | 中国联合网络通信集团有限公司 | 图像识别方法及设备 |
TWI777319B (zh) * | 2020-12-03 | 2022-09-11 | 鴻海精密工業股份有限公司 | 幹細胞密度確定方法、裝置、電腦裝置及儲存介質 |
CN112489794A (zh) * | 2020-12-18 | 2021-03-12 | 推想医疗科技股份有限公司 | 一种模型的训练方法、装置、电子终端及存储介质 |
CN116310627B (zh) * | 2023-01-16 | 2024-02-02 | 浙江医准智能科技有限公司 | 模型训练方法、轮廓预测方法、装置、电子设备及介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107563123A (zh) * | 2017-09-27 | 2018-01-09 | 百度在线网络技术(北京)有限公司 | 用于标注医学图像的方法和装置 |
CN108615237A (zh) * | 2018-05-08 | 2018-10-02 | 上海商汤智能科技有限公司 | 一种肺部图像处理方法及图像处理设备 |
CN110276411A (zh) * | 2019-06-28 | 2019-09-24 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、设备、存储介质和医疗电子设备 |
CN110827294A (zh) * | 2019-10-31 | 2020-02-21 | 北京推想科技有限公司 | 网络模型训练方法及装置、病灶区域确定方法及装置 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9218524B2 (en) * | 2012-12-06 | 2015-12-22 | Siemens Product Lifecycle Management Software Inc. | Automatic spatial context based multi-object segmentation in 3D images |
CN110363768B (zh) * | 2019-08-30 | 2021-08-17 | 重庆大学附属肿瘤医院 | 一种基于深度学习的早期癌病灶范围预测辅助系统 |
-
2019
- 2019-10-31 CN CN201911049680.5A patent/CN110827294A/zh active Pending
-
2020
- 2020-05-27 WO PCT/CN2020/092570 patent/WO2021082416A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107563123A (zh) * | 2017-09-27 | 2018-01-09 | 百度在线网络技术(北京)有限公司 | 用于标注医学图像的方法和装置 |
CN108615237A (zh) * | 2018-05-08 | 2018-10-02 | 上海商汤智能科技有限公司 | 一种肺部图像处理方法及图像处理设备 |
CN110276411A (zh) * | 2019-06-28 | 2019-09-24 | 腾讯科技(深圳)有限公司 | 图像分类方法、装置、设备、存储介质和医疗电子设备 |
CN110827294A (zh) * | 2019-10-31 | 2020-02-21 | 北京推想科技有限公司 | 网络模型训练方法及装置、病灶区域确定方法及装置 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113782221A (zh) * | 2021-09-16 | 2021-12-10 | 平安科技(深圳)有限公司 | 基于自训练学习的疾病预测装置、设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN110827294A (zh) | 2020-02-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021082416A1 (zh) | 网络模型训练方法及装置、病灶区域确定方法及装置 | |
Zhang et al. | Joint craniomaxillofacial bone segmentation and landmark digitization by context-guided fully convolutional networks | |
CN110766701B (zh) | 网络模型训练方法及装置、区域划分方法及装置 | |
US10949970B2 (en) | Methods and apparatus for the application of machine learning to radiographic images of animals | |
CN110992376A (zh) | 基于ct图像的肋骨分割方法、装置、介质及电子设备 | |
US20220301154A1 (en) | Medical image analysis using navigation processing | |
WO2021151302A1 (zh) | 基于机器学习的药品质控分析方法、装置、设备及介质 | |
US11475568B2 (en) | Method for controlling display of abnormality in chest x-ray image, storage medium, abnormality display control apparatus, and server apparatus | |
CN111524109A (zh) | 头部医学影像的评分方法和装置、电子设备及存储介质 | |
CN111476772B (zh) | 基于医学影像的病灶分析方法和装置 | |
CN111340209A (zh) | 网络模型训练方法、图像分割方法、病灶定位方法 | |
Monsi et al. | XRAY AI: Lung Disease Prediction using machine learning | |
JP2023175011A (ja) | 文書作成支援装置、方法およびプログラム | |
Kara et al. | Identification and localization of endotracheal tube on chest radiographs using a cascaded convolutional neural network approach | |
US20230377149A1 (en) | Learning apparatus, learning method, trained model, and program | |
CN115053296A (zh) | 使用机器学习的改进的手术报告生成方法及其设备 | |
JP7007469B2 (ja) | 医療文書作成支援装置、方法およびプログラム、学習済みモデル、並びに学習装置、方法およびプログラム | |
Ibrahim et al. | Lung Segmentation Using ResUnet++ Powered by Variational Auto Encoder-Based Enhancement in Chest X-ray Images | |
Hsu et al. | Development of a deep learning model for chest X-ray screening | |
KR20240048294A (ko) | 인공지능을 활용한 의료용 증강현실영상 제공 방법, 장치 및 시스템 | |
JP2024054748A (ja) | 言語特徴抽出モデルの生成方法、情報処理装置、情報処理方法及びプログラム | |
EP4309121A1 (en) | Detecting abnormalities in an x-ray image | |
JP2023020145A (ja) | 解析装置、解析方法及びプログラム | |
DE102021201912A1 (de) | Verfahren zur Bereitstellung eines Metadaten-Attributs, das mit einem medizinischen Bild assoziiert ist | |
KR102136107B1 (ko) | 뼈 감쇄된 x-선 영상의 정합 장치 및 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20881390 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20881390 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 05/09/2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20881390 Country of ref document: EP Kind code of ref document: A1 |