WO2021032062A1 - 图像处理模型生成方法、图像处理方法、装置及电子设备 - Google Patents
图像处理模型生成方法、图像处理方法、装置及电子设备 Download PDFInfo
- Publication number
- WO2021032062A1 WO2021032062A1 PCT/CN2020/109611 CN2020109611W WO2021032062A1 WO 2021032062 A1 WO2021032062 A1 WO 2021032062A1 CN 2020109611 W CN2020109611 W CN 2020109611W WO 2021032062 A1 WO2021032062 A1 WO 2021032062A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- layer
- image
- lesion
- image processing
- training sample
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 247
- 238000000034 method Methods 0.000 title claims abstract description 92
- 238000003672 processing method Methods 0.000 title claims description 24
- 230000003902 lesion Effects 0.000 claims abstract description 354
- 238000012549 training Methods 0.000 claims abstract description 187
- 201000010099 disease Diseases 0.000 claims abstract description 108
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 claims abstract description 108
- 238000011176 pooling Methods 0.000 claims description 99
- 238000013528 artificial neural network Methods 0.000 claims description 54
- 238000005070 sampling Methods 0.000 claims description 35
- 238000004590 computer program Methods 0.000 claims description 31
- 230000008569 process Effects 0.000 claims description 30
- 230000006870 function Effects 0.000 claims description 26
- 230000004913 activation Effects 0.000 claims description 22
- 238000002372 labelling Methods 0.000 claims description 19
- 230000015654 memory Effects 0.000 claims description 17
- 238000000605 extraction Methods 0.000 claims description 12
- 238000010586 diagram Methods 0.000 description 9
- 238000001514 detection method Methods 0.000 description 5
- 230000009471 action Effects 0.000 description 4
- 238000013527 convolutional neural network Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 206010035664 Pneumonia Diseases 0.000 description 3
- 208000019622 heart disease Diseases 0.000 description 3
- 102200048773 rs2224391 Human genes 0.000 description 3
- 230000006403 short-term memory Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000007635 classification algorithm Methods 0.000 description 2
- 238000003745 diagnosis Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000000750 progressive effect Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 238000012706 support-vector machine Methods 0.000 description 2
- 206010016654 Fibrosis Diseases 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 230000007882 cirrhosis Effects 0.000 description 1
- 208000019425 cirrhosis of liver Diseases 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 125000004122 cyclic group Chemical group 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 210000005036 nerve Anatomy 0.000 description 1
- 238000003909 pattern recognition Methods 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/20—ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/70—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for mining of medical data, e.g. analysing previous cases of other patients
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30096—Tumor; Lesion
Definitions
- the present disclosure relates to the field of computer technology, in particular to an image processing model generation method, an image processing method, an image processing model generation device, an image processing device, a computer-readable storage medium, and an electronic device .
- an image processing model generation method includes: inputting at least one training sample focus image of a known disease type into an initial image processing model, wherein the initial image processing model includes a classification layer and an annotation layer, and the training The sample lesion image includes the initial center point coordinates, initial length, and initial width corresponding to the lesion area; the classification layer is called to classify the training sample lesion image to obtain the training sample lesion image corresponding to each known disease category Classification probability; calling the labeling layer to process the training sample lesion image to obtain the predicted center point coordinates, predicted length, and predicted width of the lesion area contained in the training sample lesion image; according to the at least one training The classification probability, the initial center point coordinates, the initial length, the initial width, the predicted center point coordinates, the predicted length, and the predicted width corresponding to the sample lesion image, and the at least one training The loss value of the sample lesion image in the initial image processing model; determine whether the loss value is within the preset range, and if the loss value is not within the prese
- the invoking the classification layer to perform classification processing on the training sample lesion image to obtain the classification probability of the training sample lesion image corresponding to each known disease category includes: At least one set of one-dimensional feature maps of the image is input to the classification layer; the classification layer is called, the one-dimensional feature maps are classified, and the classification probabilities of the training sample lesion images corresponding to each known disease category are output.
- the calling the annotation layer to process the training sample lesion image to obtain the predicted center point coordinates, predicted length, and predicted width of the lesion area contained in the training sample lesion image includes : Input the processed groups of two-dimensional feature maps of the training sample lesion image into the annotation layer; call the annotation layer to determine the largest feature point in the two-dimensional feature map, and the largest feature point corresponding The maximum feature value and two-dimensional coordinates of the; according to the maximum feature point, the maximum feature value, the two-dimensional coordinates and a preset feature threshold, determine the prediction center corresponding to the focus area contained in the training sample focus image Point coordinates, predicted length and predicted width.
- the two-dimensional coordinates include a first coordinate value in the horizontal axis direction and a second coordinate value in the vertical axis direction.
- the width includes: determining the two-dimensional coordinates as the predicted center point coordinates; calculating the absolute value of the characteristic difference between the maximum characteristic value and the preset characteristic threshold; according to the absolute value of the characteristic difference And the two-dimensional coordinates of the largest feature point, acquiring the first feature point and the second feature point in the horizontal axis direction in the two-dimensional feature map, and the third feature point in the vertical axis direction And a fourth feature point; obtaining a first coordinate value of the first feature point in the horizontal axis direction, and a second coordinate value of the second feature point in the horizontal axis direction; obtaining the first feature point The third coordinate value of the three characteristic points in the longitudinal axis direction, and the fourth coordinate value of the fourth characteristic point in the longitudinal axis direction.
- the predicted width is calculated; based on the third coordinate value and the fourth coordinate value, the predicted length is calculated.
- the initial image processing model further includes: a global average pooling layer.
- the method further includes: inputting multiple sets of two-dimensional feature maps of the training sample lesion image into the global average pooling layer; invoking the global average The pooling layer performs global average pooling on the multiple sets of two-dimensional feature maps to obtain multiple sets of one-dimensional feature maps corresponding to the multiple sets of two-dimensional feature maps.
- the initial image processing model further includes: a fully connected layer. After the invoking the global average pooling layer, and before the invoking the classification layer, the method further includes: invoking the fully connected layer, and comparing the multiple groups of one obtained after the global average pooling. Perform feature extraction on the one-dimensional feature map to obtain at least one set of one-dimensional feature maps of the training sample lesion image, and input the at least one set of one-dimensional feature maps into the classification layer.
- the initial image processing model further includes: a feature weighted summation layer and an upsampling layer.
- the method further includes: converting multiple sets of two-dimensional feature maps of the training sample lesion image, and after the global average pooling The multiple sets of one-dimensional feature maps obtained later are input to the feature weighted summation layer; the feature weighted summation layer is called, and the multiple sets of one-dimensional feature maps are compared to the multiple sets of two-dimensional feature maps.
- the initial image processing model further includes a neural network.
- the neural network includes at least one layer, and each layer includes a convolutional layer, an activation function, and a pooling layer in turn.
- the method further includes: inputting the training sample lesion image into the neural network; calling the neural network, The training sample lesion image sequentially passes through the convolutional layer, activation function, and pooling layer of each layer of the neural network to obtain multiple sets of two-dimensional feature maps of the training sample lesion image, and the multiple sets of two
- the dimensional feature map is input to the global average pooling layer; in the case that the initial image processing model further includes a feature weighted summation layer, the multiple sets of two-dimensional feature maps are also input to the feature weighted summation layer.
- the classification probability, the initial center point coordinates, the initial length, the initial width, the predicted center point coordinates, the The prediction length and the prediction width, and obtaining the loss value of the at least one training sample lesion image in the initial image processing model includes: calculating the classification loss value according to the classification probability corresponding to the training sample lesion image Calculate the position loss value according to the initial center point coordinates, the initial length, the initial width, the predicted center point coordinates, the predicted length, and the predicted width corresponding to the training sample lesion image; According to the classification loss value and the location loss value, a loss value of the training sample lesion image in the initial image processing model is obtained.
- an image processing method including: inputting an image of a lesion to be processed into a target image processing model, the target image processing model being trained by the methods described in some of the above embodiments, the target image processing model including classification Layer and labeling layer; call the classification layer to classify the image of the lesion to be processed to obtain the disease classification probability corresponding to the image of the lesion to be processed; call the label layer to process the image of the lesion to be processed , Obtain the regional center point coordinates, region length, and region width of the lesion area included in the to-be-processed lesion image; determine the type of disease corresponding to the to-be-processed lesion image according to the disease classification probability; according to the regional center point The coordinates, the length of the area and the width of the area determine the lesion marking area in the image of the lesion to be processed.
- the invoking the classification layer to classify the image of the to-be-processed lesion to obtain the disease classification probability corresponding to the image of the to-be-processed lesion includes: comparing at least one of the image of the to-be-processed lesion Group one-dimensional feature maps into the classification layer; call the classification layer, perform classification processing on the one-dimensional feature maps, and output disease classification probabilities corresponding to the to-be-processed lesion image.
- the call the annotation layer to process the image of the lesion to be processed to obtain the coordinates of the center point of the area, the length of the area, and the width of the area corresponding to the lesion area contained in the image of the lesion to be processed.
- the method includes: inputting the processed multiple sets of two-dimensional feature maps of the to-be-processed lesion image into the annotation layer; calling the annotation layer to determine the largest feature point in the two-dimensional feature map, and the largest feature point Corresponding maximum feature value and two-dimensional coordinates; according to the maximum feature point, the maximum feature value, the two-dimensional coordinates and a preset feature threshold, determine the area corresponding to the lesion area contained in the to-be-processed lesion image Center point coordinates, area length and area width.
- the two-dimensional coordinates include a first coordinate value in the horizontal axis direction and a second coordinate value in the vertical axis direction. Said determining, according to the maximum feature point, the maximum feature value, the two-dimensional coordinates and a preset feature threshold, the coordinates of the center point of the area, the length of the area, and the width of the area of the lesion area included in the image of the to-be-processed lesion , Including: determining the two-dimensional coordinates as the coordinates of the center point of the area; calculating the absolute value of the characteristic difference between the maximum characteristic value and the preset characteristic threshold; and according to the sum of the absolute value of the characteristic difference.
- an image processing model generation device including: a sample image input component, a classification probability acquisition component, a prediction data acquisition component, a loss value acquisition component, and a target model generation component.
- the sample image input component is configured to input at least one training sample lesion image of a known disease type into an initial image processing model; wherein, the initial image processing model includes a classification layer and an annotation layer, and the training sample lesion image Including the initial center point coordinates, initial length and initial width corresponding to the lesion area.
- the classification probability acquisition component is configured to call the classification layer, perform classification processing on the training sample lesion image, and obtain the classification probability of the training sample lesion image corresponding to each known disease category.
- the prediction data acquisition component is configured to call the annotation layer to process the training sample lesion image to obtain the predicted center point coordinates, predicted length, and predicted width of the lesion area included in the training sample lesion image.
- the loss value acquiring component is configured to, according to the classification probability, the initial center point coordinates, the initial length, the initial width, the predicted center point coordinates, and the corresponding to the at least one training sample lesion image
- the prediction length and the prediction width are obtained, and the loss value of the at least one training sample lesion image in the initial image processing model is acquired.
- the target model generating component is configured to determine whether the loss value is within a preset range, and if the loss value is not within the preset range, update the parameters of the initial image processing model, and update the parameterized image
- the processing model is used as the initial image processing model for the next training; continue to train the image processing model with updated parameters until the loss value is within the preset range, and the image processing model trained last time is used as the target image processing model .
- the initial image processing model further includes a neural network, a global average pooling layer, a fully connected layer, a feature weighted summation layer, and an upsampling layer; wherein the neural network includes at least one layer, each The layers include convolutional layer, activation function and pooling layer in turn.
- the image processing model generation device further includes: a first two-dimensional feature map acquisition component coupled to the sample image input component, configured to invoke the neural network, and sequentially pass the training sample lesion image through the nerve
- the convolutional layer, activation function and pooling layer of each layer of the network obtain multiple sets of two-dimensional feature maps of the training sample lesion image.
- the image processing model generation device further includes: a first global average pooling component coupled to the first two-dimensional feature map acquisition component, configured to call the global average pooling layer, and perform a comparison of the two groups of two Performing global average pooling on the dimensional feature map to obtain multiple sets of one-dimensional feature maps corresponding to the multiple sets of two-dimensional feature maps;
- the image processing model generating device further includes: a first fully connected component coupled to the first global average pooling component, the first fully connected component is also coupled to the classification probability acquisition component, and the second A fully connected component is configured to call the fully connected layer, perform feature extraction on the multiple sets of one-dimensional feature maps to obtain at least one set of one-dimensional feature maps of the training sample lesion image, and combine the at least one set of The one-dimensional feature map is input to the classification probability acquisition component.
- the image processing model generating device further includes: a first feature weighted summation component coupled to the first two-dimensional feature map acquisition component and the first global average pooling component, and is configured to invoke the feature weighting
- the summation layer performs feature weighted summation on each group of two-dimensional feature maps in the multiple sets of two-dimensional feature maps according to the multiple sets of one-dimensional feature maps.
- the image processing model generation device further includes: a first up-sampling component coupled to the first feature weighted summation component, the first up-sampling component is also coupled to the predicted data acquisition component, the first An up-sampling component is configured to call the up-sampling layer, perform up-sampling processing on the multiple sets of two-dimensional feature maps that have undergone the feature weighted summation, and input the multiple sets of two-dimensional feature maps that have undergone the up-sampling processing The prediction data acquisition component.
- an image processing device which includes: an image input component to be processed, a disease classification probability acquisition component, a focus area data acquisition component, a disease type determination component, and a focus area determination component.
- the to-be-processed image input component is configured to input the to-be-processed lesion image into a target image processing model; wherein, the target image processing model is obtained by training the methods described in some of the above examples, and the target image processing model includes classification Layer and label layer.
- the disease classification probability acquisition component is configured to call the classification layer to perform classification processing on the to-be-processed lesion image to obtain the disease classification probability corresponding to the to-be-processed lesion image.
- the lesion area data acquisition component is configured to call the annotation layer to process the to-be-processed lesion image to obtain the area center point coordinates, the area length and the area width of the lesion area contained in the to-be-processed lesion image.
- the disease type determining component is configured to determine the disease type corresponding to the to-be-processed focus image according to the disease classification probability.
- the lesion area determining component is configured to determine the lesion marking area in the image of the to-be-processed lesion according to the coordinates of the center point of the area, the length of the area, and the width of the area.
- the target image processing model further includes a neural network, a global average pooling layer, a fully connected layer, a feature-weighted summation layer, and an upsampling layer; wherein the neural network includes at least one layer, each The layers include convolutional layer, activation function and pooling layer in turn.
- the image processing device further includes: a second two-dimensional feature map acquiring component coupled to the to-be-processed image input component, configured to call the neural network, and sequentially pass the to-be-processed lesion image through the neural network
- the convolutional layer, activation function and pooling layer of each layer of obtain multiple sets of two-dimensional feature maps of the image of the to-be-processed lesion.
- the image processing device further includes: a second global average pooling component coupled to the second two-dimensional feature map acquiring component, configured to call the global average pooling layer, and compare the multiple sets of two-dimensional features
- the graph is globally averaged pooled to obtain multiple sets of one-dimensional feature maps corresponding to the multiple sets of two-dimensional feature maps.
- the image processing device further includes: a second fully connected component coupled to the second global average pooling component, the second fully connected component is also coupled to the disease classification probability acquisition component, the second The fully connected component is configured to call the fully connected layer, perform feature extraction on the multiple sets of one-dimensional feature maps, obtain at least one set of one-dimensional feature maps of the to-be-processed lesion image, and combine the at least one set of one-dimensional feature maps.
- the dimensional feature map is input to the disease classification probability acquisition component.
- the image processing device further includes: a second feature weighted summation unit coupled to the second two-dimensional feature map acquisition unit and the second global average pooling unit, and is configured to call the feature weighted summation Layer, according to the multiple sets of one-dimensional feature maps, perform feature weighted summation on each set of two-dimensional feature maps in the multiple sets of two-dimensional feature maps.
- a second feature weighted summation unit coupled to the second two-dimensional feature map acquisition unit and the second global average pooling unit, and is configured to call the feature weighted summation Layer, according to the multiple sets of one-dimensional feature maps, perform feature weighted summation on each set of two-dimensional feature maps in the multiple sets of two-dimensional feature maps.
- the image processing device further includes a second up-sampling component coupled to the second feature weighted sum component, the second up-sampling component is also coupled to the lesion area data acquisition component, and the second The up-sampling component is configured to call the up-sampling layer, perform up-sampling processing on the multiple sets of two-dimensional feature maps that have undergone the feature weighted summation, and input the multiple sets of two-dimensional feature maps that have undergone the up-sampling processing into the office.
- the data acquisition component of the lesion area is configured to call the up-sampling layer, perform up-sampling processing on the multiple sets of two-dimensional feature maps that have undergone the feature weighted summation, and input the multiple sets of two-dimensional feature maps that have undergone the up-sampling processing into the office.
- the data acquisition component of the lesion area is configured to call the up-sampling layer, perform up-sampling processing on the multiple sets of two-dimensional feature maps that have undergone the feature weighted summ
- a computer-readable storage medium stores computer program instructions.
- the processor executes some of the above-mentioned embodiments.
- a computer program product in another aspect, includes computer program instructions, and when the computer program instructions are executed on a computer, the computer program instructions cause the computer to execute one or more steps in the image processing model generation method described in some of the above embodiments, And/or, one or more steps in the image processing method described in some of the above embodiments.
- a computer program When the computer program is executed on a computer, the computer program causes the computer to execute one or more steps in the image processing model generation method described in some of the above embodiments, and/or, as described in some of the above embodiments One or more steps in the image processing method.
- an electronic device which includes a processor, a memory, and a computer program stored on the memory and capable of being run on the processor.
- the processor executes the computer program, it implements one or more steps in the image processing model generation method described in some of the above embodiments, and/or, as in the image processing method described in some of the above embodiments One or more steps.
- Fig. 1 is a flowchart of an image processing model generation method according to some embodiments
- Figure 2 is a schematic diagram of a training sample lesion image according to some embodiments.
- Figure 3 is a structural diagram of an initial image processing model according to some embodiments.
- Fig. 4 is a flowchart of another image processing model generation method according to some embodiments.
- 5 to 8 are flowcharts of S2, S3, S33, and S4 in the image processing model generation method according to some embodiments;
- Fig. 9 is a structural diagram of a target image processing model according to some embodiments.
- Fig. 10 is a flowchart of an image processing method according to some embodiments.
- FIG. 11 is a schematic diagram of a focus image to be processed according to some embodiments.
- FIG. 12 is a flowchart of another image processing method according to some embodiments.
- FIG. 13-15 are flowcharts of S200, S300, and S303 in image processing methods according to some embodiments.
- Fig. 16 is a structural diagram of an image processing model generating apparatus according to some embodiments.
- FIG. 17 is a structural diagram of another image processing model generating apparatus according to some embodiments.
- Fig. 18 is a structural diagram of an image processing apparatus according to some embodiments.
- Fig. 19 is a structural diagram of another image processing apparatus according to some embodiments.
- first and second are only used for descriptive purposes, and cannot be understood as indicating or implying relative importance or implicitly indicating the number of indicated technical features. Thus, the features defined with “first” and “second” may explicitly or implicitly include one or more of these features. In the description of the embodiments of the present disclosure, unless otherwise specified, “plurality” means two or more.
- the expressions “coupled” and “connected” and their extensions may be used.
- the term “connected” may be used when describing some embodiments to indicate that two or more components are in direct physical or electrical contact with each other.
- the term “coupled” may be used when describing some embodiments to indicate that two or more components have direct physical or electrical contact.
- the term “coupled” or “communicatively coupled” may also mean that two or more components are not in direct contact with each other, but still cooperate or interact with each other.
- the embodiments disclosed herein are not necessarily limited to the content herein.
- a and/or B includes the following three combinations: A only, B only, and the combination of A and B.
- the term “if” is optionally interpreted to mean “when” or “when” or “in response to determination” or “in response to detection.”
- the phrase “if it is determined" or “if [the stated condition or event] is detected” is optionally interpreted to mean “when determining" or “in response to determining" Or “when [the stated condition or event] is detected” or “in response to the detection of [stated condition or event]”.
- classification algorithms based on deep learning can well diagnose the types of diseases.
- the inventors of the present disclosure have discovered in research that these classification algorithms have certain differences between the regions of interest when identifying a picture as a certain type of disease in the process of medical image processing. Although some algorithms can be used to roughly mark the location of the lesion on medical images, this marking method is too rough to reflect the specific location of the lesion on the diseased tissue.
- some embodiments of the present disclosure provide a method for generating an image processing model. As shown in FIG. 1, the method includes the following S1 to S7.
- S1 Input at least one training sample focus image of a known disease type into the initial image processing model 1 (as shown in FIG. 2).
- the initial image processing model 1 includes a classification layer 1007 (Classification) and a label layer 1010 (Attention Location).
- the training sample lesion image 1001 includes the initial center point B coordinates (G x , G y ), initial length G h and initial width G w corresponding to the lesion area.
- the initial image processing model 1 refers to an image processing model that has not been trained or is in the process of training.
- the known disease category refers to the disease category corresponding to the training sample lesion image 1001.
- the types of known diseases may be diseases such as pneumonia, heart disease, cirrhosis, etc., or other types of diseases, depending on specific needs.
- the initial image processing model 1 includes a classification layer 1007 and an annotation layer 1010, and both the classification layer 1007 and the annotation layer 1010 are network layers that can process lesion images.
- the training sample lesion image 1001 is used to train the initial image processing model 1, at least one training sample lesion image 1001 of a known disease type can be obtained, and at least one training sample lesion image 1001 is input into the initial image processing model 1 to correct The initial image processing model 1 is trained.
- the number of training sample lesion images 1001 input to the initial image processing model 1 can be one or more, for example, 500, 800, or 1000, etc., and can also be determined according to the actual situation. No specific restrictions.
- the training sample lesion image 1001 includes a lesion area, and the lesion area refers to the position of the lesion in the training sample lesion image 1001, and its position is represented in a certain area range in the training sample lesion image 1001.
- the lesion area may be represented as a rectangular area, and the lesion area may be a real lesion area previously marked in the training sample lesion image 1001 by a doctor or the like.
- the real lesion area in the training sample lesion image 1001 is the area shown in the real box TB (Truth box).
- the real box TB is based on the initial center point B coordinates (G x , G y ), initial length G h and The initial width G w is obtained.
- the initial center point B coordinates (G x , G y ) are the coordinates of the center point corresponding to the real lesion area in the image coordinate system corresponding to the training sample lesion image 1001; the initial length G h and the initial width G w are real respectively The length and width of the lesion area.
- the training sample lesion image 1001 can be sequentially input to the initial image processing model 1 to train the initial image processing model 1 and execute S2.
- S2 Call the classification layer 1007, perform classification processing on the training sample lesion image 1001, and obtain the classification probability of the training sample lesion image 1001 corresponding to each known disease category.
- the classification layer 1007 is a network layer that can be used to classify the disease type corresponding to the lesion image.
- the classification layer 1007 can obtain the classification probability of each disease type corresponding to the lesion image, thereby obtaining the disease type corresponding to the lesion image, such as Pneumonia, heart disease, etc.
- the training sample lesion image 1001 may only include lesions of one disease, or may include lesions of multiple diseases.
- the training sample lesion image 1001 includes lesions of three diseases, and the classification layer 1002 can obtain training The classification probabilities of the three diseases corresponding to the sample lesion image 1001 are determined according to the classification probabilities of each disease to determine the types of diseases included in the training sample lesion image 1001.
- S3 Call the annotation layer 1010, process the training sample lesion image 1001, and obtain the predicted center point coordinates, predicted length, and predicted width of the lesion area contained in the training sample lesion image 1001.
- the labeling layer 1010 is a network layer that can be used to label the specific location of the lesion on the diseased tissue.
- the labeling layer 1010 can add a labeling frame to the lesion area on the lesion image.
- the labeling frame can be a rectangular frame or a circle. Shape frame and so on.
- the training sample lesion image 1001 is input to the initial image processing model 1, and the label layer 1010 is processed to output a prediction box PB (Pred box).
- the prediction box PB is based on the coordinates of the predicted center point A (P x , P y ) , Predicted length P h and predicted width P w .
- the predicted center point A coordinates (P x , P y ) are the coordinates of the center point corresponding to the predicted lesion area in the image coordinate system corresponding to the training sample lesion image 1001;
- the predicted length P h and predicted width P w are respectively the predicted The length and width of the lesion area.
- S2 and S3 can be executed simultaneously; S2 and S3 can also be executed first, or S3 and S2 can be executed first.
- the classification probability of the disease type of the training sample lesion image 1001 and the prediction data of the lesion area are obtained from the above S2 and S3, and S4 can be executed.
- the loss value L can be based on the classification probability corresponding to the training sample lesion image 1001, the initial center point B coordinates (G x , G y ), the initial length G h , the initial width G w , and the predicted center point A coordinates (P x , P y ) , The predicted length P h and the predicted width P w are calculated.
- a training sample lesion image 1001 is input in a training process, and a training sample lesion image 1001 can obtain a loss value L for one disease. If a training sample lesion image 1001 includes multiple diseases, Then, according to the weights of multiple diseases and the corresponding loss values, the loss values of multiple diseases in a training sample lesion image 1001 are averaged to obtain the loss value L of a training sample lesion image 1001.
- multiple training sample lesion images 1001 are input in one training process, and the loss value L of each training sample lesion image 1001 is obtained by the method of the above embodiment, and the loss value of multiple training sample lesion images 1001 L is averaged to obtain the loss value L of multiple training sample lesion images 1001.
- the loss values of the training sample lesion images are obtained from the above S4, and S5 can be executed, see FIG. 3.
- S5 Determine whether the loss value L is within the preset range; if the loss value L is not within the preset range, execute S6; if the loss value L is within the preset range, execute S7.
- S6 Update the parameters of the initial image processing model 1, and use the updated image processing model as the initial image processing model 1 for the next training.
- S7 Use the currently trained image processing model (that is, the last trained image processing model) as the target image processing model 2 after training.
- the preset range refers to the range to be compared with the loss value L, which is a numerical range set in advance.
- loss value L is within the preset range, it means that the training result of the current training image processing model has reached the expected result, and the currently trained image processing model (that is, the last trained image processing model) can be used as the target after training Image processing model 2.
- the loss value L is not within the preset range, it means that the training result of the initial image processing model 1 did not meet expectations. Based on the result of the loss value L, the parameters of the initial image processing model 1 need to be updated, and the image processing after the updated parameters The model is used as the initial image processing model 1 for the next training, and continues to input at least one training sample lesion image 1001 of a known disease type to train the initial image processing model 1.
- the image processing model generated by the image processing model generation method provided by the embodiments of the present disclosure can accurately identify the lesion area in the lesion image, and add corresponding annotation coordinates, which can reflect the specific location of the lesion on the lesion tissue, where
- the classification layer 1007 and the annotation layer 1010 train the area of interest (ie the marked lesion area) of the image processing model to make the position of the marked lesion area more accurate.
- the trained image processing model can automatically identify the medical images contained The type of disease and the location of the lesion area of each type of disease are marked, which saves manpower and improves efficiency, and the generated image processing model has high detection accuracy for classifying and marking the type of disease.
- the initial image processing model 1 further includes: Neural Networks 1000 (Neural Networks), the neural network 1000 includes at least one layer, and each layer in turn includes a convolutional layer 1002 (Convolutional Layer), Activation function 1003 (Activation Function) and pooling layer 1007 (Pooling Layer).
- Neural Networks 1000 Neural Networks
- the neural network 1000 includes at least one layer, and each layer in turn includes a convolutional layer 1002 (Convolutional Layer), Activation function 1003 (Activation Function) and pooling layer 1007 (Pooling Layer).
- the image processing model generation method in the embodiment of the present disclosure further includes: performing S11 after S1.
- S11 Call the neural network 1000, pass the training sample lesion image 1001 through the convolutional layer 1002, activation function 1003, and pooling layer 1004 of each layer of the neural network 1000 in turn to obtain multiple sets of two-dimensional feature maps of the training sample lesion 1001 image .
- the convolutional layer 1002 is used to extract features; the activation function 1003 is used to activate the extracted features; the pooling layer 1004 is used to down-sample the activated features; through multi-layer superposition, training sample lesions can be obtained Multiple sets of two-dimensional feature maps of the image 1001.
- the neural network 1000 in the above-mentioned embodiments of the present disclosure may be a convolutional neural network (Convolutional Neural Networks, CNN for short), a Recurrent Neural Networks (RNN for short), etc.
- CNN convolutional Neural Networks
- RNN Recurrent Neural Networks
- the neural network 1000 in the above-mentioned embodiments of the present disclosure may be a long short-term memory network (Long Short-Term Memory, LSTM), artificial neural network (Artificial Neural Networks, ANNs), etc.
- LSTM Long Short-Term Memory
- ANNs Artificial Neural Networks
- the initial image processing model 1 further includes: a global average pooling layer 1005 (Global Average Pooling, GAP for short).
- GAP Global Average Pooling
- the image processing model generation method in the embodiment of the present disclosure further includes: performing S11A after S11.
- S11A Call the global average pooling layer 1005 to perform global average pooling on multiple sets of two-dimensional feature maps to obtain multiple sets of one-dimensional feature maps corresponding to multiple sets of two-dimensional feature maps.
- the multiple sets of two-dimensional feature maps obtained by the training sample lesion image 1001 through the neural network 1000 are input into the global average pooling layer 1005, and the global average pooling layer 1005 is called to perform global average pooling on multiple sets of two-dimensional feature maps to obtain the corresponding multiple Multiple groups of one-dimensional feature maps of two-dimensional feature maps.
- the initial image processing model 1 further includes: a fully connected layer 1006 (Fully Connected Layers).
- the image processing model generation method in the embodiment of the present disclosure further includes: performing S12 after S11A.
- S12 Call the fully connected layer 1006 to perform feature extraction on multiple sets of one-dimensional feature maps obtained after global average pooling, to obtain at least one set of one-dimensional feature maps of the training sample lesion image 1001.
- the multiple sets of one-dimensional feature maps obtained by processing the training sample lesion image 1001 through the global average pooling layer 1005 are input into the fully connected layer 1006, the fully connected layer 1006 is called, and the multiple sets of one-dimensional feature maps obtained after the global average pooling are processed By feature extraction, at least one set of one-dimensional feature maps of the training sample lesion image 1001 is obtained, and the at least one set of one-dimensional feature maps are input into the classification layer 1007.
- S2 calls the classification layer 1007 to classify the training sample lesion image 1001 to obtain the classification probability of the training sample lesion image 1001 corresponding to each known disease category, including:
- S21 Input at least one set of one-dimensional feature maps of the training sample lesion image 1001 into the classification layer 1007.
- S22 Call the classification layer 1007, perform classification processing on the one-dimensional feature map, and output the classification probability of the training sample lesion image 1001 corresponding to each known disease type.
- the initial image processing model 1 further includes: a feature weighted summation layer 1008 (English name: FWS) and an upsampling layer 1009 (Upsampling Layers).
- FWS feature weighted summation layer
- upsampling layer 1009 Upsampling Layers
- the image processing model generation method in the embodiment of the present disclosure further includes: performing S11B after S11.
- S11B Call the feature weighted summation layer 1008, and perform feature weighted summation on each group of two-dimensional feature maps in the multiple sets of two-dimensional feature maps according to multiple sets of one-dimensional feature maps.
- the multiple sets of two-dimensional feature maps of the training sample lesion image 1001 and the multiple sets of one-dimensional feature maps obtained after global average pooling are input to the feature weighted summation layer 1008, and the feature weighted summation layer 1008 is called.
- One-dimensional feature map which performs feature weighted summation on each group of two-dimensional feature maps in multiple sets of two-dimensional feature maps.
- S13 is executed after S11B.
- S13 Invoke the up-sampling layer 1009, perform up-sampling processing on multiple sets of two-dimensional feature maps that have been feature-weighted and summed, to obtain multiple sets of two-dimensional feature maps processed by the training sample lesion image 1001.
- S3 calls the annotation layer 1010 to process the training sample lesion image 1001 to obtain the predicted center point A coordinate (P) of the lesion area contained in the training sample lesion image 1001 x , P y ), the predicted length P h and the predicted width P w , including S31 to S33.
- S31 Input the processed multiple sets of two-dimensional feature maps of the training sample lesion image 1001 into the annotation layer 1010;
- S32 Call the labeling layer 1010 to determine the largest feature point in the two-dimensional feature map, and the largest feature value and two-dimensional coordinates corresponding to the largest feature point.
- S33 Determine the predicted center point A coordinates (P x , P y ) and predicted length P corresponding to the lesion area contained in the training sample lesion image 1001 according to the maximum feature point, maximum feature value, two-dimensional coordinates and preset feature threshold h and the predicted width P w .
- the preset feature threshold can be adjusted according to the expected range of the predicted length P h and the predicted width P w of the lesion area. Generally, when the preset feature threshold is larger, the obtained lesion area prediction The greater the length P h and the predicted width P w , and vice versa.
- S33 determines the predicted center point A coordinate (P) corresponding to the lesion area contained in the training sample lesion image 1001 according to the maximum feature point, the maximum feature value, two-dimensional coordinates, and a preset feature threshold.
- x , P y ), prediction length P h, and prediction width P w including S331 to S335.
- S331 Determine the two-dimensional coordinates of the largest feature point as the predicted center point A coordinates (P x , P y ).
- S332 Calculate the absolute value of the characteristic difference between the maximum characteristic value and the preset characteristic threshold.
- S334 Obtain the first coordinate value P x -t1 of the first feature point A1 in the horizontal axis X direction , and the second coordinate value P x +t2 of the second feature point A2 in the horizontal axis X direction, where t1, t2 None is less than zero.
- S4 is based on the classification probability corresponding to at least one training sample lesion image 1001, the initial center point B coordinate (G x , G y ), the initial length G h , the initial width G w , and the prediction
- the center point A coordinates (P x , P y ), the predicted length P h and the predicted width P w , and the loss value L of at least one training sample lesion image 1001 in the initial image processing model 1 is obtained, including S41-S43.
- the loss value L includes a classification loss value and a location loss value.
- the classification loss value means that the initial image processing model 1 classifies the training sample lesion image 1001 to obtain the classification loss with the classification probability, and the classification loss value is recorded as L cls .
- the classification loss value L cls can be obtained according to the classification probabilities.
- softmax and SVM full English name: Support Vector Machine, Chinese name: The classification loss value L cls is calculated by means of support vector machine) or sigmoid.
- the position loss value means that the initial image processing model 1 processes the training sample lesion image 1001 to determine the predicted center point A coordinates (P x , P y ) corresponding to the lesion area, the predicted length P h and the predicted width P w and the initial center Point B coordinates (G x , G y ), the initial length G h and the initial width G w exist position loss, the value of the position loss is recorded as L al .
- the position loss value L al can be based on the initial center point B coordinates (G x , G y ), initial length G h , initial width G w , predicted center point A coordinates (P x , P y ), predicted length P h , and predicted The width P w is calculated.
- the position loss value L al can be calculated according to the following formula (1):
- the calculation method of the position loss value L al may also adopt the L1 norm or the L2 norm.
- Some embodiments of the present disclosure provide an image processing method. As shown in FIG. 10, the method includes S100-S500.
- S100 Input the to-be-processed lesion image 2001 into the target image processing model 2 (as shown in FIG. 9).
- the target image processing model 2 is obtained by training the image processing model generation method of some of the above embodiments, and it can be used to process the lesion image in the medical field to determine the corresponding image of the lesion.
- the focus image 2001 to be processed refers to an image in the medical field, such as an image taken by a medical instrument.
- the target image processing model 2 includes a classification layer 2007 and an annotation layer 2010, where the classification layer 2007 can be used to classify the to-be-processed lesion image 2001 to obtain the classification probability corresponding to the to-be-processed lesion image 2001 to determine the lesion to be processed
- the disease corresponding to the image 2001; the annotation layer 2010 can be used to determine the area where the lesion in the lesion image 2001 to be processed is located in the lesion image. This process will be described in detail in the following steps.
- the to-be-processed lesion image 2001 After the to-be-processed lesion image 2001 is obtained, the to-be-processed lesion image 2001 can be input into the target image processing model 2, and S200 is executed.
- S200 Call the classification layer 2007, perform classification processing on the lesion image 2001 to be processed, and obtain the disease classification probability corresponding to the lesion image 2001 to be processed.
- the classification layer 2007 is a network layer that can be used to classify the disease type corresponding to the lesion image.
- the classification layer 2007 can obtain the classification probability of each disease type corresponding to the lesion image, thereby obtaining the disease type corresponding to the lesion image, such as Pneumonia, heart disease, etc.
- S300 Invoke the label layer 2010, process the lesion image 2001 to be processed, and obtain the coordinates of the center point of the area, the length of the area, and the width of the lesion area included in the lesion image 2001 to be processed.
- the labeling layer 2010 is a network layer that can be used to label the specific location of the lesion on the diseased tissue.
- the labeling layer 2010 can add a labeling frame to the lesion area on the lesion image.
- the labeling frame can be a rectangular frame or a circle. Shape frame and so on.
- the to-be-processed lesion image 2001 is input into the target image processing model 2, and is processed by the annotation layer 2010 to output the lesion frame Z.
- the lesion frame Z is based on the coordinates of the center point B of the area (D x , D y ) and the area length D h and the area width D w . Then the coordinates of the center point B of the area (D x , D y ) are the coordinates of the center point corresponding to the focus area in the image coordinate system corresponding to the focus image 2001 to be processed; the area length D h and the area width D w are the focus area respectively Corresponding length and width.
- S200 and S300 may be executed at the same time; S200 may be executed first, and then S300 may be executed, or S300 may be executed first, and then S200 may be executed.
- the classification probability of the disease type of the to-be-processed focus image 2001 and the area data of the focus area are obtained respectively, and S400 may be executed.
- S400 Determine the type of disease corresponding to the focus image 2001 to be processed according to the disease classification probability.
- the disease classification probability is obtained in S200. If the disease classification probability for a specific disease type in the to-be-processed focus image 2001 is within a certain range, it means that the to-be-processed focus image 2001 contains the specific disease type. If the disease classification probability for a specific disease type in the to-be-processed focus image 2001 is not within a certain range, it means that the to-be-processed focus image 2001 does not include the specific disease type.
- S500 Determine the lesion marking area in the lesion image 2001 to be processed according to the coordinates (D x , D y ) of the area center point D, the area length D h and the area width D w.
- the coordinates (D x , D y ) of the area center point B obtained in S300 are the coordinates of the center point corresponding to the focus area in the image coordinate system corresponding to the focus image 2001 to be processed.
- the area length D h and the area width D w are respectively Is the length and width corresponding to the lesion area, and the lesion marking area is marked in the lesion image 2001 to be processed, as shown in the lesion frame Z shown in FIG. 11.
- the image processing method provided by the embodiments of the present disclosure utilizes a trained image processing model to automatically identify the types of diseases contained in medical images and to mark the location of the lesion area of each type of disease, which saves manpower and improves efficiency.
- the image processing model has high detection accuracy for classification and labeling of diseases.
- the target image processing model 2 further includes: a neural network 2000, the neural network 2000 includes at least one layer, and each layer includes a convolutional layer 2002, an activation function 2003, and a pooling layer 2004 in turn .
- the image processing method in the embodiment of the present disclosure further includes: performing S101 after S100.
- S101 Invoke the neural network 2000, pass the to-be-processed lesion image 2001 through the convolutional layer 2002, activation function 2003, and pooling layer 2004 of each layer of the neural network 2000 in turn to obtain multiple sets of two-dimensional feature maps of the to-be-processed lesion image 2001 .
- the convolutional layer 2002 is used to extract features; the activation function 2003 is used to activate the extracted features; the pooling layer 2004 is used to down-sample the activated features; through multi-layer superposition, the to-be-treated lesions can be obtained Multiple sets of two-dimensional feature maps of the image 2001.
- the neural network 2000 in the foregoing embodiments of the present disclosure may be a convolutional neural network, a cyclic neural network, or the like.
- the neural network 2000 in the above-mentioned embodiments of the present disclosure may be a Long Short-Term Memory (LSTM), artificial neural networks (Artificial Neural Networks, ANNs), etc.
- LSTM Long Short-Term Memory
- ANNs Artificial Neural Networks
- the target image processing model 2 further includes: a global average pooling layer.
- the image processing method in the embodiment of the present disclosure further includes: performing S101A after S101.
- S101A Call the global average pooling layer 2005 to perform global average pooling on multiple sets of two-dimensional feature maps to obtain multiple sets of one-dimensional feature maps corresponding to multiple sets of two-dimensional feature maps.
- the target image processing model 2 further includes: a fully connected layer 2006.
- the image processing method in the embodiment of the present disclosure further includes: performing S102 after S101A.
- S102 Call the fully connected layer 2006 to perform feature extraction on multiple sets of one-dimensional feature maps obtained after global average pooling, to obtain at least one set of one-dimensional feature maps of the lesion image 2001 to be processed.
- S200 calls the classification layer 2007 to classify the to-be-processed lesion image 2001 to obtain the disease classification probability corresponding to the to-be-processed lesion image 2001, including:
- S201 Input at least one set of one-dimensional feature maps of the lesion image 2001 to be processed into the classification layer 2007.
- S202 Call the classification layer 2007, perform classification processing on the one-dimensional feature map, and output the disease classification probability corresponding to the focus image 2001 to be processed.
- the target image processing model 2 further includes: a feature weighted summation layer 2007 and an upsampling layer 2009.
- the image processing method in the embodiment of the present disclosure further includes: performing S101B after S101.
- S101B Call the feature weighted summation layer 2008, and perform feature weighted summation on each group of two-dimensional feature maps in the multiple sets of two-dimensional feature maps according to multiple sets of one-dimensional feature maps.
- the multiple sets of two-dimensional feature maps of the lesion image 2001 to be processed and the multiple sets of one-dimensional feature maps obtained after global average pooling are input to the feature weighted summation layer 2008, and the feature weighted summation layer 2008 is called.
- One-dimensional feature map which performs feature weighted summation on each group of two-dimensional feature maps in multiple sets of two-dimensional feature maps.
- S103 is executed after S101B.
- S103 Call the up-sampling layer 2009 to perform up-sampling processing on the multiple sets of two-dimensional feature maps that have been feature-weighted and summed to obtain the processed multiple sets of two-dimensional feature maps of the to-be-processed lesion image 2001, and combine the processed multiple sets of two The dimensional feature map is input to the annotation layer 2010.
- S300 calls the annotation layer 2010 to process the to-be-processed lesion image 2001 to obtain the regional center point D coordinate (D x , D y ) of the lesion area contained in the to-be-processed lesion image 2001 , The area length D h and the area width D w , including S301 to S303.
- S301 Input the processed multiple sets of two-dimensional feature maps of the to-be-processed lesion image 2001 into the annotation layer 2010.
- S302 Call the annotation layer 2010 to determine the maximum feature point in the two-dimensional feature map, and the maximum feature value and two-dimensional coordinates corresponding to the maximum feature point.
- the preset feature threshold can be adjusted according to the desired range of the area length D h and the area width D w of the lesion area.
- the preset feature threshold is larger, the obtained area of the lesion area The greater the length D h and the area width D w , and vice versa.
- S303 according to the maximum feature point, the maximum feature value, the two-dimensional coordinates, and the preset feature threshold, determine the regional center point D corresponding to the lesion area contained in the lesion image 2001 to be processed
- the coordinates (D x , D y ), area length D h and area width D w include S3031 to S3035.
- S3031 Use the two-dimensional coordinates of the largest feature point as the D coordinates (D x , D y ) of the center point of the area.
- S3032 Calculate the absolute value of the characteristic difference between the maximum characteristic value and the preset characteristic threshold.
- S3034 Obtain the first coordinate value D x- t11 of the first feature point D1 in the horizontal axis X direction , and the second coordinate value D x + t22 of the second feature point D2 in the horizontal axis X direction, where t11 and t22 None is less than zero.
- some embodiments of the present disclosure provide an image processing model generating device 3, and the image processing model generating device 3 includes:
- the sample image input component 10 is configured to input at least one training sample lesion image 1001 of a known disease type into the initial image processing model 1.
- the initial image processing model 1 includes a classification layer 1007 and an annotation layer 1010, and the training sample lesion image 1001 includes the initial center point B coordinates (G x , G y ), initial length G h and initial width G w corresponding to the lesion area.
- the classification probability acquisition component 20 is coupled with the sample image input component 10 and is configured to call the classification layer 1007 to classify the training sample lesion image 1001 to obtain the classification probability of the training sample lesion image 1001 corresponding to each known disease category.
- the prediction data acquisition component 30, coupled with the sample image input component 10, is configured to call the annotation layer 1010 to process the training sample lesion image 1001 to obtain the predicted center point A coordinate of the lesion area contained in the training sample lesion image 1001 ( P x , P y ), predicted length P h, and predicted width P w .
- the loss value acquisition component 40 is coupled to the classification probability acquisition component 20 and the prediction data acquisition component 30, and is configured according to the classification probability corresponding to the at least one training sample lesion image 1001, the initial center point B coordinates (G x , G y ), Initial length G h , initial width G w , predicted center point A coordinates, (P x , P y ) predicted length P h and predicted width P w , and obtain at least one training sample lesion image 1001. Loss in the initial image processing model 1 Value L.
- the target model generating component 50 is coupled to the loss value acquiring component 40, and is configured to determine whether the loss value L is within a preset range, and if the loss value L is not within the preset range, update the parameters of the initial image processing model 1 , The image processing model after the updated parameters is used as the initial image processing model 1 for the next training; continue to train the image processing model after the updated parameters until the loss value L is within the preset range, and the last trained image processing model As the target image processing model 2.
- the initial image processing model 1 further includes a neural network 1000, a global average pooling layer 1005, a fully connected layer 1006, a feature weighted summation layer 1008, and an upsampling layer 1009;
- the network 1000 includes at least one layer, and each layer includes a convolutional layer 1002, an activation function 1003, and a pooling layer 1004 in turn.
- the image processing model generating device 3 in the embodiment of the present disclosure further includes:
- the first two-dimensional feature map acquiring component 11 is coupled to the sample image input component 10, and is configured to call the neural network 1000, and sequentially pass the training sample lesion image 1001 through the convolutional layer 1002 of each layer of the neural network 1000, and the activation function 1003 and the pooling layer 1004 to obtain multiple sets of two-dimensional feature maps of the training sample lesion image 1001.
- the first global average pooling component 11A is coupled to the first two-dimensional feature map acquiring component 11, and is configured to call the global average pooling layer 1005 to perform global average pooling on multiple sets of two-dimensional feature maps to obtain corresponding multiple groups Multiple sets of one-dimensional feature maps of two-dimensional feature maps.
- the first fully connected component 12 is coupled to the first global average pooling component 11A, the first fully connected component 12 is also coupled to the classification probability acquisition component 20, the first fully connected component 12 is configured to call the fully connected layer 1006, Perform feature extraction on multiple sets of one-dimensional feature maps to obtain at least one set of one-dimensional feature maps of the training sample lesion image 1001, and input the at least one set of one-dimensional feature maps to the classification probability acquisition component 20.
- the first feature weighted summation component 11B is coupled to the first two-dimensional feature map acquisition component 11 and the first global average pooling component 11A, and is configured to call the feature weighted summation layer 1008, according to multiple sets of one-dimensional feature maps, Perform feature weighted summation on each set of two-dimensional feature maps in multiple sets of two-dimensional feature maps.
- the first upsampling component 13 is coupled to the first feature weighted summation component 11B, the first upsampling component 13 is also coupled to the prediction data acquisition component 30, and the first upsampling component 13 is configured to call the upsampling layer 1009,
- the up-sampling process is performed on the multiple sets of two-dimensional feature maps that have undergone feature weighted summation, and the multiple sets of up-sampling two-dimensional feature maps are input to the prediction data acquisition component 30.
- the functions of the components included in the image processing model generating device 3 provided in the foregoing embodiment of the present disclosure may refer to the description of the corresponding steps in the image processing model generating method described in the foregoing embodiment.
- the image processing model generating device 3 provided in the above-mentioned embodiments of the present disclosure can generate a target image processing model 2, which can be applied to the medical image recognition process to enable the target image processing model 2 to determine the type of disease and the location of the lesion in the medical image.
- Automatic detection saves manpower and improves the efficiency of disease diagnosis, and the identification of the type of disease in the medical image and the labeling of the lesion location through the target image processing model 2 are more accurate.
- some embodiments of the present disclosure further provide an image processing device 4, and the image processing device 4 includes:
- the to-be-processed image input component 100 is configured to input the to-be-processed lesion image 2001 into the target image processing model 2.
- the target image processing model 2 is obtained by training the image processing model generation method in some embodiments of the disclosure above, and the target image processing model 2 includes the classification layer 2007 and the label layer 2010 (as shown in Figure 9).
- the disease classification probability acquisition component 200 is coupled to the to-be-processed image input component 100 and is configured to call the classification layer 2007 to perform classification processing on the to-be-processed focus image 2001 to obtain the disease classification probability corresponding to the to-be-processed focus image 2001.
- the lesion area data acquisition component 300 is coupled to the to-be-processed image input component 100, and is configured to call the annotation layer 2010 to process the to-be-processed lesion image 2001 to obtain the coordinate of the regional center point D of the lesion area contained in the to-be-processed lesion image 2001 (D x , D y ), area length D h and area width D w (as shown in Figure 11).
- the disease type determination component 400 is coupled to the disease classification probability acquisition component 200 and is configured to determine the disease type corresponding to the focus image 2001 to be processed according to the disease classification probability.
- the lesion area determining component 500 is coupled to the lesion area data acquiring component 300, and is configured to determine the focus image to be processed according to the coordinates (D x , D y ) of the area center point D, the area length D h and the area width D w Mark the area of the lesion.
- the target image processing model 2 further includes a neural network 2000, a global average pooling layer 2005, a fully connected layer 2006, a feature weighted summation layer 2008, and an upsampling layer 2009;
- the network 2000 includes at least one layer, and each layer includes a convolutional layer 2002, an activation function 2003, and a pooling layer 2004 in turn.
- the image processing device 4 in the embodiment of the present disclosure further includes:
- the second two-dimensional feature map acquiring component 101 is coupled to the image input component 100 to be processed, and is configured to call the neural network 2000, and sequentially pass the to-be-processed lesion image 2001 through the convolutional layer 2002 of each layer of the neural network 2000 and activate Function 2003 and pooling layer 2004 to obtain multiple sets of two-dimensional feature maps of the focus image 2001 to be processed.
- the second global average pooling component 101A is coupled to the second two-dimensional feature map acquiring component 101, and is configured to call the global average pooling layer 2005 to perform global average pooling on multiple sets of two-dimensional feature maps to obtain corresponding multiple groups Multiple sets of one-dimensional feature maps of two-dimensional feature maps.
- the second fully connected component 102 is coupled to the second global average pooling component 101A and the disease classification probability acquisition component 200, and is configured to call the fully connected layer 2006 to perform feature extraction on multiple sets of one-dimensional feature maps to obtain the pending processing At least one set of one-dimensional feature maps of the lesion image 2001 is inputted into the disease classification probability acquisition component 200.
- the second feature weighted summation component 101B is coupled to the second two-dimensional feature map acquisition component 101 and the second global average pooling component 101A, and is configured to call the feature weighted summation layer 2008, according to multiple sets of one-dimensional feature maps, Perform feature weighted summation on each set of two-dimensional feature maps in multiple sets of two-dimensional feature maps.
- the second up-sampling component 103 is coupled to the second feature-weighted summation component 101B and the lesion area data acquisition component 300, and is configured to call the up-sampling layer 2009 to perform on the multiple sets of feature-weighted summation of the two-dimensional feature maps. Sampling processing is performed, and multiple sets of two-dimensional feature maps that have undergone up-sampling processing are input to the lesion area data acquiring component 300.
- the image processing device 4 provided in the above-mentioned embodiment of the present disclosure applies the target image processing model 2 to the recognition of medical images, so that the target image processing model 2 can automatically detect the disease type and focus position of the medical image, saving Manpower improves the efficiency of disease diagnosis, and the identification of disease types in medical images and the labeling of lesion locations are more accurate through the target image processing model 2.
- Some embodiments of the present disclosure provide a computer-readable storage medium (for example, a non-transitory computer-readable storage medium) in which computer program instructions are stored, and when the computer program instructions run on a processor , Causing the processor to execute one or more steps in the image processing model generation method described in some of the above embodiments, and/or, one of the image processing methods described in some of the above embodiments Or multiple steps.
- a computer-readable storage medium for example, a non-transitory computer-readable storage medium
- the foregoing computer-readable storage medium may include, but is not limited to: magnetic storage devices (for example, hard disks, floppy disks, or tapes, etc.), optical disks (for example, CD (Compact Disk), DVD (Digital Versatile Disk, Digital universal disk), etc.), smart cards and flash memory devices (for example, EPROM (Erasable Programmable Read-Only Memory), cards, sticks or key drives, etc.).
- magnetic storage devices for example, hard disks, floppy disks, or tapes, etc.
- optical disks for example, CD (Compact Disk), DVD (Digital Versatile Disk, Digital universal disk), etc.
- smart cards and flash memory devices for example, EPROM (Erasable Programmable Read-Only Memory), cards, sticks or key drives, etc.
- the various computer-readable storage media described in this disclosure may represent one or more devices and/or other machine-readable storage media for storing information.
- the term "machine-readable storage medium” may include, but is not limited to, wireless channels and various other media capable of storing, containing, and/or carrying instructions and/or data.
- Some embodiments of the present disclosure also provide a computer program product.
- the computer program product includes computer program instructions, and when the computer program instructions are executed on a computer, the computer program instructions cause the computer to execute one or more steps in the image processing model generation method described in some of the above embodiments, And/or, one or more steps in the image processing method described in some of the above embodiments.
- Some embodiments of the present disclosure also provide a computer program.
- the computer program When the computer program is executed on a computer, the computer program causes the computer to execute one or more steps in the image processing model generation method described in some of the above embodiments, and/or, as some of the above embodiments One or more steps in the image processing method described in the embodiment.
- Some embodiments of the present disclosure also provide an electronic device, including: a processor, a memory, and a computer program stored on the memory and capable of running on the processor.
- the processor executes the computer program when the computer program is executed.
- the processor is used to support the image processing model generation device 3, and/or the image processing device 4 executes one or more steps in the image processing model generation method described in some of the above embodiments, and/or, as One or more steps in the image processing method described in some of the above embodiments, and/or other processes used in the techniques described herein.
- the processor can be a central processing unit (Central Processing Unit, CPU for short), other general-purpose processors, digital signal processors (DSP), application-specific integrated circuits (ASIC), field programmable gate arrays (FPGA), or other Programming logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- CPU Central Processing Unit
- DSP digital signal processors
- ASIC application-specific integrated circuits
- FPGA field programmable gate arrays
- Programming logic devices discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor.
- the memory is used to store the program code and data of the image processing model generating device 3 and/or the program code and data of the image processing device 4 provided by the embodiment of the present disclosure.
- the processor can execute various functions of the image processing model generating device 3 and/or various functions of the image processing device 4 by running or executing a software program stored in the memory and calling data stored in the memory.
- the memory can be a read-only memory (Read-Only Memory, ROM) or other types of static storage devices that can store static information and instructions, random access memory (Random Access Memory, RAM), or other types that can store information and instructions
- Dynamic storage devices can also be Electrically Erasable Programmable Read-Only Memory (EEPROM), CD-ROM (Compact Disc Read-Only Memory, CD-ROM) or other optical disc storage, optical disc storage ( Including compact discs, laser discs, optical discs, digital universal discs, Blu-ray discs, etc.), magnetic disk storage media or other magnetic storage devices, or can be used to carry or store desired program codes in the form of instructions or data structures and can be stored by a computer Any other media taken, but not limited to this.
- the memory can exist independently and is connected to the processor through a communication bus.
- the memory can also be integrated with the processor.
Abstract
Description
Claims (19)
- 一种图像处理模型生成方法,包括:将已知疾病种类的至少一个训练样本病灶图像输入初始图像处理模型;其中,所述初始图像处理模型包括分类层和标注层,所述训练样本病灶图像包括病灶区域对应的初始中心点坐标、初始长度和初始宽度;调用所述分类层,对所述训练样本病灶图像进行分类处理,得到所述训练样本病灶图像对应各已知疾病种类的分类概率;调用所述标注层,对所述训练样本病灶图像进行处理,得到所述训练样本病灶图像中所包含的病灶区域的预测中心点坐标、预测长度和预测宽度;根据所述至少一个训练样本病灶图像对应的所述分类概率、所述初始中心点坐标、所述初始长度、所述初始宽度、所述预测中心点坐标、所述预测长度和所述预测宽度,获取所述至少一个训练样本病灶图像在所述初始图像处理模型中的损失值;判断所述损失值是否处于预设范围内,若所述损失值不在预设范围内,则对所述初始图像处理模型的参数进行更新,将更新参数后的图像处理模型作为下一次训练的初始图像处理模型;重复上述步骤,直至所述损失值处于所述预设范围内,将最后一次训练的图像处理模型作为训练后的目标图像处理模型。
- 根据权利要求1所述的方法,其中,所述调用所述分类层,对所述训练样本病灶图像进行分类处理,得到所述训练样本病灶图像对应各已知疾病种类的分类概率,包括:将所述训练样本病灶图像的至少一组一维特征图输入所述分类层;调用所述分类层,对所述一维特征图进行分类处理,输出所述训练样本病灶图像对应各已知疾病种类的分类概率。
- 根据权利要求1或2所述的方法,其中,所述调用所述标注层,对所述训练样本病灶图像进行处理,得到所述训练样本病灶图像中所包含的病灶区域的预测中心点坐标、预测长度和预测宽度,包括:将所述训练样本病灶图像的经过处理的多组二维特征图输入所述标注层;调用所述标注层,确定所述二维特征图中的最大特征点,及所述最大特征点对应的最大特征值和二维坐标;根据所述最大特征点、所述最大特征值、所述二维坐标和预设特征阈值,确定所述训练样本病灶图像中所包含的病灶区域对应的预测中心点坐标、预测长度和预测宽度。
- 根据权利要求3所述的方法,其中,所述二维坐标包括横轴方向上的第一坐标值和纵轴方向上的第二坐标值;所述根据所述最大特征点、所述最大特征值、所述二维坐标和预设特征阈值,确定所述训练样本病灶图像中所包含的病灶区域对应的预测中心点坐标、预测长度和预测宽度,包括:将所述二维坐标确定为所述预测中心点坐标;计算得到所述最大特征值和所述预设特征阈值之间的特征差值绝对值;根据所述特征差值绝对值和所述最大特征点的二维坐标,获取所述二维特征图中在所述横轴方向上的第一特征点和第二特征点,及在所述纵轴方向上的第三特征点和第四特征点;获取所述第一特征点在所述横轴方向上的第一坐标值,及所述第二特征点在所述横轴方向上的第二坐标值;获取所述第三特征点在所述纵轴方向上的第三坐标值,及所述第四特征点在所述纵轴方向上的第四坐标值;基于所述第一坐标值和所述第二坐标值,计算得到所述预测宽度;基于所述第三坐标值和所述第四坐标值,计算得到所述预测长度。
- 根据权利要求2~4中任一项所述的方法,其中,所述初始图像处理模型还包括:全局平均池化层;在所述调用所述分类层和在所述调用所述标注层之前,还包括:将所述训练样本病灶图像的多组二维特征图输入所述全局平均池化层;调用所述全局平均池化层,对所述多组二维特征图进行全局平均池化,得到对应所述多组二维特征图的多组一维特征图。
- 根据权利要求5所述的方法,其中,所述初始图像处理模型还包括:全连接层;在所述调用所述全局平均池化层之后,以及在所述调用所述分类层之前,还包括:调用所述全连接层,对经过所述全局平均池化后得到的所述多组一维特征图进行特征提取,得到所述训练样本病灶图像的至少一组一维特征图,并将所述至少一组一维特征图输入所述分类层。
- 根据权利要求5所述的方法,其中,所述初始图像处理模型还包括:特征加权求和层和上采样层;在所述调用所述全局平均池化层之后,以及在所述调用所述标注层之前,还包括:将所述训练样本病灶图像的多组二维特征图,以及经过所述全局平均池化后得到的所述多组一维特征图,输入所述特征加权求和层;调用所述特征加权求和层,根据所述多组一维特征图,对所述多组二维特征图中的每组二维特征图进行特征加权求和;调用所述上采样层,对经过所述特征加权求和的多组二维特征图进行上采样处理,得到所述训练样本病灶图像经过处理的多组二维特征图,并将所述经过处理的多组二维特征图输入所述标注层。
- 根据权利要求5~7中任一项所述的方法,其中,所述初始图像处理模型还包括:神经网络;所述神经网络包括至少一个层,每个层依次包括卷积层、激活函数和池化层;在所述将所述训练样本病灶图像的多组二维特征图输入所述全局平均池化层之前,还包括:将所述训练样本病灶图像输入所述神经网络;调用所述神经网络,将所述训练样本病灶图像依次通过所述神经网络的每个层的卷积层、激活函数和池化层,得到所述训练样本病灶图像的多组二维特征图,并将所述多组二维特征图输入所述全局平均池化层;在所述初始图像处理模型还包括特征加权求和层的情况下,所述多组二维特征图还被输入至所述特征加权求和层。
- 根据权利要求1~8中任一项所述的方法,其中,所述根据所述至少一个训练样本病灶图像对应的所述分类概率、所述初始中心点坐标、所述初始长度、所述初始宽度、所述预测中心点坐标、所述预测长度和所述预测宽度,获取所述至少一个训练样本病灶图像在所述初始图像处理模型中的损失值,包括:根据所述训练样本病灶图像对应的所述分类概率,计算得到分类损失值;根据所述训练样本病灶图像对应的所述初始中心点坐标、所述初始长度、所述初始宽度、所述预测中心点坐标、所述预测长度和所述预测宽度,计算得到位置损失值;根据所述分类损失值和所述位置损失值,获取所述训练样本病灶图像在所述初始图像处理模型中的损失值。
- 一种图像处理方法,包括:将待处理病灶图像输入目标图像处理模型;所述目标图像处理模型通过如权利要求1~9中任一项所述的方法训练得到,所述目标图像处理模型包括分类层和标注层;调用所述分类层,对所述待处理病灶图像进行分类处理,得到所述待处理病灶图像对应的疾病分类概率;调用所述标注层,对所述待处理病灶图像进行处理,得到所述待处理病灶图像中所包含病灶区域的区域中心点坐标、区域长度和区域宽度;根据所述疾病分类概率,确定所述待处理病灶图像对应的疾病种类;根据所述区域中心点坐标、所述区域长度和所述区域宽度,确定所述待处理病灶图像中的病灶标注区域。
- 根据权利要求10所述的方法,其中,所述调用所述分类层,对所述待处理病灶图像进行分类处理,得到所述待处理病灶图像对应的疾病分类概率,包括:将所述待处理病灶图像的至少一组一维特征图输入所述分类层;调用所述分类层,对所述一维特征图进行分类处理,输出所述待处理病灶图像对应的疾病分类概率。
- 根据权利要求10或11所述的方法,其中,所述调用所述标注层,对所述待处理病灶图像进行处理,得到所述待处理病灶图像中所包含的病灶区域对应的区域中心点坐标、区域长度和区域宽度,包括:将所述待处理病灶图像的经过处理的多组二维特征图输入所述标注层;调用所述标注层,确定所述二维特征图中的最大特征点,及所述最大特征点对应的最大特征值和二维坐标;根据所述最大特征点、所述最大特征值、所述二维坐标和预设特征阈值,确定所述待处理病灶图像中所包含的病灶区域对应的区域中心点坐标、区域长度和区域宽度。
- 根据权利要求12所述的方法,其中,所述二维坐标包括横轴方向上的第一坐标值和纵轴方向上的第二坐标值;所述根据所述最大特征点、所述最大特征值、所述二维坐标和预设特征阈值,确定所述待处理病灶图像中所包含的病灶区域的区域中心点坐标、区域长度和区域宽度,包括:将所述二维坐标确定为所述区域中心点坐标;计算得到所述最大特征值和所述预设特征阈值之间的特征差值绝对值;根据所述特征差值绝对值和所述最大特征点的二维坐标,获取所述二维特征图中在所述横轴方向上的第一特征点和第二特征点,及在所述纵轴方向上的第三特征点和第四特征点;获取所述第一特征点在所述横轴方向上的第一坐标值,及所述第二特征 点在所述横轴方向上的第二坐标值;获取所述第三特征点在所述纵轴方向上的第三坐标值,及所述第四特征点在所述纵轴方向上的第四坐标值;基于所述第一坐标值和所述第二坐标值,计算得到所述区域宽度;基于所述第三坐标值和所述第四坐标值,计算得到所述区域长度。
- 一种图像处理模型生成装置,包括:样本图像输入部件,被配置为将已知疾病种类的至少一个训练样本病灶图像输入初始图像处理模型;其中,所述初始图像处理模型包括分类层和标注层,所述训练样本病灶图像包括病灶区域对应的初始中心点坐标、初始长度和初始宽度;分类概率获取部件,被配置为调用所述分类层,对所述训练样本病灶图像进行分类处理,得到所述训练样本病灶图像对应各已知疾病种类的分类概率;预测数据获取部件,被配置为调用所述标注层,对所述训练样本病灶图像进行处理,得到所述训练样本病灶图像中所包含病灶区域的预测中心点坐标、预测长度和预测宽度;损失值获取部件,被配置为根据所述至少一个训练样本病灶图像对应的所述分类概率、所述初始中心点坐标、所述初始长度、所述初始宽度、所述预测中心点坐标、所述预测长度和所述预测宽度,获取所述至少一个训练样本病灶图像在所述初始图像处理模型中的损失值;目标模型生成部件,被配置为判断所述损失值是否处于预设范围内,若所述损失值不在预设范围内,则对所述初始图像处理模型的参数进行更新,将更新参数后的图像处理模型作为下一次训练的初始图像处理模型;继续对更新参数后的图像处理模型进行训练,直至所述损失值处于所述预设范围内,将最后一次训练的图像处理模型作为目标图像处理模型。
- 根据权利要求14所述的装置,其中,所述初始图像处理模型还包括神经网络、全局平均池化层、全连接层、特征加权求和层和上采样层;其中,所述神经网络包括至少一个层,每个层依次包括卷积层、激活函数和池化层;所述图像处理模型生成装置还包括:与所述样本图像输入部件耦接的第一二维特征图获取部件,被配置为调用所述神经网络,将所述训练样本病灶图像依次通过所述神经网络的每个层的卷积层、激活函数和池化层,得到所述训练样本病灶图像的多组二维特征图;与所述第一二维特征图获取部件耦接的第一全局平均池化部件,被配置为调用所述全局平均池化层,对所述多组二维特征图进行全局平均池化,得到对应所述多组二维特征图的多组一维特征图;与所述第一全局平均池化部件耦接的第一全连接部件,所述第一全连接部件还与所述分类概率获取部件耦接,所述第一全连接部件被配置为调用所述全连接层,对所述多组一维特征图进行特征提取,得到所述训练样本病灶图像的至少一组一维特征图,并将所述至少一组一维特征图输入所述分类概率获取部件;与所述第一二维特征图获取部件以及所述第一全局平均池化部件耦接的第一特征加权求和部件,被配置为调用所述特征加权求和层,根据所述多组一维特征图,对所述多组二维特征图中的每组二维特征图进行特征加权求和;与所述第一特征加权求和部件耦接的第一上采样部件,所述第一上采样部件还与所述预测数据获取部件耦接,所述第一上采样部件被配置为调用所述上采样层,对经过所述特征加权求和的多组二维特征图进行上采样处理,并将经过所述上采样处理的多组二维特征图输入所述预测数据获取部件。
- 一种图像处理装置,包括:待处理图像输入部件,被配置为将待处理病灶图像输入目标图像处理模型;其中,所述目标图像处理模型通过如权利要求1~9中任一项所述的方法训练得到,所述目标图像处理模型包括分类层和标注层;疾病分类概率获取部件,被配置为调用所述分类层,对所述待处理病灶图像进行分类处理,得到所述待处理病灶图像对应的疾病分类概率;病灶区域数据获取部件,被配置为调用所述标注层,对所述待处理病灶图像进行处理,得到所述待处理病灶图像中所包含病灶区域的区域中心点坐标、区域长度和区域宽度;疾病种类确定部件,被配置为根据所述疾病分类概率,确定所述待处理病灶图像对应的疾病种类;病灶区域确定部件,被配置为根据所述区域中心点坐标、所述区域长度和所述区域宽度,确定所述待处理病灶图像中的病灶标注区域。
- 根据权利要求16所述的装置,其中,所述目标图像处理模型还包括神经网络、全局平均池化层、全连接层、特征加权求和层和上采样层;其中,所述神经网络包括至少一个层,每个层依次包括卷积层、激活函数和池化层;所述图像处理装置还包括:与所述待处理图像输入部件耦接的第二二维特征图获取部件,被配置为 调用所述神经网络,将所述待处理病灶图像依次通过所述神经网络的每个层的卷积层、激活函数和池化层,得到所述待处理病灶图像的多组二维特征图;与所述第二二维特征图获取部件耦接的第二全局平均池化部件,被配置为调用所述全局平均池化层,对所述多组二维特征图进行全局平均池化,得到对应所述多组二维特征图的多组一维特征图;与所述第二全局平均池化部件耦接的第二全连接部件,所述第二全连接部件还与所述疾病分类概率获取部件耦接,所述第二全连接部件被配置为调用所述全连接层,对所述多组一维特征图进行特征提取,得到所述待处理病灶图像的至少一组一维特征图,并将所述至少一组一维特征图输入所述疾病分类概率获取部件;与所述第二二维特征图获取部件以及所述第二全局平均池化部件耦接的第二特征加权求和部件,被配置为调用所述特征加权求和层,根据所述多组一维特征图,对所述多组二维特征图中的每组二维特征图进行特征加权求和;与所述第二特征加权求和部件耦接的第二上采样部件,所述第二上采样部件还与所述病灶区域数据获取部件耦接,所述第二上采样部件被配置为调用所述上采样层,对经过所述特征加权求和的多组二维特征图进行上采样处理,并将经过所述上采样处理的多组二维特征图输入所述病灶区域数据获取部件。
- 一种计算机可读存储介质,所述计算机可读存储介质中存储有计算机程序指令,所述计算机程序指令在处理器上运行时,使得所述处理器执行如权利要求1~9中任一项所述的图像处理模型生成方法,和/或,如权利要求10~13中任一项所述的图像处理方法。
- 一种电子设备,包括处理器、存储器、以及存储在所述存储器上并可在所述处理器上运行的计算机程序;所述处理器执行所述计算机程序时,实现如权利要求1~9中任一项所述的图像处理模型生成方法,和/或,如权利要求10~13中任一项所述的图像处理方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/423,439 US11887303B2 (en) | 2019-08-22 | 2020-08-17 | Image processing model generation method, image processing method and device, and electronic device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910778807.0A CN110490262B (zh) | 2019-08-22 | 2019-08-22 | 图像处理模型生成方法、图像处理方法、装置及电子设备 |
CN201910778807.0 | 2019-08-22 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021032062A1 true WO2021032062A1 (zh) | 2021-02-25 |
Family
ID=68552908
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/109611 WO2021032062A1 (zh) | 2019-08-22 | 2020-08-17 | 图像处理模型生成方法、图像处理方法、装置及电子设备 |
Country Status (3)
Country | Link |
---|---|
US (1) | US11887303B2 (zh) |
CN (1) | CN110490262B (zh) |
WO (1) | WO2021032062A1 (zh) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110490262B (zh) * | 2019-08-22 | 2022-06-03 | 京东方科技集团股份有限公司 | 图像处理模型生成方法、图像处理方法、装置及电子设备 |
CN112926612A (zh) * | 2019-12-06 | 2021-06-08 | 中移(成都)信息通信科技有限公司 | 病理图像分类模型训练方法、病理图像分类方法及装置 |
CN112331314A (zh) * | 2020-11-25 | 2021-02-05 | 中山大学附属第六医院 | 一种图像标注方法、装置、存储介质和电子设备 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005339075A (ja) * | 2004-05-26 | 2005-12-08 | Nippon Telegr & Teleph Corp <Ntt> | 画像探索装置、画像探索方法、この方法のプログラム及びこのプログラムを記録した記録媒体 |
JP2010067252A (ja) * | 2008-08-15 | 2010-03-25 | Fuji Xerox Co Ltd | オブジェクト領域抽出装置及びオブジェクト領域抽出プログラム |
CN108564567A (zh) * | 2018-03-15 | 2018-09-21 | 中山大学 | 一种超高分辨率病理图像癌变区域可视化方法 |
CN108961296A (zh) * | 2018-07-25 | 2018-12-07 | 腾讯科技(深圳)有限公司 | 眼底图像分割方法、装置、存储介质和计算机设备 |
CN109363698A (zh) * | 2018-10-16 | 2019-02-22 | 杭州依图医疗技术有限公司 | 一种乳腺影像征象识别的方法及装置 |
CN110136809A (zh) * | 2019-05-22 | 2019-08-16 | 腾讯科技(深圳)有限公司 | 一种医疗图像处理方法、装置、电子医疗设备和存储介质 |
CN110490262A (zh) * | 2019-08-22 | 2019-11-22 | 京东方科技集团股份有限公司 | 图像处理模型生成方法、图像处理方法、装置及电子设备 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070049915A1 (en) * | 2005-08-26 | 2007-03-01 | Dieter Haemmerich | Method and Devices for Cardiac Radiofrequency Catheter Ablation |
EP2443992A2 (en) * | 2010-10-25 | 2012-04-25 | Fujifilm Corporation | Diagnosis support apparatus, diagnosis support method, lesioned part detection apparatus, and lesioned part detection method |
TWI460679B (zh) * | 2010-12-23 | 2014-11-11 | Nat Univ Chung Hsing | Section length enhancement unit and method |
KR101993716B1 (ko) * | 2012-09-28 | 2019-06-27 | 삼성전자주식회사 | 카테고리별 진단 모델을 이용한 병변 진단 장치 및 방법 |
KR20150002284A (ko) * | 2013-06-28 | 2015-01-07 | 삼성전자주식회사 | 병변 검출 장치 및 방법 |
US10049451B2 (en) * | 2015-12-02 | 2018-08-14 | The Cleveland Clinic Foundation | Automated lesion segmentation from MRI images |
CN106709929A (zh) * | 2016-12-30 | 2017-05-24 | 上海联影医疗科技有限公司 | 医学图像感兴趣区域显示方法及装置 |
US10593051B2 (en) * | 2017-12-20 | 2020-03-17 | International Business Machines Corporation | Medical image registration guided by target lesion |
US10468142B1 (en) * | 2018-07-27 | 2019-11-05 | University Of Miami | Artificial intelligence-based system and methods for corneal diagnosis |
CN109671053A (zh) * | 2018-11-15 | 2019-04-23 | 首都医科大学附属北京友谊医院 | 一种胃癌图像识别系统、装置及其应用 |
CN110110808B (zh) | 2019-05-16 | 2022-04-15 | 京东方科技集团股份有限公司 | 一种对图像进行目标标注的方法、装置及计算机记录介质 |
-
2019
- 2019-08-22 CN CN201910778807.0A patent/CN110490262B/zh active Active
-
2020
- 2020-08-17 WO PCT/CN2020/109611 patent/WO2021032062A1/zh active Application Filing
- 2020-08-17 US US17/423,439 patent/US11887303B2/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2005339075A (ja) * | 2004-05-26 | 2005-12-08 | Nippon Telegr & Teleph Corp <Ntt> | 画像探索装置、画像探索方法、この方法のプログラム及びこのプログラムを記録した記録媒体 |
JP2010067252A (ja) * | 2008-08-15 | 2010-03-25 | Fuji Xerox Co Ltd | オブジェクト領域抽出装置及びオブジェクト領域抽出プログラム |
CN108564567A (zh) * | 2018-03-15 | 2018-09-21 | 中山大学 | 一种超高分辨率病理图像癌变区域可视化方法 |
CN108961296A (zh) * | 2018-07-25 | 2018-12-07 | 腾讯科技(深圳)有限公司 | 眼底图像分割方法、装置、存储介质和计算机设备 |
CN109363698A (zh) * | 2018-10-16 | 2019-02-22 | 杭州依图医疗技术有限公司 | 一种乳腺影像征象识别的方法及装置 |
CN110136809A (zh) * | 2019-05-22 | 2019-08-16 | 腾讯科技(深圳)有限公司 | 一种医疗图像处理方法、装置、电子医疗设备和存储介质 |
CN110490262A (zh) * | 2019-08-22 | 2019-11-22 | 京东方科技集团股份有限公司 | 图像处理模型生成方法、图像处理方法、装置及电子设备 |
Also Published As
Publication number | Publication date |
---|---|
CN110490262B (zh) | 2022-06-03 |
CN110490262A (zh) | 2019-11-22 |
US11887303B2 (en) | 2024-01-30 |
US20220114724A1 (en) | 2022-04-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021032062A1 (zh) | 图像处理模型生成方法、图像处理方法、装置及电子设备 | |
CN114401666A (zh) | 基于深度学习的3d点云的对象检测和实例分割 | |
KR20180025093A (ko) | 약한 지도 학습 기반의 기계 학습 방법 및 그 장치 | |
Lei et al. | Attention-guided multi-branch convolutional neural network for mitosis detection from histopathological images | |
US11449717B2 (en) | System and method for identification and localization of images using triplet loss and predicted regions | |
CN110838108A (zh) | 基于医疗图像的预测模型构建方法、预测方法及装置 | |
WO2022057309A1 (zh) | 肺部特征识别方法、装置、计算机设备及存储介质 | |
JP6756406B2 (ja) | 画像処理装置、画像処理方法および画像処理プログラム | |
CN112132815B (zh) | 肺结节检测模型训练方法、检测方法及装置 | |
JP6892606B2 (ja) | 位置特定装置、位置特定方法及びコンピュータプログラム | |
CN114445356A (zh) | 基于多分辨率的全视野病理切片图像肿瘤快速定位方法 | |
CN115457492A (zh) | 目标检测方法、装置、计算机设备及存储介质 | |
Oloko-Oba et al. | Pre-trained convolutional neural network for the diagnosis of tuberculosis | |
CN112488178B (zh) | 网络模型的训练方法及装置、图像处理方法及装置、设备 | |
CN111985616B (zh) | 一种图像特征提取方法、图像检索方法、装置及设备 | |
CN112967293A (zh) | 一种图像语义分割方法、装置及存储介质 | |
CN110929731B (zh) | 一种基于探路者智能搜索算法的医疗影像处理方法及装置 | |
CN115527083B (zh) | 图像标注方法、装置和电子设备 | |
CN113111879B (zh) | 一种细胞检测的方法和系统 | |
CN112750124B (zh) | 模型生成、图像分割方法、装置、电子设备及存储介质 | |
CN113192085A (zh) | 三维器官图像分割方法、装置及计算机设备 | |
CN114912502B (zh) | 一种基于表情与语音的双模态深度半监督情感分类方法 | |
US20230281843A1 (en) | Generating depth images for image data | |
US20230196752A1 (en) | Information processing apparatus, information processing method, and non-transitory computer-readable storage medium | |
CN113706449B (zh) | 基于病理图像的细胞分析方法、装置、设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20854077 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20854077 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20854077 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 14.02.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20854077 Country of ref document: EP Kind code of ref document: A1 |