WO2021057174A1 - 图像处理方法及装置、电子设备、存储介质和计算机程序 - Google Patents
图像处理方法及装置、电子设备、存储介质和计算机程序 Download PDFInfo
- Publication number
- WO2021057174A1 WO2021057174A1 PCT/CN2020/100692 CN2020100692W WO2021057174A1 WO 2021057174 A1 WO2021057174 A1 WO 2021057174A1 CN 2020100692 W CN2020100692 W CN 2020100692W WO 2021057174 A1 WO2021057174 A1 WO 2021057174A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- processing
- sub
- result
- classification network
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/50—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for simulation or modelling of medical disorders
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30061—Lung
Definitions
- the embodiments of the present application relate to the field of computer vision technology, and relate to, but are not limited to, an image processing method and device, electronic equipment, computer storage media, and computer programs.
- the rapid development of deep learning has made remarkable achievements in the field of image processing.
- the current image processing technology based on deep learning requires multiple processing procedures to obtain a prediction of the extent of lesion development, which leads to cumbersome processing, and the accuracy of the resulting prediction results is not high.
- the embodiment of the application proposes an image processing method and device, electronic equipment, computer storage medium, and computer program.
- the embodiment of the application provides an image processing method, including:
- the change data is predicted to obtain the prediction result; the change data characterizes: the change situation of the target object in the first image processed based on the area distribution in the second image.
- the classification result can be obtained by using a target classification network, thereby simplifying the image processing process, and the classification result is obtained by inputting the first image and the second image into the target classification network jointly. , And predict the change of the target object in the first image based on the regional distribution in the second image according to the classification result, thereby improving the accuracy of the prediction result.
- the method further includes:
- the trained classification network is used as the target classification network.
- the trained classification network is used as the target classification network, and the target classification network can be used to obtain the classification result.
- the target object in the first image is based on the target object in the second image.
- the regional distribution is processed to predict the changes obtained, thereby improving the accuracy of the prediction results.
- the training the classification network to obtain the trained classification network includes:
- an image to be processed is obtained; the first image and the second image are different types of image data;
- the image to be processed is input into the classification network for training as a training sample, and the trained classification network is obtained.
- the classification network is trained based on the image to be processed after the image superposition processing of the first image and the second image is performed. Since these two types of image data are comprehensively considered and training is performed in the classification network , Therefore, the accuracy of training can be obtained.
- the obtaining the image to be processed includes:
- the first sub-image data and the second sub-image data are used as the image to be processed.
- the first image and the second image are respectively cut according to the contour of the target object, and the cut first sub-image data and second sub-image data are obtained and used for training of the classification network , Can improve training efficiency.
- the first sub-image data and the second sub-image data are image data of the same size.
- the image data of the same size is used, and the pixel position alignment may not be required in the image superposition process, and the first sub-image data and the second sub-image data can be directly used to realize the image superposition, thereby improving The processing efficiency of image overlay.
- the inputting the to-be-processed image as a training sample into the classification network for training to obtain the trained classification network includes:
- the image data is converted into the corresponding histogram, and the histogram equalization process is performed, and the histogram distribution of the image can be changed to an approximately uniform distribution, thereby enhancing the contrast of the image and making the image More clear.
- the inputting the to-be-processed image as a training sample into the classification network for training to obtain the trained classification network includes:
- the distribution of the pixels can be summarized, so that the pixels that need to be processed can be limited to a preset range after being normalized.
- the normalization process is to make the subsequent series of processes more convenient and quicker, which is beneficial to accelerate the convergence speed of the classification network training.
- the classification network includes at least one classification processing module
- the inputting the to-be-processed image as a training sample into the classification network for training to obtain the trained classification network includes:
- the image to be processed is subjected to feature extraction, dimensionality reduction processing, and global average pooling processing through at least one classification processing module to obtain a loss function, and train the classification according to the back propagation of the loss function Network, thus, the target classification network is trained.
- each classification processing module includes at least a convolutional layer
- the performing feature extraction, dimensionality reduction processing, and global average pooling processing on the image to be processed through the at least one classification processing module to obtain a loss function includes:
- the loss function is obtained.
- the second processing result can be obtained according to the first processing result obtained after the dimensionality reduction processing, and the loss function can be obtained according to the second processing result and the manual labeling result, so as The back propagation of the function trains the classification network, thereby training the target classification network.
- each residual module when the classification processing module is a residual module, each residual module includes: a convolutional layer, a regularization layer, and an activation layer;
- the method further includes:
- the structure of the module includes: a convolutional layer, a regularization layer, and an activation layer, which are obtained after feature extraction corresponding to the convolutional layer by the residual module
- the second extraction result is obtained.
- the third extraction result can be obtained, so that according to the third extraction result
- the first processing result for calculating the loss function is obtained, and after the loss function is obtained, the classification network can be trained according to the back propagation of the loss function, thereby training the target classification network.
- the performing dimensionality reduction processing to obtain the first processing result includes:
- the third extraction result can be subjected to dimensionality reduction processing to obtain the first processing result for calculating the loss function.
- the training station can be trained according to the back propagation of the loss function.
- the classification network is described, thus, the target classification network is obtained by training.
- An embodiment of the present application also provides an image processing device, which includes:
- the classification part is configured to input the first image and the second image into the target classification network to obtain a classification result
- the prediction part is configured to predict the change data according to the classification result to obtain the prediction result; the change data characterization: the target object in the first image is processed based on the area distribution in the second image The change situation.
- the device further includes a training part configured to:
- the trained classification network is used as the target classification network.
- the training part includes:
- the superposition sub-part is configured to perform image superposition processing on the first image and the second image to obtain an image to be processed; the first image and the second image are different types of image data;
- the training sub-part is configured to input the to-be-processed image as a training sample into the classification network for training to obtain the trained classification network.
- the training part further includes:
- the cutting sub-part is configured to perform image cutting on the first image and the second image respectively according to the contour of the target object to obtain the first sub-image data and the second sub-image data after being cut;
- the first sub-image data and the second sub-image data are used as the image to be processed.
- the first sub-image data and the second sub-image data are image data of the same size.
- the training part further includes:
- the equalization processing sub-part is configured to convert the first sub-image data and the second sub-image data into corresponding histograms respectively, and perform equalization processing of the histograms to obtain an equalization processing result; based on the The equalization processing result trains the classification network to obtain the trained classification network.
- the training part further includes:
- the normalization processing sub-part is configured to perform normalization processing on the corresponding pixels contained in the first sub-image data and the second sub-image data to obtain a normalization processing result; based on the normalization
- the processing result trains the classification network to obtain the trained classification network.
- the classification network includes at least one classification processing module
- the training sub-part is configured as:
- the image to be processed is subjected to feature extraction, dimensionality reduction processing, and global average pooling processing through the at least one classification processing module to obtain a loss function; the classification network is trained according to the back propagation of the loss function to obtain the Describe the classification network after training.
- each classification processing module includes at least a convolutional layer
- the training sub-part is configured as:
- each residual module when the classification processing module is a residual module, each residual module includes: a convolution layer, a regularization layer, and an activation layer;
- the training sub-part is configured as:
- the first extraction result obtained by performing feature extraction on the image to be processed through the corresponding convolutional layer in at least one residual module is processed by the regularization layer and the activation layer to obtain the second extraction result; according to the second
- the extraction result and the image to be processed are extracted to obtain a third extraction result.
- the training sub-part is configured to perform dimensionality reduction processing according to the third extraction result to obtain the first processing result.
- An embodiment of the application also provides an electronic device, including:
- a memory configured to store executable instructions of the processor
- the processor is configured to execute any one of the above-mentioned image processing methods.
- An embodiment of the present application also provides a computer-readable storage medium on which computer program instructions are stored, and when the computer program instructions are executed by a processor, any one of the above-mentioned image processing methods is implemented.
- the embodiment of the present application also provides a computer program, including computer-readable code, when the computer-readable code runs in an electronic device, the processor in the electronic device executes any one of the above-mentioned image processing method.
- the first image and the second image are input into the target classification network to obtain the classification result; according to the classification result, the target object in the first image is based on the area distribution in the second image
- the changes obtained by the processing are predicted, and the predicted results are obtained. Since the classification result can be obtained by using a target classification network, the technical solution of the embodiment of the present application simplifies the image processing process, because the classification result is obtained by jointly inputting the first image and the second image into the target classification network , And predict the change of the target object in the first image based on the regional distribution in the second image according to the classification result. Therefore, the technical solution of the embodiment of the application is adopted to improve the prediction The accuracy of the results.
- FIG. 1 is a schematic flowchart of an image processing method provided by an embodiment of the application
- Figure 2 is a schematic diagram of an application scenario of an embodiment of the application
- FIG. 3 is a schematic diagram of a training process of a classification network provided by an embodiment of the application.
- FIG. 4 is a schematic diagram of a classification network architecture for implementing an image processing method provided by an embodiment of the application
- FIG. 5 is a schematic structural diagram of an image processing device provided by an embodiment of the application.
- FIG. 6 is a schematic structural diagram of an electronic device according to an embodiment of the application.
- FIG. 7 is a schematic structural diagram of another electronic device according to an embodiment of the application.
- the prediction result may be a prediction result of the degree of influence on the designated area when the lesions in the designated area are processed.
- the lesions can be in the abdomen, lungs, kidneys, brain, heart, etc. If the designated area is the lungs, the location where the lesions exist in the lungs needs to be surgically treated to affect the lungs (such as severity or No severity, etc.) to make predictions.
- Radiation therapy to the lungs may cause radiation pneumonitis.
- Radiation pneumonitis is caused by damage to normal lung tissue in the radiation field after radiotherapy (such as lung cancer, breast cancer, esophageal cancer, malignant lymphoma or other malignant tumors of the chest).
- the inflammation caused by it Mild cases are asymptomatic, and inflammation can dissipate on its own; in severe cases, extensive fibrosis occurs in the lungs, leading to respiratory damage and even respiratory failure.
- the degree of inflammatory response is closely related to the radiation dose and the state of the lesion before radiotherapy. It is necessary to predict the severity of radiation pneumonia after radiotherapy for lung cancer.
- the extraction of image features can be achieved through radiomics.
- the extraction of image features through radioomics is to extract image features through radiographic imaging methods, and then study the relationship between the image features and clinical symptoms (such as the prediction of the degree of influence of the designated area).
- the support vector machine (SVM) and other classifiers can be used to predict the degree of impact (such as severity or non-severity, etc.) on the specified area. It can be seen that the entire image processing process includes multiple stages, which is not only cumbersome, but also the accuracy of the prediction results obtained therefrom is not high. The accuracy is not high because:
- the prediction of radiation dose used in radiotherapy is carried out separately from the prediction of the image processing process.
- the dose constant can be obtained through the method of averaging the radiation dose in the entire lung to achieve the above prediction process.
- gray ( Gray, Gy) units are used to measure the absorbed dose of radiation.
- Doctors can count the percentage of tissues whose absorbed dose in the lung exceeds a certain value to the entire lung as the dose constant.
- V20 is the percentage of tissue volume whose absorbed dose in the lung exceeds 20Gy to the total lung volume.
- the treatment method using this dose constant is too general, and does not consider the different dose sizes at different lesions.
- the lesions are in different areas such as the abdomen, lungs, kidneys, brain, heart, etc., and the dose sizes used are different.
- radiation The impact afterwards is also different.
- the internal radiation dose of the entire lung is relatively small, and the statistical constant is also relatively small, when the radiation hits key organs, such as the main trachea, blood vessels, and heart, it will also cause serious consequences.
- the constant used in the treatment method of the dose constant is only a statistic, and does not take into account the spatial distribution of the radiation in different regions. As a result, the prediction accuracy rate obtained by the treatment method of the dose constant is also inferior. not tall.
- the task of predicting the severity of pneumonia after radiotherapy is mainly solved by radioomics, which has disadvantages such as low efficiency, low robustness, failure to consider radiation distribution, and low accuracy.
- radioomics methods give ideal accuracy, the optimization of feature selection, dose constants, and hyperparameter selection of SVM in the process make the method less robust and difficult to apply to other data sets.
- the radiation dose can be constant, that is, the radiation dose in the entire lung or cancer area is counted as a constant, but this calculation loses the radiation distribution characteristics.
- the target classification network (such as a classification neural network, which may be three-dimensional) obtained after deep learning training can combine the lung image with
- the ray distribution image (both images can be three-dimensional images) is input into the target classification network at the same time, so that the target classification network can be used to comprehensively obtain the image and the image of the designated area where the lesion is located or the associated area at each location.
- Ray distribution to improve the prediction accuracy, and through the classification of the target classification network, the severity of pneumonia that will occur after radiotherapy can be directly output in one step.
- the embodiment of this application not only considers the distribution of radiation dose in the image processing prediction process, but also can be widely used on similar task data sets.
- the embodiment of this application can be directly applied Predict the severity of radiation pneumonia without changing any parameters and structures, and the application scenarios are not limited to the prediction of lesions in different areas or related areas such as the abdomen, lungs, kidneys, brain, heart, etc., and can be quickly obtained Accurate prediction results.
- FIG. 1 is a schematic flow chart of an image processing method provided by an embodiment of the application.
- the image processing method is applied to an image processing apparatus.
- the image processing apparatus can be executed by a terminal device or a server or other processing device, where the terminal device can be a user Equipment (UE, User Equipment), mobile devices, cellular phones, cordless phones, personal digital processing (PDA, Personal Digital Assistant), handheld devices, computing devices, in-vehicle devices, wearable devices, etc.
- the image processing method may be implemented by a processor calling computer-readable instructions stored in a memory. As shown in Figure 1, the process includes:
- Step S101 Input the first image and the second image into a target classification network to obtain a classification result.
- the first image may be an image with a lesion (it may be a CT image of the area where the lesion is located), such as the lesion in different areas such as the abdomen, lung, kidney, brain, heart, etc. or related
- the image of the area may be a radiation dose distribution map used to perform radiotherapy on the area where the lesion is located or the associated area.
- the two images, the image with the lesion and the radiation dose distribution map can be jointly input into the target classification network to obtain the classification result.
- Step S102 Predict the change data according to the classification result to obtain the prediction result.
- the change data represents: the change of the target object in the first image based on the area distribution in the second image.
- the target object may be the organ where the lesion is located, such as the abdomen, lung, kidney, brain, heart, etc.
- the regional distribution can be the distribution of radiation doses used for different lesions in different regions.
- the change may be the severity (such as the probability of being severe or the probability of not being serious) that may cause inflammation to the organ (such as the lung) where the lesion is located once radiotherapy is performed on the lesion.
- the severity of inflammation in the lungs in the image with the lesion based on the radiation distribution in the radiation dose distribution map can be predicted to obtain the prediction result.
- prediction can be realized only by classification according to the target classification network, and the prediction result can be obtained end-to-end in one step, without the need for multi-stage cumbersome operations.
- the different effects of different radiation doses at different locations improve the accuracy of prediction.
- the target classification network does not require artificial hyperparameters to control, but can use the target classification network obtained after deep learning training, which realizes the adaptive control of prediction in the entire image processing process, which helps to improve the prediction accuracy.
- FIG. 2 is a schematic diagram of an application scenario of an embodiment of the present application, as shown in FIG. 2, a CT image of a lung cancer patient 201 is the above first image, and the radiation dose distribution map 202 is the above second image.
- the image processing method described in the foregoing embodiment can be used to predict the severity of radiation pneumonia after radiotherapy. The severity of radiation pneumonia will be predicted. Therefore, it can help doctors predict the postoperative risk in advance, take precautions in advance or revise the radiotherapy plan.
- a classification network may be trained to obtain a trained classification network; the trained classification network is used as the target classification network.
- FIG. 3 is a schematic diagram of the training process of the classification network provided by an embodiment of the application.
- the training process of the classification network can be implemented based on an image processing device.
- the image processing device can be executed by a terminal device or a server or other processing device. It can be a UE, a mobile device, a cellular phone, a cordless phone, a PDA, a handheld device, a computing device, a vehicle-mounted device, a wearable device, etc.
- the image processing method may be implemented by a processor calling computer-readable instructions stored in a memory.
- the training process of the classification network may include:
- Step S301 After performing image superposition processing on the first image and the second image, an image to be processed is obtained.
- the first image and the second image may be different types of image data
- image cutting may be performed on the first image and the second image respectively according to the contour of the target object, Obtain the cut first sub-image data and second sub-image data.
- the outline of the target object is not the outline of the lesion, but the outline of the area where the lesion is located or the associated area, such as the entire lung outline of the lung where the lesion is located, or the outline of the heart or kidney where the lesion is located, depending on the difference. Area, using different doses.
- the first sub-image data and the second sub-image data may be image data of the same size. For example, according to the lung contour, the three-dimensional first image and the second image are cut, and the two are the same size. Then, the first sub-image data and the second sub-image data can be used as the image to be processed.
- the first sub-image data and the second sub-image data may also be image data of different sizes.
- the first sub-image data and the second sub-image data may be combined
- the corresponding pixels contained in the second sub-image data are subjected to pixel position alignment processing to obtain the aligned first sub-image data and the aligned second sub-image data, and the aligned first sub-image data and the aligned second sub-image data are obtained.
- the image to be processed is obtained.
- Step S302 Input the to-be-processed image as a training sample into the classification network for training, to obtain the trained classification network.
- the classification network is trained through step S301 to step S302, and a trained classification network can be obtained.
- the trained classification network is used as the target classification network.
- inputting the to-be-processed image as a training sample into the classification network for training, and obtaining the trained classification network may include: combining the first sub-image data and the second sub-image data Converting into corresponding histograms respectively, and performing equalization processing of the histograms to obtain equalization processing results; training the classification network based on the equalization processing results to obtain the trained classification network.
- inputting the to-be-processed image as a training sample into the classification network for training, and obtaining the trained classification network may include: combining the first sub-image data and the second sub-image data Normalization processing is performed on the corresponding pixels contained in to obtain a normalization processing result; the classification network is trained based on the normalization processing result to obtain the trained classification network.
- the first sub-image data and the second sub-image data may be respectively converted into corresponding histograms, and the histogram equalization processing is performed to obtain the equalization processing result. Then, the corresponding pixel points contained in the first sub-image data and the second sub-image data in the equalization processing result can be normalized. For example, after performing histogram equalization on the two sub-images, the two sub-images are normalized and then combined into image data represented by a two-channel four-dimensional matrix. The image data is input into the classification network, and the image data is extracted layer by layer through the convolutional layer of the classification network and the dimensionality is reduced. Finally, through the processing of the fully connected layer, the severity of radiation inflammation after radiotherapy is obtained. Probability.
- the classification network may include at least one classification processing module
- inputting the to-be-processed image as a training sample into the classification network for training to obtain the target classification network may include: extracting the features of the to-be-processed image through the at least one classification processing module (such as by The convolutional layer performs feature extraction), dimensionality reduction processing (such as pooling processing), and global average pooling processing to obtain the loss function; training the classification network according to the back propagation of the loss function (for example, according to the loss The error calculated by the function is back-propagated) to obtain the trained classification network.
- the convolutional layer performs feature extraction
- dimensionality reduction processing such as pooling processing
- global average pooling processing global average pooling processing
- each classification processing module includes at least a convolutional layer; the image to be processed is subjected to feature extraction, dimensionality reduction processing, and global average pooling processing through the at least one classification processing module to obtain a loss
- the function may include: performing layer-by-layer feature extraction of the image to be processed through the corresponding convolutional layer in at least one classification processing module, and then performing layer-by-layer dimensionality reduction processing to obtain the first processing result; and performing the global processing result on the first processing result. After averaging pooling, it is input to the fully connected layer to obtain the second processing result.
- the second processing result is the prediction result output by the classification network, which is used to characterize the predicted change of the extracted features; according to the second processing result and the manual annotation result (such as The actual changes marked by the doctor), the loss function is obtained.
- the loss function can be obtained according to the prediction result output by the classification network and the actual changes marked by the doctor. If the error between the predicted change reflected by the loss function and the real situation is within the preset range (for example, the error is zero), it means that the difference between the generated predicted change and the real situation has reached the convergence condition, and the classification network After the training is over, the target classification network after training is obtained.
- each residual module may include: a convolution layer, a regularization layer, and an activation layer.
- the second extraction result is obtained after the regularization layer and the activation layer are processed; according to the second extraction result and the image to be processed, the third extraction result is obtained, and the third extraction result is used for dimensionality reduction deal with. That is, the input of the residual module is the "image to be processed", and the final extraction result obtained by adding the input of the residual module and the output of the last active layer in the residual module is the third extraction result.
- layer-by-layer dimensionality reduction processing can also be performed to obtain the first processing result. For example, performing layer-by-layer dimensionality reduction processing according to the third extraction result to obtain the first processing result.
- FIG. 4 is a schematic diagram of a classification network architecture for implementing an image processing method provided by an embodiment of the application.
- the classification network (such as a classification neural network) may include at least one classification processing module 11.
- the classification processing module 11 may be a residual module 12 and may also include a fully connected layer 13.
- Each residual module 12 may include: at least one convolutional layer 121, at least one regularization layer 122, and at least one activation layer 123; the classification network automatically learns useful features in the extracted image, and uses these features to make predictions, Instead of extracting features and then selecting features, compared with related technologies, the prediction accuracy is improved.
- the lung image and the radiation distribution image (the two images can be three-dimensional images) are cut into the same according to the lung contour. Two sub-images of the size. The two sub-images are histogram equalized, and the two sub-images are normalized and then combined into a two-channel four-dimensional matrix.
- the four-dimensional matrix is input into the classification network, and the image is processed by the convolutional layer of the classification processing module 11 (specifically, the convolutional layer 121, the regularization layer 122, and the activation layer 123 in each residual module 12).
- the layer performs feature extraction and dimensionality reduction processing.
- the fully connected layer 13 obtains the probability of the severity of radiation pneumonia after radiotherapy.
- the process of training a classification network may include:
- the three-dimensional image of the lungs and the radiation dose distribution image are cropped into images of the same size (200x240x240) according to the lung contours, and then the image size (100x120x120) is obtained by downsampling to fit the video memory, and the lungs
- the down-sampled image corresponding to the partial image and the down-sampled image corresponding to the ray distribution image are combined into a four-dimensional matrix (2x100x120x120).
- Three-dimensional convolutional neural networks such as Res-Net (ResNeXt50 as shown in Figure 4), Dense-Net and other structures can be used to perform convolution, regularization and activation operations on the connected four-dimensional matrix, and change the feature channel from 2 One is increased to 2048, and then a one-dimensional vector is obtained by global average pooling of the features, and the one-dimensional vector is input into the fully connected layer to output two values (probability of serious or not serious), and finally through the softmax function Get the final prediction result (prediction probability).
- Res-Net Res-Net
- Dense-Net Dense-Net and other structures can be used to perform convolution, regularization and activation operations on the connected four-dimensional matrix, and change the feature channel from 2 One is increased to 2048, and then a one-dimensional vector is obtained by global average pooling of the features, and the one-dimensional vector is input into the fully connected layer to output two values (probability of serious or not serious), and finally through the softmax function Get the final prediction result (prediction
- the classification network may adopt a serialized and modularized neural network.
- Serialization refers to the sequential processing of data input to the neural network according to the serialized modules in the neural network (such as a four-dimensional matrix obtained by cutting two sub-images of the same size and connecting them), and modularity refers to the modules in the neural network can be free
- modules in the neural network can be free
- the image obtained by joining the two sub-images cut into the same size is equivalent to a four-dimensional matrix (it can be a four-dimensional matrix with two channels).
- the feature extraction through the convolution layer can be to use at least one convolution kernel to convolve the input four-dimensional matrix, and the output channel is the four-dimensional matrix with the number of convolution kernels. As the number of convolution kernels increases, the matrix The number of channels has also increased until 2048. Regularization is performed through the regularization layer, and the four-dimensional matrix can be regularized using formula (1):
- X is a four-dimensional matrix
- u is the mean of the matrix
- v is the variance of the matrix
- the activation function can be used to add non-linear factors to improve the expressive ability of the neural network.
- Global average pooling is to average the three-dimensional matrix of each channel to obtain a one-dimensional vector whose length is the number of channels. After the global average pooling of the features is performed to obtain a one-dimensional vector, the one-dimensional vector can be calculated through the neural network through the fully connected layer, and finally two values (probability of serious or not serious) are obtained, and finally the multi-classified by softmax The output value (probability of serious or not serious) is transformed into relative probability and used as the final prediction result.
- the deep learning optimizer (such as the Adam optimizer) calculates the updated difference and adds it to the original parameters to achieve the update of the classification network parameters. This process is continuously iterated until the error is within the preset range (for example, the error is zero), and the classification network reaches Convergence. In this way, the trained target classification network can be obtained.
- the writing order of the steps does not mean a strict execution order but constitutes any limitation on the implementation process.
- the specific execution order of each step should be based on its function and possibility.
- the inner logic is determined.
- embodiments of the present application also provide image processing apparatuses, electronic equipment, computer-readable storage media, and programs, all of which can be used to implement any image processing method provided in the embodiments of the present application.
- image processing apparatuses electronic equipment, computer-readable storage media, and programs, all of which can be used to implement any image processing method provided in the embodiments of the present application.
- the corresponding technical solutions and descriptions and refer to the methods Part of the corresponding records will not be repeated here.
- FIG. 5 is a schematic structural diagram of an image processing device provided by an embodiment of the application.
- the image processing device of an embodiment of the application may include: a classification part 31 configured to input the first image and the second image into the target The classification network obtains the classification result; the prediction part 32 is configured to predict the change data according to the classification result to obtain the prediction result; the change data characterization: the target object in the first image is based on the second image The regional distribution in the processing of the changes obtained.
- the device further includes a training part configured to:
- the trained classification network is used as the target classification network.
- the training part includes:
- the superposition sub-part is configured to perform image superposition processing on the first image and the second image to obtain an image to be processed; the first image and the second image are different types of image data;
- the training sub-part is configured to input the to-be-processed image as a training sample into the classification network for training to obtain the trained classification network.
- the training part further includes:
- the cutting sub-part is configured to perform image cutting on the first image and the second image respectively according to the contour of the target object to obtain the first sub-image data and the second sub-image data after being cut;
- the first sub-image data and the second sub-image data are used as the image to be processed.
- the first sub-image data and the second sub-image data are image data of the same size.
- the training part further includes:
- the equalization processing sub-part is configured to convert the first sub-image data and the second sub-image data into corresponding histograms respectively, and perform equalization processing of the histograms to obtain an equalization processing result; based on the The equalization processing result trains the classification network to obtain the trained classification network.
- the training part further includes:
- the normalization processing sub-part is configured to perform normalization processing on the corresponding pixels contained in the first sub-image data and the second sub-image data to obtain a normalization processing result; based on the normalization
- the processing result trains the classification network to obtain the trained classification network.
- the classification network includes at least one classification processing module
- the training sub-part is configured to perform feature extraction, dimensionality reduction processing, and global average pooling processing on the image to be processed through the at least one classification processing module to obtain a loss function; back propagation training according to the loss function The classification network to obtain the trained classification network.
- each of the classification processing modules includes at least a convolutional layer
- the training sub-part is configured to perform feature extraction on the image to be processed through the corresponding convolutional layer in the at least one classification processing module, and then perform dimensionality reduction processing to obtain a first processing result;
- the result is input to the fully connected layer after the global average pooling process, and the second processing result is obtained, and the second processing result is used to characterize the predicted change of the extracted features; according to the second processing result and the manual labeling result, the result is obtained.
- the loss function is configured to perform feature extraction on the image to be processed through the corresponding convolutional layer in the at least one classification processing module, and then perform dimensionality reduction processing to obtain a first processing result;
- the result is input to the fully connected layer after the global average pooling process, and the second processing result is obtained, and the second processing result is used to characterize the predicted change of the extracted features; according to the second processing result and the manual labeling result, the result is obtained.
- the loss function is configured to perform feature extraction on the image to be processed through the corresponding convolutional layer in the at
- each residual module when the classification processing module is a residual module, each residual module includes: a convolution layer, a regularization layer, and an activation layer;
- the training sub-part is configured to perform feature extraction on the image to be processed through the corresponding convolutional layer in at least one residual module to obtain the first extraction result, after processing the regularization layer and the activation layer, to obtain the second extraction result.
- Extraction result According to the second extraction result and the image to be processed, a third extraction result is obtained.
- the training sub-part is configured to perform dimensionality reduction processing according to the third extraction result to obtain the first processing result.
- the functions or modules contained in the apparatus provided in the embodiments of the present application can be used to execute the methods described in the above method embodiments.
- the functions or modules contained in the apparatus provided in the embodiments of the present application can be used to execute the methods described in the above method embodiments.
- the embodiment of the present application also proposes a computer-readable storage medium on which computer program instructions are stored, and when the computer program instructions are executed by a processor, any one of the above-mentioned image processing methods is implemented.
- the computer-readable storage medium may be a non-volatile computer-readable storage medium.
- An embodiment of the present application also proposes an electronic device, including: a processor; a memory configured to store executable instructions of the processor; wherein the processor is configured to be any of the foregoing image processing methods.
- the electronic device can be a terminal, a server, or other types of devices.
- An embodiment of the present application also proposes a computer program, including computer-readable code, and when the computer-readable code runs in an electronic device, a processor in the electronic device executes any one of the above-mentioned image processing methods.
- FIG. 6 is a schematic structural diagram of an electronic device according to an embodiment of the application.
- the electronic device 800 can be a mobile phone, a computer, a digital broadcasting terminal, a messaging device, a game console, a tablet device, a medical device, and a fitness device.
- Terminals such as equipment, personal digital assistants, etc.
- the electronic device 800 may include one or more of the following components: a first processing component 802, a first memory 804, a first power supply component 806, a multimedia component 808, an audio component 810, a first input/output (Input Output, I/O) interface 812, sensor component 814, and communication component 816.
- a first processing component 802 a first memory 804, a first power supply component 806, a multimedia component 808, an audio component 810, a first input/output (Input Output, I/O) interface 812, sensor component 814, and communication component 816.
- the first processing component 802 generally controls the overall operations of the electronic device 800, such as operations associated with display, telephone calls, data communication, camera operations, and recording operations.
- the first processing component 802 may include one or more processors 820 to execute instructions to complete all or part of the steps of the foregoing method.
- the first processing component 802 may include one or more modules to facilitate the interaction between the first processing component 802 and other components.
- the first processing component 802 may include a multimedia module to facilitate the interaction between the multimedia component 808 and the first processing component 802.
- the first memory 804 is configured to store various types of data to support operations in the electronic device 800. Examples of these data include instructions for any application or method operating on the electronic device 800, contact data, phone book data, messages, pictures, videos, etc.
- the first memory 804 can be implemented by any type of volatile or non-volatile storage device or a combination thereof, such as static random access memory (Static Random-Access Memory, SRAM), electrically erasable programmable read-only memory (Electrically Erasable Programmable Read Only Memory, EEPROM), Erasable Programmable Read-Only Memory (Electrical Programmable Read Only Memory, EPROM), Programmable Read-Only Memory (Programmable Read-Only Memory, PROM), Read-Only Memory (Read-Only Memory) Only Memory, ROM), magnetic memory, flash memory, magnetic disk or optical disk.
- SRAM static random access memory
- SRAM static random access memory
- EEPROM Electrically erasable programmable read-only memory
- EEPROM Electrically Erasable Programmable
- the first power supply component 806 provides power for various components of the electronic device 800.
- the first power supply component 806 may include a power management system, one or more power supplies, and other components associated with generating, managing, and distributing power for the electronic device 800.
- the multimedia component 808 includes a screen that provides an output interface between the electronic device 800 and the user.
- the screen may include a liquid crystal display (LCD) and a touch panel (TP). If the screen includes a touch panel, the screen may be implemented as a touch screen to receive input signals from the user.
- the touch panel includes one or more touch sensors to sense touch, sliding, and gestures on the touch panel. The touch sensor can not only sense the boundary of a touch or slide action, but also detect the duration and pressure related to the touch or slide operation.
- the multimedia component 808 includes a front camera and/or a rear camera. When the electronic device 800 is in an operation mode, such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front camera and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
- the audio component 810 is configured to output and/or input audio signals.
- the audio component 810 includes a microphone (MIC), and when the electronic device 800 is in an operation mode, such as a call mode, a recording mode, and a voice recognition mode, the microphone is configured to receive an external audio signal.
- the received audio signal may be further stored in the first memory 804 or transmitted via the communication component 816.
- the audio component 810 further includes a speaker for outputting audio signals.
- the first input/output interface 812 provides an interface between the first processing component 802 and a peripheral interface module.
- the peripheral interface module may be a keyboard, a click wheel, a button, and the like. These buttons may include, but are not limited to: home button, volume button, start button, and lock button.
- the sensor component 814 includes one or more sensors for providing the electronic device 800 with various aspects of state evaluation.
- the sensor component 814 can detect the on/off status of the electronic device 800 and the relative positioning of the components.
- the component is the display and the keypad of the electronic device 800.
- the sensor component 814 can also detect the electronic device 800 or the electronic device 800.
- the position of the component changes, the presence or absence of contact between the user and the electronic device 800, the orientation or acceleration/deceleration of the electronic device 800, and the temperature change of the electronic device 800.
- the sensor component 814 may include a proximity sensor configured to detect the presence of nearby objects when there is no physical contact.
- the sensor component 814 may also include a light sensor, such as a complementary metal oxide semiconductor (Complementary Metal Oxide Semiconductor, CMOS) or a charge coupled device (Charge Coupled Device, CCD) image sensor for use in imaging applications.
- CMOS Complementary Metal Oxide Semiconductor
- CCD Charge Coupled Device
- the sensor component 814 may also include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
- the communication component 816 is configured to facilitate wired or wireless communication between the electronic device 800 and other devices.
- the electronic device 800 can access a wireless network based on a communication standard, such as WiFi, 2G, or 3G, or a combination thereof.
- the communication component 816 receives a broadcast signal or broadcast related information from an external broadcast management system via a broadcast channel.
- the communication component 816 further includes a Near Field Communication (NFC) module to facilitate short-range communication.
- the NFC module can be based on Radio Frequency Identification (RFID) technology, Infrared Data Association (Infrared Data Association, IrDA) technology, Ultra Wide Band (UWB) technology, Bluetooth (Bluetooth, BT) technology and other technologies. Technology to achieve.
- RFID Radio Frequency Identification
- IrDA Infrared Data Association
- UWB Ultra Wide Band
- Bluetooth Bluetooth
- the electronic device 800 may be used by one or more application specific integrated circuits (ASIC), digital signal processors (Digital Signal Processor, DSP), and digital signal processing equipment (Digital Signal Process, DSPD), programmable logic device (Programmable Logic Device, PLD), Field Programmable Gate Array (Field Programmable Gate Array, FPGA), controller, microcontroller, microprocessor or other electronic components to implement the above methods .
- ASIC application specific integrated circuits
- DSP Digital Signal Processor
- DSPD digital signal processing equipment
- PLD programmable logic device
- FPGA Field Programmable Gate Array
- controller microcontroller, microprocessor or other electronic components to implement the above methods .
- a non-volatile computer-readable storage medium is also provided, such as the first memory 804 including computer program instructions, which can be executed by the processor 820 of the electronic device 800 to accomplish any of the foregoing.
- An image processing method An image processing method.
- FIG. 7 is a schematic structural diagram of another electronic device according to an embodiment of the application.
- the electronic device 900 may be provided as a server.
- the electronic device 900 includes a second processing component 922, which further includes one or more processors, and a memory resource represented by the second memory 932, for storing instructions that can be executed by the second processing component 922, For example, applications.
- the application program stored in the second memory 932 may include one or more modules each corresponding to a set of instructions.
- the second processing component 922 is configured to execute instructions to perform the above-mentioned method.
- the electronic device 900 may also include a second power supply component 926 configured to perform power management of the electronic device 900, a wired or wireless network interface 950 configured to connect the electronic device 900 to the network, and a second input and output (I/O ) Interface 958.
- the electronic device 900 may operate based on an operating system stored in the second storage 932, such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM or the like.
- a non-volatile computer-readable storage medium is also provided, such as a second memory 932 including computer program instructions, which can be executed by the second processing component 922 of the electronic device 900 to complete The above method.
- the embodiments of this application may be systems, methods and/or computer program products.
- the computer program product may include a computer-readable storage medium loaded with computer-readable program instructions for enabling a processor to implement various aspects of the present application.
- the computer-readable storage medium may be a tangible device that can hold and store instructions used by the instruction execution device.
- the computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- Non-exhaustive list of computer-readable storage media include: portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable programmable read-only memory (EPROM) Or flash memory), static random access memory (SRAM), portable compact disk read-only memory (CD-ROM), digital versatile disk (Digital Video Disc, DVD), memory stick, floppy disk, mechanical encoding device, such as storage on it Commanded punch card or raised structure in the groove, and any suitable combination of the above.
- RAM random access memory
- ROM read-only memory
- EPROM erasable programmable read-only memory
- flash memory flash memory
- SRAM static random access memory
- CD-ROM compact disk read-only memory
- DVD digital versatile disk
- memory stick floppy disk
- mechanical encoding device such as storage on it Commanded punch card or raised structure in the groove, and any suitable combination of the above.
- the computer-readable storage medium used here is not interpreted as the instantaneous signal itself, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through waveguides or other transmission media (for example, light pulses through fiber optic cables), or through wires Transmission of electrical signals.
- the computer-readable program instructions described herein can be downloaded from a computer-readable storage medium to various computing/processing devices, or downloaded to an external computer or external storage device via a network, such as the Internet, a local area network, a wide area network, and/or a wireless network.
- the network may include copper transmission cables, optical fiber transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers.
- the network adapter card or network interface in each computing/processing device receives computer-readable program instructions from the network, and forwards the computer-readable program instructions for storage in the computer-readable storage medium in each computing/processing device .
- the computer program instructions used to perform the operations of the embodiments of the present application may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine-related instructions, microcode, firmware instructions, state setting data, or one or more programming Source code or object code written in any combination of languages, the programming language includes object-oriented programming languages such as Smalltalk, C++, etc., and conventional procedural programming languages such as "C" language or similar programming languages.
- Computer-readable program instructions can be executed entirely on the user's computer, partly on the user's computer, executed as a stand-alone software package, partly on the user's computer and partly executed on a remote computer, or entirely on the remote computer or server carried out.
- the remote computer can be connected to the user's computer through any kind of network-including Local Area Network (LAN) or Wide Area Network (WAN)-or it can be connected to an external computer (for example, Use an Internet service provider to connect via the Internet).
- electronic circuits such as programmable logic circuits, FPGAs, or programmable logic arrays (Programmable Logic Array, PLA), can be customized by using the status information of computer-readable program instructions. Read the program instructions to realize all aspects of the embodiments of the present application.
- These computer-readable program instructions can be provided to the processor of a general-purpose computer, a special-purpose computer, or other programmable data processing device, thereby producing a machine that makes these instructions when executed by the processor of the computer or other programmable data processing device , A device that implements the functions/actions specified in one or more blocks in the flowcharts and/or block diagrams is produced. It is also possible to store these computer-readable program instructions in a computer-readable storage medium. These instructions make computers, programmable data processing apparatuses, and/or other devices work in a specific manner. Thus, the computer-readable medium storing the instructions includes An article of manufacture, which includes instructions for implementing various aspects of the functions/actions specified in one or more blocks in the flowcharts and/or block diagrams.
- each block in the flowchart or block diagram may represent a module, program segment, or part of an instruction, and the module, program segment, or part of an instruction contains one or more components for realizing the specified logical function.
- Executable instructions may also occur in a different order than the order marked in the drawings. For example, two consecutive blocks can actually be executed substantially in parallel, or they can sometimes be executed in the reverse order, depending on the functions involved.
- each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart can be implemented by a dedicated hardware-based system that performs the specified functions or actions Or it can be realized by a combination of dedicated hardware and computer instructions.
- the embodiment of the application provides an image processing method and device, electronic equipment, computer storage medium, and computer program.
- the method includes: inputting a first image and a second image into a target classification network to obtain a classification result; As a result, the change data is predicted to obtain the prediction result; the change data characterizes: the change situation of the target object in the first image based on the area distribution in the second image.
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Medical Informatics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Databases & Information Systems (AREA)
- Pathology (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (25)
- 一种图像处理方法,所述方法包括:将第一图像和第二图像输入目标分类网络,得到分类结果;根据所述分类结果,对变化数据进行预测,得到预测结果;所述变化数据表征:所述第一图像中的目标对象基于所述第二图像中的区域分布进行处理所得到的变化情况。
- 根据权利要求1所述的方法,其中,所述方法还包括:对分类网络进行训练,得到训练后的分类网络;将所述训练后的分类网络作为所述目标分类网络。
- 根据权利要求2所述的方法,其中,所述对分类网络进行训练,得到训练后的分类网络,包括:将第一图像和第二图像进行图像叠加处理后,得到待处理图像;所述第一图像和所述第二图像为不同种类的图像数据;将所述待处理图像作为训练样本输入所述分类网络进行训练,得到所述训练后的分类网络。
- 根据权利要求3所述的方法,其中,所述得到待处理图像包括:根据所述目标对象的轮廓,对所述第一图像及所述第二图像分别进行图像切割,得到切割后的第一子图像数据和第二子图像数据;将所述第一子图像数据和所述第二子图像数据作为所述待处理图像。
- 根据权利要求4所述的方法,其中,所述第一子图像数据和所述第二子图像数据,为相同尺寸的图像数据。
- 根据权利要求4或5所述的方法,其中,所述将所述待处理图像作为训练样本输入所述分类网络进行训练,得到所述训练后的分类网络,包括:将所述第一子图像数据和所述第二子图像数据分别转换为对应的直方图,并进行直方图的均衡化处理,得到均衡化处理结果;基于所述均衡化处理结果对所述分类网络进行训练,得到所述训练后的分类网络。
- 根据权利要求4或5所述的方法,其中,所述将所述待处理图像作为训练样本输入所述分类网络进行训练,得到所述训练后的分类网络,包括:将所述第一子图像数据和所述第二子图像数据中包含的对应像素点进行归一化处理,得到归一化处理结果;基于所述归一化处理结果对所述分类网络进行训练,得到所述训练后的分类网络。
- 根据权利要求3-7任一项所述的方法,其中,所述分类网络包括至少一个分类处理模块;所述将所述待处理图像作为训练样本输入所述分类网络进行训练,得到所述训练后的分类网络,包括:将所述待处理图像通过所述至少一个分类处理模块进行特征提取、降维处理及全局平均池化处理,得到损失函数;根据所述损失函数的反向传播训练所述分类网络,以得到所述训练后的分类网络。
- 根据权利要求8所述的方法,其中,每个所述分类处理模块至少包括卷积层;所述将所述待处理图像通过所述至少一个分类处理模块进行特征提取、降维处理及全局平均池化处理,得到损失函数,包括:将所述待处理图像通过所述至少一个分类处理模块中的对应卷积层进行特征提取后,进行降维处理,得到第一处理结果;将所述第一处理结果进行全局平均池化处理后输入全连接层,得到第二处理结果,所述第二处理结果用于表征所提取特征的预测变化情况;根据所述第二处理结果和手工标注结果,得到所述损失函数。
- 根据权利要求9所述的方法,其中,在所述分类处理模块为残差模块的情况下,每个所述残差模块包括:卷积层、正则化层和激活层;所述将所述待处理图像通过所述至少一个分类处理模块中的对应卷积层进行特征提取后,还包括:将所述待处理图像通过至少一个残差模块中的对应卷积层进行特征提取后得到的第一提取结果,经过正则化层和激活层的处理后得到第二提取结果;根据所述第二提取结果和所述待处理图像,得到第三提取结果。
- 根据权利要求10所述的方法,其中,所述进行降维处理,得到第一处理结果,包括:根据所述第三提取结果,进行降维处理,得到所述第一处理结果。
- 一种图像处理装置,所述装置包括:分类部分,配置为将第一图像和第二图像输入目标分类网络,得到分类结果;预测部分,配置为根据所述分类结果,对变化数据进行预测,得到预测结果;所述变化数据表征:所述第一图像中的目标对象基于所述第二图像中的区域分布进行处理所得到的变化情况。
- 根据权利要求12所述的装置,其中,所述装置还包括训练部分,配置为:对分类网络进行训练,得到训练后的分类网络;将所述训练后的分类网络作为所述目标分类网络。
- 根据权利要求13所述的装置,其中,所述训练部分,包括:叠加子部分,配置为将第一图像和第二图像进行图像叠加处理后,得到待处理图像;所述第一图像和所述第二图像为不同种类的图像数据;训练子部分,配置为将所述待处理图像作为训练样本输入所述分类网络进行训练,得到所述训练后的分类网络。
- 根据权利要求14所述的装置,其中,所述训练部分,还包括:切割子部分,配置为根据所述目标对象的轮廓,对所述第一图像及所述第二图像分别进行图像切割,得到切割后的第一子图像数据和第二子图像数据;将所述第一子图像数据和所述第二子图像数据作为所述待处理图像。
- 根据权利要求15所述的装置,其中,所述第一子图像数据和所述第二子图像数据,为相同尺寸的图像数据。
- 根据权利要求15或16所述的装置,其中,所述训练部分,还包括:均衡化处理子部分,配置为将所述第一子图像数据和所述第二子图像数据分别转换为对应的直方图,并进行直方图的均衡化处理,得到均衡化处理结果;基于所述均衡化处理结果对所述分类网络进行训练,得到所述训练后的分类网络。
- 根据权利要求15或16所述的装置,其中,所述训练部分,还包括:归一化处理子部分,配置为将所述第一子图像数据和所述第二子图像数据中包含的对应像素点进行归一化处理,得到归一化处理结果;基于所述归一化处理结果对所述分类网络进行训练,得到所述训练后的分类网络。
- 根据权利要求14至18任一项所述的装置,其中,所述分类网络包括至少一个分类处理模块;所述训练子部分,配置为将所述待处理图像通过所述至少一个分类处理模块进行特征提取、降维处理及全局平均池化处理,得到损失函数;根据所述损失函数的反向传播训练所述分类网络,以得到所述训练后的分类网络。
- 根据权利要求19所述的装置,其中,每个所述分类处理模块至少包括卷积层;所述训练子部分,配置为将所述待处理图像通过所述至少一个分类处理模块中的对应卷积层进行特征提取后,进行降维处理,得到第一处理结果;将所述第一处理结果进行全局平均池化处理后输入全连接层,得到第二处理结果,所述第二处理结果用于表征所提取特征的预测变化情况;根据所述第二处理结果和手工标注结果,得到所述损失函数。
- 根据权利要求20所述的装置,其中,在所述分类处理模块为残差模块的情况下,每个残差模块包括:卷积层、正则化层和激活层;所述训练子部分,配置为将所述待处理图像通过至少一个残差模块中的对应卷积层进行特征提取后得到的第一提取结果,经过正则化层和激活层的处理,得到第二提取结果;根据所述第二提取结果和所述待处理图像,得到第三提取结果。
- 根据权利要求21所述的装置,其中,所述训练子部分,配置为根据所述第三提取结果,进行降维处理,得到所述第一处理结果。
- 一种电子设备,包括:处理器;配置为存储处理器可执行指令的存储器;其中,所述处理器被配置为:执行权利要求1至11中任意一项所述的方法。
- 一种计算机可读存储介质,其上存储有计算机程序指令,所述计算机程序指令被处理器执行时实现权利要求1至11中任意一项所述的方法。
- 一种计算机程序,包括计算机可读代码,当所述计算机可读代码在电子设备中运行时,所述电子设备中的处理器执行用于实现权利要求1至11中的任一权利要求所述的方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020217037806A KR20210153700A (ko) | 2019-09-26 | 2020-07-07 | 이미지 처리 방법 및 장치, 전자 기기, 저장 매체 및 컴퓨터 프로그램 |
JP2021563034A JP2022530413A (ja) | 2019-09-26 | 2020-07-07 | 画像処理方法及び装置、電子機器、記憶媒体並びにコンピュータプログラム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910918450.1 | 2019-09-26 | ||
CN201910918450.1A CN110705626A (zh) | 2019-09-26 | 2019-09-26 | 一种图像处理方法及装置、电子设备和存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021057174A1 true WO2021057174A1 (zh) | 2021-04-01 |
Family
ID=69197459
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/100692 WO2021057174A1 (zh) | 2019-09-26 | 2020-07-07 | 图像处理方法及装置、电子设备、存储介质和计算机程序 |
Country Status (5)
Country | Link |
---|---|
JP (1) | JP2022530413A (zh) |
KR (1) | KR20210153700A (zh) |
CN (1) | CN110705626A (zh) |
TW (1) | TW202113871A (zh) |
WO (1) | WO2021057174A1 (zh) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113140012A (zh) * | 2021-05-14 | 2021-07-20 | 北京字节跳动网络技术有限公司 | 图像处理方法、装置、介质及电子设备 |
CN113706583A (zh) * | 2021-09-01 | 2021-11-26 | 上海联影医疗科技股份有限公司 | 图像处理方法、装置、计算机设备和存储介质 |
CN114120420A (zh) * | 2021-12-01 | 2022-03-01 | 北京百度网讯科技有限公司 | 图像检测方法和装置 |
CN114298177A (zh) * | 2021-12-16 | 2022-04-08 | 广州瑞多思医疗科技有限公司 | 一种适用于深度学习训练数据的扩充增强方法、系统及可读存储介质 |
CN117152128A (zh) * | 2023-10-27 | 2023-12-01 | 首都医科大学附属北京天坛医院 | 神经影像的病灶识别方法、装置、电子设备和存储介质 |
CN113140012B (zh) * | 2021-05-14 | 2024-05-31 | 北京字节跳动网络技术有限公司 | 图像处理方法、装置、介质及电子设备 |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110705626A (zh) * | 2019-09-26 | 2020-01-17 | 北京市商汤科技开发有限公司 | 一种图像处理方法及装置、电子设备和存储介质 |
CN111368923B (zh) * | 2020-03-05 | 2023-12-19 | 上海商汤智能科技有限公司 | 神经网络训练方法及装置、电子设备和存储介质 |
CN113689355B (zh) * | 2021-09-10 | 2022-07-08 | 数坤(北京)网络科技股份有限公司 | 图像处理方法、装置、存储介质及计算机设备 |
KR20230125412A (ko) * | 2022-02-21 | 2023-08-29 | 계명대학교 산학협력단 | 컴퓨터 단층 촬영 영상의 라디오믹스 파라미터를 이용한 요로결석의 요산석 분류 방법 및 분석장치 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180247410A1 (en) * | 2017-02-27 | 2018-08-30 | Case Western Reserve University | Predicting immunotherapy response in non-small cell lung cancer with serial radiomics |
CN108766563A (zh) * | 2018-05-25 | 2018-11-06 | 戴建荣 | 基于剂量组学的放射治疗结果预测方法和系统 |
CN108830028A (zh) * | 2017-05-04 | 2018-11-16 | 戴立言 | 一种设备及非诊断性确定对象功能等效均匀剂量的方法 |
CN109966662A (zh) * | 2019-04-30 | 2019-07-05 | 四川省肿瘤医院 | 一种验证放射治疗剂量的方法及系统 |
CN110705626A (zh) * | 2019-09-26 | 2020-01-17 | 北京市商汤科技开发有限公司 | 一种图像处理方法及装置、电子设备和存储介质 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH02294786A (ja) * | 1989-05-09 | 1990-12-05 | Hitachi Ltd | 画像輝度レベル値の自動正規化方法 |
KR102020383B1 (ko) * | 2016-11-28 | 2019-09-10 | 고려대학교산학협력단 | 데이터마이닝을 이용한 방사선 치료 계획의 검증 방법 |
US11517768B2 (en) * | 2017-07-25 | 2022-12-06 | Elekta, Inc. | Systems and methods for determining radiation therapy machine parameter settings |
CN107403201A (zh) * | 2017-08-11 | 2017-11-28 | 强深智能医疗科技(昆山)有限公司 | 肿瘤放射治疗靶区和危及器官智能化、自动化勾画方法 |
CN108717866B (zh) * | 2018-04-03 | 2022-10-11 | 中国医学科学院肿瘤医院 | 一种预测放疗计划剂量分布的方法、装置、设备及存储介质 |
CN109166613A (zh) * | 2018-08-20 | 2019-01-08 | 北京东方瑞云科技有限公司 | 基于机器学习的放射治疗计划评估系统及方法 |
CN109523532B (zh) * | 2018-11-13 | 2022-05-03 | 腾讯医疗健康(深圳)有限公司 | 图像处理方法、装置、计算机可读介质及电子设备 |
CN110211664B (zh) * | 2019-04-25 | 2022-11-04 | 安徽大学 | 一种基于机器学习自动设计放射治疗方案的系统 |
-
2019
- 2019-09-26 CN CN201910918450.1A patent/CN110705626A/zh active Pending
-
2020
- 2020-07-07 KR KR1020217037806A patent/KR20210153700A/ko active Search and Examination
- 2020-07-07 JP JP2021563034A patent/JP2022530413A/ja active Pending
- 2020-07-07 WO PCT/CN2020/100692 patent/WO2021057174A1/zh active Application Filing
- 2020-09-09 TW TW109130908A patent/TW202113871A/zh unknown
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180247410A1 (en) * | 2017-02-27 | 2018-08-30 | Case Western Reserve University | Predicting immunotherapy response in non-small cell lung cancer with serial radiomics |
CN108830028A (zh) * | 2017-05-04 | 2018-11-16 | 戴立言 | 一种设备及非诊断性确定对象功能等效均匀剂量的方法 |
CN108766563A (zh) * | 2018-05-25 | 2018-11-06 | 戴建荣 | 基于剂量组学的放射治疗结果预测方法和系统 |
CN109966662A (zh) * | 2019-04-30 | 2019-07-05 | 四川省肿瘤医院 | 一种验证放射治疗剂量的方法及系统 |
CN110705626A (zh) * | 2019-09-26 | 2020-01-17 | 北京市商汤科技开发有限公司 | 一种图像处理方法及装置、电子设备和存储介质 |
Non-Patent Citations (1)
Title |
---|
ZHANG JUNQIAN ,ZHANG YUAN ,YIN YONG , ZHU JIAN ,LI BAOSHENG: "A review of machine learning in tumor radiotherapy", JOURNAL OF BIOMEDICAL ENGINEERING, vol. 36, no. 5, 16 September 2019 (2019-09-16), pages 879 - 884, XP055794284, ISSN: 1001-5551, DOI: 10.7507/1001-5515.201810051 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113140012A (zh) * | 2021-05-14 | 2021-07-20 | 北京字节跳动网络技术有限公司 | 图像处理方法、装置、介质及电子设备 |
CN113140012B (zh) * | 2021-05-14 | 2024-05-31 | 北京字节跳动网络技术有限公司 | 图像处理方法、装置、介质及电子设备 |
CN113706583A (zh) * | 2021-09-01 | 2021-11-26 | 上海联影医疗科技股份有限公司 | 图像处理方法、装置、计算机设备和存储介质 |
CN113706583B (zh) * | 2021-09-01 | 2024-03-22 | 上海联影医疗科技股份有限公司 | 图像处理方法、装置、计算机设备和存储介质 |
CN114120420A (zh) * | 2021-12-01 | 2022-03-01 | 北京百度网讯科技有限公司 | 图像检测方法和装置 |
CN114120420B (zh) * | 2021-12-01 | 2024-02-13 | 北京百度网讯科技有限公司 | 图像检测方法和装置 |
CN114298177A (zh) * | 2021-12-16 | 2022-04-08 | 广州瑞多思医疗科技有限公司 | 一种适用于深度学习训练数据的扩充增强方法、系统及可读存储介质 |
CN117152128A (zh) * | 2023-10-27 | 2023-12-01 | 首都医科大学附属北京天坛医院 | 神经影像的病灶识别方法、装置、电子设备和存储介质 |
CN117152128B (zh) * | 2023-10-27 | 2024-02-27 | 首都医科大学附属北京天坛医院 | 神经影像的病灶识别方法、装置、电子设备和存储介质 |
Also Published As
Publication number | Publication date |
---|---|
JP2022530413A (ja) | 2022-06-29 |
CN110705626A (zh) | 2020-01-17 |
TW202113871A (zh) | 2021-04-01 |
KR20210153700A (ko) | 2021-12-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021057174A1 (zh) | 图像处理方法及装置、电子设备、存储介质和计算机程序 | |
WO2021147257A1 (zh) | 网络训练、图像处理方法及装置、电子设备和存储介质 | |
JP6134446B2 (ja) | 画像分割方法、画像分割装置、画像分割デバイス、プログラム及び記録媒体 | |
CN111310616B (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN109978886B (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN107798669B (zh) | 图像去雾方法、装置及计算机可读存储介质 | |
CN109871883B (zh) | 神经网络训练方法及装置、电子设备和存储介质 | |
EP3869456A1 (en) | Image segmentation method and apparatus, computer device and storage medium | |
WO2020078268A1 (zh) | 图像分割方法、装置、计算机设备及存储介质 | |
CN112767329B (zh) | 图像处理方法及装置、电子设备 | |
WO2021135145A1 (zh) | 图像分割方法及装置、电子设备和存储介质 | |
JP7152598B2 (ja) | 画像処理方法及び装置、電子機器並びに記憶媒体 | |
WO2020134866A1 (zh) | 关键点检测方法及装置、电子设备和存储介质 | |
CN105554389B (zh) | 拍摄方法及装置 | |
CN110674719A (zh) | 目标对象匹配方法及装置、电子设备和存储介质 | |
CN112541928A (zh) | 网络训练方法及装置、图像分割方法及装置和电子设备 | |
JP2021503659A (ja) | 生体検出方法、装置及びシステム、電子機器並びに記憶媒体 | |
TW202203153A (zh) | 圖像處理方法、電子設備和電腦可讀儲存介質 | |
TW202209254A (zh) | 圖像分割方法、電子設備和電腦可讀儲存介質 | |
WO2022156235A1 (zh) | 神经网络训练和图像处理方法及装置、电子设备及存储介质 | |
US20200195905A1 (en) | Method and apparatus for obtaining image, storage medium and electronic device | |
CN114494442A (zh) | 图像处理方法、装置及设备 | |
TWI765386B (zh) | 神經網路訓練及圖像分割方法、電子設備和電腦儲存介質 | |
CN115512116A (zh) | 图像分割模型优化方法、装置、电子设备及可读存储介质 | |
KR102374400B1 (ko) | 이미지 처리 방법 및 장치, 전자 기기와 저장 매체 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20869215 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021563034 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20217037806 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20869215 Country of ref document: EP Kind code of ref document: A1 |