CN115187571A - Medical image segmentation method for introducing prior knowledge based on reward function - Google Patents
Medical image segmentation method for introducing prior knowledge based on reward function Download PDFInfo
- Publication number
- CN115187571A CN115187571A CN202210893407.6A CN202210893407A CN115187571A CN 115187571 A CN115187571 A CN 115187571A CN 202210893407 A CN202210893407 A CN 202210893407A CN 115187571 A CN115187571 A CN 115187571A
- Authority
- CN
- China
- Prior art keywords
- reward function
- reward
- segmentation
- introducing
- pixel
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/194—Segmentation; Edge detection involving foreground-background segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30084—Kidney; Renal
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Databases & Information Systems (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Image Analysis (AREA)
Abstract
The invention provides a medical image segmentation method introducing prior knowledge based on a reward function, which comprises the following steps: acquiring a published medical CT image dataset; designing and obtaining a reward function based on the image data set; introducing a foreground-background ratio of the CT image in the image data set into the reward function to obtain an improved reward function; introducing the approximate difficulty of a priori knowledge example into the improved reward function to obtain a final reward function; and replacing the reward function in the reinforcement learning network with the final reward function, performing reinforcement network training, and finally outputting a final segmentation result graph through the segmentation network. The method adopts the design of introducing external prior knowledge into the reward function, so that the model learning process is greatly influenced towards a beneficial direction, each pixel in the image not only considers the state of the pixel, but also considers the states of other adjacent pixels, the strategy is updated in the constraint direction, and the segmentation result is similar to the real lesion.
Description
Technical Field
The invention belongs to the technical field of medical image segmentation, and particularly relates to a medical image segmentation method based on introduction of prior knowledge into a reward function.
Background
Medical images play a crucial role in the diagnosis of many important diseases, such as structural and functional analysis, diagnosis and treatment, etc., where tumor lesion segmentation is of great significance. From the perspective of practical clinical applications, medical images are aimed at assisting physicians in clinical diagnosis. However, a model that can only predict whether there is a lesion in a CT image does not satisfy the requirements of a doctor. In addition, there is a need to present the slice location of the lesion, the confidence of the single slice segmentation, and whether benign or malignant tumors can be distinguished without the puncture test. The traditional manual segmentation and annotation method has low efficiency of segmenting the medical image focus, and the performance of the traditional manual segmentation and annotation method depends on the experience of an annotator and the knowledge in the medical field to a great extent.
In recent years, deep learning rises rapidly in the field of computer vision and is gradually closely related to medical image segmentation, but medical image segmentation not only requires knowledge in the field of computer vision, but also requires prior knowledge in the field of medicine as theoretical support for model design. Most existing methods are learned from natural image segmentation methods, and medical features are ignored. Furthermore, the team annotating the medical image includes experts, trained personnel and laypersons, which can affect the reliability of the data annotation and limit the effectiveness of the segmentation model in real scenarios. Aiming at the problem, the invention provides a medical image segmentation method introducing prior knowledge based on a reward function so as to solve the problems in the prior art.
Disclosure of Invention
In order to solve the technical problems, the invention provides a medical image segmentation method introducing prior knowledge based on a reward function, the medical image segmentation method introducing the prior knowledge based on the reward function adopts the design of introducing external prior knowledge into the reward function, so that the external prior knowledge greatly influences the model learning process towards a beneficial direction, each pixel in an image not only considers the state of the pixel, but also considers the states of other adjacent pixels, and thus each pixel which is regarded as an intelligent body is more like the judgment of an analog expert and updates the strategy in the constraint direction.
In order to achieve the above object, the present invention provides a medical image segmentation method based on a reward function and introducing prior knowledge, comprising the following steps:
s1, acquiring a public medical CT image data set;
s2, designing and obtaining a reward function based on the image data set;
s3, introducing a foreground-background ratio of the CT image in the image data set into the reward function to obtain an improved reward function;
s4, introducing the approximate difficulty of a priori knowledge example into the improved reward function to obtain a final reward function;
and S5, replacing the reward function in the reinforcement learning network with the final reward function, carrying out reinforcement network training, and finally outputting a final segmentation result graph through the segmentation network.
Preferably, in S2, in the process of designing and obtaining the reward function, the reward P is given i Designed as the difference between the predicted value and the actual label for each pixel in the image data set,
preferably, the method for designing and obtaining the reward function comprises:
treating each pixel in the image dataset as an agent;
designing and obtaining local rewards of the single agents based on cross entropy gain;
deriving the reward function based on the local reward.
Preferably, the feedback r of a single said agent is i Is defined as follows
Wherein gamma denotes a discount factor, V(s) (t) ) Representing inputs s of a network (t) Expected total reward of, R i Indicating the initial reward, P, for a single agent i Indicating the redefined prize, y i E (-1, + 1) represents the pixel label, t represents the step t, r i Represented as the final actual reward for each agent.
Preferably, in S3, the reward function after introducing the foreground-background ratio is improved based on the following formula
Wherein R is (t) Representing the average of the total expected reward for all pixels,represents the reward of the step t, and N represents the total pixel number;
since the medical image segmentation can be regarded as a pixel binary classification, the number of pixels in the region of interest is far less than that of other parts, and the ratio of foreground to background affects the final segmentation effect, the above formula is replaced by an improved reward function formula, that is, the formula is replaced by the improved reward function formula
Wherein alpha is a hyper-parameter (-1, 0), N Front Foreground region representing a prediction map, N Back Representing the background area of the prediction map.
Preferably, in S4, the approximation difficulty of the prior knowledge example is introduced as the prior knowledge based on a setting of a reinforcement learning action space, where a specific action should be set to { mask, retain }, so as to cover parts that have a bad influence on the segmentation, retain parts that are beneficial to the segmentation, and allow the agent to adjust action selection to different degrees under different conditions.
Preferably, the method for introducing the approximate difficulty of the prior knowledge instance into the improved reward function comprises:
setting the image dataset (= { (x) {) 1 ,y 1 ),(x 2 ,y 2 ),…,(x N ,y N )},x i Representing an input pixel, y i E { -1, +1} represents the pixel label, input pixel x for each instance i (i =1,2, \8230;, N) corresponds to a partially receptive window region F i ={(x p ,y p ) I P =1,2, \ 8230;, P }, then the ADI function is expressed as
if F i Tags and x of all examples in (1) i Same, thenIf F is true i Class label and x for all instances in i In a different wayIs established, therefore, willThe replacement is as follows
Where gamma represents a discount factor, where gamma represents,the prize, V(s), representing the step t (t) ) Input s representing a network (t) The expected total prize.
Preferably, in S5, the method for obtaining the final segmentation result map includes:
continuously taking the output of the reinforced network training as the input of the segmentation network to form a segmentation model;
and combining the segmentation model by a CNN-Transformer network as an encoder, and cascading an upsampler to perform positioning and segmentation on the CT image to obtain a final segmentation result graph.
Compared with the prior art, the invention has the following advantages and technical effects:
the method adopts the design of introducing external prior knowledge into the reward function, so that the reward function greatly influences the model learning process towards a beneficial direction, and each pixel in the image not only considers the state of the pixel, but also considers the states of other adjacent pixels; each pixel considered as an agent is more like an analog expert judgment and the strategy is updated in the direction of constraint, and the segmentation result of the method is similar to the real lesion.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this application, illustrate embodiments of the application and, together with the description, serve to explain the application and are not intended to limit the application. In the drawings:
FIG. 1 is a flow chart of a medical image segmentation method based on reward function to introduce prior knowledge according to the present invention;
FIG. 2 is a schematic diagram of a segmentation model network structure according to the present invention;
FIG. 3 is a schematic diagram of a verification experiment process according to the present invention.
Detailed Description
It should be noted that, in the present application, the embodiments and features of the embodiments may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
It should be noted that the steps illustrated in the flowcharts of the figures may be performed in a computer system such as a set of computer-executable instructions and that, although a logical order is illustrated in the flowcharts, in some cases, the steps illustrated or described may be performed in an order different than presented herein.
According to fig. 1 and fig. 2, the present embodiment provides a medical image segmentation method based on a reward function and introducing a priori knowledge, including the following steps:
step one, a public medical CT image data set (such as a kidney cancer CT data set KiTS 19) is obtained.
And step two, designing and obtaining a reward function based on the image data set of the previous step.
In this embodiment, in designing the reward function, each pixel is regarded as an agent, and the local reward design logic for a single agent is based on the cross-entropy gain.
Specifically, P will be awarded i Designed as the difference between the predicted value of the pixel and the actual label,
to update the model in the direction of the constraint, the aforementioned scores are referred to as compare and override, and for the benchmark, the agent's probability is closer to the true pixel label and the agent will obtain a positive reward, and vice versa. Feedback r of each agent i Is defined as follows
Then there is
Where gamma denotes a discount factor, V(s) (t) ) Input s representing a network (t) Expected total reward of, R i Indicating the initial reward, P, for a single agent i Indicating the redefined prize, y i E (-1, + 1) represents the pixel label, t represents the step t, r i Represented as the final actual reward for each agent.
And step three, introducing the foreground-background ratio of the CT image in the image data set into the reward function to obtain an improved reward function.
Rather than modifying the loss function during the network fitting process, in this embodiment, external knowledge is introduced into the feedback (reward) function design of the reinforcement learning network. In order to relieve the problem of unbalanced area ratio of a foreground area and a background area, the foreground-background ratio of a medical image is introduced into a global reward function as prior knowledge, and the reward function after the introduction of the foreground-background ratio of the image is improved based on the following formula
Wherein R is (t) Represents the average of the total expected reward for all pixels,indicates the reward for the step t and N indicates the total number of pixels.
Since the medical image segmentation can be regarded as a pixel binary classification, the number of pixels in the region of interest is much smaller than that of other parts, and the ratio of foreground to background affects the final segmentation effect. Based on this, the above formula is replaced by an improved reward function formula, i.e. by
Where α ∈ (-1, 0) represents a hyperparameter, we use the function f (x) = x +0.2/x, which is a unimodal function with a minimum value ofN Front Foreground region representing a prediction map, N Back Representing the background area of the prediction map to encourage the lesion area to account for 40% -50% of the total image, and as the representation learning of the pixel level RL (reinforcement learning network), after the patch which is not beneficial to segmentation or irrelevant to the task is shielded, the foreground-background ratio of the prediction map will gradually approach the ratio which is most beneficial to segmentation;
and step four, introducing the approximate difficulty of the prior knowledge example into the improved reward function to obtain a final reward function.
In the embodiment, based on the setting of the reinforcement learning action space, the approximation difficulty of the example is introduced as the prior knowledge to explore the consideration degree of the adjacent pixel values, so that each agent can not only notice itself, but also notice the strategies and values of other agents in the adjacent range.
The approximation difficulty of the example is introduced as the priori knowledge, and is based on the setting of a reinforcement learning action space, the specific action should be set to { mask, retain }, the parts which have bad influence on the segmentation are covered, the parts which are beneficial to the segmentation are reserved, and the intelligent agent is allowed to adjust the action selection to different degrees under different conditions. For example, when the predicted value of the current patch is closer to the tag, a better action may be considered for selection.
Secondly, the way of introducing the approximate difficulty of the prior knowledge example into the reward function is as follows:
setting a CT image dataset D = { (x) 1 ,y 1 ),(x 2 ,y 2 ),…,(x N ,y N )},x i Representing an input pixel, y i E { -1, +1} represents the pixel label, input pixel x for each instance i (i =1,2, \8230;, N) corresponds to a partial window area F i ={(x p ,y p ) P =1,2, \8230 |, P }. ADI function expressed as
|F i | represents the number of pixels in the partial receptive field. Here, the first and second liquid crystal display panels are,obviously, if F i Tags and x of all examples in (1) i Are the same as each other, thenIf F is true i Class label and x for all instances in i In a different wayThis is true. Therefore, the present embodiment will be describedThe replacement is as follows
Where gamma represents a discount factor, where gamma represents,the prize, V(s), representing the step t (t) ) Input s representing a network (t) The expected total reward of;
and step five, replacing the reward function in the reinforcement learning network with the final reward function, performing reinforcement network training, and finally outputting a final segmentation result graph through the segmentation network.
In the embodiment, the reward function in the reinforcement learning network is replaced by the final reward function introducing external priori knowledge, the reinforcement network is guided to be trained by the reward function, and then the final segmentation result graph is output through the segmentation network.
Specifically, after an image is input, after the image is trained by an enhanced network, the output of the image is continuously used as the input of a segmentation network, a segmentation model formed by combining a CNN-Transformer network is used as an encoder and a cascade upsampler to realize accurate positioning and accurate segmentation of a focus, a final segmentation result graph is obtained, and the specific structure of the segmentation model is shown in FIG. 2.
As shown in fig. 3, a validation test of a medical image segmentation method based on a reward function and introducing a priori knowledge is provided for the embodiment, and the validation test includes:
step one, selecting a data set. Two-dimensional renal tumor CT data sets are selected for experiments, namely KiTS19 and JLUKT data sets.
KiTS19 challenge data: 210 cases of renal tumors with clinical background, CT semantic segmentation and surgical outcome. During 2010-2018, all patients who received nephrectomy or radical nephrectomy at the Minnesota university medical center were included in the database, from which comprehensive clinical outcomes of 300 patients who received nephrectomy were randomly drawn. We segmented only the whole kidney tumor, the number of training samples was 168, and the number of test samples was 42.
JLUKT renal tumor data: in 2009-2020, white claim en second clinical hospital, jilin university, performed CT enhancement examinations on different renal cancer patients and included their data in the database. CT data for 61 renal tumors are provided. The number of training samples is 49, and the number of test samples is 12.
And step two, preprocessing data. In the acquired two-dimensional renal tumor CT data set, each case comprises a plurality of pieces of continuous slice data, most of the slices do not contain the focus, only a few slices have the focus, and in addition, each CT slice has one corresponding label data. Before the experiment, preprocessing and data enhancement operation are required to be carried out on the data set to ensure that the experiment has higher accuracy, including operations of image binaryzation, expansion and corrosion, histogram equalization, random rotation and translation and the like;
step three, setting up an experiment. The experiment is completed on a hardware platform with an Ubuntu operating system display card NVIDIA RTX3090, the most popular deep learning framework pytorch at present is adopted, tool packages such as matchlotlib, re and pydicom are mainly used, and the final experiment is completed by combining pycharm. The parameters in the experiment were set as follows:
iteration times are as follows: 100 epochs
An optimizer: adam
Learning rate: 10 -2
Batch size processing: 10
The number of rounds is as follows: 50
Length per round t _ max:4
Action space n _ action:2
Discount factor γ:0.95
Parameters of the initial pure segmentation model are used for initializing the reinforced model in the experiment, so that the model is converged more quickly. A down-sampling part of the segmentation model is designed by a transform coder, an up-sampling part outputs a final segmentation mask by multi-step decoding of hidden features, a skip-connection is also used for forming a U-shaped structure, so that an MSA structure and a CNN structure based on the transform are combined to bring out the best of the two, for different data sets, an Nvidia RTX3090 GPU is used, and the training time is varied from several hours to three days. In our model, the predicted inference time of each step is 422ms;
and step four, evaluating indexes. The Dice similarity coefficient is mainly used as a main evaluation index of the tumor segmentation performance of the kidney cancer CT image, and the Hausdorff coefficient is also used as an auxiliary evaluation index of the kidney cancer CT image.
The nature of the Dice coefficient is to measure the similarity of two set samples, which is a set similarity measurement function. The Dice coefficient originally originated from the two-classification problem, and was then widely used in medical image segmentation. In the medical image segmentation task, the overlapping proportion between the final segmented result and the real labeling area (Ground Truth) is mainly calculated, the value range is 0-1, the segmentation result is 1 in the best case, and is 0 in the worst case. If the area enclosed by the segmentation result is denoted as a and the area enclosed by the group Truth is denoted as B, the evaluation index Dice coefficient can be expressed as:
the Hausdorff index, hausdorff distance, is a measure describing the degree of similarity between two sets of points. The Dice coefficient is sensitive to the partitioned internal filling, while the Hausdorffdistance is sensitive to the partitioned boundary, and the smaller the calculated value is, the better the calculated value is, the definition of the factor is as follows:
and step five, evaluating the segmentation result. The loss of training different models is compared. ADI was found to accelerate the early convergence speed of the model, which takes into account the states of neighboring pixels. Secondly, combining foreground-background ratio with ADI increases the model exploration space. This results in a slower convergence rate of the model early in the training, but the segmentation results of the experiment can be improved. The performance of reinforcement learning using ADI is superior to the version without ADI because the average reward of each agent is redefined, the local reward of the model is clearer after the ADI function is introduced, and the score of the model is increased along with the embedding of external knowledge, which also means that the final segmentation result is more accurate and more suitable for the situation of a real lesion.
The above description is only for the preferred embodiments of the present application, but the scope of the present application is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present application should be covered within the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.
Claims (8)
1. A medical image segmentation method for introducing prior knowledge based on a reward function is characterized by comprising the following steps:
s1, acquiring a public medical CT image data set;
s2, designing and obtaining a reward function based on the image data set;
s3, introducing a foreground-background ratio of the CT image in the image data set into the reward function to obtain an improved reward function;
s4, introducing the approximate difficulty of a priori knowledge example into the improved reward function to obtain a final reward function;
and S5, replacing the reward function in the reinforcement learning network with the final reward function, carrying out reinforcement network training, and finally outputting a final segmentation result graph through the segmentation network.
2. A medical image segmentation method introducing a-priori knowledge based on a reward function according to claim 1,
in the S2, in the process of designing and obtaining the reward function, the reward P is given i Designed as the difference between the predicted value and the actual label for each pixel in the image data set,
3. a medical image segmentation method introducing a-priori knowledge based on a reward function according to claim 2,
the method for designing and obtaining the reward function comprises the following steps:
treating each pixel in the image dataset as an agent;
designing and obtaining local rewards of the single agent based on the cross entropy gain;
the reward function is derived based on the local reward.
4. A medical image segmentation method introducing prior knowledge based on a reward function according to claim 3,
feedback r of individual said agents i Is defined as follows
Where gamma denotes a discount factor, V(s) (t) ) Input s representing a network (t) Expected total reward of, R i Indicating the initial reward, P, for a single agent i Indicating the redefined prize, y i E (-1, + 1) denotes the pixel label, t denotes the step t, r i Represented as the final actual reward for each agent.
5. A medical image segmentation method with a priori knowledge introduced based on a reward function according to claim 1,
in S3, the reward function after introducing the foreground-background ratio is improved based on the following formula
Wherein R is (t) Represents the average of the total expected reward for all pixels,represents the reward of the step t, and N represents the total pixel number;
since the medical image segmentation can be regarded as a pixel binary classification, the number of pixels in the region of interest is far less than that of other parts, and the ratio of foreground to background affects the final segmentation effect, the above formula is replaced by an improved reward function formula, that is, the formula is replaced by the improved reward function formula
Wherein alpha is a hyper-parameter (-1, 0), N Front Foreground region representing a prediction map, N Back Representing the background area of the prediction map.
6. A medical image segmentation method with a priori knowledge introduced based on a reward function according to claim 1,
in S4, the approximation difficulty introduced into the priori knowledge example as the priori knowledge is set based on a reinforcement learning action space, and a specific action should be set to { mask, retain }, so as to cover parts that do not affect the segmentation well, retain parts that are beneficial to the segmentation, and allow the agent to adjust the action selection to different degrees under different conditions.
7. A medical image segmentation method with a priori knowledge introduced based on a reward function according to claim 6,
the method for introducing the approximate difficulty of the prior knowledge example into the improved reward function comprises the following steps:
setting the image dataset D = { (x) 1 ,y 1 ),(x 2 ,y 2 ),...,(x N ,y N )},x i Representing an input pixel, y i E { -1, +1} represents the pixel label, input pixel x for each instance i (i =1,2,. Ang., N) corresponds to a partial receptive window region F i ={(x p ,y p )|p=1,2, P }, then the ADI function is expressed as
if F i Tags and x of all examples in (1) i Same, thenIf F is true i Class labels and x for all instances in i In a different wayIs established, therefore, willThe replacement is as follows
8. A medical image segmentation method introducing a-priori knowledge based on a reward function according to claim 1,
in S5, the method for obtaining the final segmentation result map includes:
continuously taking the output of the reinforced network training as the input of the segmentation network to form a segmentation model;
combining the segmentation model by a CNN-Transformer network as an encoder, and cascading an upsampler to perform positioning and segmentation of the CT image to obtain a final segmentation result graph.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210893407.6A CN115187571A (en) | 2022-07-27 | 2022-07-27 | Medical image segmentation method for introducing prior knowledge based on reward function |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210893407.6A CN115187571A (en) | 2022-07-27 | 2022-07-27 | Medical image segmentation method for introducing prior knowledge based on reward function |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115187571A true CN115187571A (en) | 2022-10-14 |
Family
ID=83521105
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210893407.6A Pending CN115187571A (en) | 2022-07-27 | 2022-07-27 | Medical image segmentation method for introducing prior knowledge based on reward function |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115187571A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117036910A (en) * | 2023-09-28 | 2023-11-10 | 合肥千手医疗科技有限责任公司 | Medical image training method based on multi-view and information bottleneck |
-
2022
- 2022-07-27 CN CN202210893407.6A patent/CN115187571A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117036910A (en) * | 2023-09-28 | 2023-11-10 | 合肥千手医疗科技有限责任公司 | Medical image training method based on multi-view and information bottleneck |
CN117036910B (en) * | 2023-09-28 | 2024-01-12 | 合肥千手医疗科技有限责任公司 | Medical image training method based on multi-view and information bottleneck |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Yeung et al. | Unified focal loss: Generalising dice and cross entropy-based losses to handle class imbalanced medical image segmentation | |
Sun et al. | Segmentation of the multimodal brain tumor image used the multi-pathway architecture method based on 3D FCN | |
Ranjbarzadeh et al. | Brain tumor segmentation of MRI images: A comprehensive review on the application of artificial intelligence tools | |
Budak et al. | Cascaded deep convolutional encoder-decoder neural networks for efficient liver tumor segmentation | |
Xie et al. | A context hierarchical integrated network for medical image segmentation | |
Wang et al. | Multimodal brain tumor image segmentation using WRN-PPNet | |
Rai et al. | Automatic and accurate abnormality detection from brain MR images using a novel hybrid UnetResNext-50 deep CNN model | |
Zhang et al. | Cerebrovascular segmentation from TOF-MRA using model-and data-driven method via sparse labels | |
Lv et al. | 2.5 D lightweight RIU-Net for automatic liver and tumor segmentation from CT | |
JP2023540910A (en) | Connected Machine Learning Model with Collaborative Training for Lesion Detection | |
Samudrala et al. | Semantic segmentation of breast cancer images using DenseNet with proposed PSPNet | |
Zhang et al. | Intra-domain task-adaptive transfer learning to determine acute ischemic stroke onset time | |
Hsiao et al. | A deep learning-based precision and automatic kidney segmentation system using efficient feature pyramid networks in computed tomography images | |
Chen et al. | Efficient two-step liver and tumour segmentation on abdominal CT via deep learning and a conditional random field | |
Banerjee et al. | Glioma classification using deep radiomics | |
CN113706486A (en) | Pancreas tumor image segmentation method based on dense connection network migration learning | |
Chen et al. | Generative adversarial network based cerebrovascular segmentation for time-of-flight magnetic resonance angiography image | |
CN115187571A (en) | Medical image segmentation method for introducing prior knowledge based on reward function | |
Yuan et al. | ResD-Unet research and application for pulmonary artery segmentation | |
Yang et al. | A shape-guided deep residual network for automated CT lung segmentation | |
Kong et al. | Data enhancement based on M2-Unet for liver segmentation in Computed Tomography | |
Kumar et al. | Development of an enhanced U-Net model for brain tumor segmentation with optimized architecture | |
Murmu et al. | A novel Gateaux derivatives with efficient DCNN-Resunet method for segmenting multi-class brain tumor | |
Tian et al. | Fully-automated functional region annotation of liver via a 2.5 D class-aware deep neural network with spatial adaptation | |
Zhang et al. | Dial/Hybrid cascade 3DResUNet for liver and tumor segmentation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |