CN113793345B - Medical image segmentation method and device based on improved attention module - Google Patents

Medical image segmentation method and device based on improved attention module Download PDF

Info

Publication number
CN113793345B
CN113793345B CN202111042489.5A CN202111042489A CN113793345B CN 113793345 B CN113793345 B CN 113793345B CN 202111042489 A CN202111042489 A CN 202111042489A CN 113793345 B CN113793345 B CN 113793345B
Authority
CN
China
Prior art keywords
feature matrix
target
image
attention
determining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111042489.5A
Other languages
Chinese (zh)
Other versions
CN113793345A (en
Inventor
李郁欣
曹加旺
刘学玲
冯瑞
耿道颖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huashan Hospital of Fudan University
Original Assignee
Huashan Hospital of Fudan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huashan Hospital of Fudan University filed Critical Huashan Hospital of Fudan University
Priority to CN202111042489.5A priority Critical patent/CN113793345B/en
Publication of CN113793345A publication Critical patent/CN113793345A/en
Application granted granted Critical
Publication of CN113793345B publication Critical patent/CN113793345B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H30/00ICT specially adapted for the handling or processing of medical images
    • G16H30/20ICT specially adapted for the handling or processing of medical images for handling medical images, e.g. DICOM, HL7 or PACS
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • G06T2207/10088Magnetic resonance imaging [MRI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20112Image segmentation details
    • G06T2207/20132Image cropping
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30004Biomedical image processing
    • G06T2207/30016Brain
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Abstract

The application discloses an image processing method, an image processing device, computer equipment and a storage medium. The method comprises the following steps: acquiring target tensor data of a target image; inputting the target tensor data into a target neural network model, the target neural network model including an encoder, a decoder, and an attention module; the encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result; and carrying out segmentation feedback on the target image according to the output of the target neural network model. The accuracy of image segmentation can be improved.

Description

Medical image segmentation method and device based on improved attention module
Technical Field
The embodiment of the application relates to a machine learning technology, in particular to an image processing method, an image processing device, computer equipment and a storage medium.
Background
Medical image segmentation is an important task in computer-aided diagnosis, but the segmentation work has been challenging due to irregular shape of the target object, blurred boundaries, etc. Thus, for identifying these complex target objects of different dimensions, an accurate and reliable segmentation method is needed.
With the development of deep convolutional neural networks (Convolutional Neural Networks, CNNs), U-Net of U-shaped neural networks and a technical scheme for fusing the U-Net with an attention mechanism are presented. However, the current U-Net combined with the attention mechanism cannot accurately segment targets with different shapes and scales in the medical image, and the accuracy of image segmentation is low.
Disclosure of Invention
The application provides an image processing method, an image processing device, computer equipment and a storage medium, so as to improve the accuracy of image segmentation.
In a first aspect, an embodiment of the present application provides an image processing method, including:
acquiring target tensor data of a target image;
inputting the target tensor data into a target neural network model, the target neural network model including an encoder, a decoder, and an attention module;
the encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result;
and carrying out segmentation feedback on the target image according to the output of the target neural network model.
In a second aspect, an embodiment of the present application further provides an image processing apparatus, including:
the target image tensor acquisition module is used for acquiring target tensor data of the target image;
the segmentation module is used for inputting target tensor data into a target neural network model, and the target neural network model comprises an encoder, a decoder and an attention module;
the encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result;
and the output module is used for carrying out segmentation feedback on the target image according to the output of the target neural network model.
In a third aspect, an embodiment of the present application further provides a computer device, including a memory, a processor, and a computer program stored in the memory and capable of running on the processor, where the processor executes the program to implement an image processing method as shown in the embodiment of the present application.
In a third aspect, embodiments of the present application also provide a storage medium containing computer-executable instructions which, when executed by a computer processor, are used to perform an image processing method as shown in the embodiments of the present application.
The image processing method provided by the embodiment of the application can acquire the target tensor data of the target image; inputting target tensor data into a target neural network model, and performing downsampling by an encoder of the target neural network model according to the target tensor data to obtain a first feature matrix; the attention module of the target neural network model is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder of the target neural network model is used for up-sampling according to the third feature matrix to obtain an image segmentation result; and carrying out segmentation feedback on the target image according to the output of the target neural network model. Compared with the problem of low accuracy of U-Net image segmentation combined with an attention mechanism at present, the attention module of the target neural network model can process the first feature matrix of the target image based on hole convolution, the range of the receptive field can be improved through hole convolution, further, the region of interest of a user in the target image can be determined more accurately, segmentation is carried out according to the region, and accuracy of image segmentation is improved.
Drawings
Fig. 1 is a flowchart of an image processing method in a first embodiment of the present application;
fig. 2 is a schematic diagram of the structure of an image processing apparatus in a third embodiment of the present application;
fig. 3 is a schematic structural diagram of a computer device in a fourth embodiment of the present application.
Detailed Description
The application is described in further detail below with reference to the drawings and examples. It is to be understood that the specific embodiments described herein are merely illustrative of the application and are not limiting thereof. It should be further noted that, for convenience of description, only some, but not all of the structures related to the present application are shown in the drawings.
Example 1
Fig. 1 is a flowchart of an image processing method according to an embodiment of the present application, where the embodiment is applicable to a case of dividing an area of interest of a user in an image, especially a case of dividing an area of interest of a user such as a lesion in a medical image, where the method may be performed by a computer device that provides an image dividing function for the user, and the computer device may be a personal computer, a notebook computer, a tablet computer or a server. The method specifically comprises the following steps:
step 110, obtaining target tensor data of the target image.
The target image may be an image to be segmented, and the target tensor data is a target image expressed in a vector form. For example, the shape of the target tensor data may be 1×128×128, representing 128 pixels in the horizontal direction and 128 pixels in the vertical direction in a certain color channel of the target image.
On the basis of the above embodiment, acquiring target tensor data of a target image includes:
preprocessing a target image according to a preset preprocessing mode to obtain a preprocessed image, wherein the preprocessing mode comprises an image clipping mode and an image normalization mode; target tensor data is determined from the preprocessed image.
The image cropping mode includes cropping images of different sizes into images of the same size. The image normalization approach involves normalizing images of different formats to a uniform file format. The preprocessing image with uniform specification can be obtained through the preprocessing mode. Since the preprocessed image contains the same number of pixels in the horizontal and vertical directions and the same number of color channels, the target tensor data format obtained from the preprocessed image is also the same. The target tensor data obtained through preprocessing is suitable for the target neural network model, and the segmentation accuracy of the target neural network model can be improved.
On the basis of the above embodiment, the target image is a medical image, and the medical image includes: images are taken using computed tomography images, nuclear magnetic resonance images, high definition cameras.
The high-definition camera can be high-definition camera equipment in the medical field. For example, the high definition camera may be a DRIVE eye disease dataset image acquisition tool. Specifically, a Canon CR5 non-mydriatic 3CCD camera was used to capture images at 45 degrees of field of view. Each image was captured with 8 bits in the color plane of 768×584 pixels. The field of view of each image is circular, with a diameter of about 540 pixels.
On the basis of the above embodiment, acquiring target tensor data of a target image includes:
desensitizing the target image to obtain a desensitized image;
target tensor data for the target image is determined from the desensitized image.
In the embodiment, because the nuclear magnetic resonance scanning image related to the suspected patient with the clinical Parkinson disease is desensitized data with the format of Nii.gz after desensitization treatment, information such as the name of the patient and the like can not be read, and important privacy is not involved, so that the information safety of a user is ensured to the greatest extent.
Step 120, inputting the target tensor data into the target neural network model.
Wherein the target neural network model includes an encoder, a decoder, and an attention module.
The encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result.
The encoder includes a plurality of convolution blocks. The convolutional layer of each convolutional block is followed by an activation function (ReLU) and a Batch Normalization (BN) layer, which are then downsampled using a max-pooling operation, the downsampling step size may be 2. After each downsampling step, the number of characteristic channels will double. The first feature matrix is obtained by a plurality of donation blocks.
In one implementation, the data recall model may be constructed based on data characteristics of clinical brain nuclear magnetic resonance images used for training. Specifically: firstly, an account number and a password with partial user authority are obtained from a PACS system of a hospital, and a computer tomography image of a suspected patient in the PACS system and corresponding tag data are obtained through a communication protocol and stored.
During storage, the brain nuclear magnetic resonance scan images may have different slice numbers according to the thickness of the scan layer, and are stored as a patient-table header. And downsampling the brain nuclear magnetic resonance scanning image through an encoder to obtain a first feature matrix.
After the first feature matrix is obtained, the first feature matrix is input to an attention module (also called an AHA module). And optimizing the first feature matrix in the attention module, and outputting a third feature matrix.
The decoder and the encoder comprise a plurality of symmetrical convolution blocks, and the decoder is used for up-sampling the third feature matrix to restore the original image resolution.
In one implementation, the attention module is configured to perform hole convolution according to the first feature matrix to obtain a second feature matrix, and includes: and carrying out cavity convolution on the first feature matrix according to the plurality of cavity convolutions to obtain a plurality of second feature matrices, wherein each cavity convolution corresponds to the second feature matrix one by one. Accordingly, determining the original attention profile from the second feature matrix comprises: performing mutual matrix multiplication according to the plurality of second feature matrixes to obtain a matrix multiplication result; the original attention map is determined from the matrix multiplication result. Accordingly, determining a hole attention map from the original attention map includes: and carrying out transfer closing operation on the original attention force diagram to obtain a hole attention force diagram. Accordingly, determining a third feature matrix from the first feature matrix and the hole attention map comprises: multiplying the first feature matrix by the hole attention map to obtain a third feature matrix.
In one implementation, the generation process of the hole high order attention matrix is as follows: wherein, at the bottom of the encoder, a feature map (also called a first feature matrix) is obtained from the previous stageWherein H W C represents the height and width of the feature map, respectivelyAnd the channel number, which is sent into 4 shared weight convolution layers (r=1, r=2, r=4, r=8 are respectively set) to generate a multi-scale feature map (also called a second feature matrix) X r=1,2,4,8 . Scaling the multi-scale feature map (also called second feature matrix) by 1×1 convolution, and performing dot multiplication by the following formula one to obtain a preliminary attention matrix (also called original attention map)/(original attention map)>
Based on the theory of "transitive closure" in graph theory, one can try according to the original attentionObtaining a high-order attention moment arrayThe calculation process is as follows:
first, an original thresholded attention matrix (also called a hole attention map) is obtained according to the following equation II:
where δ is a super-parameter threshold, in the embodiment of the present application, δ may be set to 0.5.
Then, according to the transfer closure theory, carrying out transfer closure calculation on the hollow attention force diagram, and obtaining a high-order attention moment array according to the following formula III
Where M represents the high rank attention of the mth order. And weighting the first feature map through the high-order attention matrix to obtain a third feature matrix, so as to filter noise influence and strengthen useful information.
The attention module described above may be embedded into any "U" -shaped encoder-decoder neural network model for use in medical image segmentation. In order to expand the limited receptive field of local convolution of the convolution layer in the downsampling stage and the deconvolution layer in the upsampling stage, more global information can be effectively captured in a high dimension by adopting multi-scale hole convolution. In addition, the high-order attention mechanism constructs a high-order attention feature map through the transfer closure of the map model so as to acquire stronger relevant context features.
Optionally, the segmentation algorithm used in the embodiment of the application may be a threshold segmentation algorithm, so that the influence of peripheral noise can be removed, the quality of the normalized image is improved, and the robustness of the U-shaped neural network model based on the improved attention module is improved.
And 130, performing segmentation feedback of the target image according to the output of the target neural network model.
On the basis of the above embodiment, the method for performing segmentation feedback of the target image according to the output of the target neural network model includes: and outputting the target image, and determining the highlight region of the target image according to the image segmentation result.
The user can import the target image through the computer equipment, and the computer equipment displays the target image and the image segmentation result after segmenting the target image according to the method. The highlighting may be used to display the segmented user region of interest.
The image processing method provided by the embodiment of the application can acquire the target tensor data of the target image; inputting target tensor data into a target neural network model, and performing downsampling by an encoder of the target neural network model according to the target tensor data to obtain a first feature matrix; the attention module of the target neural network model is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder of the target neural network model is used for up-sampling according to the third feature matrix to obtain an image segmentation result; and carrying out segmentation feedback on the target image according to the output of the target neural network model. Compared with the problem of low accuracy of U-Net image segmentation combined with an attention mechanism at present, the attention module of the target neural network model can process the first feature matrix of the target image based on hole convolution, the range of the receptive field can be improved through hole convolution, further, the region of interest of a user in the target image can be determined more accurately, segmentation is carried out according to the region, and accuracy of image segmentation is improved.
On the basis of the above embodiment, before inputting the tensor data into the target neural network model, the method further includes:
acquiring training tensor data of training images of a training set; and training the target neural network model according to the training tensor data so as to optimize network parameters of the target neural network model.
The training images can be obtained by interfacing with a hospital image archiving and communication system (Picture Archiving and Communication Systems, PACS) and can be expanded based on the images obtained by the interfacing. The expansion mode comprises random overturn, random rotation, random translation, random clipping and Gaussian noise addition. The training tensor data obtained through expansion is used for training the target neural network model, so that the sensitivity of the target neural network model to unstable factors such as rotation, translation, gaussian noise and the like can be improved, and the robustness of the U-shaped neural network model based on the improved attention module is further improved.
Example two
An embodiment II of the present application provides an image processing method, as a further explanation of the above embodiment, including:
and E1, storing medical image information related to the suspected Parkinson patient. The medical image information includes brain magnetic resonance imaging.
The clinical head nuclear magnetic resonance imaging is standard image format data with a format of nii.gz after desensitization treatment.
And E2, preprocessing the target image according to a preset preprocessing mode to obtain a preprocessed image, and determining target tensor data according to the preprocessed image.
The pretreatment method comprises the following steps:
and step T1, obtaining a detection machine model corresponding to brain nuclear magnetic resonance imaging, regulating the nuclear magnetic resonance imaging to obtain a standard image to be detected, and normalizing to obtain the normalized image to be detected.
In this example, all magnetic resonance examinations were performed on a clinical radiology department 3.0-T nuclear magnetic resonance scanner (GE medical company discoery (TM) MR 750) equipped with an 8-channel head matrix coil.
In step T2, since the pixel occupation of the region to be segmented is relatively small, particularly, in this embodiment, the original image is cut in the center, the middle region is taken for further processing, and the nuclear magnetic resonance image of the brain to be detected is converted into Tensor data as the preprocessing data.
Wherein the shape of the Tensor data is 1×128×128.
And E3, training a neural network model to serve as a medical image segmentation model.
Training a "U" shaped neural network model based on an improved attention module includes the steps of:
step S1, constructing a data calling model according to the data characteristics of the clinical brain nuclear magnetic resonance images for training. Specifically:
firstly, an account number and a password with partial user authority are obtained from a PACS system of a hospital, and a computer tomography image of a suspected patient in the PACS system and corresponding tag data are obtained through a communication protocol and stored.
In addition, during storage, the brain nuclear magnetic resonance scan images may have different slice numbers according to the thickness of the scan layer, and are stored as a patient-table.
S2, inputting the image into a downsampling feature extraction network to obtain a first feature matrix;
and S3, correcting the first feature matrix by using a hole high-rank attention mechanism to obtain a second feature matrix.
Step S3 comprises the following sub-steps: step S3-1, carrying out cavity convolution on the first feature matrix according to the plurality of cavity convolutions to obtain a plurality of second feature matrices, wherein each cavity convolution corresponds to the second feature matrix one by one, and step S3-2, carrying out mutual matrix multiplication according to the plurality of second feature matrices to obtain a matrix multiplication result; determining an original attention map according to the matrix multiplication result; s2-3, transmitting the original attention force diagram to carry out a closing operation to obtain a hole attention force diagram; and S2-4, multiplying the first feature matrix by the hole attention map to obtain a third feature matrix.
And S4, outputting the third feature matrix obtained in the previous step to an up-sampling segmentation network to obtain a segmentation result.
And S4, training the U-shaped neural network model based on the improved attention module to obtain the trained U-shaped neural network model based on the improved attention module as a hole attention segmentation neural network model.
In step S4-1, the original brain nuclear magnetic resonance image is preprocessed and then converted into Tensor data serving as training data. Specifically:
and extracting pixel values of the brain nuclear magnetic resonance scanning image, normalizing the nuclear magnetic resonance image pixels and cutting off the pixel values, so that the influence of peripheral noise is removed, and the model robustness is improved. The center cropping method was used to preserve intermediate images of the brain with length and width of 128. The pixel matrix is converted into Tensor data by using a Tensor API in the PyTorch framework, so that GPU parallel computing is convenient to use.
And S4-2, expanding the normalized image by using a preset data expansion method to obtain an expanded image.
The data expansion method comprises random overturn, random rotation, random translation, random clipping and Gaussian noise addition.
And S4-3, inputting the expanded image and the label data as processing results into a U-shaped neural network model based on the improved attention module.
In this embodiment, the setting conditions of the relevant parameters when training the hole attention-segmentation neural network model are: the batch size is set to 48, the epoch is 100, the gradient descent method is random gradient descent optimization (SGD), the learning rate adopts an Adam optimizer strategy, the initial value is 0.001, the momentum is 0.9, and the attenuation rate is 3E-5.
And E4, dividing the user attention area according to the target image (medical image) input by the user.
In this embodiment, the activation function is a general activation function (ReLU), and the user attention area is a key division area requiring model learning division, in this embodiment, a black compact (Substantia Nigra Pars Compacta, SNpc).
And E5, displaying the nuclear magnetic resonance scanning image, and displaying a model segmentation result by using a red highlight region to assist a doctor in diagnosis.
All data obtained in this example were scanned by a 3.0T nuclear magnetic resonance scanner (model: discovery TM MR750, GE Healthcare). A total of 188 brain nuclear magnetic resonance imaging screening samples, including in particular 140 Parkinson disease patients and 48 healthy people. The important focusing area of the Parkinson disease is a substantia nigra compact part, so that the classification influence of the Parkinson patient and the healthy person is not needed to be considered in the segmentation task, and all segmentation labels are independently marked by a professional doctor of clinical experience of an eight-year imaging department.
The data are divided into two independent sets, namely training in the development process of algorithm models and other verification data for algorithm verification. Wherein the training set comprises 152 testers 'head nuclear magnetic resonance scan images and the verification set comprises 36 testers' head nuclear magnetic resonance scan images.
Verification by the method or the device is carried out to obtain a 'U' -shaped neural network model based on an improved attention module, wherein the Dairy coefficient (DSC) of the 'U' -shaped neural network model on verification data is 0.8769, the Accuracy (ACC) is 0.9992, the Sensitivity (SE) is 0.8935, the Specificity (SP) is 0.9995, and the area under a subject working characteristic curve (receiver operating characteristic curve, ROC) is 0.9465.
Meanwhile, experiments show that the test process of a batch of 32 images takes about 3 seconds, and all the tests of one patient take about 30 seconds, so that the device has higher time efficiency.
Compared with manual diagnosis, the medical image segmentation device 1 and the medical image segmentation method based on the improved attention module have high accuracy and high efficiency on the related segmentation task of the medical image, and are close to the diagnosis level of expert doctors, so that the medical image segmentation device and the medical image segmentation method based on the improved attention module are particularly suitable for the conditions of lack of basic medical resources and insufficient quantity of expert doctors. In addition, the device can also effectively assist doctors to carry out rapid diagnosis, and is beneficial to improving the diagnosis and treatment efficiency of the image department links of hospitals.
Example III
Fig. 2 is a schematic structural diagram of an image processing apparatus according to a third embodiment of the present application, where the present embodiment is applicable to a case of dividing an area of interest of a user in an image, especially a case of dividing an area of interest of a user such as a lesion in a medical image, and the method may be performed by a computer device that provides an image dividing function for the user, and the computer device may be a personal computer, a notebook computer, a tablet computer, or a server. The device comprises: a target image tensor acquisition module 210, a segmentation module 220, and an output module 230.
A target image tensor acquisition module 210, configured to acquire target tensor data of a target image;
a segmentation module 220 for inputting target tensor data into a target neural network model, the target neural network model comprising an encoder, a decoder, and an attention module;
the encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result;
and the output module 230 is used for carrying out segmentation feedback on the target image according to the output of the target neural network model.
On the basis of the above embodiment, the segmentation module 220 is configured to:
carrying out cavity convolution on the first feature matrix according to the cavity convolutions to obtain a plurality of second feature matrices, wherein each cavity convolution corresponds to the second feature matrix one by one;
performing mutual matrix multiplication according to the plurality of second feature matrixes to obtain a matrix multiplication result; determining an original attention map according to the matrix multiplication result;
transmitting the original attention force diagram to carry out a closing operation to obtain a cavity attention force diagram;
multiplying the first feature matrix by the hole attention map to obtain a third feature matrix.
On the basis of the above embodiment, the target image tensor obtaining module 210 is configured to:
preprocessing a target image according to a preset preprocessing mode to obtain a preprocessed image, wherein the preprocessing mode comprises an image clipping mode and an image normalization mode;
target tensor data is determined from the preprocessed image.
In the above embodiment, the target image is a medical image, and the medical image includes: images are taken using computed tomography images, nuclear magnetic resonance images, high definition cameras.
On the basis of the above embodiment, the output module 230 is configured to:
and outputting the target image, and determining the highlight region of the target image according to the image segmentation result.
On the basis of the above embodiment, the target image tensor obtaining module 210 is configured to:
desensitizing the target image to obtain a desensitized image;
target tensor data for the target image is determined from the desensitized image.
On the basis of the embodiment, the training device further comprises a training module. The training module is used for:
acquiring training tensor data of training images of a training set;
and training the target neural network model according to the training tensor data so as to optimize network parameters of the target neural network model.
According to the image processing device provided by the embodiment of the application, the target image tensor acquisition module 210 can acquire target tensor data of a target image; the segmentation module 220 inputs the target tensor data into a target neural network model, and an encoder of the target neural network model performs downsampling according to the target tensor data to obtain a first feature matrix; the attention module of the target neural network model is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder of the target neural network model is used for up-sampling according to the third feature matrix to obtain an image segmentation result; the output module 230 performs segmentation feedback of the target image according to the output of the target neural network model. Compared with the problem of low accuracy of U-Net image segmentation combined with an attention mechanism at present, the attention module of the target neural network model can process the first feature matrix of the target image based on hole convolution, the range of the receptive field can be improved through hole convolution, further, the region of interest of a user in the target image can be determined more accurately, segmentation is carried out according to the region, and accuracy of image segmentation is improved.
The image processing device provided by the embodiment of the application can execute the image processing method provided by any embodiment of the application, and has the corresponding functional modules and beneficial effects of the execution method.
Example IV
Fig. 3 is a schematic structural diagram of a computer device according to a fourth embodiment of the present application, and as shown in fig. 3, the computer device includes a processor 30, a memory 31, an input device 32 and an output device 33; the number of processors 30 in the computer device may be one or more, one processor 30 being taken as an example in fig. 3; the processor 30, the memory 31, the input means 32 and the output means 33 in the computer device may be connected by a bus or by other means, in fig. 3 by way of example.
The memory 31 is a computer-readable storage medium that can be used to store a software program, a computer-executable program, and modules such as program instructions/modules corresponding to the image processing method in the embodiment of the present application (for example, the target image tensor acquisition module 210, the segmentation module 220, and the output module 230 in the image processing apparatus). The processor 30 executes various functional applications of the computer device and data processing, i.e., implements the image processing method described above, by running software programs, instructions, and modules stored in the memory 31.
The memory 31 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, at least one application program required for functions; the storage data area may store data created according to the use of the terminal, etc. In addition, the memory 31 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other non-volatile solid-state storage device. In some examples, memory 31 may further comprise memory remotely located relative to processor 30, which may be connected to the computer device via a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The input device 32 is operable to receive input numeric or character information and to generate key signal inputs related to user settings and function control of the computer device. The output means 33 may comprise a display device such as a display screen.
Example five
A fifth embodiment of the present application also provides a storage medium containing computer-executable instructions, which when executed by a computer processor, are for performing an image processing method, the method comprising:
acquiring target tensor data of a target image;
inputting the target tensor data into a target neural network model, the target neural network model including an encoder, a decoder, and an attention module;
the encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix according to the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result;
and carrying out segmentation feedback on the target image according to the output of the target neural network model.
On the basis of the foregoing embodiment, the attention module is configured to perform hole convolution according to the first feature matrix to obtain a second feature matrix, and includes:
carrying out cavity convolution on the first feature matrix according to the cavity convolutions to obtain a plurality of second feature matrices, wherein each cavity convolution corresponds to the second feature matrix one by one;
accordingly, determining the original attention profile from the second feature matrix comprises:
performing mutual matrix multiplication according to the plurality of second feature matrixes to obtain a matrix multiplication result; determining an original attention map according to the matrix multiplication result;
accordingly, determining a hole attention map from the original attention map includes:
transmitting the original attention force diagram to carry out a closing operation to obtain a cavity attention force diagram;
accordingly, determining a third feature matrix from the first feature matrix and the hole attention map comprises:
multiplying the first feature matrix by the hole attention map to obtain a third feature matrix.
On the basis of the above embodiment, acquiring target tensor data of a target image includes:
preprocessing a target image according to a preset preprocessing mode to obtain a preprocessed image, wherein the preprocessing mode comprises an image clipping mode and an image normalization mode;
target tensor data is determined from the preprocessed image.
In the above embodiment, the target image is a medical image, and the medical image includes: images are taken using computed tomography images, nuclear magnetic resonance images, high definition cameras.
On the basis of the above embodiment, the method for performing segmentation feedback of the target image according to the output of the target neural network model includes:
and outputting the target image, and determining the highlight region of the target image according to the image segmentation result.
On the basis of the above embodiment, acquiring target tensor data of a target image includes:
desensitizing the target image to obtain a desensitized image;
target tensor data for the target image is determined from the desensitized image.
On the basis of the above embodiment, before inputting the tensor data into the target neural network model, the method further includes:
acquiring training tensor data of training images of a training set;
and training the target neural network model according to the training tensor data so as to optimize network parameters of the target neural network model.
Of course, the storage medium containing the computer executable instructions provided in the embodiments of the present application is not limited to the method operations described above, and may also perform the related operations in the image processing method provided in any embodiment of the present application.
From the above description of embodiments, it will be clear to a person skilled in the art that the present application may be implemented by means of software and necessary general purpose hardware, but of course also by means of hardware, although in many cases the former is a preferred embodiment. Based on such understanding, the technical solution of the present application may be embodied essentially or in a part contributing to the prior art in the form of a software product, which may be stored in a computer readable storage medium, such as a floppy disk, a Read-Only Memory (ROM), a random access Memory (Random Access Memory, RAM), a FLASH Memory (FLASH), a hard disk or an optical disk of a computer, etc., and include several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the method according to the embodiments of the present application.
It should be noted that, in the above-mentioned embodiments of the search apparatus, each unit and module included are only divided according to the functional logic, but not limited to the above-mentioned division, as long as the corresponding functions can be implemented; in addition, the specific names of the functional units are also only for distinguishing from each other, and are not used to limit the protection scope of the present application.
Note that the above is only a preferred embodiment of the present application and the technical principle applied. It will be understood by those skilled in the art that the present application is not limited to the particular embodiments described herein, but is capable of various obvious changes, rearrangements and substitutions as will now become apparent to those skilled in the art without departing from the scope of the application. Therefore, while the application has been described in connection with the above embodiments, the application is not limited to the embodiments, but may be embodied in many other equivalent forms without departing from the spirit or scope of the application, which is set forth in the following claims.

Claims (9)

1. An image processing method, comprising:
acquiring target tensor data of a target image;
inputting the target tensor data to a target neural network model, the target neural network model comprising an encoder, a decoder, and an attention module;
the encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix based on the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result;
carrying out segmentation feedback on the target image according to the output of the target neural network model;
correspondingly, the attention module is configured to perform hole convolution according to the first feature matrix to obtain a second feature matrix, and includes: carrying out cavity convolution on the first feature matrix according to the cavity convolutions to obtain a plurality of second feature matrices, wherein each cavity convolution corresponds to the second feature matrix one by one;
correspondingly, the determining the original attention map according to the second feature matrix includes: performing mutual matrix multiplication according to the plurality of second feature matrixes to obtain a matrix multiplication result; determining an original attention map according to the matrix multiplication result;
accordingly, the determining a hole attention map according to the original attention map includes: carrying out transfer closing operation on the original attention force diagram to obtain a hole attention force diagram;
correspondingly, the determining a third feature matrix according to the first feature matrix and the hole attention map comprises: multiplying the first feature matrix with the hole attention map to obtain a third feature matrix.
2. The method of claim 1, wherein obtaining target tensor data for the target image comprises:
preprocessing the target image according to a preset preprocessing mode to obtain a preprocessed image, wherein the preprocessing mode comprises an image cutting mode and an image normalization mode;
and determining target tensor data according to the preprocessed image.
3. The method of claim 1, wherein the target image is a medical image, the medical image comprising: images are taken using computed tomography images, nuclear magnetic resonance images, high definition cameras.
4. The method of claim 1, wherein the performing segmentation feedback of the target image based on the output of the target neural network model comprises:
and outputting the target image, and determining a highlight region of the target image according to the image segmentation result.
5. The method of claim 1, wherein obtaining target tensor data for the target image comprises:
desensitizing the target image to obtain a desensitized image;
and determining target tensor data of the target image according to the desensitization image.
6. The method of claim 1, further comprising, prior to inputting the tensor data into a target neural network model:
acquiring training tensor data of training images of a training set;
and training the target neural network model according to the training tensor data so as to optimize network parameters of the target neural network model.
7. An image processing apparatus, comprising:
the target image tensor acquisition module is used for acquiring target tensor data of the target image;
a segmentation module for inputting the target tensor data into a target neural network model, the target neural network model comprising an encoder, a decoder, and an attention module;
the encoder is used for carrying out downsampling according to the target tensor data to obtain a first feature matrix; the attention module is used for carrying out cavity convolution according to the first feature matrix to obtain a second feature matrix; determining an original attention map from the second feature matrix; determining a hole attention map from the original attention map; determining a third feature matrix based on the first feature matrix and the hole attention map; outputting the third feature matrix to a value decoder; the decoder is used for up-sampling according to the third feature matrix to obtain an image segmentation result;
the output module is used for carrying out segmentation feedback on the target image according to the output of the target neural network model;
correspondingly, the attention module is configured to perform hole convolution according to the first feature matrix to obtain a second feature matrix, and includes: carrying out cavity convolution on the first feature matrix according to the cavity convolutions to obtain a plurality of second feature matrices, wherein each cavity convolution corresponds to the second feature matrix one by one;
correspondingly, the determining the original attention map according to the second feature matrix includes: performing mutual matrix multiplication according to the plurality of second feature matrixes to obtain a matrix multiplication result; determining an original attention map according to the matrix multiplication result;
accordingly, the determining a hole attention map according to the original attention map includes: carrying out transfer closing operation on the original attention force diagram to obtain a hole attention force diagram;
correspondingly, the determining a third feature matrix according to the first feature matrix and the hole attention map comprises: multiplying the first feature matrix with the hole attention map to obtain a third feature matrix.
8. A computer device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, wherein the processor implements the image processing method according to any of claims 1-6 when executing the program.
9. A storage medium containing computer executable instructions which, when executed by a computer processor, are for performing the image processing method of any of claims 1-6.
CN202111042489.5A 2021-09-07 2021-09-07 Medical image segmentation method and device based on improved attention module Active CN113793345B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111042489.5A CN113793345B (en) 2021-09-07 2021-09-07 Medical image segmentation method and device based on improved attention module

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111042489.5A CN113793345B (en) 2021-09-07 2021-09-07 Medical image segmentation method and device based on improved attention module

Publications (2)

Publication Number Publication Date
CN113793345A CN113793345A (en) 2021-12-14
CN113793345B true CN113793345B (en) 2023-10-31

Family

ID=78879628

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111042489.5A Active CN113793345B (en) 2021-09-07 2021-09-07 Medical image segmentation method and device based on improved attention module

Country Status (1)

Country Link
CN (1) CN113793345B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111681273B (en) * 2020-06-10 2023-02-03 创新奇智(青岛)科技有限公司 Image segmentation method and device, electronic equipment and readable storage medium
CN114387436B (en) * 2021-12-28 2022-10-25 北京安德医智科技有限公司 Wall coronary artery detection method and device, electronic device and storage medium
CN114565628B (en) * 2022-03-23 2022-09-13 中南大学 Image segmentation method and system based on boundary perception attention
CN115731243B (en) * 2022-11-29 2024-02-09 北京长木谷医疗科技股份有限公司 Spine image segmentation method and device based on artificial intelligence and attention mechanism

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018036293A1 (en) * 2016-08-26 2018-03-01 杭州海康威视数字技术股份有限公司 Image segmentation method, apparatus, and fully convolutional network system
CN109191472A (en) * 2018-08-28 2019-01-11 杭州电子科技大学 Based on the thymocyte image partition method for improving U-Net network
CN109685813A (en) * 2018-12-27 2019-04-26 江西理工大学 A kind of U-shaped Segmentation Method of Retinal Blood Vessels of adaptive scale information
CN110059586A (en) * 2019-03-29 2019-07-26 电子科技大学 A kind of Iris Location segmenting system based on empty residual error attention structure
WO2019153908A1 (en) * 2018-02-11 2019-08-15 北京达佳互联信息技术有限公司 Image recognition method and system based on attention model
CN111612807A (en) * 2020-05-15 2020-09-01 北京工业大学 Small target image segmentation method based on scale and edge information
WO2020177651A1 (en) * 2019-03-01 2020-09-10 华为技术有限公司 Image segmentation method and image processing device
WO2020215985A1 (en) * 2019-04-22 2020-10-29 腾讯科技(深圳)有限公司 Medical image segmentation method and device, electronic device and storage medium
CN111898709A (en) * 2020-09-30 2020-11-06 中国人民解放军国防科技大学 Image classification method and device
CN112017198A (en) * 2020-10-16 2020-12-01 湖南师范大学 Right ventricle segmentation method and device based on self-attention mechanism multi-scale features
CN112017191A (en) * 2020-08-12 2020-12-01 西北大学 Method for establishing and segmenting liver pathology image segmentation model based on attention mechanism
WO2020244108A1 (en) * 2019-06-05 2020-12-10 Boe Technology Group Co., Ltd. Methods and apparatuses for semantically segmenting input image, and computer-program product
CN112101363A (en) * 2020-09-02 2020-12-18 河海大学 Full convolution semantic segmentation system and method based on cavity residual error and attention mechanism
WO2021031066A1 (en) * 2019-08-19 2021-02-25 中国科学院深圳先进技术研究院 Cartilage image segmentation method and apparatus, readable storage medium, and terminal device
CN112541918A (en) * 2020-12-23 2021-03-23 山东师范大学 Three-dimensional medical image segmentation method based on self-attention mechanism neural network
CN112750142A (en) * 2021-01-26 2021-05-04 上海工程技术大学 Ultrasonic image segmentation system and method based on side window attention mechanism
GB202104506D0 (en) * 2021-03-30 2021-05-12 Ucl Business Plc Medical Image Analysis Using Neural Networks

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8209120B2 (en) * 1997-10-22 2012-06-26 American Vehicular Sciences Llc Vehicular map database management techniques
US11195057B2 (en) * 2014-03-18 2021-12-07 Z Advanced Computing, Inc. System and method for extremely efficient image and pattern recognition and artificial intelligence platform
US10755391B2 (en) * 2018-05-15 2020-08-25 Adobe Inc. Digital image completion by learning generation and patch matching jointly
US10776671B2 (en) * 2018-05-25 2020-09-15 Adobe Inc. Joint blur map estimation and blur desirability classification from an image
US11461998B2 (en) * 2019-09-25 2022-10-04 Samsung Electronics Co., Ltd. System and method for boundary aware semantic segmentation

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018036293A1 (en) * 2016-08-26 2018-03-01 杭州海康威视数字技术股份有限公司 Image segmentation method, apparatus, and fully convolutional network system
WO2019153908A1 (en) * 2018-02-11 2019-08-15 北京达佳互联信息技术有限公司 Image recognition method and system based on attention model
CN109191472A (en) * 2018-08-28 2019-01-11 杭州电子科技大学 Based on the thymocyte image partition method for improving U-Net network
CN109685813A (en) * 2018-12-27 2019-04-26 江西理工大学 A kind of U-shaped Segmentation Method of Retinal Blood Vessels of adaptive scale information
WO2020177651A1 (en) * 2019-03-01 2020-09-10 华为技术有限公司 Image segmentation method and image processing device
CN110059586A (en) * 2019-03-29 2019-07-26 电子科技大学 A kind of Iris Location segmenting system based on empty residual error attention structure
WO2020215985A1 (en) * 2019-04-22 2020-10-29 腾讯科技(深圳)有限公司 Medical image segmentation method and device, electronic device and storage medium
WO2020244108A1 (en) * 2019-06-05 2020-12-10 Boe Technology Group Co., Ltd. Methods and apparatuses for semantically segmenting input image, and computer-program product
WO2021031066A1 (en) * 2019-08-19 2021-02-25 中国科学院深圳先进技术研究院 Cartilage image segmentation method and apparatus, readable storage medium, and terminal device
CN111612807A (en) * 2020-05-15 2020-09-01 北京工业大学 Small target image segmentation method based on scale and edge information
CN112017191A (en) * 2020-08-12 2020-12-01 西北大学 Method for establishing and segmenting liver pathology image segmentation model based on attention mechanism
CN112101363A (en) * 2020-09-02 2020-12-18 河海大学 Full convolution semantic segmentation system and method based on cavity residual error and attention mechanism
CN111898709A (en) * 2020-09-30 2020-11-06 中国人民解放军国防科技大学 Image classification method and device
CN112017198A (en) * 2020-10-16 2020-12-01 湖南师范大学 Right ventricle segmentation method and device based on self-attention mechanism multi-scale features
CN112541918A (en) * 2020-12-23 2021-03-23 山东师范大学 Three-dimensional medical image segmentation method based on self-attention mechanism neural network
CN112750142A (en) * 2021-01-26 2021-05-04 上海工程技术大学 Ultrasonic image segmentation system and method based on side window attention mechanism
GB202104506D0 (en) * 2021-03-30 2021-05-12 Ucl Business Plc Medical Image Analysis Using Neural Networks

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
A Multi-Scaled Receptive Field Learning Approach for Medical Image Segmentation;P. Guo, X. Su, H. Zhang, M. Wang and F. Bao;《ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)》;全文 *
Deep Learning with Skip Connection Attention for Choroid Layer Segmentation in OCT Images;X. Mao et al.;《2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC)》;全文 *
基于多尺度密集网络与两级残差注意力的乳腺肿瘤分割方法研究;赵引;《CNKI优秀硕士学位论文全文库》;全文 *

Also Published As

Publication number Publication date
CN113793345A (en) 2021-12-14

Similar Documents

Publication Publication Date Title
CN113793345B (en) Medical image segmentation method and device based on improved attention module
Mahapatra et al. Image super-resolution using progressive generative adversarial networks for medical image analysis
US11908580B2 (en) Image classification method, computer-readable storage medium, and computer device
US11645748B2 (en) Three-dimensional automatic location system for epileptogenic focus based on deep learning
US20190191988A1 (en) Screening method for automated detection of vision-degenerative diseases from color fundus images
WO2018120942A1 (en) System and method for automatically detecting lesions in medical image by means of multi-model fusion
US11449717B2 (en) System and method for identification and localization of images using triplet loss and predicted regions
KR20190136577A (en) Method for classifying images using deep neural network and apparatus using the same
CN113240655B (en) Method, storage medium and device for automatically detecting type of fundus image
WO2023207743A1 (en) Image detection method and apparatus, and computer device, storage medium and program product
CN113744271A (en) Neural network-based automatic optic nerve segmentation and compression degree measurement and calculation method
CN113782184A (en) Cerebral apoplexy auxiliary evaluation system based on facial key point and feature pre-learning
WO2023047118A1 (en) A computer-implemented method of enhancing object detection in a digital image of known underlying structure, and corresponding module, data processing apparatus and computer program
CN112614573A (en) Deep learning model training method and device based on pathological image labeling tool
CN116091522A (en) Medical image segmentation method, device, equipment and readable storage medium
JP2020010804A (en) Medical image diagnosis support apparatus, program, and medical image diagnosis support method
CN112862785B (en) CTA image data identification method, device and storage medium
CN112862786B (en) CTA image data processing method, device and storage medium
CN111369564B (en) Image processing method, model training method and model training device
CN112967246A (en) X-ray image auxiliary device and method for clinical decision support system
CN116258717B (en) Lesion recognition method, device, apparatus and storage medium
KR102442591B1 (en) Method, program, and apparatus for generating label
CN117392468B (en) Cancer pathology image classification system, medium and equipment based on multi-example learning
CN116580446B (en) Iris characteristic recognition method and system for vascular diseases
CN112862787B (en) CTA image data processing method, device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant