CN116758093B - Image segmentation method, model training method, device, equipment and medium - Google Patents
Image segmentation method, model training method, device, equipment and medium Download PDFInfo
- Publication number
- CN116758093B CN116758093B CN202310624615.0A CN202310624615A CN116758093B CN 116758093 B CN116758093 B CN 116758093B CN 202310624615 A CN202310624615 A CN 202310624615A CN 116758093 B CN116758093 B CN 116758093B
- Authority
- CN
- China
- Prior art keywords
- feature
- segmentation
- feature map
- generate
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 69
- 238000003709 image segmentation Methods 0.000 title claims abstract description 52
- 238000012549 training Methods 0.000 title claims abstract description 40
- 230000011218 segmentation Effects 0.000 claims abstract description 141
- 239000013598 vector Substances 0.000 claims abstract description 72
- 238000012545 processing Methods 0.000 claims abstract description 22
- 238000000605 extraction Methods 0.000 claims abstract description 10
- 230000004927 fusion Effects 0.000 claims description 32
- 238000013527 convolutional neural network Methods 0.000 claims description 25
- 238000010586 diagram Methods 0.000 claims description 18
- 238000013507 mapping Methods 0.000 claims description 15
- 238000011176 pooling Methods 0.000 claims description 15
- 238000002372 labelling Methods 0.000 claims description 13
- 238000005070 sampling Methods 0.000 claims description 12
- 230000006870 function Effects 0.000 claims description 8
- 238000004590 computer program Methods 0.000 claims description 6
- 238000005457 optimization Methods 0.000 abstract description 4
- 238000005516 engineering process Methods 0.000 description 16
- 238000012014 optical coherence tomography Methods 0.000 description 12
- 230000008569 process Effects 0.000 description 10
- 238000013473 artificial intelligence Methods 0.000 description 8
- 210000004204 blood vessel Anatomy 0.000 description 7
- 230000000694 effects Effects 0.000 description 7
- 239000011159 matrix material Substances 0.000 description 6
- 230000002490 cerebral effect Effects 0.000 description 5
- 230000002526 effect on cardiovascular system Effects 0.000 description 4
- 238000003384 imaging method Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 3
- 206010003211 Arteriosclerosis coronary artery Diseases 0.000 description 2
- 208000024172 Cardiovascular disease Diseases 0.000 description 2
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 description 2
- 208000026106 cerebrovascular disease Diseases 0.000 description 2
- 208000029078 coronary artery disease Diseases 0.000 description 2
- 208000026758 coronary atherosclerosis Diseases 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000010606 normalization Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 241000870659 Crassula perfoliata var. minor Species 0.000 description 1
- 206010028980 Neoplasm Diseases 0.000 description 1
- 208000037062 Polyps Diseases 0.000 description 1
- 208000007536 Thrombosis Diseases 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 230000000740 bleeding effect Effects 0.000 description 1
- 210000005013 brain tissue Anatomy 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 210000004351 coronary vessel Anatomy 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000003745 diagnosis Methods 0.000 description 1
- 201000010099 disease Diseases 0.000 description 1
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 description 1
- 230000036541 health Effects 0.000 description 1
- 201000011066 hemangioma Diseases 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000002513 implantation Methods 0.000 description 1
- 208000014674 injury Diseases 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 238000003333 near-infrared imaging Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000010422 painting Methods 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000004393 prognosis Methods 0.000 description 1
- 208000037803 restenosis Diseases 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 230000001225 therapeutic effect Effects 0.000 description 1
- 238000002560 therapeutic procedure Methods 0.000 description 1
- 230000008733 trauma Effects 0.000 description 1
- 230000002792 vascular Effects 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30016—Brain
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30101—Blood vessel; Artery; Vein; Vascular
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Mathematical Physics (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Quality & Reliability (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Image Analysis (AREA)
Abstract
The application relates to an image segmentation method, a model training method, a device, equipment and a medium, belongs to the technical field of computer image processing, and solves the problem that the calculation time is longer along with the increase of the scale of a problem in the existing optimization model. The technical scheme of the application mainly comprises the following steps: acquiring an image to be segmented, and carrying out feature extraction on the image to be segmented to acquire a first feature map; inputting the first feature map into a first segmentation network to generate a first segmentation result and a second feature map; generating a query vector through the second feature map; inputting the first feature map into a transducer module for high-dimensional feature coding to generate a third feature map, and simultaneously, acquiring the query vector by the transducer module to guide the high-dimensional feature coding; the third feature map is input into a second segmentation network to generate a second segmentation result.
Description
Technical Field
The application belongs to the technical field of computer image processing, and particularly relates to an image segmentation method, a model training method, a device, equipment and a medium.
Background
Worldwide, cardiovascular and cerebrovascular diseases have become one of the main diseases threatening human health, and coronary atherosclerosis is the main cause of cardiovascular and cerebrovascular diseases. At present, the coronary stent interventional therapy has become a main therapeutic scheme for treating coronary atherosclerosis due to small trauma and good effect. During treatment, stents are placed inside the coronary arteries by treatment to reduce the probability of restenosis and thrombosis.
After a cerebral vessel stent implantation operation is performed, the adherence of each stent wire is an important reference index for the operation effect and prognosis effect. In the process of measuring the adherence of the stent wire, the medical image of the intravascular stent is required to be utilized to segment the vessel wall and the stent so as to facilitate the acquisition of the respective spatial information of the vessel wall and the stent.
In the prior art, a multi-decoding multi-label segmentation model, such as a U-net network, is generally adopted for similar segmentation tasks, but the segmentation task targets for specific tasks actually have own spatial correlation, and the factor is not considered in the prior art, so that the segmentation accuracy of the segmentation task targets is still to be improved.
Disclosure of Invention
In view of the above analysis, the embodiments of the present invention are directed to an image segmentation method, a model training method, a device, a facility, and a medium, which are used for solving the problem that the segmentation accuracy is not high enough due to the fact that the existing image segmentation technology does not consider the spatial relationship existing in the segmentation target itself.
An embodiment of a first aspect of the present application provides an image segmentation method, including the steps of:
Acquiring an image to be segmented, and carrying out feature extraction on the image to be segmented to acquire a first feature map; inputting the first feature map into a first segmentation network to generate a first segmentation result and a second feature map;
Generating a query vector through the second feature map;
Inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and simultaneously, acquiring the query vector by the transducer module to guide the high-dimensional feature decoding;
the third feature map is input into a second segmentation network to generate a second segmentation result.
In some embodiments, the transducer module comprises a feature mapping unit, a transducer encoder, a transducer decoder, and a multi-headed attention module;
The inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and the transducer module obtaining the query vector to guide the high-dimensional decoding, includes:
Performing feature mapping on the first feature map to generate a feature vector, and performing position coding processing on the feature vector;
The feature vector is acquired by the transducer encoder to generate a first feature code;
the first feature code and the query vector are obtained by the transducer decoder to generate a second feature code;
the first feature code and the second feature code are acquired by the multi-headed attention module to generate a high-dimensional feature.
In some embodiments, the generating the high-dimensional feature further comprises:
and carrying out feature fusion on the high-dimensional features and the feature vectors to generate fusion features, and generating the third feature map through deconvolution according to the fusion features.
In some embodiments, the acquiring an image to be segmented, and extracting features of the image to be segmented to obtain a first feature map, includes:
Inputting the image to be segmented into a CNN convolutional neural network, wherein the CNN convolutional neural network comprises an initialization module and M first residual error modules, the initialization module is used for extracting initial characteristics of the image to be segmented, the first residual error modules comprise a pooling layer and a first convolution layer, the pooling layer is used for downsampling the output of the initialization module or the previous first residual error modules, and the first convolution layer is used for carrying out convolution operation on the output of the pooling layer.
In some embodiments, the first and second split networks have the same split network structure, the split network structure comprising:
m second residual error modules and an output module;
the second residual error module comprises an up-sampling layer, a fusion layer and a second convolution layer, wherein the up-sampling layer is used for up-sampling the acquired characteristics, the fusion layer is used for carrying out characteristic fusion on the output of the up-sampling layer and the output of the corresponding first residual error module, and the second convolution layer is used for carrying out convolution operation on the output of the fusion layer;
the output module includes a third convolution layer and a Sigmoid function.
In some embodiments, the generating a query vector from the second feature map includes:
feature mapping is carried out through a convolution kernel with the size of 1x 1;
acquiring characteristics in the channel dimension through global average pooling;
performing dimension expansion according to the number of dimensions required by the transducer module;
random noise perturbations are added to each dimension separately to form the query vector.
An embodiment of the second aspect of the present application provides a training method for an image segmentation model, including the steps of:
Constructing a training data set, wherein the training data set comprises a plurality of training images with a first segmentation target and a second segmentation target, the training images are provided with a first label and a second label, the first label is a labeling area of the first segmentation target, and the second label is a labeling area of the second segmentation target;
Extracting features of the training image to obtain a first feature map;
inputting the first feature map into a first segmentation network to generate a first segmentation result and a second feature map;
Generating a query vector through the second feature map;
Inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and simultaneously, acquiring the query vector by the transducer module to guide the high-dimensional feature decoding;
Inputting the third feature map into a second segmentation network to generate a second segmentation result;
Determining a first loss value according to the difference between the first segmentation result and the first label, determining a second loss value according to the difference between the second segmentation result and the second label, and performing joint training based on the first loss value and the second loss value to obtain the image segmentation model, wherein the image segmentation model is used for segmenting the first target and the second target in an image to be segmented.
An embodiment of a third aspect of the present application provides an image segmentation apparatus, including:
The acquisition module is used for acquiring an image to be segmented, and extracting the characteristics of the image to be segmented to obtain a first characteristic image;
The first segmentation module inputs the first feature map into a first segmentation network to generate a first segmentation result and a second feature map;
The query vector generation module is used for generating a query vector through the second feature map;
The feature coding module is used for inputting the first feature map into a transducer module to perform high-dimensional feature coding so as to generate a third feature map, and meanwhile, the transducer module is used for acquiring the query vector and guiding the high-dimensional feature decoding;
and the second segmentation module inputs the third characteristic diagram into a second segmentation network to generate a second segmentation result.
An embodiment of a fourth aspect of the present application provides an electronic device, including a memory and a processor, the memory storing a computer program which, when executed by the processor, implements the image segmentation method and/or the training method of the image segmentation model as described in any of the embodiments above.
An embodiment of a fifth aspect of the present application provides a computer-readable storage medium, on which a computer program is stored which, when executed by a processor, implements the image segmentation method and/or the training method of an image segmentation model as described in any of the embodiments above.
On one hand, the embodiment of the application considers the spatial distribution relation of the segmented targets, utilizes a transducer to carry out high-dimensional feature coding, and enhances the segmentation effect through the spatial correlation characteristic of the segmented targets. On the other hand, considering the specific spatial position relation among different segmentation targets, generating a query vector through the characteristics of the first segmentation target, inputting the query vector into a transducer module, and guiding the encoding of the second segmentation target so as to realize the accurate segmentation and positioning of the second segmentation target.
Drawings
In order to more clearly illustrate the embodiments of the present description or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the embodiments of the present description, and other drawings may be obtained according to these drawings for a person having ordinary skill in the art.
FIG. 1 is an application environment diagram of an image segmentation method according to an embodiment of the present invention;
Fig. 2 is a schematic flow chart of an image segmentation method according to an embodiment of the first aspect of the present invention;
FIG. 3 is a schematic diagram of a segmentation model architecture of an image segmentation method according to an embodiment of the first aspect of the present invention;
fig. 4 is a schematic structural diagram of a CNN convolutional neural network according to a first embodiment of the present invention;
FIG. 5 is a schematic diagram of a split network according to an embodiment of the first aspect of the present invention;
FIG. 6 is a schematic diagram of a query volume generation module according to an embodiment of the first aspect of the present invention;
FIG. 7 is a schematic diagram of a network architecture of a transducer module according to an embodiment of the first aspect of the present invention;
FIG. 8 is a flowchart of an image segmentation model training method according to a second embodiment of the present invention;
FIG. 9 is a schematic illustration of a mask for labeling an image set according to an embodiment of the second aspect of the invention;
Fig. 10 is a schematic diagram of an image segmentation apparatus according to an embodiment of the third aspect of the present invention;
Fig. 11 is a schematic diagram of an electronic device architecture according to the present invention.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present application more apparent, the technical solutions of the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present application, and it is apparent that the described embodiments are some embodiments of the present application, but not all embodiments of the present application. It should be noted that embodiments and features of embodiments in the present disclosure may be combined, separated, interchanged, and/or rearranged with one another without conflict. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting. As used herein, the singular forms "a", "an" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. Furthermore, when the terms "comprises" and/or "comprising," and variations thereof, are used in the present specification, the presence of stated features, integers, steps, operations, elements, components, and/or groups thereof is described, but the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof is not precluded. It is also noted that, as used herein, the terms "substantially," "about," and other similar terms are used as approximation terms and not as degree terms, and as such, are used to explain the inherent deviations of measured, calculated, and/or provided values that would be recognized by one of ordinary skill in the art.
Artificial intelligence (ARTIFICIAL INTELLIGENCE, AI) is the theory, method, technique, and application system that simulates, extends, and extends human intelligence using a digital computer or a machine controlled by a digital computer, perceives the environment, obtains knowledge, and uses the knowledge to obtain optimal results. In other words, artificial intelligence is an integrated technology of computer science that attempts to understand the essence of intelligence and to produce a new intelligent machine that can react in a similar way to human intelligence. Artificial intelligence, i.e. research on design principles and implementation methods of various intelligent machines, enables the machines to have functions of sensing, reasoning and decision.
The artificial intelligence technology is a comprehensive subject, and relates to the technology with wide fields, namely the technology with a hardware level and the technology with a software level. Artificial intelligence infrastructure technologies generally include technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technologies, operation/interaction systems, mechatronics, and the like. The artificial intelligence software technology mainly comprises a computer vision technology, a voice processing technology, a natural language processing technology, machine learning/deep learning and other directions.
Computer Vision (CV) is a science of studying how to "look" a machine, and more specifically, to replace a human eye with a camera and a Computer to perform machine Vision such as recognition and measurement on a target, and further perform graphic processing to make the Computer process an image more suitable for human eye observation or transmission to an instrument for detection. As a scientific discipline, computer vision research-related theory and technology has attempted to build artificial intelligence systems that can acquire information from images or multidimensional data. Computer vision techniques typically include image processing, image recognition, image semantic understanding, image retrieval, OCR, video processing, video semantic understanding, video content/behavior recognition, three-dimensional object reconstruction, 3D techniques, virtual reality, augmented reality, synchronous positioning, and map construction, among others, as well as common biometric recognition techniques such as face recognition, fingerprint recognition, and others.
Image segmentation is a technique and process of dividing an image into several specific regions with unique properties and presenting objects of interest. It is a key step from image processing to image analysis. The existing image segmentation methods are mainly divided into the following categories: a threshold-based segmentation method, a region-based segmentation method, an edge-based segmentation method, a segmentation method based on a specific theory, and the like. From a mathematical perspective, image segmentation is the process of dividing a digital image into mutually disjoint regions. The process of image segmentation is also a labeling process, i.e. pixels belonging to the same region are given the same number.
To automate the image segmentation task, a multi-decoding multi-labeled segmentation model, such as a U-net network, is typically employed for similar segmentation tasks. However, for the specific task, the space correlation exists in the segmentation task targets actually, and the specific correlation of the segmentation targets and the specific space distribution mode of the targets are not considered, so that the method has room for improvement in the accurate segmentation and positioning effects of the segmentation targets.
An embodiment of the invention provides an image segmentation method, which is based on an asymmetric multi-task transducer segmentation model, realizes full-automatic and multi-task segmentation target parallelization processing when an image is segmented, and improves the precision of segmentation tasks by utilizing the interrelationship between different segmentation targets and the self spatial distribution characteristics of a single segmentation target when the model is trained.
For ease of understanding, referring to fig. 1, fig. 1 is an application environment diagram of an image segmentation method according to an embodiment of the present application, and as shown in fig. 1, an image processing method according to an embodiment of the present application is applied to an image processing system. The image processing system includes: server and terminal equipment. The image processing apparatus may be disposed on a server or may be disposed on a terminal device, and the embodiment of the present application is described by taking the disposition on the server as an example, which should not be construed as limiting the present application. The server may be an independent physical server, a server cluster or a distributed system formed by a plurality of physical servers, or a cloud server providing cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, a content distribution network (Content Delivery Network, CDN), basic cloud computing services such as big data and an artificial intelligent platform. Terminals include, but are not limited to, cell phones, computers, intelligent voice interaction devices, intelligent appliances, vehicle terminals, aircraft, and the like. The terminal and the server may be directly or indirectly connected through wired or wireless communication, and embodiments of the present application are not limited herein. The embodiment of the application can be applied to various medical image segmentation scenes, including but not limited to segmentation of vascular stents, brain tissue segmentation, tumor segmentation or polyp segmentation, and the like.
Firstly, a server acquires an image to be segmented, and performs feature extraction on the image to be segmented to acquire a first feature map; the server then inputs the first feature map into a first segmentation network to generate a first segmentation result and a second feature map; the server generates a query vector through the second feature map; meanwhile, the server inputs the first feature map into a transducer module to perform high-dimensional feature encoding so as to generate a third feature map, and meanwhile, the transducer module acquires the query vector to guide the high-dimensional feature decoding; the server then enters the third feature map into a second segmentation network to generate a second segmentation result.
The optical coherence tomography (optical coherence tomography, OCT) in the cardiovascular and cerebrovascular vessels is a novel cardiovascular and cerebrovascular imaging technology. Taking a cerebrovascular vessel as an example, the OCT imaging catheter moves forwards along the vessel, and performs fine imaging on the vessel wall with a certain thickness through near infrared imaging to assist diagnosis and treatment (plaque, hemangioma, bleeding and the like) of the cerebrovascular cavity, so that the OCT image in the cerebral vessel consists of a series of axial slices of the cerebrovascular cavity sequentially, and the middle area is a near-circular vessel cavity. In this embodiment, the image segmentation method of this embodiment will be described by taking the task of segmenting an OCT image of a cerebrovascular stent as an example, and this embodiment may also be referred to as a blood vessel and stent segmentation method based on the OCT image.
The image segmentation method provided by the embodiment of the first aspect of the present invention will be described from the perspective of a server. Referring to fig. 2 and fig. 3, fig. 2 is a flow chart of an image segmentation method according to an embodiment of the first aspect of the present invention, and fig. 3 is a schematic view of a segmentation model architecture of the image segmentation method according to the embodiment of the first aspect of the present invention; an image segmentation method provided by an embodiment of a first aspect of the present invention includes:
Step one, obtaining an image to be segmented, and carrying out feature extraction on the image to be segmented to obtain a first feature map.
It should be understood that, in this embodiment, the segmentation task of the OCT image of the stent cerebral blood vessel is described, the image to be segmented is one of a series of OCT images of the stent cerebral blood vessel, the first segmentation target is a blood vessel lumen or a blood vessel wall in the image, the second segmentation target is a stent in the image, and the primary feature map of the image to be segmented is extracted, where the feature extraction is generally performed by using a CNN convolutional neural network.
Preferably, in some embodiments, the acquiring an image to be segmented, and extracting features of the image to be segmented to obtain a first feature map, includes:
The image to be segmented is input into a CNN convolutional neural network, as shown in fig. 4, and fig. 4 is a schematic structural diagram of the CNN convolutional neural network according to the first embodiment of the present invention. The CNN convolutional neural network comprises an initialization module and M first residual modules, wherein the initialization module is used for extracting initial characteristics of an image to be segmented, the first residual modules comprise a pooling layer and a first convolution layer, the pooling layer is used for downsampling the output of the initialization module or the previous first residual modules, and the first convolution layer is used for carrying out convolution operation on the output of the pooling layer. In this embodiment, the value of M is 4.
Specifically, the convolutional neural network used by the invention is used as a basic network skeleton, namely CNN backhaul, to perform low-level feature extraction. Firstly, extracting low-order visual features through an initialization module, inputting an original OCT image matrix into the initialization module, wherein the initialization module comprises three convolution layers, each convolution layer carries out convolution and batch normalization (Batch Normalization, BN) on the input image matrix, and activates the input image matrix by adopting a linear rectification function (RECTIFIED LINEAR Unit, reLU), outputting a feature matrix and inputting the feature matrix into the next convolution layer, and completing the extraction of the low-order visual features after passing through the three convolution layers to obtain initial features.
And then inputting the initial features into the feature extraction of depth in first residual modules after the initial modules, wherein each first residual module comprises a maximum pooling layer, two first convolution layers, a residual channel and a ReLU. And after four downsampling and convolution processes, finally outputting high-dimensional characteristics. The size of the input image matrix is 1x512x512 (CxHxW), after 4 residual convolution blocks, the output characteristic size is 512x32x32 (CxHxW), and the final image characteristic is provided for the subsequent modules to process. It is noted that the output of each first residual module is input into the subsequent first residual module, and the fourth first residual module outputs a first feature map of size 512x32x32 (CxHxW), on the one hand, and also via a jump connection, on the other hand.
In this embodiment, the extraction of the first feature map is implemented by using a CNN convolutional neural network.
The first feature map is then input to the corresponding processing modules of step two and step four for processing, respectively. It should be understood that the second step and the fourth step are only steps to be distinguished, and are not sequential.
Inputting the first feature map into a first segmentation network to generate a first segmentation result and a second feature map.
In this embodiment, the stent and the vessel lumen are segmented, wherein the segmentation task of the vessel wall is relatively simple and easy, so that the vessel lumen is segmented directly through the first segmentation network according to the first feature map, and the generated first segmentation result, that is, the vessel lumen slice, is processed through the third step. Wherein the first segmentation network and the second segmentation network following stent segmentation are both segmentation networks for a specific target.
Preferably, in some embodiments, the first split network and the second split network in this embodiment have the same split network structure, and the specific split network structure is shown in fig. 5, and fig. 5 is a schematic view of the split network structure according to the first aspect of the present invention. The split network architecture includes: m second residual error modules and an output module; the second residual error module comprises an up-sampling layer, a fusion layer and a second convolution layer, wherein the up-sampling layer is used for up-sampling the acquired characteristics, the fusion layer is used for carrying out characteristic fusion on the output of the up-sampling layer and the corresponding output of the first residual error module, and the second convolution layer is used for carrying out convolution operation on the output of the fusion layer. The output module includes a third convolution layer and a Sigmoid function.
In particular, the first segmentation network of the vessel lumen is also referred to as the first segmentation head shown in fig. 3, the second segmentation network of the stent is also referred to as the second segmentation head shown in fig. 3, the actual working of the second residual block is decoding the first feature map, and the second residual block may also be referred to as a decoding convolution block. The second residual module performs upsampling by using bilinear interpolation first, and performs feature fusion with the features from the first residual module, where the feature fusion should be that the size of the feature map output in CNN matches the size of the feature map output by the second residual module, for example, the first second residual module performs feature fusion with the output of the third first residual module after upsampling. Then processing through a 2-layer second convolution layer, simultaneously carrying out residual connection on the fusion characteristic and the outputs of the two second convolution layers, and then activating through a ReLU.
The dimension of the feature map is reduced after the feature map is decoded by the four second residual modules, the dimension is restored to the size of an input image, the feature map is input to a third convolution layer of an output module, the third convolution layer comprises a convolution layer with a convolution kernel size of 3x3 and a convolution layer with a convolution kernel size of 1x1, the probability of each pixel classification is output after the convolution is processed by a Sigmoid function, and a segmentation result is obtained after threshold processing.
In this embodiment, the precision of the segmentation result is further improved through multi-level decoding and encoding and feature fusion.
And thirdly, generating a query vector through the second feature map.
In some embodiments, preferably, the generating a query vector by the second feature map includes:
feature mapping is carried out through a convolution kernel with the size of 1x 1;
acquiring characteristics in the channel dimension through global average pooling;
performing dimension expansion according to the number of dimensions required by the transducer module;
random noise perturbations are added to each dimension separately to form the query vector.
Specifically, in order to make the segmentation of the model to the stent more accurate, query generated by the lumen segmentation is used as a Query vector and input into a Transformer Decoder decoder to guide the model to more accurately locate and segment the region where the stent exists. In the processing procedure of the query vector generation module, as shown in fig. 6, fig. 6 is a schematic architecture diagram of the query vector generation module according to the embodiment of the first aspect of the present invention. The output result of the inner cavity segmentation head is subjected to feature mapping through two-dimensional convolution with the convolution kernel size of 1, features in the channel dimension are obtained through global average pooling (Global Average Pooling, GAP), dimension expansion is carried out according to the number of query vectors, noise disturbance is added, and the final query feature vector is input into a Transformer Decoder decoder to guide stent segmentation.
In this embodiment, by generating a query vector for guiding stent segmentation according to the feature map of the vessel lumen segmentation result, the accuracy of stent segmentation and positioning can be improved.
And step four, inputting the first feature map into a transducer module for high-dimensional feature encoding so as to generate a third feature map, and simultaneously, acquiring the query vector by the transducer module to guide the high-dimensional feature decoding.
Preferably, in some embodiments, as shown in fig. 7, fig. 7 is a schematic diagram of a network architecture of a transducer module according to an embodiment of the first aspect of the present invention. The transducer module comprises a feature mapping unit, a transducer encoder, a transducer decoder and a multi-head attention module;
the inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and the transducer module obtaining the query vector to guide the high-dimensional encoding includes:
Performing feature mapping on the first feature map to generate a feature vector, and performing position coding processing on the feature vector;
The feature vector is acquired by the transducer encoder to generate a first feature code;
the first feature code and the query vector are obtained by the transducer decoder to generate a second feature code;
The first feature code and the second feature code are acquired by the multi-headed attention module to generate a high-dimensional feature;
and generating the third characteristic diagram through deconvolution according to the high-dimensional characteristics.
Specifically, as shown in fig. 7, fig. 7 is a schematic diagram of a architecture of a transducer codec module according to a first embodiment of the present invention. The basic composition of Transformer Encoder Decoder modules is 3 Transformer Encoder and 2 Tranformer Decoder. In the process of high-dimensional coding, firstly, feature images extracted from a CNN convolutional neural network are converted into image feature vectors through feature mapping, the image feature vectors are input into a Transformer Encoder module for higher-dimensional coding after feature position coding is added, the coded result is input into a Transformer Decoder module for decoding, meanwhile, a Transformer Decoder module receives query vectors generated by an inner cavity segmentation result as segmentation guide, and finally, the Transformer Decoder decoder output and the Transformer Encoder module output are subjected to multi-head attention so as to generate high-dimensional feature coding, namely high-dimensional features. Multi-head attention is calculated in parallel through a plurality of independent attention as an integrated effect, so that overfitting is prevented.
Preferably, in some embodiments, the generating the high-dimensional feature further includes:
and carrying out feature fusion on the high-dimensional features and the feature vectors to generate fusion features, and generating the third feature map through deconvolution according to the fusion features.
In this embodiment, the transform is used to perform high-dimensional feature encoding, and the segmentation effect is enhanced by the spatial correlation characteristic of the segmentation target itself. Meanwhile, the generated high-dimensional features and low-dimensional features are fused and then segmented, so that the segmentation accuracy can be further improved.
And fifthly, inputting the third characteristic diagram into a second segmentation network to generate a second segmentation result.
The structure of the second split network is similar to that of the first split network, and will not be described here again.
An embodiment of the second aspect of the present invention provides an image segmentation model training method, as shown in fig. 8, including:
Constructing a training data set, wherein the training data set comprises a plurality of training images with a first segmentation target and a second segmentation target, the training images are provided with a first label and a second label, the first label is a labeling area of the first segmentation target, and the second label is a labeling area of the second segmentation target;
Extracting features of the training image to obtain a first feature map;
inputting the first feature map into a first segmentation network to generate a first segmentation result and a second feature map;
Generating a query vector through the second feature map;
Inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and simultaneously, acquiring the query vector by the transducer module to guide the high-dimensional feature decoding;
Inputting the third feature map into a second segmentation network to generate a second segmentation result;
Determining a first loss value according to the difference between the first segmentation result and the first label, determining a second loss value according to the difference between the second segmentation result and the second label, and performing joint training based on the first loss value and the second loss value to obtain the image segmentation model, wherein the image segmentation model is used for segmenting the first target and the second target in an image to be segmented.
Specifically, the training dataset, that is, the OCT cerebrovascular dataset in this embodiment is manually labeled, and after the labeling is completed and verified, the training dataset is made into a dataset for the deep learning algorithm, the labeling software uses the open source medical labeling software ITK-SNAP, and the labeling of the cerebrovascular lumen and the stent is performed by using a painting tool (Paint Brush) in the labeling tool. The labeling samples total 15 cases, namely 15 cases of original medical Digital Imaging and Communication (DICOM) files are used. The labeled objects are a blood vessel inner cavity and a bracket, and in order to solve the problem of overlapping of the areas of the blood vessel inner cavity and the bracket, the two targets are labeled independently. The artificial labeling results of the cerebral vessel lumen and the stent are shown in fig. 9, and the areas pointed by arrows are the mask of the vessel lumen, namely the first label, and the mask of the stent, namely the second label.
In the present embodiment, when creating an image set, an OCT original image and a label are created as a data set in units of DICOM samples. The dataset amounted to 15 samples and the final dataset produced amounted to 2458 images. Randomly selecting 11 samples to train, wherein the total number of the samples is 1855, and the ratio is 75%; randomly selecting 2 samples for verification during training, wherein the total 373 pictures account for 15 percent; the rest 2 samples are taken as independent test sets, and the total number of the samples is 230, and the ratio is 10%; during the training process, successive image slices in a single DICOM sample are placed into the network for training and testing.
For the structure of the model, see the relevant description of the embodiments of the first aspect, which is not repeated here.
Preferably, the loss of the network model, including the first loss value and the second loss value, is determined by a combination of cross entropy of two classifications (Binary CrossEntropy Loss, BCELoss) and TerskyLoss, where BCELoss is used for classification of pixel points and TerskyLoss is used to balance the equalization of positive and negative samples.
In particular, the method comprises the steps of,Where y i is the label, p i is the predicted value of the model, and N is the number of total pixels.
Where y i is the label, p i is the predicted value of the model, and N is the number of total pixels. Alpha and beta are the corresponding weights.
The first loss value or the second loss value may be expressed as:
Loss=λ BCELoss + (1λ) TverskyLoss, where λ is the weight parameter.
In this embodiment, the combined optimization training is performed after the first loss value, which is the loss of the vessel lumen segmentation, and the second loss value, which is the loss of the stent segmentation, are added, the model uses a random gradient descent algorithm to perform optimization of network parameters, and the training model uses an Adam optimizer with faster convergence to perform rapid optimization of parameters of the model.
An image segmentation apparatus according to an embodiment of the third aspect of the present application, as shown in fig. 10, includes:
The acquisition module is used for acquiring an image to be segmented, and extracting the characteristics of the image to be segmented to obtain a first characteristic image;
The first segmentation module inputs the first feature map into a first segmentation network to generate a first segmentation result and a second feature map;
The query vector generation module is used for generating a query vector through the second feature map;
The feature coding module is used for inputting the first feature map into a transducer module to perform high-dimensional feature coding so as to generate a third feature map, and meanwhile, the transducer module is used for acquiring the query vector and guiding the high-dimensional feature decoding;
and the second segmentation module inputs the third characteristic diagram into a second segmentation network to generate a second segmentation result.
An embodiment of a fourth aspect of the present application provides an electronic device, including a memory and a processor, the memory storing a computer program which, when executed by the processor, implements the image segmentation method and/or the training method of the image segmentation model as described in any of the embodiments above.
An embodiment of a fifth aspect of the present application provides a computer-readable storage medium, on which a computer program is stored which, when executed by a processor, implements the image segmentation method and/or the training method of an image segmentation model as described in any of the embodiments above.
Computer-readable storage media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static Random Access Memory (SRAM), dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), read Only Memory (ROM), electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, read only compact disc read only memory (CD-ROM), digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium, which can be used to store information that can be accessed by the computing device. Computer-readable media, as defined herein, does not include transitory computer-readable media (transmission media), such as modulated data signals and carrier waves.
Those of skill would further appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative elements and steps are described above generally in terms of function in order to clearly illustrate the interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied in hardware, in a software module executed by a processor, or in a combination of the two. The software modules may be disposed in Random Access Memory (RAM), memory, read Only Memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
The foregoing description of the embodiments has been provided for the purpose of illustrating the general principles of the application, and is not meant to limit the scope of the application, but to limit the application to the particular embodiments, and any modifications, equivalents, improvements, etc. that fall within the spirit and principles of the application are intended to be included within the scope of the application.
Claims (8)
1. An image segmentation method, characterized by comprising the steps of:
Acquiring an image to be segmented, and carrying out feature extraction on the image to be segmented to acquire a first feature map;
inputting the first feature map into a first segmentation network to generate a first segmentation result and a second feature map, the first segmentation result comprising a vessel wall;
Generating a query vector through the second feature map;
Inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and simultaneously, acquiring the query vector by the transducer module to guide the high-dimensional feature decoding;
Inputting the third feature map into a second segmentation network to generate a second segmentation result, the second segmentation result comprising a stent;
The transducer module comprises a feature mapping unit, a transducer encoder, a transducer decoder and a multi-head attention module;
The inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and the transducer module obtaining the query vector to guide the high-dimensional feature decoding, includes:
Performing feature mapping on the first feature map to generate a feature vector, and performing position coding processing on the feature vector;
The feature vector is acquired by the transducer encoder to generate a first feature code;
the first feature code and the query vector are obtained by the transducer decoder to generate a second feature code;
The first feature code and the second feature code are acquired by the multi-headed attention module to generate a high-dimensional feature;
and carrying out feature fusion on the high-dimensional features and the feature vectors to generate fusion features, and generating the third feature map through deconvolution according to the fusion features.
2. The image segmentation method as set forth in claim 1, wherein: the obtaining the image to be segmented, and extracting the features of the image to be segmented to obtain a first feature map includes:
Inputting the image to be segmented into a CNN convolutional neural network, wherein the CNN convolutional neural network comprises an initialization module and M first residual error modules, the initialization module is used for extracting initial characteristics of the image to be segmented, the first residual error modules comprise a pooling layer and a first convolution layer, the pooling layer is used for downsampling the output of the initialization module or the previous first residual error modules, and the first convolution layer is used for carrying out convolution operation on the output of the pooling layer.
3. The image segmentation method as set forth in claim 2, wherein: the first and second split networks have the same split network structure, the split network structure comprising:
m second residual error modules and an output module;
the second residual error module comprises an up-sampling layer, a fusion layer and a second convolution layer, wherein the up-sampling layer is used for up-sampling the acquired characteristics, the fusion layer is used for carrying out characteristic fusion on the output of the up-sampling layer and the output of the corresponding first residual error module, and the second convolution layer is used for carrying out convolution operation on the output of the fusion layer;
the output module includes a third convolution layer and a Sigmoid function.
4. The image segmentation method as set forth in claim 1, wherein: the generating the query vector through the second feature map includes:
feature mapping is carried out through a convolution kernel with the size of 1x 1;
acquiring characteristics in the channel dimension through global average pooling;
performing dimension expansion according to the number of dimensions required by the transducer module;
random noise perturbations are added to each dimension separately to form the query vector.
5. The training method of the image segmentation model is characterized by comprising the following steps of:
Constructing a training data set, wherein the training data set comprises a plurality of training images with a first segmentation target and a second segmentation target, the training images are provided with a first label and a second label, the first label is a labeling area of the first segmentation target, and the second label is a labeling area of the second segmentation target;
Extracting features of the training image to obtain a first feature map;
inputting the first feature map into a first segmentation network to generate a first segmentation result and a second feature map, the first segmentation result comprising a vessel wall;
Generating a query vector through the second feature map;
Inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and simultaneously, acquiring the query vector by the transducer module to guide the high-dimensional feature decoding;
Inputting the third feature map into a second segmentation network to generate a second segmentation result, the second segmentation result comprising a stent;
Determining a first loss value according to the difference between the first segmentation result and the first label, determining a second loss value according to the difference between the second segmentation result and the second label, and performing joint training based on the first loss value and the second loss value to obtain an image segmentation model, wherein the image segmentation model is used for segmenting the first segmentation target and the second segmentation target in an image to be segmented;
The transducer module comprises a feature mapping unit, a transducer encoder, a transducer decoder and a multi-head attention module;
The inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and the transducer module obtaining the query vector to guide the high-dimensional feature decoding, includes:
Performing feature mapping on the first feature map to generate a feature vector, and performing position coding processing on the feature vector;
The feature vector is acquired by the transducer encoder to generate a first feature code;
the first feature code and the query vector are obtained by the transducer decoder to generate a second feature code;
The first feature code and the second feature code are acquired by the multi-headed attention module to generate a high-dimensional feature;
and carrying out feature fusion on the high-dimensional features and the feature vectors to generate fusion features, and generating the third feature map through deconvolution according to the fusion features.
6. An image dividing apparatus, comprising:
The acquisition module is used for acquiring an image to be segmented, and extracting the characteristics of the image to be segmented to obtain a first characteristic image;
A first segmentation module that inputs the first feature map into a first segmentation network to generate a first segmentation result and a second feature map, the first segmentation result comprising a vessel wall;
The query vector generation module is used for generating a query vector through the second feature map;
The feature coding module is used for inputting the first feature map into a transducer module to perform high-dimensional feature coding so as to generate a third feature map, and meanwhile, the transducer module is used for acquiring the query vector and guiding the high-dimensional feature decoding;
The second segmentation module inputs the third characteristic diagram into a second segmentation network to generate a second segmentation result, and the second segmentation result comprises a bracket;
The transducer module comprises a feature mapping unit, a transducer encoder, a transducer decoder and a multi-head attention module;
The inputting the first feature map into a transducer module for high-dimensional feature encoding to generate a third feature map, and the transducer module obtaining the query vector to guide the high-dimensional feature decoding, includes:
Performing feature mapping on the first feature map to generate a feature vector, and performing position coding processing on the feature vector;
The feature vector is acquired by the transducer encoder to generate a first feature code;
the first feature code and the query vector are obtained by the transducer decoder to generate a second feature code;
The first feature code and the second feature code are acquired by the multi-headed attention module to generate a high-dimensional feature;
and carrying out feature fusion on the high-dimensional features and the feature vectors to generate fusion features, and generating the third feature map through deconvolution according to the fusion features.
7. An electronic device comprising a memory and a processor, the memory storing a computer program which, when executed by the processor, implements the image segmentation method of any one of claims 1-4 and/or the training method of the image segmentation model of claim 5.
8. A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, implements the image segmentation method according to any one of claims 1-4 and/or the training method of the image segmentation model according to claim 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310624615.0A CN116758093B (en) | 2023-05-30 | 2023-05-30 | Image segmentation method, model training method, device, equipment and medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202310624615.0A CN116758093B (en) | 2023-05-30 | 2023-05-30 | Image segmentation method, model training method, device, equipment and medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN116758093A CN116758093A (en) | 2023-09-15 |
CN116758093B true CN116758093B (en) | 2024-05-07 |
Family
ID=87958087
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310624615.0A Active CN116758093B (en) | 2023-05-30 | 2023-05-30 | Image segmentation method, model training method, device, equipment and medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116758093B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117422848B (en) * | 2023-10-27 | 2024-08-16 | 神力视界(深圳)文化科技有限公司 | Method and device for segmenting three-dimensional model |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113191285A (en) * | 2021-05-08 | 2021-07-30 | 山东大学 | River and lake remote sensing image segmentation method and system based on convolutional neural network and Transformer |
WO2021179205A1 (en) * | 2020-03-11 | 2021-09-16 | 深圳先进技术研究院 | Medical image segmentation method, medical image segmentation apparatus and terminal device |
CN113420827A (en) * | 2021-07-08 | 2021-09-21 | 上海浦东发展银行股份有限公司 | Semantic segmentation network training and image semantic segmentation method, device and equipment |
CN115147598A (en) * | 2022-06-02 | 2022-10-04 | 粤港澳大湾区数字经济研究院(福田) | Target detection segmentation method and device, intelligent terminal and storage medium |
CN115375698A (en) * | 2022-08-22 | 2022-11-22 | 武汉理工大学重庆研究院 | Medical image segmentation method and device, electronic equipment and storage medium |
CN115761222A (en) * | 2022-09-27 | 2023-03-07 | 阿里巴巴(中国)有限公司 | Image segmentation method, remote sensing image segmentation method and device |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220148188A1 (en) * | 2020-11-06 | 2022-05-12 | Tasty Tech Ltd. | System and method for automated simulation of teeth transformation |
US11580646B2 (en) * | 2021-03-26 | 2023-02-14 | Nanjing University Of Posts And Telecommunications | Medical image segmentation method based on U-Net |
-
2023
- 2023-05-30 CN CN202310624615.0A patent/CN116758093B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2021179205A1 (en) * | 2020-03-11 | 2021-09-16 | 深圳先进技术研究院 | Medical image segmentation method, medical image segmentation apparatus and terminal device |
CN113191285A (en) * | 2021-05-08 | 2021-07-30 | 山东大学 | River and lake remote sensing image segmentation method and system based on convolutional neural network and Transformer |
CN113420827A (en) * | 2021-07-08 | 2021-09-21 | 上海浦东发展银行股份有限公司 | Semantic segmentation network training and image semantic segmentation method, device and equipment |
CN115147598A (en) * | 2022-06-02 | 2022-10-04 | 粤港澳大湾区数字经济研究院(福田) | Target detection segmentation method and device, intelligent terminal and storage medium |
CN115375698A (en) * | 2022-08-22 | 2022-11-22 | 武汉理工大学重庆研究院 | Medical image segmentation method and device, electronic equipment and storage medium |
CN115761222A (en) * | 2022-09-27 | 2023-03-07 | 阿里巴巴(中国)有限公司 | Image segmentation method, remote sensing image segmentation method and device |
Non-Patent Citations (2)
Title |
---|
Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers;Sixiao Zheng等;《arXiv:2012.15840v3》;20210725;第1-12页 * |
融合图卷积和差异性池化函数的点云数据分类分割模型;张新良;付鹏飞;赵运基;谢恒;王琬如;;中国图象图形学报;20200616(第06期);第137-144页 * |
Also Published As
Publication number | Publication date |
---|---|
CN116758093A (en) | 2023-09-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113159056A (en) | Image segmentation method, device, equipment and storage medium | |
CN113256592B (en) | Training method, system and device of image feature extraction model | |
Naval Marimont et al. | Implicit field learning for unsupervised anomaly detection in medical images | |
CN116758093B (en) | Image segmentation method, model training method, device, equipment and medium | |
CN112070781A (en) | Processing method and device of craniocerebral tomography image, storage medium and electronic equipment | |
CN109447096B (en) | Glance path prediction method and device based on machine learning | |
CN115546270A (en) | Image registration method, model training method and equipment for multi-scale feature fusion | |
CN112560710A (en) | Method for constructing finger vein recognition system and finger vein recognition system | |
CN116977663A (en) | Image data processing method, device, equipment and medium | |
CN114677349A (en) | Image segmentation method and system for edge information enhancement and attention guidance of encoding and decoding | |
CN117974693A (en) | Image segmentation method, device, computer equipment and storage medium | |
Desiani et al. | A Novelty Patching of Circular Random and Ordered Techniques on Retinal Image to Improve CNN U-Net Performance. | |
CN113850796A (en) | Lung disease identification method and device based on CT data, medium and electronic equipment | |
Upadhyay et al. | Big data analytics with deep learning based intracranial haemorrhage diagnosis and classification model | |
CN111815569B (en) | Image segmentation method, device, equipment and storage medium based on deep learning | |
CN117616467A (en) | Method for training and using deep learning algorithm to compare medical images based on reduced dimension representation | |
CN117953208A (en) | Graph-based edge attention gate medical image segmentation method and device | |
CN116013475B (en) | Method and device for sketching multi-mode medical image, storage medium and electronic equipment | |
CN117058262A (en) | Method, system and device for magnetic nanoparticle imaging reconstruction | |
CN114565964B (en) | Emotion recognition model generation method, emotion recognition device, emotion recognition medium and emotion recognition equipment | |
Mortazi et al. | Weakly supervised segmentation by a deep geodesic prior | |
CN114820636A (en) | Three-dimensional medical image segmentation model and training method and application thereof | |
CN114418955A (en) | Coronary plaque stenosis degree detection system and method | |
CN113538493A (en) | Automatic delineation method, delineation system, computing device and storage medium for brain functional region of MRI head image | |
CN116342608B (en) | Medical image-based stent adherence measurement method, device, equipment and medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |