CN117788826A - Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment - Google Patents

Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment Download PDF

Info

Publication number
CN117788826A
CN117788826A CN202311851468.7A CN202311851468A CN117788826A CN 117788826 A CN117788826 A CN 117788826A CN 202311851468 A CN202311851468 A CN 202311851468A CN 117788826 A CN117788826 A CN 117788826A
Authority
CN
China
Prior art keywords
convolution
pyramid
feature
image pair
mapping
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202311851468.7A
Other languages
Chinese (zh)
Inventor
陈加
袁科
卫阳钰
庞世燕
董石
田元
刘智
罗恒
肖克江
蔡欣芷
陈迪
童名文
左明章
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Central China Normal University
Original Assignee
Central China Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Central China Normal University filed Critical Central China Normal University
Priority to CN202311851468.7A priority Critical patent/CN117788826A/en
Publication of CN117788826A publication Critical patent/CN117788826A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses an image pair collaborative segmentation method, system and electronic equipment based on pyramid depth convolution, which comprise the following steps: s1, acquiring an image pair, preprocessing the image pair and unifying the size; s2, inputting the image pairs into a twin encoder for advanced semantic feature extraction to obtain advanced semantic feature mapping; s3, for high-level languageThe artificial feature mapping uses pooling operation to construct a feature pyramid, so as to obtain a corresponding enhanced mapping; s4, mapping the features f A And f B The corresponding enhanced mapping cascade is used for obtaining a feature map, and a common object probability map is generated by using up-sampling of a twin decoder; s5, up-sampling is used to segment a segmentation mask with the same resolution as the original image. The invention can solve the problem of multiple scales in space to a certain extent, reduce the calculation complexity and simultaneously highlight the position information of the same semantic object.

Description

Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment
Technical Field
The invention belongs to the technical field of image segmentation, and particularly relates to an image pair collaborative segmentation method, system and electronic equipment based on pyramid depth convolution.
Background
With the improvement of computer performance, deep learning is vigorously developed, and a great deal of research completely discards the traditional characteristics. A network architecture using a twin encoder-decoder is designed to address the image co-segmentation task. In the encoder, the method of the architecture uses a convolutional neural network to acquire feature mapping of the image, and introduces a cross-correlation layer to perform feature matching to acquire corresponding mapping. The decoder concatenates the feature map and the corresponding map for upsampling until the original image size is restored, thereby obtaining a co-segmentation result. One of the methods using this architecture is a method for introducing a semantic aware attention mechanism, which has three structures, namely a channel attention module, a hybrid channel attention module and a channel space attention module. The attention mechanism may focus on high response feature channels and reduce the response of no light feature channels.
The image collaborative segmentation method based on deep learning needs segmentation clues to be used as the basis for obtaining collaborative segmentation results. Segmentation cues are the basis for identifying objects to be segmented from an image, including object cues and related cues. The object clues refer to the "object" features on the present image, and the related clues are the related features of the co-segmented object acquired from the multiple images. Unlike traditional semantic segmentation, which only requires feature extraction of objects contained in a single image, collaborative segmentation of images should have at least one object of the same semantic meaning in the image pair to be segmented without interference, which is clue information that is difficult to use. How to use the cue information to construct a simple and efficient collaborative feature extraction and related cue extraction model to complete an image collaborative segmentation task is a key problem which is most needed to be solved by the image collaborative segmentation task.
There may be a large number of images in the image that contain occlusion relations, background clutter, and small target foreground objects. These images are a great challenge for task image tasks, especially for collaborative segmentation tasks of images that only use the same semantic information between images. Advanced semantic features can be obtained, which is based on the advantages of the deep learning method, but a large optimization space still exists for the accuracy and the reasoning speed of the result. How to solve the problem that the accuracy is reduced due to the complex images, so that the image collaborative segmentation effect is improved as a whole is also one of the main problems to be solved in the image collaborative segmentation field.
Image co-segmentation requires processing of two or more images, which results in a larger network architecture for processing the image co-segmentation. Some approaches avoid compressing or simplifying the network in the future, but this approach has the consequence of reduced performance. How to reasonably optimize the network model, and to achieve good balance between performance and resources, is also one of the problems to be solved in the field of image collaborative segmentation. The prior art in the field of image collaborative segmentation has insufficient high-precision mining capability for the same semantic information between image pairs, which may lead to inaccurate segmentation results and even errors of multiple segmentation or fewer segmentation. And the prior art does not handle images containing occlusions, background clutter, or small target foreground objects well enough. Finally, the prior art generally only accepts input with a fixed size, but in practical application, it is difficult to ensure that the images to be processed are all the same size, which results in limited flexibility of the prior art and difficulty in being well applied to real scenes.
Disclosure of Invention
In view of the above shortcomings of the prior art, the present invention aims to provide a pyramid depth convolution-based image pair collaborative segmentation method, system and electronic device, which can solve the problem of multiple dimensions in space to a certain extent, and then perform semantic feature correlation calculation on a channel through depth convolution, so as to highlight the position information of the same semantic object while reducing the calculation complexity.
In order to achieve the above purpose, the technical scheme of the method of the invention is as follows:
an image pair collaborative segmentation method based on pyramid depth convolution comprises the following steps:
s1, acquiring an image pair I A And I B, Pair of images pair I A And I B Preprocessing and unifying the sizes;
s2, image pair I A And I B Inputting into a twin encoder for advanced semantic feature extraction to obtain advanced semantic feature mapping f A And f B
S3 mapping f to advanced semantic features A And f B Using pooling operations, constructsCreating feature pyramids to obtain corresponding enhanced mappings F A And F B
S4, mapping the features f A And f B Corresponding enhanced mapping F A And F B Cascading to obtain a feature map F AB And F BA Generating a common object mask M using twin decoder upsampling A And M B
S5, employing up-sampling to segment the segmentation mask P with the same resolution as the original image A 、P B
Further, the twin encoder in the step S2 is provided with a plurality of convolution layers and pooling layers, and the twin encoder may use different backbolts. In one example, we employ the network structure of vgg and pre-training weights (except for the last convolutional layer) and add a ReLU activation function after each convolutional layer.
Further, in the step S3, the pooling layers with different step sizes are used for the high-level semantic features f A And f B Processing to construct a three-layer multi-scale feature pyramid f A ' and f B ’。
Further, pyramid feature f B Each channel of the ith layer of' acts as a convolution kernel for the depth convolution for f A Carrying out convolution operation on each corresponding channel to obtain a characteristic diagram with the channel number of ixc, then carrying out pixel-by-pixel addition and averaging on the characteristic diagram with the channel number of ixc, then carrying out convolution on the characteristic diagram with the channel number of c by connecting 1x1, and recording the output result as: f (F) A The mathematical description is as follows:
wherein L is 2 (. Cndot.) represents L 2 The normalization is performed so that the data of the sample,representing the depth convolution, m is the number of layers of the pyramid, +.>A convolution layer with a size of 1x1 with a ReLU activation function is shown.
Further, pyramid feature f A Each channel of the ith layer of' acts as a convolution kernel for the depth convolution for f A Carrying out convolution operation on each corresponding channel to obtain a characteristic diagram with the channel number of ixc, then carrying out pixel-by-pixel addition and averaging on the characteristic diagram with the channel number of ixc, then carrying out convolution on the characteristic diagram with the channel number of c by connecting 1x1, and recording the output result as: f (F) B The mathematical description is as follows:
wherein L is 2 (. Cndot.) represents L 2 The normalization is performed so that the data of the sample,representing the depth convolution, m is the number of layers of the pyramid, +.>A convolution layer with a size of 1x1 with a ReLU activation function is shown.
Further, in the step S4, the twin decoder in the step S4 is used to extract the foreground mask of the common object of the two input images, and the enhanced mapping F of the step S3 is first performed A And F B And corresponding high-level semantic feature map f B 、f A Connecting according to channel dimension to obtain a characteristic diagram F with the number of channels being twice as large as that of original characteristics AB 、F BA The convolution layer is then used to fuse the input feature graphs while reducing the number of channels of the advanced semantic feature graphs. The fused 256-channel feature map is then upsampled to an upsampled size equal to the output size of the second convolution block in the encoding stage, after which the low-level feature map output by the second convolution block of the twin encoder is connected to the previously upsampled high-level semantic feature map, after which the twin encoder uses two 3x3 convolution layers and one1x1 convolution layer without activation function to obtain two probability maps M A 、M B
In a second aspect, the present invention provides an image pair collaborative segmentation system based on pyramid depth convolution, including:
a preprocessing module for acquiring an image pair I A And I B, Pair of images pair I A And I B Preprocessing and unifying the sizes;
a twin encoding module for encoding a pair of images I A And I B Inputting into a twin encoder for advanced semantic feature extraction to obtain advanced semantic feature mapping f A And f B
A feature pyramid module for mapping f to high-level semantic features A And f B Using pooling operation to construct feature pyramid to obtain corresponding enhanced mapping F A And F B
A twin decoding module for mapping the features f A And f B Corresponding enhanced mapping F A And F B Cascading to obtain a feature map F AB And F BA Generating a common object mask P using twin decoder upsampling A And P B
A segmentation mask acquisition module for segmenting a segmentation mask M having the same resolution as the original image using upsampling A 、M B
The image pair collaborative segmentation system based on pyramid depth convolution is used for executing the steps in the image pair collaborative segmentation method based on pyramid depth convolution.
In a third aspect, the present invention provides an electronic device, including a memory, a processor, and a computer program stored on the memory and executable on the processor, the processor implementing the pyramid depth convolution-based image pair co-segmentation method when executing the program.
Compared with the prior art, the invention has the following beneficial effects:
the pyramid depth convolution module is designed, the position of the high response characteristic in the channel is highlighted, the problem that the high-precision mining capability of the same semantic information between the image pairs is insufficient in the prior art is solved, and the performance of collaborative segmentation of the images is improved.
The twin encoder of the method can process two images with different resolution sizes in the testing and actual segmentation processes, so that the method is more likely to be applied to practical application compared with the prior art.
Drawings
Fig. 1 is a detailed schematic diagram of a network structure according to the present invention.
Fig. 2 is a schematic diagram of a pyramid depth convolution module according to the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention more clear, the technical solutions of the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. It will be apparent that the described embodiments are some, but not all, embodiments of the invention. All other embodiments, which are obtained by a person of ordinary skill in the art without making any inventive effort, are within the scope of the present invention based on the embodiments of the present invention.
Example 1
The embodiment provides an image pair collaborative segmentation method based on pyramid depth convolution, which is characterized by comprising the following steps:
s1, acquiring an image pair I A And I B, Pair of images pair I A And I B Preprocessing and unifying the sizes;
s2, image pair I A And I B Inputting into a twin encoder for advanced semantic feature extraction to obtain advanced semantic feature mapping f A And f B
S3 mapping f to advanced semantic features A And f B Using pooling operation to construct feature pyramid to obtain corresponding enhanced mapping F A And F B
S4, mapping the features f A And f B Corresponding enhanced mapping F A And F B Cascading to obtain a feature map F AB And F BA Generating a common object probability map P using twin decoder upsampling A And P B
S5, employing up-sampling to segment the segmentation mask M with the same resolution as the original image A 、M B
In the following, with reference to specific embodiments, the invention designs an image pair collaborative segmentation method based on pyramid depth convolution, and the main innovation is a Pyramid Depth Convolution (PDC) module which firstly uses pooling operation on input advanced semantic features to construct a feature pyramid. The pooling operation can obtain multi-scale features containing different receptive fields without introducing additional parameters, so that the method can solve the problem of multi-scale in space to a certain extent, and then the semantic feature correlation calculation on the channel is performed through the deep convolution, thereby reducing the calculation complexity and simultaneously highlighting the position information of the same semantic object.
The method mainly explores the correlation relation of the same semantic object in the image pair from the aspect of the characteristic channel, and as different channels reflect different characteristics, the method tries to realize interaction among the channels by using a simple operation so as to highlight the position of the high-response characteristic in the channel, thereby improving the segmentation performance of the method. The network frame of the method is shown in figure 1, the method can segment foreground objects with the same semantics from two input images by end-to-end implementation, and in training, the input of the network of the method is an image pair I A And I B ,I A And I B The uniform size is 512x512 after pretreatment. I A And I B The corresponding high-level semantic feature map f is obtained after the processing of the twin encoder A And f B 。f A And f B Then as the input of PDC, PDC is the pooling pyramid depth convolution module provided by the method, and the output of the pooling pyramid depth convolution module is F A And F B Then f A And F B ,f B And F A Respectively spliced high-level semantic feature map F AB And F BA As input to a twin decoder, resulting in a segmentation mask M A And M B
The network structure of the method mainly comprises three parts: (1) a twin encoder. The input of this part is an image pair I A And I B The output is an advanced semantic feature map f obtained through a backbone network A And f B . (2) PDC module. Pyramid depth convolution module designed by the method and used for high-level semantic feature graph f A And f B Processing to obtain corresponding enhanced mapping F A And F B . (3) a twin decoder. The decoder first maps the features f A And f B Corresponding enhanced mapping F A And F B Cascading to obtain a feature map F AB And F BA Generating a common object mask M using twin decoder upsampling A And M B
The method completes the flow of image collaborative segmentation:
1. the image pairs are characterized by a twin encoder.
2. Features extracted by the twin encoder are input to the PDC module, which functions to perform feature interactions, converting the independent feature maps into enhanced features containing mutual features.
3. The enhancement features are input to a twin decoder, which decodes the enhancement features obtained in the PDC module to obtain the final segmentation mask.
And a coding module. The method uses a twin encoder structure as an encoding module, and the twin encoder can simultaneously extract two images I A And I B Is described. The encoder network of the method has 14 convolution layers and 4 pooling layers, the parameters of the first 13 convolution layers are consistent with VGG16 except the last convolution layer, and the ReLU activation function is added after the convolution layers. During training, all convolutions are set to a size of 3x3, with a step size of 1. The input of the twin encoder during training is two 512x512 RGB images, and the output is two 512 channel characteristic images f A And f B The size is 32x32. It should be noted that in the test and actual segmentation tasks, the method can process two images with different resolutions and trainThe data augmentation approach is used to result in a fixed input size.
The second part of the method is a pyramid depth convolution module, the main purpose of which is to calculate the high-level semantic features F A And F B Correlation between them. The input of the module is an advanced semantic feature map f of the encoder output A And f B . First construct f respectively A And f B Corresponding pyramid features, the method selects 3x3, and the pooling layers with steps of 1,2 and 3 respectively for the high-level semantic features f A And f B Processing to construct a three-layer feature pyramid f A ' and f B '. A schematic diagram of a PDC module is shown in figure 2,
the following is characterized by pyramid f B The 'ith layer' describes the computation of the depth convolution by way of example. By pyramid feature f B Each channel of the ith layer of' acts as a convolution kernel for the depth convolution for f A And carrying out convolution operation on each corresponding channel. Due to f B ' is the pyramid of the i-th layer, the depth convolution does not change the number of channels, and thus a feature map with the number of channels ixc can be obtained. Then, the characteristic map with the channel number of ixc is averaged by adding the channels pixel by pixel, and then a 1x1 convolution is connected to obtain the characteristic map with the channel number of c. The output result is recorded as: f (F) A . The mathematical description is as follows:
wherein L is 2 (. Cndot.) represents L 2 The normalization is performed so that the data of the sample,representing the depth convolution, m is the number of layers of the pyramid. />A convolution layer with a size of 1x1 with a ReLU activation function is shown. F can be obtained by the same method B:
The PDC module pools the advanced semantic feature graphs to construct pyramid features with features of different sizes. And then taking the feature pyramid as a convolution kernel to finish the deep convolution operation so as to obtain the cross-correlation information between the two pictures. The depth convolution is an operation on a channel, and the channels of the single feature images are not interfered with each other before and after the operation, so that the PDC module adopts the operation to carry out convolution operation in each channel of the high-level semantic feature image, when convolution kernels of different sizes are convolved on the same channel, the same semantic object with the same high response can be further activated, and the PDC module uses the feature pyramid as three depth convolution kernels of different sizes, so that the operation can solve the problem of multiple scales of the object to a certain extent.
The reason for the deep convolution operation of one image with the feature map of another image as a convolution check is that the paper different channels represent different semantics shows that different feature channels will represent different semantic features.
The reason for constructing pyramid features of different sizes in the PDC module is that the same semantic objects of the image pair usually have arbitrary sizes and positions, so the use of a single-size feature map as a convolution kernel will limit the acquisition of related features, and the method uses three-size pyramid features to match semantic objects of different sizes and positions.
The decoder employed in the method is also a twin structure with shared parameters. The decoding module is used to extract a foreground mask of a common object of the two input images. Result F of PDC Module A 、F B And corresponding high-level semantic feature map f B 、f A Connecting according to channel dimension to obtain 1024-channel feature diagram F AB 、F BA . The input feature maps are then fused using a convolutional layer to obtain two 256-channel advanced featuresSemantic feature graphs. The fused 256-channel signature is then upsampled, the upsampled size being equal to the output size of the second convolution block in the encoding stage. The low-level feature map output by the second convolution block of the encoder is then concatenated with the previously up-sampled high-level semantic feature map. The low-level features generally contain richer visual information, and the segmentation effect can be effectively improved by properly adding the low-level features in the decoder. The decoder then uses two 3x3 convolutional layers and one 1x1 convolutional layer without an activation function to obtain two probability maps M A 、M B . Finally, upsampling is adopted to partition a partition mask P with the same resolution as the original image A 、P B . The image collaborative segmentation only needs to segment the foreground and the background, so the image collaborative segmentation is used as a binary image marking problem, and the standard cross entropy loss function is used for training the network of the method.
Example 2
The embodiment provides an image pair collaborative segmentation system based on pyramid depth convolution, which comprises the following steps:
a preprocessing module for acquiring an image pair I A And I B, Pair of images pair I A And I B Preprocessing and unifying the sizes;
a twin encoding module for encoding a pair of images I A And I B Inputting into a twin encoder for advanced semantic feature extraction to obtain advanced semantic feature mapping f A And f B
A feature pyramid module for mapping f to high-level semantic features A And f B Using pooling operation to construct feature pyramid to obtain corresponding enhanced mapping F A And F B
A twin decoding module for mapping the features f A And f B Corresponding enhanced mapping F A And F B Cascading to obtain a feature map F AB And F BA Generating a common object probability map P using twin decoder upsampling A And P B
A segmentation mask acquisition module for employing upsamplingDividing the dividing mask M with the same resolution as the original image A 、M B
The image pair collaborative segmentation system based on pyramid depth convolution is used for executing the steps in the image pair collaborative segmentation method based on pyramid depth convolution.
Example 3
The embodiment provides an electronic device, which comprises a memory, a processor and a computer program stored on the memory and capable of running on the processor, wherein the processor realizes the image pair collaborative segmentation method based on pyramid depth convolution when executing the program.
The present invention is not limited to the above-mentioned embodiments, but any modifications, equivalents, improvements and modifications within the scope of the invention will be apparent to those skilled in the art. While preferred embodiments of the present invention have been described, additional variations and modifications in those embodiments may occur to those skilled in the art once they learn of the basic inventive concepts. It is therefore intended that the following claims be interpreted as including the preferred embodiments and all such alterations and modifications as fall within the scope of the invention.
It will be apparent to those skilled in the art that various modifications and variations can be made to the embodiments of the present invention without departing from the spirit or scope of the embodiments of the invention. Thus, if such modifications and variations of the embodiments of the present invention fall within the scope of the claims and the equivalents thereof, the present invention is intended to include such modifications and variations.
Other parts not described in detail are prior art.

Claims (8)

1. The image pair collaborative segmentation method based on pyramid depth convolution is characterized by comprising the following steps of:
s1, acquiring an image pair I A And I B Pair of images I A And I B Preprocessing and unifying the sizes;
s2, image pair I A And I B Inputting into a twin encoder for advanced semantic feature extraction to obtain advanced semantic feature mapping f A And f B
S3 mapping f to advanced semantic features A And f B Using pooling operation to construct feature pyramid to obtain corresponding enhanced mapping F A And F B
S4, mapping the features f A And f B Corresponding enhanced mapping F A And F B Cascading to obtain a feature map F AB And F BA Generating a common object probability map P using twin decoder upsampling A And P B
S5, employing up-sampling to segment the segmentation mask M with the same resolution as the original image A 、M B
2. The image pair collaborative segmentation method based on pyramid depth convolution according to claim 1, wherein the twin encoder in step S2 is provided with a plurality of convolution layers and pooling layers, and the twin encoder adopts different backbones.
3. The image pair collaborative segmentation method based on pyramid depth convolution according to claim 1, wherein the step S3 is implemented by pooling layers of different steps for advanced semantic features f A And f B Processing to construct a three-layer multi-scale feature pyramid f A ' and f B ’。
4. A pyramid depth convolution based image pair co-segmentation method according to claim 3, characterized in that pyramid feature f B Each channel of the ith layer of' acts as a convolution kernel for the depth convolution for f A Performing convolution operation on each corresponding channel to obtain a characteristic diagram with the channel number of i xc, then averaging the characteristic diagram with the channel number of ixc according to the channel pixel-by-pixel addition, then connecting a 1x1 convolution to obtain a characteristic diagram with the channel number of c, and outputting a junctionThe fruit is recorded as follows: f (F) A The mathematical description is as follows:
wherein L is 2 (. Cndot.) represents L 2 The normalization is performed so that the data of the sample,representing the depth convolution, m is the number of layers of the pyramid, +.>A convolution layer with a size of 1x1 with a ReLU activation function is shown.
5. The image pair collaborative segmentation method based on pyramid depth convolution according to claim 4, wherein pyramid feature f A Each channel of the ith layer of' acts as a convolution kernel for the depth convolution for f A Carrying out convolution operation on each corresponding channel to obtain a characteristic diagram with the channel number of i xc, then carrying out pixel-by-pixel addition and averaging on the characteristic diagram with the channel number of ixc, then carrying out convolution on the characteristic diagram with the channel number of c by connecting with 1x1, and recording the output result as follows: f (F) B The mathematical description is as follows:
wherein L is 2 (. Cndot.) represents L 2 The normalization is performed so that the data of the sample,representing the depth convolution, m is the number of layers of the pyramid, +.>A convolution layer with a size of 1x1 with a ReLU activation function is shown.
6. The image pair collaborative segmentation method based on pyramid depth convolution according to claim 3, wherein in the step S4, the twin decoder in the step S4 is used to extract the foreground mask of the common object of the two input images, and the enhancement map F of the step S3 is first A And F B And corresponding high-level semantic feature map f B 、f A Connecting according to channel dimension to obtain a characteristic diagram F with the number of channels being twice as large as that of original characteristics AB 、F BA Then the convolution layers are used to fuse the input feature images while reducing the number of channels of the two advanced semantic feature images, the fused feature images are then up-sampled to a size equal to the output size of the second convolution block in the encoding stage, then the low-level feature images output by the second convolution block of the twin encoder are connected to the previously up-sampled advanced semantic feature images, and then the twin encoder uses the two 3x3 convolution layers and one 1x1 convolution layer without an activation function to obtain two probability images P A 、P B
7. An image pair collaborative segmentation system based on pyramid depth convolution, comprising:
a preprocessing module for acquiring an image pair I A And I B, Pair of images pair I A And I B Preprocessing and unifying the sizes;
a twin encoding module for encoding a pair of images I A And I B Inputting into a twin encoder for advanced semantic feature extraction to obtain advanced semantic feature mapping f A And f B
A feature pyramid module for mapping f to high-level semantic features A And f B Using pooling operation to construct feature pyramid to obtain corresponding enhanced mapping F A And F B
A twin decoding module for mapping the features f A And f B Corresponding enhanced mapping F A And F B Cascading to obtain a feature map F AB And F BA Using twinGenerating a common object probability map P by upsampling a raw decoder A And P B
A segmentation mask acquisition module for segmenting a segmentation mask M having the same resolution as the original image using upsampling A 、M B
The pyramid depth convolution based image pair co-segmentation system is configured to perform the steps in the pyramid depth convolution based image pair co-segmentation method according to any one of claims 1-6.
8. An electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, wherein the processor implements the pyramid depth convolution-based image pair co-segmentation method according to any one of claims 1 to 6 when executing the program.
CN202311851468.7A 2023-12-28 2023-12-28 Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment Pending CN117788826A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311851468.7A CN117788826A (en) 2023-12-28 2023-12-28 Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311851468.7A CN117788826A (en) 2023-12-28 2023-12-28 Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment

Publications (1)

Publication Number Publication Date
CN117788826A true CN117788826A (en) 2024-03-29

Family

ID=90386926

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311851468.7A Pending CN117788826A (en) 2023-12-28 2023-12-28 Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment

Country Status (1)

Country Link
CN (1) CN117788826A (en)

Similar Documents

Publication Publication Date Title
CN111210443B (en) Deformable convolution mixing task cascading semantic segmentation method based on embedding balance
US10614574B2 (en) Generating image segmentation data using a multi-branch neural network
CN112308200B (en) Searching method and device for neural network
CN111325751A (en) CT image segmentation system based on attention convolution neural network
CN110689599B (en) 3D visual saliency prediction method based on non-local enhancement generation countermeasure network
CN112001914A (en) Depth image completion method and device
CN113642390B (en) Street view image semantic segmentation method based on local attention network
CN111914654B (en) Text layout analysis method, device, equipment and medium
CN113674146A (en) Image super-resolution
CN116309648A (en) Medical image segmentation model construction method based on multi-attention fusion
CN113554032B (en) Remote sensing image segmentation method based on multi-path parallel network of high perception
Oyama et al. Influence of image classification accuracy on saliency map estimation
CN111833360B (en) Image processing method, device, equipment and computer readable storage medium
CN114638994B (en) Multi-modal image classification system and method based on attention multi-interaction network
CN109766918A (en) Conspicuousness object detecting method based on the fusion of multi-level contextual information
CN112150363B (en) Convolutional neural network-based image night scene processing method, computing module for operating method and readable storage medium
CN113379606A (en) Face super-resolution method based on pre-training generation model
CN117934824A (en) Target region segmentation method and system for ultrasonic image and electronic equipment
CN112884702A (en) Polyp identification system and method based on endoscope image
Fan et al. EGFNet: Efficient guided feature fusion network for skin cancer lesion segmentation
CN117788826A (en) Image pair collaborative segmentation method and system based on pyramid depth convolution and electronic equipment
Jia et al. Learning rich information for quad bayer remosaicing and denoising
CN110489584B (en) Image classification method and system based on dense connection MobileNet model
CN114299105A (en) Image processing method, image processing device, computer equipment and storage medium
Yang et al. An end‐to‐end perceptual enhancement method for UHD portrait images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination