CN117315366A - Image processing method based on double-branch parallel neural network - Google Patents
Image processing method based on double-branch parallel neural network Download PDFInfo
- Publication number
- CN117315366A CN117315366A CN202311334063.6A CN202311334063A CN117315366A CN 117315366 A CN117315366 A CN 117315366A CN 202311334063 A CN202311334063 A CN 202311334063A CN 117315366 A CN117315366 A CN 117315366A
- Authority
- CN
- China
- Prior art keywords
- module
- input
- output
- steps
- block
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 26
- 238000003672 processing method Methods 0.000 title claims abstract description 14
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 29
- 230000004927 fusion Effects 0.000 claims abstract description 23
- 230000002457 bidirectional effect Effects 0.000 claims abstract description 22
- 238000000034 method Methods 0.000 claims abstract description 16
- 239000013598 vector Substances 0.000 claims abstract description 10
- 238000000605 extraction Methods 0.000 claims abstract description 8
- 239000000284 extract Substances 0.000 claims abstract description 5
- 238000013507 mapping Methods 0.000 claims abstract description 4
- 238000011176 pooling Methods 0.000 claims description 37
- 238000004364 calculation method Methods 0.000 claims description 7
- 241000282326 Felis catus Species 0.000 claims description 6
- 230000003044 adaptive effect Effects 0.000 claims description 5
- 230000003321 amplification Effects 0.000 claims description 3
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 3
- 239000011159 matrix material Substances 0.000 description 7
- 238000010586 diagram Methods 0.000 description 5
- 230000008901 benefit Effects 0.000 description 3
- 230000007246 mechanism Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 239000012014 frustrated Lewis pair Substances 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 238000012800 visualization Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000009792 diffusion process Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006698 induction Effects 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/86—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using syntactic or structural representations of the image or video pattern, e.g. symbolic string recognition; using graph matching
Abstract
The invention relates to the technical field of computer vision image classification, in particular to an image processing method based on a double-branch parallel neural network. The method comprises the following steps: s1: the input image is sent to a first basic convolution module and a feature embedding module to respectively conduct feature extraction and space vector mapping to obtain feature X 0 And space vector Z 0 The method comprises the steps of carrying out a first treatment on the surface of the S2: x is to be 0 And Z 0 Input into a bidirectional bridging module Block, and output X block0 And Z block0 The method comprises the steps of carrying out a first treatment on the surface of the The step S2 comprises the following steps: s2-1: x is to be 0 And Z 0 Feeding into a first fusion moduleOutput Z hid The method comprises the steps of carrying out a first treatment on the surface of the S2-2: will Z hid Input to a transducer network module, learn and extract low-frequency global features in image features, and output Z block0 The method comprises the steps of carrying out a first treatment on the surface of the S2-3: x is to be 0 And Z block0 Inputting into a Flexible CNN, acquiring local features in image features, and outputting X hid The method comprises the steps of carrying out a first treatment on the surface of the S2-4: x is to be hid And Z block0 Input into the second fusion module, output X block0 The method comprises the steps of carrying out a first treatment on the surface of the S3: taking the output of the last bidirectional bridging module Block as the input of the next bidirectional bridging module Block, repeating the step S2 for a plurality of times, and finally outputting X block And Z block 。
Description
Technical Field
The invention relates to the technical field of computer vision image classification, in particular to an image processing method based on a double-branch parallel neural network.
Background
Neural networks, as a powerful machine learning model, achieve excellent results in a variety of visual tasks such as image classification, object detection, and semantic separation by learning complex features and functions from a large amount of data. In order to enable the network to learn sufficiently rich data features and meet the requirements of various complex scenes, the depth and width of the neural network are continuously increased, and the network calculation amount and the parameter amount are increased. The lightweight neural network aims to build a lightweight network model with less parameter and lower calculation resource consumption under the condition of meeting the use performance by means of better module combination, so that the problem that the deep neural network consumes serious storage and calculation resources is essentially solved, and the popularization and the landing of deep learning are promoted. Most of the existing lightweight neural networks only optimize convolution modules, common convolution is disassembled into a 1×1 convolution combination of depth convolution for changing the size of a feature map and the number of channels for expanding the feature map, so that FLPs (floating point operation amount) of the neural network are reduced, most of the work is only optimized for the convolution modules, basic convolution operation is carried out on pixel points in a rectangular range around a convolution kernel each time, feature extraction in the network is always limited to a local range no matter how the depth and the width of the network change, long-distance connection relation of global images cannot be established, on the other hand, the capacity of the neural network is limited, and network benefits brought by increasing data samples are gradually reduced.
In order to solve the defect that convolutional neural experience is limited in visual field and global information is difficult to capture, for example, a BotNet (Bottleneck Transformers for Visual Recognition) network combines a transducer and a convolutional neural network, a Self-Attention mechanism in the transducer is integrated into a backbone network of a traditional convolutional neural network, a new module named as Bot BottleNeck is proposed, resNet-50 is used as a basic network, 3×3 convolution in the last three BottleNeck in ResNet-50 is replaced by a Multi-Head Self-Attention module (Multi-Self-Attention) similar to the transducer, which is abbreviated as MHSA, and the rest modules remain unchanged. The network takes the traditional image as input, extracts the characteristics through the original backbone network, obtains a 2048-channel characteristic diagram, inputs the characteristic diagram into the Bot BottleNeck, firstly carries out channel reduction by 1X 1 convolution, and then sends the characteristic diagram into the multi-head self-attention layer. Unlike the multi-head self-attention layer in the conventional transducer, for the Position coding part, two vectors are used as the space attention of two dimensions of the transverse and longitudinal directions in the BotNet, after the two vectors are added, the Content-Position is obtained by multiplying the query matrix, and then the Content-Position is multiplied by the Content code obtained through the query matrix and the key value matrix to obtain the similarity characteristic of the space nameplate, so that when the multi-head self-attention module focuses on a more proper area, the characteristic diagram is sent to a diffusion channel in the convolution of 1×1 again.
However, the existing solutions for combining convolutional neural networks and transformers either use convolution at the beginning or embed convolution into each transducer module, and the whole network still adopts a serial design paradigm. According to the technical scheme, the two network structures are connected in series, so that the mutual dependence among the modules is strong, the connection relation and the insertion position of different structures must be fully considered, and network modification is difficult. In addition, the existing scheme does not consider the gain ratio between the computational complexity of the model and the performance improvement caused by the model, the built network model FLPs are up to several G, and the deployment cost is high.
Disclosure of Invention
The invention aims to provide an image processing method based on a double-branch parallel neural network, which can solve the problem of high network deployment cost when a convolutional neural network is combined with a transducer.
The basic scheme provided by the invention is as follows: an image processing method based on a double-branch parallel neural network comprises the following steps:
s1: the input image is sent to a first basic convolution module and a feature embedding module to respectively conduct feature extraction and space vector mapping to obtain feature X 0 And space vector Z 0 ;
S2: x is to be 0 And Z 0 Inputting the data into a bidirectional bridging module Block, wherein the bidirectional bridging module Block comprises a parallel convolutional neural network module Flexible CNN, a parallel Transformer network module, a first fusion module and a second fusion module;
the step S2 comprises the following steps:
s2-1: x is to be 0 And Z 0 Sending the local feature and the global feature to a first fusion module, wherein the first fusion module is used for completing fusion of the local feature and the global feature and outputting Z hid ;
S2-2: will Z hid Input to a transducer network module, learn and extract low-frequency global features in image features, and output Z block0 ;
S2-3: x is to be 0 And Z block0 Inputting into a Flexible CNN, acquiring local features in image features, and outputting X hid ;
S2-4: x is to be hid And Z block0 Inputting a second fusion module, wherein the second fusion module is used for fusing the local feature and the global feature again and outputting X block0 ;
S3: x outputting previous bidirectional bridging module Block block0 And Z block0 As the input of the next bidirectional bridging module Block, repeating the step S2 for several times, and finally outputting X block And Z block 。
The principle and the advantages of the invention are as follows: the invention focuses on adjusting the structure of the network to be a double-branch parallel structure, and combining the relatively efficient computing operation of the convolutional neural network with the global feature extraction capability of the transducer model. The network consists of a lightweight convolution branch and a modularized transducer branch, and each branch only needs to pay attention to the feature extraction and the module construction of the branch, so that the dependence between two different networks is reduced. Meanwhile, a bidirectional bridging module is constructed by referring to a self-attention mechanism, the bridging module receives the characteristics from the convolution branches and the Transformer branches respectively, and the characteristics are fused and output in the module to realize information interaction of different branches, so that a double-branch parallel network structure is finally obtained, and the problem of high network deployment cost is solved.
And constructing the convolutional neural network and the transducer model neural network as parallel results, and respectively extracting input characteristics by parallel convolutional neural network branches and transducers so that the two neural networks are mutually decoupled. Meanwhile, a bidirectional bridging module is constructed by referring to a self-attention mechanism, and features extracted by two branches are subjected to bidirectional fusion interaction, so that the advantage of the convolutional neural network in terms of feature capture and global information modeling of the transform network can be fully utilized by the network.
In the prior art, the network structure is designed into a serial shape, and a transducer module is introduced at the head, middle or tail of the convolutional neural network, which means that when each module is constructed, the relevance between the preceding module and the following module must be considered, so that the flexibility of the network is insufficient. The limitation of the scheme combining the convolutional neural network and the transducer network is reexamined, the network structure is redesigned, and the modules in the network are easier to modify in a parallel mode.
Further, the method also comprises the following steps:
s4: x is to be block Inputting into a second basic convolution module, performing channel amplification, and outputting X tail ;
S5: x is to be tail Input into an adaptive pooling module, X tail The Input size of (a) is recorded as Input, the expected Output size is Output, and the self-adaptive pooling module calculates the pooling core size and the step of the pooling operation according to the Input and OutputSize, dimension reduction is carried out on the characteristics, and X is output pool 。
In order to ensure the consistency of output after pooling, a self-adaptive pooling operation is introduced to replace conventional pooling, the size of a pooling core and the step size can be automatically adjusted in the pooling process, the limitation of a network on the size of an input image is reduced, the input image with different sizes is adapted, and the flexibility of the network is improved.
Further, the step S5 includes the steps of:
s501: judging whether the Input can divide the Output or not, executing S502 when the Input can divide the Output, and executing S503 when the Input cannot divide the Output;
s502: calculating a pooling kernel size K and a step size S through the following formula;
s503: the pooling kernel size K is calculated by the following formula i Step size S i ;
i denotes the calculation from 0 for the ith pooling operation in one feature map, ceil denotes the rounding up, floor denotes the rounding down.
When Input cannot divide the Ouput evenly, the step size and the pooling core size are dynamically changed each time the pooling operation, and the pooling core size and step size are calculated by rounding up and down.
Further, the method also comprises the following steps:
s6: x is to be pool And Z block Splicing in channel dimension, outputting X cat ;
S7: x is to be cat Inputting into two layers of fully connected neural networks to obtain final output X out 。
Further, the step S3 includes the steps of:
s301: and before the output of the last bidirectional bridging module Block is output to the bidirectional bridging module Block again, parameters in the convolutional neural network module are adjusted, wherein the parameters comprise channels and step sizes.
And by adjusting parameters, the number of the characteristic map channels finally output in each step is gradually increased, and the characteristic map size is gradually reduced.
Drawings
FIG. 1 is a schematic diagram of a structure of a dual-branch network according to an embodiment of an image processing method based on a dual-branch parallel neural network of the present invention;
fig. 2 is a visualization result of a layer-by-layer class activation map of a dual-branch network in an embodiment of an image processing method based on a dual-branch parallel neural network.
Detailed Description
The following is a further detailed description of the embodiments:
an example is substantially as shown in figures 1 and 2:
an image processing method based on a double-branch parallel neural network comprises the following steps:
s1: the input image is sent to a first basic convolution module and a feature embedding module to respectively conduct feature extraction and space vector mapping to obtain feature X 0 And space vector Z 0 。
In this embodiment, S1 further includes the following steps:
s101: the input image is sent into a Base Conv module, the Base Conv module consists of three pure convolution operations, the first convolution operation is named as Stem Conv, the second convolution operation and the third convolution operation are named as Neck Conv, and basic feature extraction is mainly carried out on the input image to obtain an output X of the step 0 。
S2: x is to be 0 And Z 0 Input into a bidirectional bridging module Block, and output X block0 And Z block0 The bidirectional bridging module Block comprises a parallel convolutional neural network module Flexible CNN, a parallel convolutional neural network module and a parallel convolutional neural network module;
the step S2 comprises the following steps:
s2-1: x is to be 0 And Z 0 Sending the local feature and the global feature to a first fusion module, wherein the first fusion module is used for completing fusion of the local feature and the global feature and outputting Z hid . In this embodiment, the first fusion module is the CNN to Former module in fig. 2, which is a plurality of matrix multiplication and addition operations. Referring to the standard self-attention module in the transducer, the implementation is as follows, and matrix multiplication operations are as follows:
(1) Will Z 0 Multiplying the query matrix obtained by training;
(2) Combining the result obtained in (1) with X 0 Multiplying;
(3) Feeding the result obtained in step (2) into a Softmax function;
(4) Combining the result obtained in step (3) with X 0 Multiplying;
(5) Multiplying the result obtained in the step (4) with a transformation matrix obtained by training;
(6) Mixing the result obtained in step (5) with Z 0 Adding to obtain the final output Z hid
S2-2: will Z hid Input to a transducer network module, learn and extract low-frequency global features in image features, and output Z block0 ;
S2-3: x is to be 0 And Z block0 Inputting into a Flexible CNN, acquiring local features in image features, and outputting X hid . Specifically, local small domain context information of the feature is obtained through convolution and induction bias action to obtain an output X of the step hid 。
S2-4: x is to be hid And Z block0 Inputting a second fusion module, wherein the second fusion module is used for fusing the local feature and the global feature again and outputting X block0 ;
S3: taking the output of the last bidirectional bridging module Block as the input of the next bidirectional bridging module Block, repeating the step S2 for a plurality of times, and finally outputting X block And Z block . The second fusion module is a Former to CNN module in FIG. 2, and realizes the re-interaction fusion of the global features and the local features.
Specifically, X 0 And Z 0 Input into a Block-0 module and output X block0 And Z block0 After which X is blook0 And Z block0 Inputting into a Block-1 module, repeating the operation of step S2, outputting X block1 And Z block1 . And (3) circulating in this way, before the output of the previous Block is used as the input of the downward moving Block, adjusting the parameters of the Block module, wherein the parameters comprise a channel and a step length, and the corresponding configuration file config. Py is read when the network is constructed, and different networks are constructed according to the parameters of each layer set in the file. This approach is a common processing means in convolutional neural networks. Each channel can be regarded as different characteristics, as the characteristics are learned by the assumption of the network to be more abstract, the number of channels is increased, so that the network can characterize and cover more key characteristics, and the classification accuracy of the network is improved. The purpose of feature map reduction is to reduce the number of parameters in the network, avoid overfitting, and improve the reasoning speed of the network.
So that the number of the channels of the feature map finally output in each step is gradually increased and the size of the feature map is gradually reduced. In this embodiment, the process is repeated eight times from Block-0 to Block-7, and Block-7 is finally used as the final output X block7 And Z block7 . The visualization result of the layer-by-layer class activation map from Block-0 to Block-7 is shown in FIG. 2.
S4: x is to be block Inputting into a second basic convolution module, performing channel amplification, and outputting X tail ;
S5: x is to be tail Input into an adaptive pooling module, X tail The Input size of the (2) is recorded as Input, the expected Output size is Output, the self-adaptive pooling module calculates the pooling core size and the step size of pooling operation according to the Input and Output, reduces the dimension of the feature and outputs X pool . The size of the output feature map is directly given when the step code is realized, then the step size and the pooling core size of each pooling process are adjusted according to the integer division relation between the input feature map and the output feature map, and the calculation modes of the two parameter sizes during each pooling are given in the steps S501 and S502.
S501: judging whether the Input can divide the Output or not, executing S502 when the Input can divide the Output, and executing S503 when the Input cannot divide the Output;
s502: calculating a pooling kernel size K and a step size S through the following formula;
s503: when the Input cannot divide the Output completely, the step size and the pooling core size of each pooling operation are dynamically changed, and the pooling core size K is calculated by the following formula i Step size S i ;
i denotes the calculation from 0 for the ith pooling operation in one feature map, ceil denotes the rounding up, floor denotes the rounding down.
Assuming that the adaptive pooling is maximally pooled, the Input size input=14, and the desired Output size output=4, the adaptive pooling operation divides the above-mentioned pooling step into four intervals.
Input: [0,1,2,3,4,5,6,7,8,9,10,11,12,13]
And (3) outputting: [3,6,10,13]
Interval 1: pool core size = 4[0,1,2,3] → [3]
Step size 3
Interval 2: pool core size = 4[3,4,5,6] → [6]
Step size 4
Interval 3: pool core size = 4[7,8,9,10] → [10]
Step size 3
Interval 4: pool core size = 4[10,11,12,13] → [13].
S6: x is to be pool And Z block Splicing in channel dimension, outputting X cat ;
S7: x is to be cat Inputting into two layers of fully connected neural networks to obtain final output X out 。
The foregoing is merely exemplary of the present invention, and the specific structures and features well known in the art are not described in any way herein, so that those skilled in the art will be able to ascertain all prior art in the field, and will not be able to ascertain any prior art to which this invention pertains, without the general knowledge of the skilled person in the field, before the application date or the priority date, to practice the present invention, with the ability of these skilled persons to perfect and practice this invention, with the help of the teachings of this application, with some typical known structures or methods not being the obstacle to the practice of this application by those skilled in the art. It should be noted that modifications and improvements can be made by those skilled in the art without departing from the structure of the present invention, and these should also be considered as the scope of the present invention, which does not affect the effect of the implementation of the present invention and the utility of the patent. The protection scope of the present application shall be subject to the content of the claims, and the description of the specific embodiments and the like in the specification can be used for explaining the content of the claims.
Claims (5)
1. The image processing method based on the double-branch parallel neural network is characterized by comprising the following steps of: the method comprises the following steps:
s1: the input image is sent to a first basic convolution module and a feature embedding module to respectively conduct feature extraction and space vector mapping to obtain feature X 0 And space vector Z 0 ;
S2: x is to be 0 And Z 0 Input into a bidirectional bridging module Block, and output X block0 And Z block0 The bidirectional bridging module Block comprises a parallel convolutional neural network module Flexible CNN, a parallel convolutional neural network module and a parallel convolutional neural network module;
the step S2 comprises the following steps:
s2-1: x is to be 0 And Z 0 Sending the local feature and the global feature to a first fusion module, wherein the first fusion module is used for completing fusion of the local feature and the global feature and outputting Z hid ;
S2-2: will Z hid Input to a transducer network module, learn and extract low-frequency global features in image features, and output Z block0 ;
S2-3: x is to be 0 And Z block0 Inputting into a Flexible CNN, acquiring local features in image features, and outputting X hid ;
S2-4: x is to be hid And Z block0 Inputting a second fusion module, wherein the second fusion module is used for fusing the local feature and the global feature again and outputting X block0 ;
S3: taking the output of the last bidirectional bridging module Block as the input of the next bidirectional bridging module Block, repeating the step S2 for a plurality of times, and finally outputting X block And Z block 。
2. The image processing method based on the dual-branch parallel neural network according to claim 1, wherein: the method also comprises the following steps:
s4: x is to be block Inputting into a second basic convolution module, performing channel amplification, and outputting X tail ;
S5: x is to be tail Input into an adaptive pooling module, X tail The Input size of the (2) is recorded as Input, the expected Output size is Output, the self-adaptive pooling module calculates the pooling core size and the step size of pooling operation according to the Input and Output, reduces the dimension of the feature and outputs X pool 。
3. The image processing method based on the dual-branch parallel neural network according to claim 2, wherein:
the step S5 comprises the following steps:
s501: judging whether the Input can divide the Output or not, executing S502 when the Input can divide the Output, and executing S503 when the Input cannot divide the Output;
s502: calculating a pooling kernel size K and a step size S through the following formula;
s503: the pooling kernel size K is calculated by the following formula i Step size S i ;
i denotes the calculation from 0 for the ith pooling operation in one feature map, ceil denotes the rounding up, floor denotes the rounding down.
4. The image processing method based on the dual-branch parallel neural network according to claim 3, wherein: the method also comprises the following steps:
s6: x is to be pool And Z block Splicing in channel dimension, outputting X cat ;
S7: x is to be cat Inputting into two layers of fully connected neural networks to obtain final output X out 。
5. The image processing method based on the dual-branch parallel neural network according to claim 1, wherein: the step S3 comprises the following steps:
s301: and before the output of the last bidirectional bridging module Block is output to the bidirectional bridging module Block again, parameters in the convolutional neural network module are adjusted, wherein the parameters comprise channels and step sizes.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311334063.6A CN117315366A (en) | 2023-10-13 | 2023-10-13 | Image processing method based on double-branch parallel neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202311334063.6A CN117315366A (en) | 2023-10-13 | 2023-10-13 | Image processing method based on double-branch parallel neural network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117315366A true CN117315366A (en) | 2023-12-29 |
Family
ID=89237063
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202311334063.6A Pending CN117315366A (en) | 2023-10-13 | 2023-10-13 | Image processing method based on double-branch parallel neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117315366A (en) |
-
2023
- 2023-10-13 CN CN202311334063.6A patent/CN117315366A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108764317B (en) | Residual convolutional neural network image classification method based on multipath feature weighting | |
CN109783666B (en) | Image scene graph generation method based on iterative refinement | |
CN110222760B (en) | Quick image processing method based on winograd algorithm | |
CN111091130A (en) | Real-time image semantic segmentation method and system based on lightweight convolutional neural network | |
CN111242844B (en) | Image processing method, device, server and storage medium | |
CN110458084B (en) | Face age estimation method based on inverted residual error network | |
CN113807355A (en) | Image semantic segmentation method based on coding and decoding structure | |
CN110889416B (en) | Salient object detection method based on cascade improved network | |
CN113516133B (en) | Multi-modal image classification method and system | |
CN113706545A (en) | Semi-supervised image segmentation method based on dual-branch nerve discrimination dimensionality reduction | |
CN110569851A (en) | real-time semantic segmentation method for gated multi-layer fusion | |
CN115471670A (en) | Space target detection method based on improved YOLOX network model | |
CN113888505B (en) | Natural scene text detection method based on semantic segmentation | |
CN114743027A (en) | Weak supervision learning-guided cooperative significance detection method | |
CN110110775A (en) | A kind of matching cost calculation method based on hyper linking network | |
CN117217274A (en) | Vector processor, neural network accelerator, chip and electronic equipment | |
CN117078539A (en) | CNN-transducer-based local global interactive image restoration method | |
CN117315366A (en) | Image processing method based on double-branch parallel neural network | |
CN116524180A (en) | Dramatic stage scene segmentation method based on lightweight backbone structure | |
CN114937153B (en) | Visual characteristic processing system and method based on neural network in weak texture environment | |
CN114494284B (en) | Scene analysis model and method based on explicit supervision area relation | |
CN115861841A (en) | SAR image target detection method combined with lightweight large convolution kernel | |
CN113836319B (en) | Knowledge completion method and system for fusion entity neighbors | |
CN112529064B (en) | Efficient real-time semantic segmentation method | |
CN112733934A (en) | Multi-modal feature fusion road scene semantic segmentation method in complex environment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |