CN113033630A - Infrared and visible light image deep learning fusion method based on double non-local attention models - Google Patents

Infrared and visible light image deep learning fusion method based on double non-local attention models Download PDF

Info

Publication number
CN113033630A
CN113033630A CN202110258048.2A CN202110258048A CN113033630A CN 113033630 A CN113033630 A CN 113033630A CN 202110258048 A CN202110258048 A CN 202110258048A CN 113033630 A CN113033630 A CN 113033630A
Authority
CN
China
Prior art keywords
fusion
features
attention
infrared
visible light
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110258048.2A
Other languages
Chinese (zh)
Inventor
王志社
武园园
王君尧
邵文禹
陈彦林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Taiyuan University of Science and Technology
Original Assignee
Taiyuan University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Taiyuan University of Science and Technology filed Critical Taiyuan University of Science and Technology
Priority to CN202110258048.2A priority Critical patent/CN113033630A/en
Publication of CN113033630A publication Critical patent/CN113033630A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Biophysics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention relates to an infrared and visible light image depth learning fusion method based on a double non-local attention model. The method specifically comprises the following steps: constructing a multi-scale depth network, and extracting multi-scale depth features of two types of images by adopting a coding-decoding network model, wherein the coding network comprises a common convolution module and two multi-scale convolution modules; the fusion layer utilizes a double non-local attention model of space and channel to enhance and combine the multi-scale depth features to obtain final fusion features; and the decoding network is used for reconstructing the final fusion characteristic to obtain a fusion image. The method solves the problem that the target area and texture details in the fusion image are lost due to the fact that the existing deep learning fusion method only depends on convolution operation to extract local features and does not fully consider multi-scale characteristics and global correlation of the local features, and can be applied to the fields of remote sensing detection, medical diagnosis, intelligent driving, safety monitoring and the like.

Description

Infrared and visible light image deep learning fusion method based on double non-local attention models
Technical Field
The invention relates to an image fusion method in the field of image processing and artificial intelligence, in particular to an infrared and visible light image deep learning fusion method based on a double non-local attention model.
Background
The infrared and visible light image fusion technology is widely applied to the fields of remote sensing detection, medical diagnosis, intelligent driving, safety monitoring and the like. Due to differences in imaging mechanisms and hardware conditions, an image sensor can only capture a portion of the information. Infrared image sensors use thermal radiation imaging to reflect the radiation characteristics of background objects, but are generally lacking in structural features and texture detail. The visible light image sensor is formed by reflecting light rays, can describe scene information of the environment, and is easily influenced by changes of lighting conditions and weather changes. The purpose of image fusion is to obtain a composite image with rich scene details and prominent background objects, so as to be suitable for human visual perception and other visual tasks. Therefore, image fusion techniques are an important prerequisite for improving the level of infrared and visible image detection and recognition.
Currently, infrared and visible light image fusion techniques can be broadly classified into conventional fusion methods and deep learning fusion methods. The traditional fusion method usually adopts unified feature transformation or feature representation to solve the fusion problem, and feature extraction and feature representation of manual design are one difficulty of the traditional fusion method. The convolution operation has stronger image feature extraction capability and obtains a satisfactory fusion effect based on a deep learning method in a big data driving mode. However, the current fusion method based on deep learning has some defects. Firstly, most methods rely on the last layer of depth features for processing only due to the limitation of the size of the kernel function of the filter and the depth of the network, and cannot effectively extract the multi-scale depth features. In fact, the multi-scale characteristics of the image play an important role in machine vision, and only a single-scale depth feature cannot effectively represent the spatial information of a large target in the image. Secondly, the methods only pay attention to the extraction of the local features and do not consider the global correlation of the local features, the global correlation of the local features can effectively highlight useful information and inhibit irrelevant information, and the method plays an important role in the fusion task. Third, the current method directly uses the depth features obtained by the convolutional neural network for combined reconstruction, but these deep features are not refined and enhanced, thereby resulting in the loss of details of the target region and texture in the fused image.
In summary, there is an urgent need for a method capable of effectively extracting multi-scale depth features of an image, establishing global correlation of the multi-scale depth features, suppressing irrelevant information when enhancing useful information, and enhancing depth feature representation capability, thereby improving the fusion effect of infrared and visible light images.
Disclosure of Invention
The invention provides an infrared and visible light image deep learning fusion method based on a double non-local attention model, aiming at solving the problem that the target area and texture details in a fusion image are lost due to the fact that the existing deep learning fusion method only depends on convolution operation to extract local features and does not fully consider multi-scale characteristics and global correlation of the existing deep learning fusion method.
The invention relates to an infrared and visible light image deep learning fusion method based on a double non-local attention model, which comprises the following steps:
designing and constructing a multi-scale depth network: the multi-scale depth network comprises two sub-networks of encoding and decoding, the encoding sub-network is used for extracting multi-scale depth features of a source image, and the decoding sub-network is used for reconstructing final fusion features to obtain a fusion image;
designing a fusion layer: for input infrared and visible light images, multi-scale depth features of the two types of images are respectively extracted from an encoding sub-network, space and channel attention features of the two types of images are respectively obtained by utilizing a space and channel non-local attention module, the space attention features of the two types of images are fused to obtain a space attention fusion feature, the channel attention features of the two types of images are fused to obtain a channel attention fusion feature, and finally the space and channel attention fusion feature is weighted to obtain a final fusion feature.
The infrared and visible light image deep learning fusion method based on the double non-local attention models is characterized in that a coding sub-network for extracting multi-scale depth features is composed of a common convolution module and two multi-scale convolution modules, the common convolution module uses a 3 x 3 convolution kernel, the step length is 1, the filling is 0, the number of filters is 16, and the convolution layer is followed by a correction linear unit (ReLU) and used for extracting shallow depth features of images; the two multi-scale convolution modules are used for extracting the multi-scale depth features of the image, the scale is 4, the number of output channels is 64, and the common convolution module and the two multi-scale convolution modules are connected densely to enhance the feature characterization capability.
According to the infrared and visible light image deep learning fusion method based on the double non-local attention model, the spatial attention feature fusion process is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding spatial attention characteristics through the spatial attention module
Figure BDA0002968362700000021
Then using spatial attention characteristics
Figure BDA0002968362700000022
Calculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented as
Figure BDA0002968362700000023
And
Figure BDA0002968362700000024
finally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the spatial attention fusion features
Figure BDA0002968362700000025
In the above method for fusing infrared and visible light image deep learning based on the double non-local attention model, the fusion process of the channel attention features is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding channel attention characteristics through the channel attention module
Figure BDA0002968362700000026
Then using the channel attention feature
Figure BDA0002968362700000027
Calculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented as
Figure BDA0002968362700000031
And
Figure BDA0002968362700000032
finally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the channel attention fusion features
Figure BDA0002968362700000033
According to the infrared and visible light image deep learning fusion method based on the double non-local attention models, the spatial non-local attention model is selected to be subjected to 8 x 8 pooling operation.
Compared with the existing deep learning fusion technology, the invention has the following advantages:
1. according to the invention, a multi-scale convolution module is embedded into a coding sub-network, the coding network can effectively extract multi-scale depth features of an image, the dense connection can retain the depth features of a network middle layer, and the multi-scale features and the dense connection can effectively enhance the depth feature characterization capability;
2. according to the method, a space and channel non-local attention model is adopted, multi-scale depth features are enhanced and fused from the dimensions of the space and the channel, useful information is enhanced, irrelevant information is inhibited, and the fused image highlights infrared image target information while abundant detail information of a visible image is reserved;
3. the invention establishes the deep learning fusion method facing the double non-local attention model, obviously improves the fusion effect, can also be applied to the fusion of multi-mode images, multi-focus images and medical images, and has high application value in the field of image fusion.
Drawings
FIG. 1 is a schematic fusion diagram of the method of the present invention.
FIG. 2 is a schematic diagram of the training of the method of the present invention.
FIG. 3 is a schematic diagram of a multi-scale convolution module of the method of the present invention.
FIG. 4 is a schematic diagram of a fusion of dual non-local attention features according to the method of the present invention.
FIG. 5 is a schematic view of a spatial and channel attention model of the method of the present invention.
FIG. 6 is a first set of graphs of an infrared and visible image fusion experiment.
Fig. 7 is a second set of infrared and visible light image fusion experimental graphs.
FIG. 8 is a third set of experimental graphs of infrared and visible light image fusion.
Detailed Description
A method for deep learning and fusing infrared and visible light images based on a double non-local attention model comprises the following steps:
s1: and designing and constructing a multi-scale depth network. The multi-scale depth network comprises two sub-networks of encoding and decoding. The coding network is used for extracting the multi-scale depth features of the source image. And the decoding network is used for reconstructing the final fusion characteristics to obtain a fusion image.
S11: and (4) forming a coding network. The coding network consists of a common convolution module and two multi-scale convolution modules. The common convolution module uses a 3 × 3 convolution kernel with a step size of 1, a fill of 0, and a number of filters of 16, and the convolution layer is followed by a modified linear unit (ReLU) for extracting shallow depth features of the image. The two multi-scale convolution modules are used for extracting the image multi-scale depth features. In order to fully utilize the intermediate features of the deep network, dense connections are applied between the ordinary convolution module and the two multi-scale convolution modules.
S12: and a multi-scale convolution module. Depth feature F for normal convolution module outputxAfter 1 × 1 convolution, the depth feature is divided into s sub-features xiI is 1,2, … s, except for x1In addition, other sub-features are convolved by 3 × 3 and modified by a linear unit (ReLU), denoted as Ci() Each of Ci-1() Is added to Ci() In, then output yiCan be expressed as
Figure BDA0002968362700000041
Then to the output yiMaking a channel connection (termination) with the input FxAdding to obtain final output characteristic Fy=concat(yi)+FxWhere s represents a scale control parameter. Experiments prove that the scale control parameter s takes a value of 4, and the output channel of the multi-scale convolution module is 64.
Because of intensive connection, shallow depth features output by the common convolution module are used as input of the first multi-scale convolution module, the shallow depth features output by the common convolution module and output features of the first multi-scale convolution module are subjected to channel connection (coordination) and are used as input of the second multi-scale convolution module, and finally, the bottom depth features output by the common convolution module and the output features of the two multi-scale convolution modules are subjected to channel connection (coordination) to obtain the multi-scale depth features phi.
S13: and (5) decoding network composition. The decoding subnetwork comprises 4 general convolution modules, each convolution using a 3 × 3 convolution kernel with step size 1, padding 0, convolution layer followed by a modified linear unit (ReLU), the number of convolution filters being 112,64,32,16, respectively.
S2: and (3) fusion layer design, for input infrared and visible light images, respectively extracting multi-scale depth features of the two images in an encoding network, and obtaining final fusion features by using double non-local attention modules of space and channels.
S21: spatial non-local attention models. Spatial non-local attention establishes global feature correlation for local features from the spatial dimension. For a multiscale depth feature Φ ∈ RC×H×WFirstly, reconstructing and transposing the same to obtain phiA∈RHW ×CThen, n × n pooling operation and reconstruction are carried out on the multi-scale depth features to obtain another two depth features
Figure BDA0002968362700000042
Is connected immediately afterAlignment feature phiAAnd phiBMatrix multiplication and spatial attention mapping by SoftMax operation
Figure BDA0002968362700000051
Wherein
Figure BDA0002968362700000052
Indicating the effect of the ith position on the jth position. Finally, is aligned again
Figure BDA0002968362700000053
And phiCMatrix multiplication and reconstruction are carried out to obtain spatial non-local attention characteristics
Figure BDA0002968362700000054
In the spatial non-local attention module, the n × n pooling operation is to reduce the amount of calculation, and experiments verify that the pooling operations for 1 × 1,2 × 2, 4 × 4 and 8 × 8 do not affect the image fusion performance, so the 8 × 8 pooling operation is finally selected.
S22: the channel is a non-local attention model. Channel non-local attention establishes global feature correlation for local features from the channel dimension. For a multiscale depth feature Φ ∈ RC×H×WFirstly, reconstructing the data to obtain phi epsilon RC×HWThen for the characteristic phi epsilon RC×HWAnd transposed phi epsilon RC×HWMatrix multiplication and channel attention mapping by SoftMax operation
Figure BDA0002968362700000055
Wherein
Figure BDA0002968362700000056
Shows the effect of the ith channel on the jth channel, and finally on again
Figure BDA0002968362700000057
Multiplying the sum phi by a matrix and reconstructing to obtain the non-local attention characteristic of the channel
Figure BDA0002968362700000058
S23: spatial attention features fusion. Multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding spatial attention characteristics through the spatial attention module
Figure BDA0002968362700000059
Then using spatial attention characteristics
Figure BDA00029683627000000510
Calculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented as
Figure BDA00029683627000000511
And
Figure BDA00029683627000000512
finally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the spatial attention fusion features
Figure BDA00029683627000000513
S24: channel attention features fusion. Multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding channel attention characteristics through the channel attention module
Figure BDA00029683627000000514
Then using the channel attention feature
Figure BDA00029683627000000515
Calculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented as
Figure BDA00029683627000000516
And
Figure BDA00029683627000000517
finally, the sum of the infrared and visible imagesMultiplying the weight coefficient by the corresponding depth feature to obtain the channel attention fusion feature
Figure BDA0002968362700000061
S25: spatial and channel attention characteristics are fused. Weighting the space and channel attention fusion characteristics of the infrared image and the visible light image to obtain the final fusion characteristics
Figure BDA0002968362700000062
And finally, obtaining a fused image through the decoding network according to the fused characteristics.
S3: and (5) training a network model. Performing graying and size adjustment on an input visible light image by adopting an MS-COCO image data set, and training a network model by adopting structural similarity and root mean square error as loss functions to obtain hyper-parameters of the network model;
s31: a training data set is selected. 80000 visible light images are selected from the MS-COCO image dataset as a training set, the image grayscale range is converted to [0,255], and the size is converted to 256 × 256.
S32: and setting training parameters. The overall loss function consists of the structural similarity and the root mean square error loss, denoted LTotal=LMSE+βLSSIMWherein the structural similarity loss function LSSIM1-SSIM (O, I), root mean square error loss function LMSE=∑||O-I||2I and O are input and output images, SSIM represents a structural similarity operator, beta is a hyper-parameter for controlling network balance, and the value of beta is 1 in the invention. The batchsize and epoch sizes were 5 and 4, respectively, and the learning rate was 0.0001.

Claims (5)

1. A method for deep learning and fusion of infrared and visible light images based on a double non-local attention model is characterized by comprising the following steps: the method comprises the following steps:
designing and constructing a multi-scale depth network: the multi-scale depth network comprises two sub-networks of encoding and decoding, the encoding sub-network is used for extracting multi-scale depth features of a source image, and the decoding sub-network is used for reconstructing final fusion features to obtain a fusion image;
designing a fusion layer: for input infrared and visible light images, multi-scale depth features of the two types of images are respectively extracted from an encoding sub-network, space and channel attention features of the two types of images are respectively obtained by utilizing a space and channel non-local attention module, the space attention features of the two types of images are fused to obtain a space attention fusion feature, the channel attention features of the two types of images are fused to obtain a channel attention fusion feature, and finally the space and channel attention fusion feature is weighted to obtain a final fusion feature.
2. The infrared and visible light image deep learning fusion method based on the double non-local attention model according to claim 1, characterized in that: the coding subnetwork for extracting the multi-scale depth features consists of a convolution module and two multi-scale convolution modules, wherein the convolution module uses a 3 multiplied by 3 convolution kernel, the step length is 1, the filling is 0, the number of filters is 16, and the convolution layer is followed by a correction linear unit and used for extracting the shallow depth features of the image; the two multi-scale convolution modules are used for extracting the multi-scale depth features of the image, the scale is 4, the number of output channels is 64, and the feature characterization capability is enhanced by adopting dense connection between the convolution modules and the two multi-scale convolution modules.
3. The infrared and visible light image deep learning fusion method based on the double non-local attention model according to claim 2, characterized in that: the spatial attention feature fusion process is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding spatial attention characteristics through the spatial attention module
Figure FDA0002968362690000011
Then using spatial attention characteristics
Figure FDA0002968362690000012
Calculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented as
Figure FDA0002968362690000013
And
Figure FDA0002968362690000014
finally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the spatial attention fusion features
Figure FDA0002968362690000015
4. The infrared and visible light image deep learning fusion method based on the double non-local attention model according to claim 3, characterized in that: the channel attention feature fusion process is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding channel attention characteristics through the channel attention module
Figure FDA0002968362690000016
Then using the channel attention feature
Figure FDA0002968362690000017
Calculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented as
Figure FDA0002968362690000021
And
Figure FDA0002968362690000022
finally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the channel attention fusion features
Figure FDA0002968362690000023
5. The method for fusing infrared and visible light image deep learning based on the double non-local attention model according to claim 1,2, 3 or 4, characterized in that: the spatial non-local attention model selects an 8 x 8 pooling operation.
CN202110258048.2A 2021-03-09 2021-03-09 Infrared and visible light image deep learning fusion method based on double non-local attention models Pending CN113033630A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110258048.2A CN113033630A (en) 2021-03-09 2021-03-09 Infrared and visible light image deep learning fusion method based on double non-local attention models

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110258048.2A CN113033630A (en) 2021-03-09 2021-03-09 Infrared and visible light image deep learning fusion method based on double non-local attention models

Publications (1)

Publication Number Publication Date
CN113033630A true CN113033630A (en) 2021-06-25

Family

ID=76468676

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110258048.2A Pending CN113033630A (en) 2021-03-09 2021-03-09 Infrared and visible light image deep learning fusion method based on double non-local attention models

Country Status (1)

Country Link
CN (1) CN113033630A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113343953A (en) * 2021-08-05 2021-09-03 南京信息工程大学 FGR-AM method and system for remote sensing scene recognition
CN113487530A (en) * 2021-08-02 2021-10-08 广东工业大学 Infrared and visible light fusion imaging method based on deep learning
CN113837353A (en) * 2021-08-17 2021-12-24 中国地质大学(武汉) Convolutional neural network feature fusion algorithm based on feature screening and deep fusion
CN114972812A (en) * 2022-06-02 2022-08-30 华侨大学 Non-local attention learning method based on structural similarity
CN115311186A (en) * 2022-10-09 2022-11-08 济南和普威视光电技术有限公司 Cross-scale attention confrontation fusion method for infrared and visible light images and terminal
CN115330658A (en) * 2022-10-17 2022-11-11 中国科学技术大学 Multi-exposure image fusion method, device, equipment and storage medium
CN117115442A (en) * 2023-08-17 2023-11-24 浙江航天润博测控技术有限公司 Semantic segmentation method based on visible light-infrared photoelectric reconnaissance image fusion

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111080629A (en) * 2019-12-20 2020-04-28 河北工业大学 Method for detecting image splicing tampering
CN111563418A (en) * 2020-04-14 2020-08-21 浙江科技学院 Asymmetric multi-mode fusion significance detection method based on attention mechanism
CN111709902A (en) * 2020-05-21 2020-09-25 江南大学 Infrared and visible light image fusion method based on self-attention mechanism

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111080629A (en) * 2019-12-20 2020-04-28 河北工业大学 Method for detecting image splicing tampering
CN111563418A (en) * 2020-04-14 2020-08-21 浙江科技学院 Asymmetric multi-mode fusion significance detection method based on attention mechanism
CN111709902A (en) * 2020-05-21 2020-09-25 江南大学 Infrared and visible light image fusion method based on self-attention mechanism

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
GAO HUANG ET AL.: "Densely Connected Convolutional Networks", 《ARXIV》 *
SHANG HUA GAO ET AL.: "Res2Net: A New Multi-scale Backbone Architecture", 《ARXIV》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113487530A (en) * 2021-08-02 2021-10-08 广东工业大学 Infrared and visible light fusion imaging method based on deep learning
CN113487530B (en) * 2021-08-02 2023-06-16 广东工业大学 Infrared and visible light fusion imaging method based on deep learning
CN113343953A (en) * 2021-08-05 2021-09-03 南京信息工程大学 FGR-AM method and system for remote sensing scene recognition
CN113837353A (en) * 2021-08-17 2021-12-24 中国地质大学(武汉) Convolutional neural network feature fusion algorithm based on feature screening and deep fusion
CN114972812A (en) * 2022-06-02 2022-08-30 华侨大学 Non-local attention learning method based on structural similarity
CN115311186A (en) * 2022-10-09 2022-11-08 济南和普威视光电技术有限公司 Cross-scale attention confrontation fusion method for infrared and visible light images and terminal
CN115311186B (en) * 2022-10-09 2023-02-03 济南和普威视光电技术有限公司 Cross-scale attention confrontation fusion method and terminal for infrared and visible light images
CN115330658A (en) * 2022-10-17 2022-11-11 中国科学技术大学 Multi-exposure image fusion method, device, equipment and storage medium
CN115330658B (en) * 2022-10-17 2023-03-10 中国科学技术大学 Multi-exposure image fusion method, device, equipment and storage medium
CN117115442A (en) * 2023-08-17 2023-11-24 浙江航天润博测控技术有限公司 Semantic segmentation method based on visible light-infrared photoelectric reconnaissance image fusion

Similar Documents

Publication Publication Date Title
CN113033630A (en) Infrared and visible light image deep learning fusion method based on double non-local attention models
CN103020933B (en) A kind of multisource image anastomosing method based on bionic visual mechanism
CN110189286B (en) Infrared and visible light image fusion method based on ResNet
CN113139585B (en) Infrared and visible light image fusion method based on unified multi-scale dense connection network
CN109410135B (en) Anti-learning image defogging and fogging method
CN112967178B (en) Image conversion method, device, equipment and storage medium
CN113283444B (en) Heterogeneous image migration method based on generation countermeasure network
CN114548265B (en) Crop leaf disease image generation model training method, crop leaf disease identification method, electronic equipment and storage medium
CN115713679A (en) Target detection method based on multi-source information fusion, thermal infrared and three-dimensional depth map
CN115311186B (en) Cross-scale attention confrontation fusion method and terminal for infrared and visible light images
CN115035003A (en) Infrared and visible light image anti-fusion method for interactively compensating attention
CN115511767B (en) Self-supervised learning multi-modal image fusion method and application thereof
CN104408697B (en) Image Super-resolution Reconstruction method based on genetic algorithm and canonical prior model
CN116258936A (en) Infrared and visible light image fusion method based on multi-scale features
CN112767243A (en) Hyperspectral image super-resolution implementation method and system
CN114463172A (en) Light field image super-resolution reconstruction method oriented to view consistency
CN115546442A (en) Multi-view stereo matching reconstruction method and system based on perception consistency loss
CN116757986A (en) Infrared and visible light image fusion method and device
CN113298744B (en) End-to-end infrared and visible light image fusion method
CN114639002A (en) Infrared and visible light image fusion method based on multi-mode characteristics
CN113011438B (en) Bimodal image significance detection method based on node classification and sparse graph learning
CN117292244A (en) Infrared and visible light image fusion method based on multilayer convolution
CN116051947A (en) Multi-scale feature fusion method for multi-source heterogeneous data of transformer
CN113160104B (en) Image fusion method based on densely connected network
CN113362281A (en) Infrared and visible light image fusion method based on WSN-LatLRR

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20210625

WD01 Invention patent application deemed withdrawn after publication