CN113033630A - Infrared and visible light image deep learning fusion method based on double non-local attention models - Google Patents
Infrared and visible light image deep learning fusion method based on double non-local attention models Download PDFInfo
- Publication number
- CN113033630A CN113033630A CN202110258048.2A CN202110258048A CN113033630A CN 113033630 A CN113033630 A CN 113033630A CN 202110258048 A CN202110258048 A CN 202110258048A CN 113033630 A CN113033630 A CN 113033630A
- Authority
- CN
- China
- Prior art keywords
- fusion
- features
- attention
- infrared
- visible light
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention relates to an infrared and visible light image depth learning fusion method based on a double non-local attention model. The method specifically comprises the following steps: constructing a multi-scale depth network, and extracting multi-scale depth features of two types of images by adopting a coding-decoding network model, wherein the coding network comprises a common convolution module and two multi-scale convolution modules; the fusion layer utilizes a double non-local attention model of space and channel to enhance and combine the multi-scale depth features to obtain final fusion features; and the decoding network is used for reconstructing the final fusion characteristic to obtain a fusion image. The method solves the problem that the target area and texture details in the fusion image are lost due to the fact that the existing deep learning fusion method only depends on convolution operation to extract local features and does not fully consider multi-scale characteristics and global correlation of the local features, and can be applied to the fields of remote sensing detection, medical diagnosis, intelligent driving, safety monitoring and the like.
Description
Technical Field
The invention relates to an image fusion method in the field of image processing and artificial intelligence, in particular to an infrared and visible light image deep learning fusion method based on a double non-local attention model.
Background
The infrared and visible light image fusion technology is widely applied to the fields of remote sensing detection, medical diagnosis, intelligent driving, safety monitoring and the like. Due to differences in imaging mechanisms and hardware conditions, an image sensor can only capture a portion of the information. Infrared image sensors use thermal radiation imaging to reflect the radiation characteristics of background objects, but are generally lacking in structural features and texture detail. The visible light image sensor is formed by reflecting light rays, can describe scene information of the environment, and is easily influenced by changes of lighting conditions and weather changes. The purpose of image fusion is to obtain a composite image with rich scene details and prominent background objects, so as to be suitable for human visual perception and other visual tasks. Therefore, image fusion techniques are an important prerequisite for improving the level of infrared and visible image detection and recognition.
Currently, infrared and visible light image fusion techniques can be broadly classified into conventional fusion methods and deep learning fusion methods. The traditional fusion method usually adopts unified feature transformation or feature representation to solve the fusion problem, and feature extraction and feature representation of manual design are one difficulty of the traditional fusion method. The convolution operation has stronger image feature extraction capability and obtains a satisfactory fusion effect based on a deep learning method in a big data driving mode. However, the current fusion method based on deep learning has some defects. Firstly, most methods rely on the last layer of depth features for processing only due to the limitation of the size of the kernel function of the filter and the depth of the network, and cannot effectively extract the multi-scale depth features. In fact, the multi-scale characteristics of the image play an important role in machine vision, and only a single-scale depth feature cannot effectively represent the spatial information of a large target in the image. Secondly, the methods only pay attention to the extraction of the local features and do not consider the global correlation of the local features, the global correlation of the local features can effectively highlight useful information and inhibit irrelevant information, and the method plays an important role in the fusion task. Third, the current method directly uses the depth features obtained by the convolutional neural network for combined reconstruction, but these deep features are not refined and enhanced, thereby resulting in the loss of details of the target region and texture in the fused image.
In summary, there is an urgent need for a method capable of effectively extracting multi-scale depth features of an image, establishing global correlation of the multi-scale depth features, suppressing irrelevant information when enhancing useful information, and enhancing depth feature representation capability, thereby improving the fusion effect of infrared and visible light images.
Disclosure of Invention
The invention provides an infrared and visible light image deep learning fusion method based on a double non-local attention model, aiming at solving the problem that the target area and texture details in a fusion image are lost due to the fact that the existing deep learning fusion method only depends on convolution operation to extract local features and does not fully consider multi-scale characteristics and global correlation of the existing deep learning fusion method.
The invention relates to an infrared and visible light image deep learning fusion method based on a double non-local attention model, which comprises the following steps:
designing and constructing a multi-scale depth network: the multi-scale depth network comprises two sub-networks of encoding and decoding, the encoding sub-network is used for extracting multi-scale depth features of a source image, and the decoding sub-network is used for reconstructing final fusion features to obtain a fusion image;
designing a fusion layer: for input infrared and visible light images, multi-scale depth features of the two types of images are respectively extracted from an encoding sub-network, space and channel attention features of the two types of images are respectively obtained by utilizing a space and channel non-local attention module, the space attention features of the two types of images are fused to obtain a space attention fusion feature, the channel attention features of the two types of images are fused to obtain a channel attention fusion feature, and finally the space and channel attention fusion feature is weighted to obtain a final fusion feature.
The infrared and visible light image deep learning fusion method based on the double non-local attention models is characterized in that a coding sub-network for extracting multi-scale depth features is composed of a common convolution module and two multi-scale convolution modules, the common convolution module uses a 3 x 3 convolution kernel, the step length is 1, the filling is 0, the number of filters is 16, and the convolution layer is followed by a correction linear unit (ReLU) and used for extracting shallow depth features of images; the two multi-scale convolution modules are used for extracting the multi-scale depth features of the image, the scale is 4, the number of output channels is 64, and the common convolution module and the two multi-scale convolution modules are connected densely to enhance the feature characterization capability.
According to the infrared and visible light image deep learning fusion method based on the double non-local attention model, the spatial attention feature fusion process is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding spatial attention characteristics through the spatial attention moduleThen using spatial attention characteristicsCalculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented asAndfinally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the spatial attention fusion features
In the above method for fusing infrared and visible light image deep learning based on the double non-local attention model, the fusion process of the channel attention features is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding channel attention characteristics through the channel attention moduleThen using the channel attention featureCalculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented asAndfinally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the channel attention fusion features
According to the infrared and visible light image deep learning fusion method based on the double non-local attention models, the spatial non-local attention model is selected to be subjected to 8 x 8 pooling operation.
Compared with the existing deep learning fusion technology, the invention has the following advantages:
1. according to the invention, a multi-scale convolution module is embedded into a coding sub-network, the coding network can effectively extract multi-scale depth features of an image, the dense connection can retain the depth features of a network middle layer, and the multi-scale features and the dense connection can effectively enhance the depth feature characterization capability;
2. according to the method, a space and channel non-local attention model is adopted, multi-scale depth features are enhanced and fused from the dimensions of the space and the channel, useful information is enhanced, irrelevant information is inhibited, and the fused image highlights infrared image target information while abundant detail information of a visible image is reserved;
3. the invention establishes the deep learning fusion method facing the double non-local attention model, obviously improves the fusion effect, can also be applied to the fusion of multi-mode images, multi-focus images and medical images, and has high application value in the field of image fusion.
Drawings
FIG. 1 is a schematic fusion diagram of the method of the present invention.
FIG. 2 is a schematic diagram of the training of the method of the present invention.
FIG. 3 is a schematic diagram of a multi-scale convolution module of the method of the present invention.
FIG. 4 is a schematic diagram of a fusion of dual non-local attention features according to the method of the present invention.
FIG. 5 is a schematic view of a spatial and channel attention model of the method of the present invention.
FIG. 6 is a first set of graphs of an infrared and visible image fusion experiment.
Fig. 7 is a second set of infrared and visible light image fusion experimental graphs.
FIG. 8 is a third set of experimental graphs of infrared and visible light image fusion.
Detailed Description
A method for deep learning and fusing infrared and visible light images based on a double non-local attention model comprises the following steps:
s1: and designing and constructing a multi-scale depth network. The multi-scale depth network comprises two sub-networks of encoding and decoding. The coding network is used for extracting the multi-scale depth features of the source image. And the decoding network is used for reconstructing the final fusion characteristics to obtain a fusion image.
S11: and (4) forming a coding network. The coding network consists of a common convolution module and two multi-scale convolution modules. The common convolution module uses a 3 × 3 convolution kernel with a step size of 1, a fill of 0, and a number of filters of 16, and the convolution layer is followed by a modified linear unit (ReLU) for extracting shallow depth features of the image. The two multi-scale convolution modules are used for extracting the image multi-scale depth features. In order to fully utilize the intermediate features of the deep network, dense connections are applied between the ordinary convolution module and the two multi-scale convolution modules.
S12: and a multi-scale convolution module. Depth feature F for normal convolution module outputxAfter 1 × 1 convolution, the depth feature is divided into s sub-features xiI is 1,2, … s, except for x1In addition, other sub-features are convolved by 3 × 3 and modified by a linear unit (ReLU), denoted as Ci() Each of Ci-1() Is added to Ci() In, then output yiCan be expressed asThen to the output yiMaking a channel connection (termination) with the input FxAdding to obtain final output characteristic Fy=concat(yi)+FxWhere s represents a scale control parameter. Experiments prove that the scale control parameter s takes a value of 4, and the output channel of the multi-scale convolution module is 64.
Because of intensive connection, shallow depth features output by the common convolution module are used as input of the first multi-scale convolution module, the shallow depth features output by the common convolution module and output features of the first multi-scale convolution module are subjected to channel connection (coordination) and are used as input of the second multi-scale convolution module, and finally, the bottom depth features output by the common convolution module and the output features of the two multi-scale convolution modules are subjected to channel connection (coordination) to obtain the multi-scale depth features phi.
S13: and (5) decoding network composition. The decoding subnetwork comprises 4 general convolution modules, each convolution using a 3 × 3 convolution kernel with step size 1, padding 0, convolution layer followed by a modified linear unit (ReLU), the number of convolution filters being 112,64,32,16, respectively.
S2: and (3) fusion layer design, for input infrared and visible light images, respectively extracting multi-scale depth features of the two images in an encoding network, and obtaining final fusion features by using double non-local attention modules of space and channels.
S21: spatial non-local attention models. Spatial non-local attention establishes global feature correlation for local features from the spatial dimension. For a multiscale depth feature Φ ∈ RC×H×WFirstly, reconstructing and transposing the same to obtain phiA∈RHW ×CThen, n × n pooling operation and reconstruction are carried out on the multi-scale depth features to obtain another two depth featuresIs connected immediately afterAlignment feature phiAAnd phiBMatrix multiplication and spatial attention mapping by SoftMax operationWhereinIndicating the effect of the ith position on the jth position. Finally, is aligned againAnd phiCMatrix multiplication and reconstruction are carried out to obtain spatial non-local attention characteristicsIn the spatial non-local attention module, the n × n pooling operation is to reduce the amount of calculation, and experiments verify that the pooling operations for 1 × 1,2 × 2, 4 × 4 and 8 × 8 do not affect the image fusion performance, so the 8 × 8 pooling operation is finally selected.
S22: the channel is a non-local attention model. Channel non-local attention establishes global feature correlation for local features from the channel dimension. For a multiscale depth feature Φ ∈ RC×H×WFirstly, reconstructing the data to obtain phi epsilon RC×HWThen for the characteristic phi epsilon RC×HWAnd transposed phi epsilon RC×HWMatrix multiplication and channel attention mapping by SoftMax operationWhereinShows the effect of the ith channel on the jth channel, and finally on againMultiplying the sum phi by a matrix and reconstructing to obtain the non-local attention characteristic of the channel
S23: spatial attention features fusion. Multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding spatial attention characteristics through the spatial attention moduleThen using spatial attention characteristicsCalculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented asAndfinally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the spatial attention fusion features
S24: channel attention features fusion. Multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding channel attention characteristics through the channel attention moduleThen using the channel attention featureCalculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented asAndfinally, the sum of the infrared and visible imagesMultiplying the weight coefficient by the corresponding depth feature to obtain the channel attention fusion feature
S25: spatial and channel attention characteristics are fused. Weighting the space and channel attention fusion characteristics of the infrared image and the visible light image to obtain the final fusion characteristicsAnd finally, obtaining a fused image through the decoding network according to the fused characteristics.
S3: and (5) training a network model. Performing graying and size adjustment on an input visible light image by adopting an MS-COCO image data set, and training a network model by adopting structural similarity and root mean square error as loss functions to obtain hyper-parameters of the network model;
s31: a training data set is selected. 80000 visible light images are selected from the MS-COCO image dataset as a training set, the image grayscale range is converted to [0,255], and the size is converted to 256 × 256.
S32: and setting training parameters. The overall loss function consists of the structural similarity and the root mean square error loss, denoted LTotal=LMSE+βLSSIMWherein the structural similarity loss function LSSIM1-SSIM (O, I), root mean square error loss function LMSE=∑||O-I||2I and O are input and output images, SSIM represents a structural similarity operator, beta is a hyper-parameter for controlling network balance, and the value of beta is 1 in the invention. The batchsize and epoch sizes were 5 and 4, respectively, and the learning rate was 0.0001.
Claims (5)
1. A method for deep learning and fusion of infrared and visible light images based on a double non-local attention model is characterized by comprising the following steps: the method comprises the following steps:
designing and constructing a multi-scale depth network: the multi-scale depth network comprises two sub-networks of encoding and decoding, the encoding sub-network is used for extracting multi-scale depth features of a source image, and the decoding sub-network is used for reconstructing final fusion features to obtain a fusion image;
designing a fusion layer: for input infrared and visible light images, multi-scale depth features of the two types of images are respectively extracted from an encoding sub-network, space and channel attention features of the two types of images are respectively obtained by utilizing a space and channel non-local attention module, the space attention features of the two types of images are fused to obtain a space attention fusion feature, the channel attention features of the two types of images are fused to obtain a channel attention fusion feature, and finally the space and channel attention fusion feature is weighted to obtain a final fusion feature.
2. The infrared and visible light image deep learning fusion method based on the double non-local attention model according to claim 1, characterized in that: the coding subnetwork for extracting the multi-scale depth features consists of a convolution module and two multi-scale convolution modules, wherein the convolution module uses a 3 multiplied by 3 convolution kernel, the step length is 1, the filling is 0, the number of filters is 16, and the convolution layer is followed by a correction linear unit and used for extracting the shallow depth features of the image; the two multi-scale convolution modules are used for extracting the multi-scale depth features of the image, the scale is 4, the number of output channels is 64, and the feature characterization capability is enhanced by adopting dense connection between the convolution modules and the two multi-scale convolution modules.
3. The infrared and visible light image deep learning fusion method based on the double non-local attention model according to claim 2, characterized in that: the spatial attention feature fusion process is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding spatial attention characteristics through the spatial attention moduleThen using spatial attention characteristicsCalculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented asAndfinally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the spatial attention fusion features
4. The infrared and visible light image deep learning fusion method based on the double non-local attention model according to claim 3, characterized in that: the channel attention feature fusion process is as follows: multi-scale depth feature phi for infrared and visible light imagesI、ΦVObtaining corresponding channel attention characteristics through the channel attention moduleThen using the channel attention featureCalculating to obtain weighting coefficients of the infrared image and the visible light image respectively represented asAndfinally, multiplying the weighting coefficients of the infrared and visible light images with the corresponding depth features to obtain the channel attention fusion features
5. The method for fusing infrared and visible light image deep learning based on the double non-local attention model according to claim 1,2, 3 or 4, characterized in that: the spatial non-local attention model selects an 8 x 8 pooling operation.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110258048.2A CN113033630A (en) | 2021-03-09 | 2021-03-09 | Infrared and visible light image deep learning fusion method based on double non-local attention models |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110258048.2A CN113033630A (en) | 2021-03-09 | 2021-03-09 | Infrared and visible light image deep learning fusion method based on double non-local attention models |
Publications (1)
Publication Number | Publication Date |
---|---|
CN113033630A true CN113033630A (en) | 2021-06-25 |
Family
ID=76468676
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110258048.2A Pending CN113033630A (en) | 2021-03-09 | 2021-03-09 | Infrared and visible light image deep learning fusion method based on double non-local attention models |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113033630A (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113343953A (en) * | 2021-08-05 | 2021-09-03 | 南京信息工程大学 | FGR-AM method and system for remote sensing scene recognition |
CN113487530A (en) * | 2021-08-02 | 2021-10-08 | 广东工业大学 | Infrared and visible light fusion imaging method based on deep learning |
CN113837353A (en) * | 2021-08-17 | 2021-12-24 | 中国地质大学(武汉) | Convolutional neural network feature fusion algorithm based on feature screening and deep fusion |
CN114972812A (en) * | 2022-06-02 | 2022-08-30 | 华侨大学 | Non-local attention learning method based on structural similarity |
CN115311186A (en) * | 2022-10-09 | 2022-11-08 | 济南和普威视光电技术有限公司 | Cross-scale attention confrontation fusion method for infrared and visible light images and terminal |
CN115330658A (en) * | 2022-10-17 | 2022-11-11 | 中国科学技术大学 | Multi-exposure image fusion method, device, equipment and storage medium |
CN117115442A (en) * | 2023-08-17 | 2023-11-24 | 浙江航天润博测控技术有限公司 | Semantic segmentation method based on visible light-infrared photoelectric reconnaissance image fusion |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111080629A (en) * | 2019-12-20 | 2020-04-28 | 河北工业大学 | Method for detecting image splicing tampering |
CN111563418A (en) * | 2020-04-14 | 2020-08-21 | 浙江科技学院 | Asymmetric multi-mode fusion significance detection method based on attention mechanism |
CN111709902A (en) * | 2020-05-21 | 2020-09-25 | 江南大学 | Infrared and visible light image fusion method based on self-attention mechanism |
-
2021
- 2021-03-09 CN CN202110258048.2A patent/CN113033630A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111080629A (en) * | 2019-12-20 | 2020-04-28 | 河北工业大学 | Method for detecting image splicing tampering |
CN111563418A (en) * | 2020-04-14 | 2020-08-21 | 浙江科技学院 | Asymmetric multi-mode fusion significance detection method based on attention mechanism |
CN111709902A (en) * | 2020-05-21 | 2020-09-25 | 江南大学 | Infrared and visible light image fusion method based on self-attention mechanism |
Non-Patent Citations (2)
Title |
---|
GAO HUANG ET AL.: "Densely Connected Convolutional Networks", 《ARXIV》 * |
SHANG HUA GAO ET AL.: "Res2Net: A New Multi-scale Backbone Architecture", 《ARXIV》 * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113487530A (en) * | 2021-08-02 | 2021-10-08 | 广东工业大学 | Infrared and visible light fusion imaging method based on deep learning |
CN113487530B (en) * | 2021-08-02 | 2023-06-16 | 广东工业大学 | Infrared and visible light fusion imaging method based on deep learning |
CN113343953A (en) * | 2021-08-05 | 2021-09-03 | 南京信息工程大学 | FGR-AM method and system for remote sensing scene recognition |
CN113837353A (en) * | 2021-08-17 | 2021-12-24 | 中国地质大学(武汉) | Convolutional neural network feature fusion algorithm based on feature screening and deep fusion |
CN114972812A (en) * | 2022-06-02 | 2022-08-30 | 华侨大学 | Non-local attention learning method based on structural similarity |
CN115311186A (en) * | 2022-10-09 | 2022-11-08 | 济南和普威视光电技术有限公司 | Cross-scale attention confrontation fusion method for infrared and visible light images and terminal |
CN115311186B (en) * | 2022-10-09 | 2023-02-03 | 济南和普威视光电技术有限公司 | Cross-scale attention confrontation fusion method and terminal for infrared and visible light images |
CN115330658A (en) * | 2022-10-17 | 2022-11-11 | 中国科学技术大学 | Multi-exposure image fusion method, device, equipment and storage medium |
CN115330658B (en) * | 2022-10-17 | 2023-03-10 | 中国科学技术大学 | Multi-exposure image fusion method, device, equipment and storage medium |
CN117115442A (en) * | 2023-08-17 | 2023-11-24 | 浙江航天润博测控技术有限公司 | Semantic segmentation method based on visible light-infrared photoelectric reconnaissance image fusion |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113033630A (en) | Infrared and visible light image deep learning fusion method based on double non-local attention models | |
CN103020933B (en) | A kind of multisource image anastomosing method based on bionic visual mechanism | |
CN110189286B (en) | Infrared and visible light image fusion method based on ResNet | |
CN113139585B (en) | Infrared and visible light image fusion method based on unified multi-scale dense connection network | |
CN109410135B (en) | Anti-learning image defogging and fogging method | |
CN112967178B (en) | Image conversion method, device, equipment and storage medium | |
CN113283444B (en) | Heterogeneous image migration method based on generation countermeasure network | |
CN114548265B (en) | Crop leaf disease image generation model training method, crop leaf disease identification method, electronic equipment and storage medium | |
CN115713679A (en) | Target detection method based on multi-source information fusion, thermal infrared and three-dimensional depth map | |
CN115311186B (en) | Cross-scale attention confrontation fusion method and terminal for infrared and visible light images | |
CN115035003A (en) | Infrared and visible light image anti-fusion method for interactively compensating attention | |
CN115511767B (en) | Self-supervised learning multi-modal image fusion method and application thereof | |
CN104408697B (en) | Image Super-resolution Reconstruction method based on genetic algorithm and canonical prior model | |
CN116258936A (en) | Infrared and visible light image fusion method based on multi-scale features | |
CN112767243A (en) | Hyperspectral image super-resolution implementation method and system | |
CN114463172A (en) | Light field image super-resolution reconstruction method oriented to view consistency | |
CN115546442A (en) | Multi-view stereo matching reconstruction method and system based on perception consistency loss | |
CN116757986A (en) | Infrared and visible light image fusion method and device | |
CN113298744B (en) | End-to-end infrared and visible light image fusion method | |
CN114639002A (en) | Infrared and visible light image fusion method based on multi-mode characteristics | |
CN113011438B (en) | Bimodal image significance detection method based on node classification and sparse graph learning | |
CN117292244A (en) | Infrared and visible light image fusion method based on multilayer convolution | |
CN116051947A (en) | Multi-scale feature fusion method for multi-source heterogeneous data of transformer | |
CN113160104B (en) | Image fusion method based on densely connected network | |
CN113362281A (en) | Infrared and visible light image fusion method based on WSN-LatLRR |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20210625 |
|
WD01 | Invention patent application deemed withdrawn after publication |