CN114187221A - Infrared and visible light image fusion method based on adaptive weight learning - Google Patents

Infrared and visible light image fusion method based on adaptive weight learning Download PDF

Info

Publication number
CN114187221A
CN114187221A CN202111513619.9A CN202111513619A CN114187221A CN 114187221 A CN114187221 A CN 114187221A CN 202111513619 A CN202111513619 A CN 202111513619A CN 114187221 A CN114187221 A CN 114187221A
Authority
CN
China
Prior art keywords
image
visible light
infrared
discriminator
fusion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111513619.9A
Other languages
Chinese (zh)
Inventor
董毅
翟佳
彭实
陈�峰
郭单
谢晓丹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Environmental Features
Original Assignee
Beijing Institute of Environmental Features
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Environmental Features filed Critical Beijing Institute of Environmental Features
Priority to CN202111513619.9A priority Critical patent/CN114187221A/en
Publication of CN114187221A publication Critical patent/CN114187221A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/251Fusion techniques of input or preprocessed data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10048Infrared image

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to an infrared and visible light image fusion method based on adaptive weight learning, which is characterized in that a depth feature adaptive extraction module is constructed based on pixel level attention mechanism and fusion weight adaptive generation, and a generator network is constructed based on the depth feature adaptive extraction module in a cross-hierarchy cascade mode; based on twin network thought, a dual-channel discriminator network is established; performing primary connection operation on the infrared image and the visible light image, inputting the infrared image and the visible light image into a generator to generate a fused image, and finishing the fusion generation of the infrared image and the visible light image through game countermeasure of the generator and a discriminator; the invention integrates a pixel level attention mechanism and a level cascade thought, can enhance the network depth feature extraction capability under the condition of reducing the network parameter number, improves the fusion image generation quality of the infrared and visible light images, and realizes the fusion enhancement of the infrared image and the visible light image.

Description

Infrared and visible light image fusion method based on adaptive weight learning
Technical Field
The invention relates to the technical field of computer vision, in particular to an infrared and visible light image fusion method based on adaptive weight learning.
Background
The infrared image is obtained by imaging the thermal radiation characteristic of an object and is not influenced by illumination conditions, but the scene detail information capturing capability of the infrared image is not strong and the target edge information is fuzzy; the visible light image is obtained by imaging the reflection characteristic of an object, the detail texture information of a target can be captured, and the scene restoring capability is strong. The imaging information of the infrared image and the visible light image has high complementarity, so that the infrared image and the visible light image are fused, the defect of insufficient information acquisition capability of a single sensor can be overcome, the environmental adaptability is improved, and high-quality scene imaging information is obtained.
At present, the image fusion enhancement method based on infrared and visible light is mainly applied to the aspects of face recognition, target detection, target tracking, defect detection and the like. However, the infrared and visible light images have the problems of complex background environment, complex combination mode between targets and environments, and the like, which brings great challenges to image fusion, and a satisfactory image fusion effect cannot be obtained by means of an artificial intelligence technology; for example, the existing antagonistic neural network fusion-based algorithm has the problems that the utilization efficiency of the depth features is low, and the texture information of the fused image is not sufficiently restored.
Aiming at the defects, the invention provides an infrared and visible light image fusion method based on adaptive weight learning, which can enhance the network depth feature extraction capability under the condition of reducing the network parameter quantity and improve the fusion image generation quality of the infrared and visible light images.
Disclosure of Invention
The invention aims to solve the technical problems that the depth feature utilization efficiency of the existing image fusion algorithm based on the countermeasure type neural network is low, and the reduction of the detail information of the texture of the fused image is insufficient; the infrared and visible light image fusion method based on adaptive weight learning integrates a pixel level attention mechanism and a hierarchical cascade thought, can enhance the network depth feature extraction capability under the condition of reducing the network parameter quantity, and improves the fusion image generation quality of the infrared and visible light images.
In order to solve the technical problem, the invention provides an infrared and visible light image fusion method based on adaptive weight learning, which comprises the following steps: constructing a generator and a discriminator based on a generation countermeasure mechanism; the source data enters a generator network after connection operation, and a fusion image is generated through depth feature extraction and hierarchy feature fusion; the fused image enters a double-channel discriminator for discrimination; and outputting a fused image result through the counterstudy of the generator and the discriminator.
Preferably, the step of constructing the generator and the discriminator based on the generation countermeasure includes the following processes: and constructing a generator, namely constructing a depth feature self-adaptive extraction module by adopting a pixel-level attention mechanism and a fusion weight self-adaptive learning mechanism, and constructing a generator network by adopting a cross-layer cascade mode and a plurality of depth feature self-adaptive extraction modules.
Preferably, the depth feature adaptive extraction module comprises an attention subnetwork, a non-attention subnetwork and a fusion weight adaptive generation subnetwork.
Preferably, the attention subnetwork and the non-attention subnetwork extract and compress the image features, fuse the weight self-adaptive generation subnetwork to self-adaptively generate the weight factors, and perform channel recovery and adaptive feature reconstruction on the features output by the attention subnetwork and the non-attention subnetwork.
Preferably, the step of entering the generator network after the source data is subjected to connection operation and generating the fusion image through depth feature extraction and hierarchical feature fusion specifically comprises the following steps: performing connection operation on source data to obtain an input image; after extracting gradients from the input image, inputting the input image to a generator as a whole; the fused image is obtained through a generator network.
Preferably, the source data includes an infrared image and a visible light image.
Preferably, the step of constructing the generator and the discriminator based on the generation countermeasure includes the following processes: and (4) constructing a discriminator, namely constructing a double-path discriminator by adopting a twin network.
Preferably, the two path discriminators are a visible light discriminator and an infrared discriminator, respectively, the visible light discriminator is used for discriminating the fusion image and the visible light image, and the infrared discriminator is used for discriminating the fusion image and the infrared image.
Preferably, the loss function of the generator is LGThe function of the penalty is LadvGradient loss function of LgradThe loss function of structural similarity is LssimThen, there is,
LG=Ladv+Lgrad+Lssim
Figure BDA0003406076000000031
Figure BDA0003406076000000032
Lssim-1-SSIM,
in the formula, N represents the number of fused images,
Figure BDA0003406076000000033
represents the result of the classification of the fused image,
Figure BDA0003406076000000034
representing spurious data values, β, that the generator wishes the arbiter to trust1、β2Is a constant parameter, If、Iir、IvisRespectively representing a fused image, an infrared image, a visible light image, H and W are input image height and width,
Figure BDA0003406076000000035
respectively representing the gradient value of the fused image, the gradient value of the infrared image and the gradient value of the visible light image, wherein SSIM is a calculated value by adopting an image multi-scale structure similarity method.
Preferably, the loss function of the infrared discriminator is LD_irThe loss function of the visible light discriminator is LD_visThen, there are:
Figure BDA0003406076000000036
Figure BDA0003406076000000037
in the formula, N represents the number of fused images, a, b and c represent the truth value of the infrared image, the truth value of the fused image and the truth value of the visible light image respectively, and DIR(Iir) Indicating the result of discrimination of the infrared image by the infrared discriminator, DIR(If) Representing the result of discrimination of the fused image by the infrared discriminator, Dvis(Ivis) Indicating the result of discrimination of the visible light image by the visible light discriminator, Dvis(If) And showing the discrimination result of the visible light discriminator on the fused image.
The implementation of the infrared and visible light image fusion method based on the adaptive weight learning has the following beneficial effects: a depth feature self-adaptive extraction module is constructed on the basis of pixel level attention mechanism and fusion weight self-adaptive generation, and a generator network is constructed on the basis of the depth feature self-adaptive extraction module in a cross-hierarchy cascade mode; based on twin network thought, a dual-channel discriminator network is established; performing primary connection operation on the infrared image and the visible light image, inputting the infrared image and the visible light image into a generator to generate a fused image, and finishing the fusion generation of the infrared image and the visible light image through game countermeasure of the generator and a discriminator; the invention integrates a pixel level attention mechanism and a level cascade thought, can enhance the network depth feature extraction capability under the condition of reducing the network parameter number, improves the generation quality of the fusion image of the infrared image and the visible light image, realizes the fusion enhancement of the infrared image and the visible light image, and has great application value in the aspects of face recognition, target detection, target tracking, defect detection and the like.
Drawings
FIG. 1 is a flowchart of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention;
FIG. 2 is a network framework diagram of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention;
FIG. 3 is a structural framework diagram of a generator of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention;
FIG. 4 is a schematic structural diagram of a depth feature adaptive extraction module of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention;
FIG. 5 is a structural framework diagram of a discriminator of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention;
in the figure, 1: source data; 11: a visible light image; 12: infrared image; 2: a generator; 21: a depth feature adaptive extraction module; 211: an attention subnetwork; 212: a gratuitous subnetwork; 213: fusing weight self-adaptive generation sub-networks; 22: a cascade operation module; 23: a DCB module; 3: fusing the images; 41: a visible light discriminator; 42: an infrared discriminator.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, are within the scope of the present invention.
FIG. 1 is a flowchart of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention; as shown in fig. 1, the method for fusing infrared and visible light images based on adaptive weight learning according to the embodiment of the present invention includes the following steps; step S01: constructing a generator and a discriminator based on a generation countermeasure mechanism; step S02: the source data enters a generator network after connection operation, and a fusion image is generated through depth feature extraction and hierarchy feature fusion; step S03: the fused image enters a double-channel discriminator for discrimination; step S04: and outputting a fused image result through the counterstudy of the generator and the discriminator.
FIG. 2 is a network framework diagram of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention; as shown in fig. 2, the network structure of the infrared and visible light image fusion method based on adaptive weight learning is as follows: the generator 2 and the discriminator are constructed based on a generation countermeasure mechanism, wherein in the construction of the generator 2, a depth feature adaptive extraction module 21(DFAFE module) is constructed by adopting a pixel level attention mechanism and a fusion weight adaptive learning mechanism, and a generator network is constructed by adopting a cross-layer cascade mode through a cascade operation module 22 and a plurality of depth feature adaptive extraction modules 21. In the construction of the discriminator, a twin network is adopted to construct the two-way discriminator. The method comprises the steps that source data 1 (including an infrared image 12 and a visible light image 11) are connected firstly and then enter a generator network, a fused image 3 is generated through depth feature extraction and hierarchical feature fusion, then the fused image enters a dual-channel discriminator for discrimination, and finally fusion generation of high-quality infrared and visible light images 11 is achieved through counterstudy of a generator 2 and the discriminator.
In the infrared and visible light image fusion method based on adaptive weight learning provided by the embodiment of the present invention, the depth feature adaptive extraction module 21 includes three sub-networks: attention subnetwork 211, inattention subnetwork 212, and fusion weight adaptive generation subnetwork 213. The attention sub-network 211 and the non-attention sub-network 212 can extract image features and compress channels, the fusion weight adaptive generation sub-network 213 can adaptively generate weight factors, and can perform channel recovery and adaptive feature reconstruction on the features extracted by the attention sub-network 211 and the non-attention sub-network 212, so that the generalization capability of network representation is improved on the premise of only increasing few parameters. The infrared and visible light image fusion method based on adaptive weight learning provided by the embodiment of the invention adopts a hierarchical cascade mode, and a plurality of depth feature adaptive extraction modules 21 are connected in a cross-hierarchy mode to form a multi-hierarchy feature extraction network, so that the extraction and fusion operation of the hierarchy features is realized.
According to the infrared and visible light image fusion method based on adaptive weight learning provided by the embodiment of the invention, under the guidance of a countermeasure type neural network design framework, a depth feature adaptive extraction module 21 based on a pixel level attention mechanism and fusion weight adaptive generation is designed, image features are efficiently extracted, and then a plurality of depth feature adaptive extraction modules 21 are subjected to cross-layer cascade in a hierarchical cascade mode to further improve the depth feature extraction and hierarchical feature fusion capability so as to construct a generator network. Then, a two-way neural network-based construction discriminator set is constructed. And finally, connecting the infrared image 12 and the visible light image 11, inputting the connected images into a network, and finishing the fusion generation of the infrared image and the visible light image through game countermeasure of the generator 2 and the discriminator. Experimental results show that the fused image 3 generated by the method provided by the invention reserves more detailed information of the visible light image 11 while reserving the target information of the infrared image 12, and has better performance in both subjective evaluation and objective evaluation.
Fig. 3 is a structural framework diagram of the generator 2 of the infrared and visible light image fusion method based on adaptive weight learning according to the embodiment of the present invention, and in the infrared and visible light image fusion method based on adaptive weight learning according to the embodiment of the present invention, a generator network is constructed in a multilayer cascade manner with a depth feature adaptive extraction module 21 as a basic unit, so that hierarchical features can be effectively fused, and the overall feature characterization capability of the network is improved. The data flow of the generator network is that the infrared image 12 and the visible light image 11 are firstly connected to obtain an input image, then the connected input image is input into the generator 2 as a whole after gradient extraction, and then the fusion image 3 is obtained through a feature extraction network constructed by a depth feature adaptive extraction module 21 through cross-layer cascade. In this case, the DCB module 23 (dense connection module) in the network includes 2 convolutional layers, each of which is activated by using the ReLU function.
Fig. 4 is a schematic structural diagram of the depth feature adaptive extraction module 21 of the infrared and visible light image fusion method based on adaptive weight learning according to the embodiment of the present invention; in the infrared and visible light image fusion method based on adaptive weight learning provided by the embodiment of the present invention, the depth feature adaptive extraction module 21 is mainly used for efficiently extracting feature information, and is composed of an attention subnetwork 211, a non-attention subnetwork 212, and a fusion weight adaptive generation subnetwork 213. The attention subnetwork 211 is composed of two 3 × 3 convolutions and one 1 × 1 convolution, and a pixel level attention mechanism is added to the first 3 × 3 convolution, and the pixel level attention mechanism can assign corresponding weights to different channels, so that the network's ability to extract important information can be improved. The gratuitous subnetwork 212, in order to preserve the original information to the maximum, a 3 x 3 convolution is used for feature mapping and a 1 x 1 convolution is used for channel reconstruction.
In order to filter redundant information in the feature network and reduce the number of parameters, a fusion weight adaptive generation sub-network 213 is provided in the depth feature adaptive network. The network can dynamically assign weights to attention subnetwork 211 and inattentive subnetwork 212 through a self-learning process, thereby weakening the ineffective attention characteristics and enabling attention subnetwork 211 and inattentive subnetwork 212 to achieve adaptive balancing.
In addition, the feature x is inputn-1After passing through attention subnetwork 211 and non-attention subnetwork 212, respectively, the number of channels included in the features obtained at the respective outputs is reduced to half the number of input channels, x'n、x″n. And x'n、x″nAfter entering the fusion weight adaptive generation sub-network 213 as the input features, x 'is convolved by 1 × 1'n、x″nChannel lifting is carried out so that x'n、x″nNumber of channels and XnKept consistent then x'n、x″nAnd gamma1、γ2Corresponding multiplication and element addition are carried out, and finally, the input is carried out after 1 × 1 convolution. By means of the feature channel transformation and the self-adaptive reconstruction operation, the generalization capability of network representation is improved on the premise of only increasing few parameters, the depth feature self-adaptive extraction capability is improved, the capture of complex background and target detail information in infrared and visible light images is facilitated, and efficient high-quality fusion generation of the infrared and visible light images is achieved.
FIG. 5 is a structural framework diagram of a discriminator of an infrared and visible light image fusion method based on adaptive weight learning according to an embodiment of the present invention; according to the infrared and visible light image fusion method based on adaptive weight learning, the discriminator is used for forming a countermeasure game with the generator 2. The invention adopts a double-Discriminator structure, namely a visible light Discriminator 41(Discriminator-VIS) and an infrared Discriminator 42(Discriminator-IR), which are respectively used for distinguishing the fused image from the visible light and the infrared image. The discriminator adopts a twin network structure and shares network parameters. The single-branch network adopts a three-layer convolution structure, the step length is set to be 2, and scalar, namely two-classification results are output.
In the infrared and visible light image fusion method based on adaptive weight learning provided by the embodiment of the invention, the loss function of the generator 2 is LGThe function of the penalty is LadvGradient loss function of LgradThe loss function of structural similarity is LssimThen, there is,
LG=Ladv|Lgrad|Lssim
Figure BDA0003406076000000081
Figure BDA0003406076000000082
Lssim=1-SSIM,
in the formula, N represents the number of fused images,
Figure BDA0003406076000000083
represents the result of the classification of the fused image,
Figure BDA0003406076000000084
representing spurious data values, β, that the generator wishes the arbiter to trust1、β2Is a constant parameter, If、Iir、IvisRespectively representing a fused image, an infrared image, a visible light image, H and W are input image height and width,
Figure BDA0003406076000000085
respectively representing the gradient value of the fused image, the gradient value of the infrared image and the gradient value of the visible light image, wherein SSIM is a calculated value by adopting an image multi-scale structure similarity method. In order to enable the countergames between the generator and the discriminator to reach a balanced state and realize the mutual promotion of synchronous learning evolution between the generator and the discriminator, the embodiment of the invention provides a loss function L of a generator 2GComprises three parts, one is a function for resisting loss LadvThe fusion image generation device is used for controlling the countermeasure process between the generator and the discriminator so that the fusion image can obtain more infrared and visible light detail information; second is the gradient loss function LgradThe learning ability of the generator for the thermal radiation information of the infrared image and the texture detail information of the visible light image is controlled; third, the loss function L of structural similarityssimFor controlling the structural similarity of the fused image with the infrared and visible light images.
In the infrared and visible light image fusion method based on adaptive weight learning provided by the embodiment of the invention, the loss function of the infrared discriminator 42 is LD_irThe loss function of the visible light discriminator 41 is LD_visThen, there are:
Figure BDA0003406076000000091
Figure BDA0003406076000000092
in the formula, N represents the number of fused images, a, b and c represent the truth value of the infrared image, the truth value of the fused image and the truth value of the visible light image respectively, and DIR(Iir) Indicating the result of discrimination of the infrared image by the infrared discriminator, DIR(If) Representing the result of discrimination of the fused image by the infrared discriminator, Dvis(Ivis) Indicating the result of discrimination of the visible light image by the visible light discriminator, Dvis(If) And showing the discrimination result of the visible light discriminator on the fused image. The embodiment of the invention adopts a double discriminator design, and can effectively capture the thermal radiation information of the infrared image and the texture detail information of the visible light image.
The infrared and visible light image fusion method based on adaptive weight learning is adopted to carry out experiment verification on the infrared and visible light image fusion performance of the method, an image fusion public data set TNO is adopted in a verification experiment, about 20 groups of infrared and visible light image pairs are randomly extracted from the image fusion public data set TNO, and the data are expanded to 10200 pairs of images in a data augmentation mode. Wherein the training data set and the verification data set are constructed according to the ratio of 7: 3. In addition, 10 pairs of infrared and visible images were randomly extracted from the TNO dataset to construct a test dataset.
In the experiment, the contrast algorithm also adopts a neural network image fusion algorithm based on a countermeasure mechanism, namely FusionGAN (fusion generation countermeasure network), DDcGAN (condition generation countermeasure network) and fusionDN (fusion generation countermeasure improvement network).
In order to objectively evaluate the fused images obtained by the different fusion methods, 4 indexes of information Entropy (EN), Standard Deviation (SD), Mutual Information (MI) and multi-scale structure similarity (MS-SSIM) are selected for objectively evaluating, and the larger the EN is, the larger the information amount in the fused images is, the more image details are kept. The larger the SD, the higher the quality of the image and the clearer it. MI measures the similarity degree between the images, and the larger MI indicates that the fused image has more information of the source image, and the quality is better. MS-SSIM measures the similarity of an image to a source image.
The statistics of the experimental results are shown in the table, and it can be seen from the table that the infrared and visible light image fusion algorithm provided by the invention is improved in comparison with a comparison algorithm on multiple indexes, thereby illustrating the effectiveness of the algorithm provided by the invention.
Figure BDA0003406076000000101
In conclusion, by implementing the infrared and visible light image fusion method based on adaptive weight learning, the generated fusion image retains the target information of the infrared image and more detailed information of the visible light image, and obtains better performance in both subjective evaluation and objective evaluation.
Finally, it should be noted that: the above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (10)

1. An infrared and visible light image fusion method based on adaptive weight learning is characterized by comprising the following steps:
constructing a generator and a discriminator based on a generation countermeasure mechanism;
the source data enters a generator network after connection operation, and a fusion image is generated through depth feature extraction and hierarchy feature fusion;
the fused image enters a double-channel discriminator for discrimination;
and outputting a fused image result through the counterstudy of the generator and the discriminator.
2. The infrared and visible light image fusion method based on adaptive weight learning as claimed in claim 1, wherein the step of constructing the generator and the discriminator based on the generation countermeasure comprises the following processes: and constructing the generator, constructing a depth feature self-adaptive extraction module by adopting a pixel-level attention mechanism and a fusion weight self-adaptive learning mechanism, and constructing the generator network by adopting a plurality of depth feature self-adaptive extraction modules in a cross-layer cascade mode.
3. The infrared and visible light image fusion method based on adaptive weight learning of claim 2, wherein the depth feature adaptive extraction module comprises an attention subnetwork, a non-attention subnetwork and a fusion weight adaptive generation subnetwork.
4. The infrared and visible light image fusion method based on adaptive weight learning as claimed in claim 3, wherein the attention subnetwork and the inattentive subnetwork extract and channel compress image features, the fusion weight adaptive generation subnetwork adaptively generates weight factors, and performs channel restoration and adaptive feature reconstruction on features output by the attention subnetwork and the inattentive subnetwork.
5. The infrared and visible light image fusion method based on adaptive weight learning according to claim 3, wherein the step of connecting the source data, entering a generator network, and generating a fusion image through depth feature extraction and hierarchical feature fusion specifically comprises the following steps:
performing connection operation on the source data to obtain an input image;
inputting the input image into a generator as a whole after extracting gradients;
and obtaining a fused image through the generator network.
6. The infrared and visible light image fusion method based on adaptive weight learning according to any one of claims 1-5, characterized in that the source data comprises an infrared image and a visible light image.
7. The infrared and visible light image fusion method based on adaptive weight learning as claimed in claim 6, wherein the step of constructing the generator and the discriminator based on the generation countermeasure comprises the following processes: and constructing the discriminator, and constructing a double-path discriminator by adopting a twin network.
8. The infrared and visible light image fusion method based on adaptive weight learning of claim 7, wherein the two-way discriminator is a visible light discriminator and an infrared discriminator, the visible light discriminator is used for discriminating the fusion image from the visible light image, and the infrared discriminator is used for discriminating the fusion image from the infrared image.
9. The infrared and visible light image fusion method based on adaptive weight learning of claim 8, wherein the loss function of the generator is LGThe function of the penalty is LadvGradient loss function of LgradThe loss function of structural similarity is LssimThen, there is,
LG=Ladv|Lgrad|Lssim
Figure FDA0003406075990000021
Figure FDA0003406075990000022
Lssim=1-SSIM,
in the formula, N represents the number of fused images,
Figure FDA0003406075990000023
represents the result of the classification of the fused image,
Figure FDA0003406075990000024
representing spurious data values, β, that the generator wishes the arbiter to trust1、β2Is a constant parameter, If、Iir、IvisRespectively representing a fused image, an infrared image, a visible light image, H and W are input image height and width,
Figure FDA0003406075990000025
respectively representing the gradient value of the fused image, the gradient value of the infrared image and the gradient value of the visible light image, wherein SSIM is a calculated value by adopting an image multi-scale structure similarity method.
10. The infrared and visible light image fusion method based on adaptive weight learning of claim 8, wherein the loss function of the infrared discriminator is LD_irThe loss function of the visible light discriminator is LD_visThen, there are:
Figure FDA0003406075990000031
Figure FDA0003406075990000032
in the formula, N represents the number of fused images, a, b and c represent the truth value of the infrared image, the truth value of the fused image and the truth value of the visible light image respectively, and DIR(Iir) Indicating the result of discrimination of the infrared image by the infrared discriminator, DIR(If) Representing the result of discrimination of the fused image by the infrared discriminator, Dvis(Ivis) Indicating the result of discrimination of the visible light image by the visible light discriminator, Dvis(If) And showing the discrimination result of the visible light discriminator on the fused image.
CN202111513619.9A 2021-12-13 2021-12-13 Infrared and visible light image fusion method based on adaptive weight learning Pending CN114187221A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111513619.9A CN114187221A (en) 2021-12-13 2021-12-13 Infrared and visible light image fusion method based on adaptive weight learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111513619.9A CN114187221A (en) 2021-12-13 2021-12-13 Infrared and visible light image fusion method based on adaptive weight learning

Publications (1)

Publication Number Publication Date
CN114187221A true CN114187221A (en) 2022-03-15

Family

ID=80543335

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111513619.9A Pending CN114187221A (en) 2021-12-13 2021-12-13 Infrared and visible light image fusion method based on adaptive weight learning

Country Status (1)

Country Link
CN (1) CN114187221A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114898429A (en) * 2022-05-10 2022-08-12 电子科技大学 Thermal infrared-visible light cross-modal face recognition method
CN115393679A (en) * 2022-08-01 2022-11-25 国网江苏省电力有限公司南通供电分公司 RGB-infrared power transmission line defect image feature fusion method and system
CN116843588A (en) * 2023-06-20 2023-10-03 大连理工大学 Infrared and visible light image fusion method for target semantic hierarchy mining

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114898429A (en) * 2022-05-10 2022-08-12 电子科技大学 Thermal infrared-visible light cross-modal face recognition method
CN114898429B (en) * 2022-05-10 2023-05-30 电子科技大学 Thermal infrared-visible light cross-modal face recognition method
CN115393679A (en) * 2022-08-01 2022-11-25 国网江苏省电力有限公司南通供电分公司 RGB-infrared power transmission line defect image feature fusion method and system
CN116843588A (en) * 2023-06-20 2023-10-03 大连理工大学 Infrared and visible light image fusion method for target semantic hierarchy mining
CN116843588B (en) * 2023-06-20 2024-02-06 大连理工大学 Infrared and visible light image fusion method for target semantic hierarchy mining

Similar Documents

Publication Publication Date Title
CN114187221A (en) Infrared and visible light image fusion method based on adaptive weight learning
CN109543640B (en) Living body detection method based on image conversion
Do et al. Forensics face detection from GANs using convolutional neural network
CN112288663A (en) Infrared and visible light image fusion method and system
CN113592736B (en) Semi-supervised image deblurring method based on fused attention mechanism
CN111915525B (en) Low-illumination image enhancement method capable of generating countermeasure network based on improved depth separation
Li et al. Deep dehazing network with latent ensembling architecture and adversarial learning
CN103020933B (en) A kind of multisource image anastomosing method based on bionic visual mechanism
CN111179208B (en) Infrared-visible light image fusion method based on saliency map and convolutional neural network
CN106295501A (en) The degree of depth based on lip movement study personal identification method
Zhang et al. CNN cloud detection algorithm based on channel and spatial attention and probabilistic upsampling for remote sensing image
CN110082821A (en) A kind of no label frame microseism signal detecting method and device
CN111914758A (en) Face in-vivo detection method and device based on convolutional neural network
Zhu et al. Towards automatic wild animal detection in low quality camera-trap images using two-channeled perceiving residual pyramid networks
Wang et al. Suspect multifocus image fusion based on sparse denoising autoencoder neural network for police multimodal big data analysis
Te et al. Exploring hypergraph representation on face anti-spoofing beyond 2d attacks
Balamurali et al. Multiple parameter algorithm approach for adult image identification
Yang et al. Research on digital camouflage pattern generation algorithm based on adversarial autoencoder network
Khan et al. Efficient person re-identification for IoT-assisted cyber-physical systems
CN114708173A (en) Image fusion method, computer program product, storage medium, and electronic device
WO2021258284A1 (en) Edge processing data de-identification
Aparna Swarm intelligence for automatic video image contrast adjustment
CN116311434A (en) Face counterfeiting detection method and device, electronic equipment and storage medium
CN116091844A (en) Image data processing method and system based on edge calculation
Feng et al. Coal mine image dust and fog clearing algorithm based on deep learning network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination