CN115841438A - Infrared image and visible light image fusion method based on improved GAN network - Google Patents

Infrared image and visible light image fusion method based on improved GAN network Download PDF

Info

Publication number
CN115841438A
CN115841438A CN202211300180.6A CN202211300180A CN115841438A CN 115841438 A CN115841438 A CN 115841438A CN 202211300180 A CN202211300180 A CN 202211300180A CN 115841438 A CN115841438 A CN 115841438A
Authority
CN
China
Prior art keywords
image
layer
fusion
detail
visible light
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211300180.6A
Other languages
Chinese (zh)
Inventor
吴杰
高策
余毅
张艳超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Changchun Institute of Optics Fine Mechanics and Physics of CAS
Original Assignee
Changchun Institute of Optics Fine Mechanics and Physics of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Changchun Institute of Optics Fine Mechanics and Physics of CAS filed Critical Changchun Institute of Optics Fine Mechanics and Physics of CAS
Priority to CN202211300180.6A priority Critical patent/CN115841438A/en
Publication of CN115841438A publication Critical patent/CN115841438A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Image Processing (AREA)

Abstract

The invention relates to an infrared image and visible light image fusion method based on an improved GAN network, which comprises the steps of firstly decomposing an infrared image and a visible light image into a basic layer image and a detail layer image through a guide filter respectively; then, enhancing the basic layer image of the visible light image in a histogram mapping mode, and fusing the basic layer image by using an improved LT algorithm; secondly, extracting characteristic information of a source image by combining a generator and an encoder in an improved GAN algorithm, and guiding the fusion of the detail layer images of the infrared and visible light images in a counterstudy mode; and finally, setting relevant parameters for the fusion image of the basic layer and the fusion image of the detail layer, and obtaining the final fusion image by adopting a weighted average method. The method effectively avoids the defect that the traditional fusion method excessively depends on complex fusion rules, and retains the global structural features and local textures of the image, so that the image has natural impression.

Description

Infrared image and visible light image fusion method based on improved GAN network
Technical Field
The invention relates to the technical field of image processing, in particular to an infrared image and visible light image fusion method based on an improved GAN network.
Background
Infrared and visible image fusion are the core technologies in multi-sensor image fusion. The infrared image is a thermal radiation image of an object, can generate clear foreground contour information under the influence of environmental factors such as illumination and the like, and has the defect of low resolution. The visible light image has the advantages of high resolution and rich target texture details, and the defect that the visible light image is easily influenced by environmental factors such as haze, illumination and the like. The two kinds of images can be fused to effectively concentrate advantages and make up respective adverse factors to generate clear images with large information quantity, so that the technology has wide application in the fields of target detection, object identification, military detection and the like.
The image fusion is an information technology for fusing the characteristics of details, outlines and the like of different images into a new image. The infrared and visible light fused image not only retains the clear texture details of the visible light image, but also has the anti-interference characteristic of the infrared image. Image fusion methods are roughly classified into three types. One is the conventional method using transform domain, which mainly includes laplacian pyramid fusion method and non-subsampled Contourlet transform (NSCT). The algorithm analyzes the image by a filter according to the characteristics of the image so as to realize fusion, and has the defect of over-dependence on complex fusion rules. And the other is a fusion method based on a spatial domain, which mainly comprises a principal component analysis algorithm and has the defect of larger operation amount by reducing dimension of the source image information of the data set. And thirdly, a fusion mode based on deep learning is mainly combined with a convolutional neural network and image processing to better extract the image depth characteristics and prepare for a subsequent fusion stage. Beginning in 2018, GAN networks are gradually applied to the field of image fusion, and the method is roughly divided into two stages: in the network stage, the images to be fused are used as a training set to train the network, the network comprises a generator and a discriminator and is used for extracting features and guiding to generate the images, and the images are fused through the trained model. The GAN algorithm solves the problem that the fusion rule of the traditional algorithm is complex, however, the result of the antagonistic generation stage has uncertainty due to random network initialization parameters, and the extracted image features have a fuzzy effect, so that noise and artifacts exist in the fused image.
Disclosure of Invention
In order to solve the problems in the prior art, the invention provides an infrared image and visible light image fusion method based on an improved GAN network, which can increase the detail texture information of an image after the infrared image and the visible light image are fused, effectively reduce artifacts and noise and keep high contrast.
In order to achieve the purpose, the invention adopts the following technical scheme:
an infrared image and visible light image fusion method based on an improved GAN network comprises the following steps:
(1) Respectively solving infrared images I by utilizing guide filters 1 Base layer image of
Figure BDA0003904276730000021
And a visible light image I 2 Based on the base layer image->
Figure BDA0003904276730000022
And respectively calculate to obtain the corresponding detail layer image->
Figure BDA0003904276730000023
And detail layer image->
Figure BDA0003904276730000024
(2) Mapping base layer image by histogram
Figure BDA0003904276730000025
Performing enhancement to obtain an enhanced base layer image->
Figure BDA0003904276730000026
The base layer image is then based on the improved LT algorithm>
Figure BDA0003904276730000027
And enhancementBase layer image->
Figure BDA0003904276730000028
Fusing to obtain a base layer fused image F b And ^ ing on the detail layer image>
Figure BDA0003904276730000029
And detail layer image->
Figure BDA00039042767300000210
Performing fusion model training on the improved GAN network as a training set to obtain a detail layer fusion image F d Wherein the improved GAN network comprises a generator, an encoder, D z A discriminator and a target discriminator, the encoder will input the detail layer image->
Figure BDA00039042767300000211
Mapping to a low-dimensional feature vector z through convolution operation, feeding the low-dimensional feature vector z and a label vector l into a generator, mapping the low-dimensional image to a high-dimensional image in the generator in a mode of 2 deconvolution, connecting the label vector l to the low-dimensional feature vector z, and obtaining a new vector [ z, l ]]Feedback to the generator and output the detail layer fusion image F d Pick up the detail layer image->
Figure BDA00039042767300000212
As target image, image->
Figure BDA00039042767300000213
And the detail layer image->
Figure BDA00039042767300000214
Between D z Discriminator, said D z The discriminator is used for forcing the distribution of the generated low-dimensional feature vector z to be gradually close to the prior and the image is/is based on the detail level>
Figure BDA00039042767300000215
Fusing image F with detail layer d Inter-construction object discriminationA target discriminator for fusing the detail layer into an image F d And detail layer image>
Figure BDA00039042767300000216
Performing antagonistic learning;
(3) Fusing base layer to image F b Fused image F with detail layer d And carrying out weighted addition on the corresponding pixel points to obtain a final fusion image.
The invention provides an infrared image and visible light image fusion method based on an improved GAN Network, which is a layered image fusion method based on an improved generation type countermeasure Network (GAN) and Laplace Transformation (LT), wherein the method comprises the steps of firstly decomposing an infrared image and a visible light image into a basic layer image and a detail layer image through a guide filter respectively; then, enhancement processing is carried out on the basic layer image of the visible light image in a histogram mapping mode, the contour effect after fusion is improved, and an improved LT algorithm is used for fusing the basic layer image; secondly, extracting characteristic information of a source image by combining a generator and an encoder in an improved GAN algorithm, and guiding the fusion of detail layer images of infrared and visible light images in a counterstudy mode to make the fused image detail information richer; and finally, setting relevant parameters for the fusion image of the basic layer and the fusion image of the detail layer, and obtaining the final fusion image by adopting a weighted average method. The method effectively avoids the defect that the traditional fusion method excessively depends on complex fusion rules, and retains the global structural features and local textures of the image, so that the image has natural impression. The invention has the following beneficial effects:
1) By utilizing the end-to-end characteristic of the GAN network, the traditional registration method is prevented from excessively depending on manual design weight distribution and complex fusion rules, so that the algorithm performance is improved;
2) Enhancing the basic layer image of the low-illumination visible light image in a histogram mapping mode, and fusing the basic layer image of the low-illumination visible light image with the basic layer image of the infrared image by an improved LT algorithm to ensure that the fused image has good contrast and overall appearance;
3) The target image and the generated image are identified according to the double discriminators and the encoder, and the guidance of the generated image is realized by restricting network parameters in a low-rank prior decomposition mode, so that the anti-interference capability is improved, and the stability and the continuity are better;
4) The generator is combined with the depth migration module to enhance the extraction of the detail information of the image, the fault tolerance capability is strong, the information loss can be well prevented, and the noise and the artifacts of the fused image are reduced.
Drawings
FIG. 1 is a schematic diagram of an infrared image and visible light image fusion method based on an improved GAN network according to the present invention;
FIG. 2 is a schematic diagram of a histogram mapping algorithm;
FIG. 3 is a base layer image
Figure BDA0003904276730000047
And enhancing the base layer image->
Figure BDA0003904276730000048
A fusion schematic;
fig. 4 is a block diagram of an improved GAN network;
FIG. 5 is a diagram of a densely connected network architecture for a depth feature migration module;
fig. 6 is a schematic view of fused image reconstruction.
Detailed Description
The technical solution of the present invention will be described in detail with reference to the accompanying drawings and preferred embodiments.
The invention provides an infrared image and visible light image fusion method based on an improved GAN network, as shown in figure 1, the method comprises the following steps:
(1) Guided filtering based image decomposition
The invention relates to a novel visible light and infrared image fusion method, and the fusion principle is shown in figure 1. Firstly, respectively solving an infrared image I by utilizing a guide filter 1 Base layer image of
Figure BDA0003904276730000041
And a visible light image I 2 Based on the base layer image->
Figure BDA0003904276730000042
And respectively calculate to obtain the corresponding detail layer image->
Figure BDA0003904276730000043
And detail layer image->
Figure BDA0003904276730000044
Compared with the traditional method, the multi-scale decomposition analysis mode adopting the guide filter can separate the overlapped features in the space. Let I k (k =1,2) is the input image, then I 1 And I 2 Respectively infrared image and visible image, for each input image I k By solving equation (1), the corresponding base layer image ≥ can be obtained>
Figure BDA0003904276730000045
Equation (1) is as follows:
Figure BDA0003904276730000046
wherein, g x =[-1,1]Representing horizontal gradient operator, g y =[-1,1] T Denotes the vertical gradient operator and λ denotes the scale factor, in the above equation the λ parameter is set to 5.
Detail layer image
Figure BDA0003904276730000051
Subtracting the base layer image part from the source image, and calculating according to the following formula:
Figure BDA0003904276730000052
wherein the base layer image contains large-scale contour information and the detail layer image contains small-scale texture information and edge information.
(2) Base layer image fusion and detail layer image fusion
When the base layer image fusion is carried out, the base layer image of the visible light image is firstly processed
Figure BDA0003904276730000053
Increasing contrast in low-light images using a sub-histogram mapping algorithm resulting in an enhanced base layer image +>
Figure BDA0003904276730000054
Then carries out->
Figure BDA0003904276730000055
And
Figure BDA0003904276730000056
the fusion of (1). The histogram mapping algorithm is to divide the image S into m × n subintervals by using local minimum values, S = [ S ] 1 ,s 2 ,…,s m×n ]Enhancing the reconstructed image T, T = [ T ] for the base image by the subinterval histogram 1 ,t 2 ,…,t m×n ]As shown in fig. 2. The invention uses the local entropy to control the mapping range of the sub-histogram, thereby avoiding the generation of excessive histogram peaks in the traditional method, wherein the local entropy is used for measuring the texture richness of the image block, if the local entropy is smaller, the smaller the local entropy is, the less the texture is represented, and the remapping range is enlarged. The mapping range controlled by the local entropy is:
Figure BDA0003904276730000057
wherein, gamma is j ∈[0,0.8]For the purpose of normalization, the method is used,
Figure BDA0003904276730000058
is a sub-histogram interval [ m j ,m j+1 ]Wherein p () represents a probability density and u represents a sub-histogram bin.
Histogram areaWhen inter-grain is small, R j More mapping intervals are allocated. Thus, the resulting enhanced base layer image after the final enhancement process
Figure BDA0003904276730000059
The calculation formula is as follows:
Figure BDA00039042767300000510
wherein R is t The local entropy for each partition j.
Enhancement of base layer images by improving LT algorithm (i.e. fusion rule of Laplacian pyramid transform)
Figure BDA0003904276730000061
And base layer image>
Figure BDA0003904276730000062
The process of performing fusion includes the steps of:
as shown in fig. 3, the base layer image will be enhanced
Figure BDA0003904276730000063
And the base layer image->
Figure BDA0003904276730000064
Respectively decomposed into->
Figure BDA0003904276730000065
And &>
Figure BDA0003904276730000066
Where l.epsilon. (1,2,3,4). Wherein a region average gradient is calculated over a top-level image by means of an m x n matrix of the region size centered on each pixel in the image>
Figure BDA0003904276730000067
And &>
Figure BDA0003904276730000068
The fusion is carried out, and the formula is as follows:
Figure BDA0003904276730000069
wherein, delta I x And Δ I y The first order differences of the pixel f (x, y) on the x-axis and the y-axis are respectively represented by G 1 (i, j) and G 2 (i, j) represents the average gradient of the region of each pixel point at the top layer, which can reflect the definition of the image, and the top layer image fusion can be represented as:
Figure BDA00039042767300000610
bottom three-layer LT fusion requires calculation of regional energy of each layer of infrared and visible light images
Figure BDA00039042767300000611
The calculation formula is as follows:
Figure BDA00039042767300000612
wherein p and q are 1,w which is a 3 × 3 matrix, when l is greater than 0 and less than 4, the fusion result of the LT image pyramid of l layers is shown as the following formula:
Figure BDA00039042767300000613
based on a fusion algorithm of the infrared image and the visible light image of the base layer of the improved LT, the enhancement processing is carried out on the base layer of the low-light visible light image in a histogram mapping mode, and the low-light visible light image is fused with the base layer of the infrared image through the improved LT algorithm, so that the fused image has good contrast and overall appearance.
When the detail layer image fusion is carried out, the detail layer image of the infrared image is used
Figure BDA00039042767300000614
And a detail layer image of the visible light image->
Figure BDA00039042767300000615
Performing fusion model training on the improved GAN network as a training set, optimizing and guiding the image fusion process in a counterstudy mode to obtain a detail layer fusion image F d . Wherein the improved GAN network comprises a generator, an encoder and two discriminators, as shown in FIG. 4, wherein the encoder comprises five convolution layers and a BN layer, the generator comprises three dense connection modules and three deconvolution modules, D z The discriminator and the target discriminator are respectively composed of a convolution layer, a BN layer, a Leaky Relu layer and a full connection layer. The improved GAN network of the present invention differs from the traditional generation of countermeasure networks GAN in two ways: firstly, a discriminator and an encoder are added for providing a prior optimization model to guide the generation of a fusion image; and secondly, a depth feature migration module is added in the generator network, so that the network can better extract the depth features of the source image and obtain the best fusion effect. Compared with the traditional GAN, an encoder and a discriminator are added, the fusion algorithm of the infrared image and the visible light image of the detail layer based on the improved GAN restrains network parameters in a low-rank prior decomposition mode, guides the generated fusion image to be closer to a target image, designs a loss function of a network and ensures the stability of a network model.
The purpose of the encoder is to reduce the dimension of the source image while providing a priori initialization constraint model. Detail layer image
Figure BDA0003904276730000071
And the input x is sent into a newly added encoder E as an input x, the encoder E maps the input x to a low-dimensional feature vector z through convolution operation, the low-dimensional feature vector z is spliced with a label vector o, and network parameters are constrained in a low-rank prior decomposition mode. Wherein:
E(x)=z∈R n (9)
wherein R is n Representing an n-dimensional space, the output preserves the texture features of the input x.
The generator aims to extract more detail information in a source image, generate a fusion image with rich textures, and introduce a depth feature migration module to realize feature migration and dimension ascending of a low-dimensional feature vector z, as shown in fig. 3. The depth feature migration module extracts features of real semantic information in a dense connection mode, and assumes that the network has L layers and x layers in common 0 As input to the network, x l For output at layer I in the network, x l-1 Is the output of layer l-1, H l () For the nonlinear transformation acting on the l-th layer, the dense connection network structure of the depth feature migration module is shown in fig. 5, where Input represents the Input of an image, BN-ReLU-Conv represents a regularization layer, an activation layer and a convolution layer which are connected in sequence, and the relationship is shown in formula (10). The depth feature migration module integrates the features of each stage in a dense connection mode, and adds the output features of the previous layer to each input end, so that the dimension expansion is achieved. Mapping the low-dimensional image into a high-dimensional image in a generator by using 2 deconvolution modes, connecting a label vector l to a low-dimensional feature vector z, and connecting a new vector [ z, l]Fed back to the generator to output a detail layer fusion image F d Is formula (11):
x l =H l ([x 0 ,x 1 ,...,x l-1 ]) (10)
F d =G(z,l)=G(E(x),l) (11)
combining with the generator network design of the depth feature migration module, integrating the features of each stage in a jump connection mode, adding the output features of the previous layer at each input end, and mapping the low-dimensional image into a high-dimensional image by combining with deconvolution, so that the fused image has richer detail information, and the artifacts and the noise are effectively reduced.
The purpose of the discriminator is to distinguish the generated image from the target image and to transfer information to the fusion process of the generator. Detail layer image with richer texture features
Figure BDA0003904276730000081
As target images, in each case a detail layer image->
Figure BDA0003904276730000082
And the detail layer image->
Figure BDA0003904276730000083
Between D z Discriminator and image on detail layer->
Figure BDA0003904276730000084
Fusing image F with detail layer d And constructing a target discriminator. Wherein D z The discriminator is used to force the distribution of the generated low-dimensional feature vector z to be progressively closer to a priori. The object discriminator is used to fuse the detail layers into an image F d And the detail layer image->
Figure BDA0003904276730000085
Performing counterstudy to fuse detail layers into an image F d And is more true. Still see FIG. 4,D z The discriminator and the target discriminator both comprise three convolution modules, a normalization layer, an activation function layer and a full-link layer, wherein the normalization layer uses a BN layer to normalize data, the activation function layer takes Leaky Relu as an activation function, and the full-link layer predicts the data.
The training loss function of encoder E and generator G in the present invention is as follows:
Figure BDA0003904276730000086
/>
where L () represents a norm, and x is an input, i.e., a detail layer image
Figure BDA0003904276730000087
G () and E () represent the generator output and the encoder output, respectively, and l is the label vector.
D z Loss function of discriminator
Figure BDA0003904276730000088
And the loss function L of the target discriminator D D Respectively as follows:
Figure BDA0003904276730000089
Figure BDA0003904276730000091
wherein b and c represent detail layer images, respectively
Figure BDA0003904276730000092
And detail layer image->
Figure BDA0003904276730000093
The real label of (1), the value ranges are 0.4-0.6 and 0.7-1, respectively; d represents a true label of the fused detail layer, and the value thereof ranges from 0 to 0.3; n represents the number of fused images; />
Figure BDA0003904276730000094
And &>
Figure BDA0003904276730000095
Respectively represent->
Figure BDA0003904276730000096
And F d The classification result of (1).
(3) Fused image reconstruction
Through the above steps (1) and (2), the base layer fusion image F can be obtained b Fused image F with detail layer d Fusing the image F with the base layer b Fused image F with detail layer d To obtain the final fused image. Fusing base layer to image F b Fused image F with detail layer d The corresponding pixel points in (b) are weighted and added as shown in formula (14), so as to obtain a final fusion image, as shown in fig. 6:
F(x,y)=αF b (x,y)+βF d (x,y) (14)
wherein, (x, y) represent the pixel coordinates of the fusion image, wherein α is the fusion parameter of the base layer, β is the fusion parameter of the detail layer, α =0.6, and β =0.4 in this patent.
The technical features of the embodiments described above may be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the embodiments described above are not described, but should be considered as being within the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
The above-mentioned embodiments only express several embodiments of the present invention, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the inventive concept, which falls within the scope of the present invention. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (6)

1. An infrared image and visible light image fusion method based on an improved GAN network is characterized by comprising the following steps:
(1) Respectively solving infrared images I by utilizing guide filters 1 Base layer image of
Figure FDA0003904276720000011
And a visible light image I 2 Base layer image of
Figure FDA0003904276720000012
And respectively calculating to obtain corresponding detail layer images
Figure FDA0003904276720000013
And detail layer image
Figure FDA0003904276720000014
(2) Mapping base layer image by histogram
Figure FDA0003904276720000015
Enhancing to obtain enhanced base layer image
Figure FDA0003904276720000016
Enhancement of the base layer image by improving the LT algorithm
Figure FDA0003904276720000017
And base layer image
Figure FDA0003904276720000018
Fusing to obtain a base layer fused image F b And with detail layer images
Figure FDA0003904276720000019
And detail layer image
Figure FDA00039042767200000110
Performing fusion model training on the improved GAN network as a training set to obtain a detail layer fusion image F d Wherein the improved GAN network comprises a generator, an encoder, D z Discriminator and object discriminator, encoder inputting detail layer image
Figure FDA00039042767200000111
Mapping to a low-dimensional feature vector z through convolution operation, feeding the low-dimensional feature vector z and a label vector l into a generator, mapping the low-dimensional image to a high-dimensional image in the generator in a mode of 2 deconvolution, connecting the label vector l to the low-dimensional feature vector z, and obtaining a new vector [ z, l ]]Feedback to the generator and output the detail layer fusion image F d Image of detail layer
Figure FDA00039042767200000112
As the target image, in the detail layer image
Figure FDA00039042767200000113
With detail layer images
Figure FDA00039042767200000114
Between D z Discriminator, said D z The discriminator is used for forcing the distribution of the generated low-dimensional feature vector z to be gradually close to the prior and the image at the detail layer
Figure FDA00039042767200000115
Fusing image F with detail layer d A target discriminator is constructed for fusing the detail layers into an image F d With detail layer images
Figure FDA00039042767200000116
Performing antagonistic learning;
(3) Fusing base layer to image F b Fused image F with detail layer d And carrying out weighted addition on the corresponding pixel points to obtain a final fusion image.
2. The method as claimed in claim 1, wherein the histogram mapping algorithm is used to blend the base layer image with the infrared image based on the improved GAN network
Figure FDA00039042767200000117
When enhancement is carried out, the local entropy is used for controlling the mapping range of the sub-histogram, and the mapping range controlled by the local entropy is as follows:
Figure FDA00039042767200000118
wherein, γ j ∈[0,0.8]For the purpose of normalization, the method is used,
Figure FDA0003904276720000021
is a sub-histogram interval [ m j ,m j+1 ]Wherein p () represents a probability density and u represents a sub-histogram bin;
enhanced base layer image obtained after enhancement processing
Figure FDA0003904276720000022
The calculation formula is as follows:
Figure FDA0003904276720000023
wherein R is t The local entropy for each partition j.
3. The method for fusing infrared image and visible image based on improved GAN network as claimed in claim 1, wherein the enhancement of the base layer image is performed by improving LT algorithm
Figure FDA0003904276720000024
And base layer image
Figure FDA0003904276720000025
When fusion is carried out, the basic layer image is enhanced
Figure FDA0003904276720000026
And base layer image
Figure FDA0003904276720000027
Respectively decomposing the L-layer pyramid into
Figure FDA0003904276720000028
And
Figure FDA0003904276720000029
wherein l is ∈ (1,2,3,4); calculating the average gradient of the region by using an m multiplied by n matrix of the region size taking each pixel in the image as the center, and comparing the average gradient with the top layer image
Figure FDA00039042767200000210
And
Figure FDA00039042767200000211
the fusion is carried out, and the formula is as follows:
Figure FDA00039042767200000212
wherein, delta I x And Δ I y The first order differences of the pixel f (x, y) on the x-axis and the y-axis are respectively represented by G 1 (i, j) and G 2 (i, j) represents the regional average gradient of each pixel point at the top level, and the top-level image fusion can be represented as:
Figure FDA00039042767200000213
bottom three-layer LT fusion requires calculation of regional energy of each layer of infrared and visible light images
Figure FDA00039042767200000214
The calculation formula is as follows:
Figure FDA00039042767200000215
wherein p and q are 1,w which is a 3 × 3 matrix, when l is greater than 0 and less than 4, the fusion result of the LT image pyramid of l layers is shown as the following formula:
Figure FDA0003904276720000031
4. the method as claimed in claim 1, wherein the improved GAN network incorporates a depth feature migration module, the depth feature migration module integrates features of each stage by adopting a jump connection manner, adds output features of a previous layer to each input end, and maps a low-dimensional image into a high-dimensional image by combining deconvolution.
5. The method for fusing infrared image and visible light image based on improved GAN network as claimed in claim 1, wherein D is z The discriminator and the target discriminator both comprise three convolution modules, a BatchNorm normalization layer, a Leaky Relu activation function layer and a full link layer.
6. The method for fusing infrared images and visible light images based on the improved GAN network as claimed in claim 1, wherein the training loss function of the encoder (E) and the generator (G) is as follows:
Figure FDA0003904276720000032
where L () represents a norm, and x is an input, i.e., a detail layer image
Figure FDA0003904276720000033
G () and E () represent the generator output and the encoder output, respectively, l being the label vector;
D z loss function of discriminator
Figure FDA0003904276720000034
And the loss function L of the target discriminator D D Respectively as follows:
Figure FDA0003904276720000035
Figure FDA0003904276720000036
wherein b and c represent detail layer images
Figure FDA0003904276720000037
And detail layer image
Figure FDA0003904276720000038
The real label of (1), the value ranges are 0.4-0.6 and 0.7-1, respectively; d represents a true label of the fused detail layer, and the value thereof ranges from 0 to 0.3; n represents the number of fused images;
Figure FDA0003904276720000039
and
Figure FDA00039042767200000310
respectively representing detail layer images
Figure FDA00039042767200000311
Detail layer image
Figure FDA00039042767200000312
Fused image F with detail layer d The classification result of (1).
CN202211300180.6A 2022-10-24 2022-10-24 Infrared image and visible light image fusion method based on improved GAN network Pending CN115841438A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211300180.6A CN115841438A (en) 2022-10-24 2022-10-24 Infrared image and visible light image fusion method based on improved GAN network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211300180.6A CN115841438A (en) 2022-10-24 2022-10-24 Infrared image and visible light image fusion method based on improved GAN network

Publications (1)

Publication Number Publication Date
CN115841438A true CN115841438A (en) 2023-03-24

Family

ID=85576426

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211300180.6A Pending CN115841438A (en) 2022-10-24 2022-10-24 Infrared image and visible light image fusion method based on improved GAN network

Country Status (1)

Country Link
CN (1) CN115841438A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116152120A (en) * 2023-04-20 2023-05-23 南京大学 Low-light image enhancement method and device integrating high-low frequency characteristic information

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116152120A (en) * 2023-04-20 2023-05-23 南京大学 Low-light image enhancement method and device integrating high-low frequency characteristic information

Similar Documents

Publication Publication Date Title
CN111462126B (en) Semantic image segmentation method and system based on edge enhancement
CN110109060A (en) A kind of radar emitter signal method for separating and system based on deep learning network
CN111539247B (en) Hyper-spectrum face recognition method and device, electronic equipment and storage medium thereof
CN112967178B (en) Image conversion method, device, equipment and storage medium
CN109684922A (en) A kind of recognition methods based on the multi-model of convolutional neural networks to finished product dish
CN113160062B (en) Infrared image target detection method, device, equipment and storage medium
Jian et al. Infrared and visible image fusion based on deep decomposition network and saliency analysis
CN110852199A (en) Foreground extraction method based on double-frame coding and decoding model
CN112669249A (en) Infrared and visible light image fusion method combining improved NSCT (non-subsampled Contourlet transform) transformation and deep learning
CN114782298A (en) Infrared and visible light image fusion method with regional attention
CN112884758A (en) Defective insulator sample generation method and system based on style migration method
CN115223017B (en) Multi-scale feature fusion bridge detection method based on depth separable convolution
CN115937774A (en) Security inspection contraband detection method based on feature fusion and semantic interaction
CN115841438A (en) Infrared image and visible light image fusion method based on improved GAN network
CN110148083B (en) Image fusion method based on rapid BEMD and deep learning
Zhou et al. MSAR‐DefogNet: Lightweight cloud removal network for high resolution remote sensing images based on multi scale convolution
CN108764287B (en) Target detection method and system based on deep learning and packet convolution
Qu et al. Low illumination enhancement for object detection in self-driving
CN114387195A (en) Infrared image and visible light image fusion method based on non-global pre-enhancement
CN112330562B (en) Heterogeneous remote sensing image transformation method and system
CN113627481A (en) Multi-model combined unmanned aerial vehicle garbage classification method for smart gardens
CN107529647B (en) Cloud picture cloud amount calculation method based on multilayer unsupervised sparse learning network
CN112686830A (en) Super-resolution method of single depth map based on image decomposition
CN110852255A (en) Traffic target detection method based on U-shaped characteristic pyramid
CN116128766A (en) Improved Retinex-Net-based infrared image enhancement method for power equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination