Multispectral image fusion method based on interactive feature embedding
Technical Field
The invention belongs to the field of computer vision, and particularly relates to multispectral image fusion based on interactive feature embedding.
Background
The multispectral image fusion is to integrate the image characteristics of the same scene captured by the multispectral detector, so as to more comprehensively and accurately describe scene information. Multispectral image fusion is part of the image fusion task and has wide application in many areas, such as scene monitoring [1], target recognition, geological exploration, military and the like.
Deep learning techniques play an important role in image fusion. The existing image fusion method based on deep learning is mainly divided into two types: a convergence method based on a countermeasure network and a convergence method based on a non-countermeasure network. The fusion method based on the countermeasure network aims at fusing main features of a source image through designing a loss function in the countermeasure training process. However, this type of method has the following limitations: it is difficult for the network to optimize and to design a loss function that contains all the important information of the source image. In the fusion method based on the non-countermeasure network, the feature extraction process is often realized in an unsupervised mode, and the feature extraction is difficult to guarantee. Therefore, regardless of the counterlearning based on the loss function design or the unsupervised learning, ignoring any important information in the source image (such as gradient, edge, texture, intensity and contrast) will result in the loss of important features from the fusion result.
Therefore, the feature extraction capability of the network plays a key role in multi-source image fusion. In order to improve the network feature extraction capability, the invention provides an interactive feature-embedded multispectral image fusion network based on self-supervision learning, breaks through the technical bottleneck of comprehensively extracting the source image features in the existing fusion network, and has important significance for promoting more deep application of multispectral images in other fields.
Disclosure of Invention
The invention aims to improve the network feature extraction capability and provides a multispectral image fusion method based on interactive feature embedding.
The technical scheme of the invention is as follows:
a multispectral image fusion method based on interactive feature embedding comprises the following steps:
the method comprises the following steps: making a multi-spectral image fusion dataset
1) Acquiring a multispectral image dataset, a source image I1And a source image I2;
2) For the multispectral source image I in the step 1)1,I2Adjusted to a height and a width ofSo that;
3) for the source images I with the same size in the step 2)1,I2Sliding from left to right to obtain image blocks from top to bottom according to a window with a fixed size and step length;
4) turning over and mirroring the image pair obtained in the step 3), and enlarging the size of the training data set sample;
step two: designing an interactive feature-embedded multispectral image fusion network for self-supervision learning to realize multi-focus image fusion
1) Designing a self-supervision feature extraction module, wherein the module comprises two branches with the same structure; each branch consists of a plurality of convolution layers, and the parameter of convolution kernel of each layer is 3 x f, wherein f is the number of convolution kernels; the hierarchical feature extracted from the convolutional layer is represented by F'
m、F”
mM is denoted as the mth layer, ranging from {1, 2.., M }; the two branches input a source image I with width W and height H
1、I
2The output result is a source image reconstruction result
Loss function L of the module
1Expressed as:
where MSE represents the mean square error, I
nFor the source image I
1、I
2,
Representing a source image I
1、I
2Corresponding reconstructed result
And
2) designing an interactive feature embedding module, which is composed of a plurality of convolution layers, wherein the convolution kernel parameter of each layer is 3 x fWherein f is the number of convolution kernels; the hierarchy features extracted for the convolutional layer are denoted as Fm(ii) a Wherein the hierarchical features of the first layer are derived from the source image I1、I2Obtaining the hierarchical characteristics F from the second layer to the M layers after convolutionmHierarchical feature F 'extracted by self-supervised feature extraction module'm、F”mThe process expression obtained by the convolution operation is:
wherein, C2For 2 convolution operations, C44 convolution operations; cat represents concat operation; from the above formula, it can be observed that the layer of the intermediate layer and the feature FmIs a hierarchical feature F 'extracted by a self-supervised feature extraction module'm、F”mDerived therefrom, this ensures FmAnd F'm,F”mSharing low, medium and high-grade characteristics to further serve fusion tasks;
hierarchical feature F 'extracted by self-supervision feature extraction module on the other hand'm、F”mAlso derived from the hierarchical features FmFrom FmObtained after a convolution operation, expressed as:
F'm,F”m=C(Fm),M≥m≥1 (3)
in view of feature F 'for reconstructing the source image'm,F”mFrom FmThis also ensures FmThe method comprises the main characteristics of a source image, and further serves a fusion task;
3) outputting a fusion result; fusion result IfThe final output result weight W of the source image and the interactive feature embedding module is multiplied to obtain:
If=I1*W+I2*(1-W) (4)
wherein W is a weight map represented by FMObtained by a convolution operation:
W=C4(FM) (5)
wherein C is4Represents four convolution operations;
step three: network training, wherein the network training process is a process of optimizing a loss function; the self-supervision learning interactive feature embedded multispectral image fusion network loss function provided by the method consists of two parts: loss of self-supervised training, i.e. L1(ii) a Loss of fusion, i.e. Lf(ii) a Network training is the process of minimizing the loss function L,
L=L1+Lf (6)
in particular, LfIs a loss function based on SSIM;
step four: a testing stage; inputting two multispectral images I with width W and height H
1、I
2Output the corresponding reconstruction result
And final fusion result I
f。
The invention has the beneficial effects that: compared with the prior art, the invention has the following beneficial effects: the invention provides a multispectral image fusion method for self-supervision learning, which can effectively improve the network feature extraction capability through a self-supervision mechanism. The invention provides an interactive feature embedding structure which can be used as a bridge connection image fusion and reconstruction task, and can gradually embed key information acquired by self-supervision learning into the fusion task, so that the fusion performance is improved finally.
Drawings
FIG. 1 is a schematic diagram of the basic structure of the process of the present invention.
Fig. 2 is a schematic diagram of the fusion result of the present embodiment.
Detailed Description
The specific embodiment of the multispectral image fusion method based on interactive feature embedding is explained in detail as follows:
the method comprises the following steps: the multispectral image fusion data set production specifically comprises the following steps:
1) acquiring a multi-spectral image dataset, a source mapLike I1And a source image I2;
2) For the multispectral source image I in the step 1)1,I2Adjusting to be consistent in height and width;
3) for the source images I with the same size in the step 2)1,I2And sliding the image blocks from left to right from top to bottom in a window with a fixed size and step length.
4) Turning over and mirroring the image pair obtained in the step 3), and enlarging the size of the training data set sample;
step two: as shown in fig. 1, designing a multispectral image fusion network with interactive feature embedding for self-supervised learning to implement multispectral image fusion includes:
1) and designing a self-supervision characteristic extraction module. As shown in fig. 1, the module comprises two structurally identical branches. In this embodiment, each branch is composed of M (M ═ 3) convolution layers, each layer having convolution kernel parameters of 3 × f (f is the number of convolution kernels). The number of convolution kernels in the first layer is 64, the number of convolution kernels in the second layer is 128, and the number of convolution kernels in the third layer is 256. The hierarchical feature extracted from the convolutional layer is represented by F'
m,F”
m(m is denoted as the mth layer, ranging from {1,2,3 }). The two branches input a source image I with width W and height H
1、I
2The output result is a source image reconstruction result
Loss function L of the module
1Expressed as:
where MSE represents the mean square error, I
nFor the source image I
1、I
2,
Representing a source image I
1、I
2Corresponding reconstructed result
And
2) interactive feature embedding module design. As shown in fig. 1, in this embodiment, the module is composed of M +1(M ═ 3) convolutional layers, and the convolution kernel parameter of each layer is 3 × f (f is the number of convolution kernels). The number of convolution kernels in the first layer is 64, the number of convolution kernels in the second layer is 128, the number of convolution kernels in the third layer is 256, and the number of convolution kernels in the fourth layer is 1. The hierarchy features extracted for the convolutional layer are denoted as Fm. Wherein the hierarchical feature F of the first layer1From a source image I1、I2Obtaining the hierarchical characteristics F from the second layer to the M layers after convolutionmHierarchical feature F 'extracted by self-supervised feature extraction module'm,F”mThe process expression obtained by the convolution operation is:
wherein C is2For 2 convolution operations, C4Is 4 convolution operations. Cat represents the concat operation. From the above formula, it can be observed that the layer of the intermediate layer and the feature FmIs a hierarchical feature F 'extracted by a self-supervised feature extraction module'm,F”mDerived therefrom, this ensures FmCan be reacted with F'm,F”mSharing low, medium and high level features to serve fusion tasks.
Hierarchical feature F 'extracted by self-supervision feature extraction module on the other hand'm,F”mAlso derived from the hierarchical features FmFrom FmObtained after a convolution operation, expressed as:
F'm,F”m=C(Fm),M≥m≥1 (3)
in view of feature F 'for reconstructing the source image'm,F”mFrom FmThis also ensures FmThe method comprises the main characteristics of the source image, and further serves a fusion task. Thus, interact withThe self-supervision mechanism can be fully utilized by the formula characteristic embedding mechanism, so that important characteristics are prevented from being lost in the fusion result.
3) And outputting a fusion result. As shown in FIG. 1, fusion result IfThe final output result weight W of the source image and the interactive feature embedding module is multiplied to obtain:
If=I1*W+I2*(1-W) (4)
wherein W is a weight map represented by FMObtained by a convolution operation:
W=C4(FM) (5)
wherein C is4Representing four convolution operations.
Step three: and (5) network training. The network training process is a process that optimizes a loss function. The interactive feature embedded multispectral image fusion network loss function provided by the invention consists of two parts: loss of self-supervised training, i.e. L1(shown in formula 1); loss of fusion, i.e. Lf. Network training is the process of minimizing the loss function L,
L=L1+Lf (6)
in particular, LfIs a loss function based on SSIM.
The parameters in the network training process are set as follows:
base _ lr:1 e-4/learning rate
momentum of 0.9/momentum
weight _ decay:5 e-3/weight decay
batch size 1/batch size
solution _ mode GPU/example training Using GPU
Step four: and (5) a testing stage. Inputting two multispectral images I with width W and height H
1、I
2The model of the invention outputs its corresponding reconstructed result
And final fusion result I
f. As shown in fig. 2, compared to other fusion methodsThe fusion result obtained by the method can better retain the main characteristics in the source image, including the brightness characteristic and the texture characteristic.