CN114897751A - Infrared and visible light image perception fusion method based on multi-scale structural decomposition - Google Patents

Infrared and visible light image perception fusion method based on multi-scale structural decomposition Download PDF

Info

Publication number
CN114897751A
CN114897751A CN202210381391.0A CN202210381391A CN114897751A CN 114897751 A CN114897751 A CN 114897751A CN 202210381391 A CN202210381391 A CN 202210381391A CN 114897751 A CN114897751 A CN 114897751A
Authority
CN
China
Prior art keywords
image
infrared
scale
visible light
contrast
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210381391.0A
Other languages
Chinese (zh)
Inventor
周志强
费二芳
缪玲娟
崔赛佳
叶何
李家琪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Original Assignee
Beijing Institute of Technology BIT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT filed Critical Beijing Institute of Technology BIT
Priority to CN202210381391.0A priority Critical patent/CN114897751A/en
Publication of CN114897751A publication Critical patent/CN114897751A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10048Infrared image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)

Abstract

The invention relates to an infrared and visible light image perception fusion method based on multi-scale structural decomposition, and belongs to the technical field of multi-sensor image fusion. The method fully considers relevant characteristics of a Human Visual System (HVS), and can help solve potential defects of current fusion research in visual information perception. Compared with other algorithms, the method constructs a multi-scale structure decomposition method based on scale perception edge preservation, and can obtain image structures with different scales, wherein edge information is kept in each layer, and small-scale details can be regarded as structures with fine spatial scales. In addition, the method fully considers the significant information of the pixel level and the large-scale structural information in the fusion process, so that a fusion image with rich information and good visual perception effect can be obtained.

Description

Infrared and visible light image perception fusion method based on multi-scale structural decomposition
Technical Field
The invention relates to an infrared and visible light image perception fusion method based on multi-scale structural decomposition, and belongs to the technical field of multi-sensor image fusion.
Background
The image fusion technology has important significance in image processing and computer vision, and is widely applied to the fields of military affairs, remote sensing, medical image processing, industrial detection and the like. Among them, infrared and visible image fusion has become one of the most studied branches due to its uniqueness in application. Visible light images are generally of higher resolution and contain important detailed information of the scene, but their imaging quality is susceptible to external factors such as weather, light, etc. In contrast, infrared images contain hidden information that is lost in visible light images and can reflect the thermal radiation information of the scene, but their detail information is often poor. Therefore, the information of the visible light image and the infrared image are complementary to a certain extent, and a relatively complete scene mapping can be obtained by fusing the infrared image and the visible light image. One basic principle of infrared and visible image fusion is to preserve as much salient information as possible in the infrared and visible images. In addition, it is desirable that the fused image introduces less artifacts and has good visual perception.
In general, infrared and visible image fusion includes three important steps, namely feature extraction, fusion strategy formulation, and image reconstruction. Depending on the analysis tools used in the above process, existing infrared and visible light image fusion algorithms can be classified into six categories: multi-scale transform based methods, subspace based methods, sparse representation based methods, saliency based methods, deep learning based methods, and hybrid methods. Among the methods, the most widely studied and applied method is a fusion method based on multi-scale transformation, which first decomposes a source image by using a transformation technology to obtain multi-scale information, and then fuses each scale information one by using a certain fusion strategy. Among them, the laplacian pyramid is a classical transformation technique commonly used for multi-scale decomposition, and thus multi-scale decomposition techniques such as a contrast pyramid, a steerable pyramid, and a morphological pyramid are derived. Wavelet transformation, another important multi-scale decomposition tool, provides a method for decomposing an image into a low-pass layer image and detail layer images in different directions, so that noise in a fused image can be reduced. On the basis of the above, researchers have proposed improved analysis tools such as discrete wavelet transform, contourlet transform, shear wave transform, etc., which have better decomposition performance. In addition, many edge-preserving filters, such as bilateral filters, guided filters, are proposed and widely used for multi-scale decomposition of images, which can preserve the spatial continuity of the image structure and reduce the generation of halos and artifacts. In the fusion process, the fusion weight is often determined by adopting a maximum value selection and weighted average strategy. Toet et al uses contrast pyramid transformation to decompose the source image and then selects the maximum contrast value as the fusion coefficient. Adu et al propose to use a weighted average strategy to calculate the weight coefficients of the decomposed images, and then to fuse the images of the same scale by the weight coefficients.
Existing fusion methods focus more on preserving significant information or avoiding artifacts in infrared and visible images, and take less into account perceptual issues and the characteristics of the Human Visual System (HVS). Considering the mechanism of the HVS not only can produce what we generally consider visually pleasing fusion results, but more importantly it can help address potential drawbacks of the current fusion framework. Generally, the image fusion process may involve using visual features from different source images, in particular by comparison between them to determine fusion weights or to obtain an appropriate information fusion strategy. However, visual features are susceptible to external physical conditions (e.g., ambient lighting, characteristics of different sensors, etc.), which means that these features are not placed on an equal and unambiguous basis when compared and fused, which can affect fusion quality, especially for infrared and visible image fusion, the response characteristics of the two sensors vary greatly, and the visual information in the visible spectrum can be severely affected by changes in external lighting conditions.
Disclosure of Invention
The technical problem to be solved by the invention is as follows: the method transforms the physical strength of source images into a visual response space of an HVS (high voltage sequence transformation) so that input information from different source images can be compared and fused in a unified human visual response space, and all features in the space are in the same perception state, thereby eliminating external physical factors which can influence the fusion process, finally generating a fusion image with rich information and good visual perception effect, solving the current potential defects and improving the fusion effect.
The technical scheme of the invention is as follows:
an infrared and visible light image perception fusion method based on multi-scale structural decomposition comprises the following steps:
step 1, carrying out multi-scale structural decomposition on infrared and visible light images of the same scene based on a scale-aware edge preserving (SAEP) filtering algorithm to obtain infrared and visible light multi-scale filtering images
Figure BDA0003591984540000031
And
Figure BDA0003591984540000032
wherein j is 0,1, …, N; n is the number of scale layers;
step 2, the infrared and visible light multi-scale filtering image obtained in the step 1 is processed
Figure BDA0003591984540000033
And
Figure BDA0003591984540000034
converting the image into a visual response space of an HVS (high Voltage stereo System), and obtaining the multi-scale perception contrast of the infrared image
Figure BDA0003591984540000035
And multi-scale perceptual contrast of visible light images
Figure BDA0003591984540000036
The conversion method comprises the following steps:
(1) calculating multi-scale adaptive contrast of infrared and visible light images according to contrast sensitivity and local adaptive mechanism of HVS
Figure BDA0003591984540000037
And
Figure BDA0003591984540000038
Figure BDA0003591984540000039
Figure BDA00035919845400000310
wherein the content of the first and second substances,
Figure BDA00035919845400000311
is the j-th layer low-pass image of the infrared image,
Figure BDA00035919845400000312
is a jth layer band-pass image of the infrared image,
Figure BDA00035919845400000313
is the jth layer low pass image of the visible image,
Figure BDA00035919845400000314
the j-th layer bandpass image of the visible light image, t is an adaptive parameter, and is a set value, preferably, t is 1, α is an adjustment parameter, preferably, α is 0.8,
Figure BDA00035919845400000315
(2) calculating initial value of multi-scale perception contrast of infrared and visible light images according to nonlinear conversion mechanism of HVS
Figure BDA00035919845400000316
And
Figure BDA00035919845400000317
when in use
Figure BDA00035919845400000318
When the positive value and the 0 value are taken,
Figure BDA0003591984540000041
when in use
Figure BDA0003591984540000042
When the negative value is taken as the value,
Figure BDA0003591984540000043
when in use
Figure BDA0003591984540000044
When the positive value and the 0 value are taken,
Figure BDA0003591984540000045
when in use
Figure BDA0003591984540000046
When the negative value is taken as the value,
Figure BDA0003591984540000047
wherein h is a threshold, preferably, h is 0.5; c is a constant, and is set to be 21.3, p values are different in different scale layers, and when j is increased from small to large, and when N is set to be 4, the values of p are respectively as follows: 1.40,1.15,1.04, 1.15;
(3) for the initial value of the multi-scale perception contrast obtained in the step (2)
Figure BDA0003591984540000048
And
Figure BDA0003591984540000049
carrying out noise and intensity saturation suppression to obtain the final multi-scale perception contrast of the infrared image
Figure BDA00035919845400000410
And multi-scale perceptual contrast of visible light images
Figure BDA00035919845400000411
The noise suppression method is as follows:
Figure BDA0003591984540000051
Figure BDA0003591984540000052
where th is a threshold to distinguish between noise and useful information,
Figure BDA0003591984540000053
the average gray value of the source image is obtained; since noise usually occurs in a small-scale layer, it is common to set
Figure BDA0003591984540000054
The strength was suppressed as follows:
Figure BDA0003591984540000055
Figure BDA0003591984540000056
unlike noise, overexposure often occurs at the large scale layer, rAs overexposure suppression parameter, I 0 Representing a source image;
step 3, according to the visual characteristics of human eyes, the lowest layer low-pass image of the infrared and visible light images is processed
Figure BDA0003591984540000057
And
Figure BDA0003591984540000058
self-adaptive adjustment is carried out, fusion weight is determined based on a significance strategy, and then infrared and visible light bottom layer low-pass fusion images are obtained
Figure BDA0003591984540000059
The specific method comprises the following steps:
(1) for the bottom low-pass image of infrared and visible light images
Figure BDA00035919845400000510
And
Figure BDA00035919845400000511
performing self-adaptive adjustment to obtain an adjusted bottommost layer low-pass image
Figure BDA00035919845400000512
And
Figure BDA00035919845400000513
Figure BDA00035919845400000514
Figure BDA00035919845400000515
wherein l is a threshold value which reflects the average background brightness of the adjusted lowest layer low-pass image, and is set to be 128;
(2) determining visible light according to significance fusion strategyLowest layer low pass image
Figure BDA0003591984540000061
Fusion weight w of (c):
Figure BDA0003591984540000062
wherein
Figure BDA0003591984540000063
Which represents a gaussian filtering operation, is shown,
Figure BDA0003591984540000064
and
Figure BDA0003591984540000065
the saliency maps of the lowest low-pass image of the visible and infrared images, respectively, are calculated as follows:
Figure BDA0003591984540000066
Figure BDA0003591984540000067
wherein the content of the first and second substances,
Figure BDA0003591984540000068
and
Figure BDA0003591984540000069
respectively representing the gray values of a pixel point n and an adjacent pixel point k in an infrared image bottom layer image area omega,
Figure BDA00035919845400000610
and
Figure BDA00035919845400000611
respectively representing pixel point n and adjacent pixel point k in visible light image bottom layer image area omegaThe gray value of (a);
thus, the infrared and visible bottom layer low-pass fusion image
Figure BDA00035919845400000612
Is composed of
Figure BDA00035919845400000613
Step 4, determining the multi-scale perception contrast of the infrared image by using a bidirectional significance polymerization strategy
Figure BDA00035919845400000614
And multi-scale perceptual contrast of visible light images
Figure BDA00035919845400000615
The fusion weight of (1), wherein one direction is the combination of pixel level significance from top to bottom, and the other direction is the aggregation of the opposite directions of the structure significance, and further a fusion image of the j-th layer perception contrast of the infrared and visible light images is obtained;
the method comprises the following specific steps:
(1) for pixel level saliency, contrast will be perceived
Figure BDA00035919845400000616
And
Figure BDA00035919845400000617
respectively polymerizing from small scale to large scale to obtain the perception contrast of the jth layer
Figure BDA00035919845400000618
Pixel level saliency of
Figure BDA00035919845400000619
And j-th layer perceived contrast
Figure BDA00035919845400000620
Pixel level saliency of
Figure BDA00035919845400000621
Figure BDA00035919845400000622
Figure BDA00035919845400000623
(2) For structural significance, contrast will be perceived
Figure BDA0003591984540000071
And
Figure BDA0003591984540000072
aggregating from large scale to small scale separately, and also taking into account the adjusted lowest low-pass image
Figure BDA0003591984540000073
And
Figure BDA0003591984540000074
thereby obtaining the j-th layer perception contrast
Figure BDA0003591984540000075
Structural significance of
Figure BDA0003591984540000076
And j-th layer perceived contrast
Figure BDA0003591984540000077
Structural significance of
Figure BDA0003591984540000078
Figure BDA0003591984540000079
Figure BDA00035919845400000710
Where sf is the structural significance function, which is calculated as follows:
Figure BDA00035919845400000711
gamma is a balance parameter, and gamma is 0.1; s 1 、s 2 In relation to the eigenvalues of the gradient covariance matrix C, it is given by:
Figure BDA00035919845400000712
wherein I x (X) and I y (X) respectively represent partial windows W i The gradient of pixel points X in the X and y directions.
(3) Computing the j-th layer perceived contrast
Figure BDA00035919845400000713
Overall significance of
Figure BDA00035919845400000714
And j-th layer perceived contrast
Figure BDA00035919845400000715
Overall significance of
Figure BDA00035919845400000716
Figure BDA00035919845400000717
Figure BDA00035919845400000718
Wherein beta is a balance parameter, beta is 5,
Figure BDA00035919845400000719
a saliency map for layer j of an infrared image,
Figure BDA00035919845400000720
the saliency adjustment map of the jth layer of the visible light image has the following values:
Figure BDA00035919845400000721
Figure BDA00035919845400000722
wherein the content of the first and second substances,
Figure BDA00035919845400000723
in the form of a source infrared image,
Figure BDA00035919845400000724
representing the average gray value of the infrared image in the neighborhood omega; sg denotes a sigmoid function which is,
Figure BDA0003591984540000081
u is a control parameter, and when u takes different values, the sigmoid function has different shapes, and here, u is set to 5.
(4) Overall saliency from perceived contrast of layers of visible and infrared images
Figure BDA0003591984540000082
And
Figure BDA0003591984540000083
the fusion weight of each layer of perception contrast of the visible light image is as follows:
Figure BDA0003591984540000084
wherein, in different scale layers, u has different values, and u is equal to0.1*2 4-j
Further, a fusion image of the j-th layer perception contrast of the infrared and visible light images
Figure BDA0003591984540000085
Is composed of
Figure BDA0003591984540000086
And 5, obtaining a final fusion image through inverse transformation and reconstruction processes:
Figure BDA0003591984540000087
wherein the content of the first and second substances,
Figure BDA0003591984540000088
is the lowest-level low-pass fused image,
Figure BDA0003591984540000089
resulting from the inverse transformation process in step 2:
Figure BDA00035919845400000810
Figure BDA00035919845400000811
a contrast image is perceived for the fused layers.
Advantageous effects
1. The invention provides a novel sensing framework based on multi-scale structural decomposition, which is used for fusing infrared and visible light images. The proposed framework fully considers relevant characteristics of the HVS and can help solve potential drawbacks of current fusion studies in visual information perception.
2. The invention constructs a multi-scale structure decomposition method based on an SAEP filter to design a perception fusion framework. Compared with other algorithms, the method has excellent edge retention and scale perception characteristics, can obtain image structures of different scales, wherein edge information is kept in each layer, and small-scale details can be regarded as structures with fine spatial scales.
3. The invention provides a novel bidirectional significance aggregation algorithm for determining the fusion weight of multi-scale perception contrast, and the algorithm fully considers pixel-level significance information and large-scale structure information, so that a fusion image with rich information and good visual perception effect can be obtained.
4. The framework proposed by the present invention combines some key characteristics of the HVS, including multi-scale processing channels, contrast sensitivity, local adaptation, and supra-threshold characteristics. All relevant key features of the HVS are integrated synthetically into the proposed framework to simulate human visual response in complex scenes, creating a visual response space in the HVS that is representative of multi-scale perceptual contrast.
5. The present invention constructs a multi-scale structural decomposition by utilizing a scale-aware edge preservation (SAEP) filter that has good scale separation and edge preservation characteristics. By decomposition, an image structure of different scales is obtained, with edges remaining in each layer, and the details therein can be regarded as an image structure with a fine spatial scale.
6. In the fusion process, the invention proposes a two-way saliency aggregation strategy to fuse the perceptual contrast of each scale, one direction is aggregated from top to bottom in a scale space to obtain pixel-level saliency, and the other direction is aggregated reversely to calculate structural saliency. The two types of saliency are then combined and fusion weights are calculated according to the sigmoid function.
Drawings
FIG. 1 is a sigmoid function with u taking different values;
FIG. 2 is a flow chart of the fusion framework of the present invention;
FIG. 3 is a comparison of fused images of infrared and visible light images obtained by different methods. The image processing method comprises the following steps of (a) obtaining an infrared image, (b) obtaining a visible light image, (c) obtaining a fused image of the infrared image and the visible light image obtained by a WLS method, (d) obtaining a fused image of the infrared image and the visible light image obtained by a U2Fusion method, (e) obtaining a fused image of the infrared image and the visible light image obtained by an IFCNN method, and (f) obtaining the fused image of the infrared image and the visible light image obtained by the method.
Detailed Description
The present invention will be described in detail below with reference to the accompanying drawings.
The invention provides an infrared and visible light image perception fusion framework based on multi-scale structural decomposition, which converts a source image into a visual response space of a Human Visual System (HVS) for comparison and fusion by taking the reference of a relevant mechanism of the HVS.
Based on this, the specific embodiments of the present invention are:
suppose the input infrared and visible images are I respectively r And I v As shown in fig. 2, the fusion steps are as follows:
step 1: according to the characteristics of a multi-scale processing channel of the HVS, the invention constructs multi-scale structure decomposition based on a scale perception edge preserving (SAEP) filtering algorithm to obtain infrared and visible light multi-scale filtering images
Figure BDA0003591984540000101
And
Figure BDA0003591984540000102
I j =SAEP(I j-1j ,r j,0 ,r j,1 ),j=1,2,…,N
wherein I 0 λ is the global smoothing weight, λ 1 =0.1,λ j+1 =λ j + 0.9; r is a scale parameter, the scale is [ r ] j,0 ,r j,1 ]The image structure in between will be smoothed. In this example, r 1,0 =0,r 1,1 =4,r j+1,1 =2r j,1 ,r j+1,0 =r j,1 The filtering number N is 4.
Further, infrared and visible light layer band pass images
Figure BDA0003591984540000103
And
Figure BDA0003591984540000104
is obtained by the following formula:
B j =I j-1 -I j ,j=1,2,…,N
step 2: based on the correlation mechanisms such as contrast sensitivity, local adaptation and super-threshold characteristics of the HVS, the band-pass and low-pass images obtained by the multi-scale decomposition are converted into the visual response space of the HVS to obtain the multi-scale perception contrast of the infrared image
Figure BDA0003591984540000105
And multi-scale perceptual contrast of visible light images
Figure BDA0003591984540000106
(1) Calculating multi-scale adaptive contrast of infrared and visible light images according to contrast sensitivity and local adaptive mechanism of HVS
Figure BDA0003591984540000107
And
Figure BDA0003591984540000108
Figure BDA0003591984540000109
wherein, I j And B j A jth layer low pass image and a band pass image of the infrared or visible light image, respectively; t is a self-adaptive parameter, and different values are taken according to the characteristics of the infrared image and the visible light image; α is a regulation parameter, and preferably, α is 0.8.
(2) The adaptive contrast resulting from (1) adapts to some extent to human vision, but it is still not the perceptual contrast in the visual response space. The HVS presents a non-linear transfer function that helps to achieve perceptual contrast in this unified space. Obtaining multi-scale perception contrast of infrared and visible light images according to HVS nonlinear conversion mechanism
Figure BDA0003591984540000111
And
Figure BDA0003591984540000112
Figure BDA0003591984540000113
wherein R is j Taking a positive value, if the positive value is negative, taking the absolute value as an input to calculate, and inverting the output of the absolute value; h is a threshold, preferably, h is 0.5; c is a constant, and is 21.3; in different scale layers, the p values are different, and when the scale is increased from small to large, the values are respectively as follows: 1.40,1.15,1.04,1.15,1.35,1.93.
(3) Further carrying out noise and intensity saturation suppression on the obtained perception contrast to obtain the final multi-scale perception contrast of the infrared image
Figure BDA0003591984540000114
And multi-scale perceptual contrast of visible light images
Figure BDA0003591984540000115
In general, smaller scale layers contain more noise, while larger scale layers have more over-exposure. Therefore, different suppression methods are applied in different frequency layers.
The noise suppression method is as follows:
Figure BDA0003591984540000116
where th is a threshold to distinguish between noise and useful information,
Figure BDA0003591984540000117
the average gray value of the source image is obtained; in this example, let
Figure BDA0003591984540000118
The suppression method of intensity saturation is as follows:
Figure BDA00035919845400001113
wherein r is an overexposure suppression parameter, I 0 A source image is represented.
And step 3: the lowest low-pass layers of the infrared and visible images reflect background information of the scene. According to the visual characteristics of human eyes, the lowest layer low-pass image of the infrared image and the visible light image is subjected to
Figure BDA0003591984540000119
And
Figure BDA00035919845400001110
carrying out self-adaptive adjustment, and determining fusion weight based on a significance strategy:
(1) for the bottom low-pass image of infrared and visible light images
Figure BDA00035919845400001111
And
Figure BDA00035919845400001112
performing self-adaptive adjustment to obtain an adjusted bottommost layer low-pass image
Figure BDA0003591984540000121
And
Figure BDA0003591984540000122
Figure BDA0003591984540000123
wherein l is a threshold value which reflects the average background brightness of the adjusted lowest layer low-pass image, and is set to be 128; α is a regulation parameter, and preferably, α is 0.8.
(2) Fusion strategy according to significanceCalculating the low-pass image of the bottom layer of visible light
Figure BDA0003591984540000124
Fusion weight w of (c):
Figure BDA0003591984540000125
wherein
Figure BDA0003591984540000126
Which represents a gaussian filtering operation, is shown,
Figure BDA0003591984540000127
and
Figure BDA0003591984540000128
the saliency maps of the lowest low-pass image of the visible and infrared images, respectively, are calculated as follows:
Figure BDA0003591984540000129
wherein A is N (n) and A N (k) Respectively representing the gray values of the pixel point n and the adjacent pixel point k in the bottom layer image area omega.
Thus, the infrared and visible bottom layer low-pass fusion image
Figure BDA00035919845400001210
Is composed of
Figure BDA00035919845400001211
And 4, step 4: in the visual response space of the HVS, the perceived contrast typically contains fine pixel-level saliency information and structural information of the image. Thus, multi-scale perceptual contrast for infrared images
Figure BDA00035919845400001212
And it can be seenMulti-scale perceptual contrast of light images
Figure BDA00035919845400001213
The invention proposes a two-way significance aggregation strategy to fully aggregate these features and determine fusion weights based on this. One of which is a top-down combination of pixel-level saliency, the other is a polymerization of the structural saliency in the opposite direction.
(1) For pixel level significance, the method superposes the perception contrast from small scale to large scale to obtain the perception contrast C of the jth layer j Pixel level saliency of D j
Figure BDA00035919845400001214
The j-th layer pixel level significance comprises fine-grained information of a current layer and a smaller-scale layer, and more complete details can be reserved, so that a final fused image is finer and smoother.
(2) For structural significance, the invention aggregates the perceived contrast from large scale to small scale, and in addition, due to the adjusted lowest low-pass layer image A N Contains the basic structural information of the source image and we take this into account to obtain relatively complete structural saliency. Layer j perceived contrast C j Structural significance of (1) G j Comprises the following steps:
Figure BDA0003591984540000131
wherein sf is a structural significance function, can reflect structural information such as corners and the like in the image, and is calculated as follows:
Figure BDA0003591984540000132
α is a balance parameter, and in the present invention, α is made 0.1; s 1 、s 2 With respect to the eigenvalues of the gradient covariance matrix C, it is obtained by the following equation:
Figure BDA0003591984540000133
wherein I x (X) and I y (X) respectively denote partial windows w i The gradient of the inner pixel point X in the X and y directions.
(3) Calculating the j-th layer perception contrast C j Overall significance of S j
S j =M j *(D j +β*G j )
Wherein beta is a balance parameter, and beta is 5; denotes element-by-element multiplication operations; m j The map is adjusted for saliency, which can help the fused image capture more highlight target information and less noise from the infrared image. For infrared and visible images, the values are as follows:
Figure BDA0003591984540000134
Figure BDA0003591984540000135
wherein the content of the first and second substances,
Figure BDA0003591984540000136
in the form of a source infrared image,
Figure BDA0003591984540000137
representing the average gray value of the infrared image in the neighborhood omega; sg denotes the sigmoid function and,
Figure BDA0003591984540000141
u is a control parameter, and when u takes different values, the sigmoid function has different shapes, as shown in fig. 1, where u is 5.
(4) Overall saliency from perceived contrast of layers of visible and infrared images
Figure BDA0003591984540000142
And
Figure BDA0003591984540000143
the fusion weight of each layer of perceived contrast of the visible light image is as follows:
Figure BDA0003591984540000144
wherein, in different scale layers, u takes different values, in this example, u is 0.1 × 2 4-j
Further, the fusion process of the infrared and visible image perception contrast layers can be described as
Figure BDA0003591984540000145
Wherein the content of the first and second substances,
Figure BDA0003591984540000146
and the fusion image is the j-th layer of perceived contrast.
And 5: lowest-layer low-pass fusion image obtained in visual response space of HVS based on step 3 and step 4
Figure BDA0003591984540000147
And fused, individual scale perceptual contrast images
Figure BDA0003591984540000148
Obtaining a final fused image through inverse transformation and reconstruction processes:
Figure BDA0003591984540000149
wherein the content of the first and second substances,
Figure BDA00035919845400001410
resulting from the inverse transformation process in step 2:
Figure BDA00035919845400001411
FIG. 3 is a comparison of fused images and other methods in accordance with the teachings of the present invention. Wherein, (a) is an infrared image, (b) is a visible light image, and (c), (d), (e) and (f) are Fusion results of the WLS method, the U2Fusion method, the IFCNN method and the method of the invention respectively. It can be seen that the inventive framework can achieve better fusion results by performing fusion in a consistent and well-defined visual response space, since the relevant properties of the human visual system are fully taken into account.

Claims (10)

1. The infrared and visible light image perception fusion method based on multi-scale structural decomposition is characterized by comprising the following steps:
step 1, performing multi-scale structural decomposition on infrared and visible light images of the same scene to obtain infrared and visible light multi-scale filtering images
Figure FDA0003591984530000011
And
Figure FDA0003591984530000012
wherein j is 0,1, …, N; n is the number of scale layers;
step 2, the infrared and visible light multi-scale filtering image obtained in the step 1 is processed
Figure FDA0003591984530000013
And
Figure FDA0003591984530000014
converting the image into a visual response space of an HVS (high Voltage stereo System), and obtaining the multi-scale perception contrast of the infrared image
Figure FDA0003591984530000015
And multi-scale perceptual contrast of visible light images
Figure FDA0003591984530000016
Step 3, the lowest layer low-pass image of the infrared and visible light images
Figure FDA0003591984530000017
And
Figure FDA0003591984530000018
self-adaptive adjustment is carried out, fusion weight is determined based on a significance strategy, and an infrared and visible light bottom layer low-pass fusion image is obtained
Figure FDA0003591984530000019
Step 4, determining the multi-scale perception contrast of the infrared image
Figure FDA00035919845300000110
And multi-scale perceptual contrast of visible light images
Figure FDA00035919845300000111
Obtaining a fusion image of the j-th layer perception contrast of the infrared and visible light images;
step 5, low-pass fusion image of the infrared and visible light bottom layer obtained in the step 3
Figure FDA00035919845300000112
And 4, obtaining the final fusion image by the fusion image of the infrared image and the visible light image with each layer perception contrast through inverse transformation and reconstruction processes.
2. The infrared and visible image perception fusion method based on multi-scale structural decomposition according to claim 1, characterized in that:
in the step 1, multi-scale structural decomposition is performed on the infrared and visible light images of the same scene based on a scale-aware edge preservation (SAEP) filtering algorithm.
3. The infrared and visible image perception fusion method based on multi-scale structural decomposition according to claim 1, characterized in that:
in the step 2, the infrared and visible light multi-scale filtering images are obtained
Figure FDA00035919845300000113
And
Figure FDA00035919845300000114
converting the image into a visual response space of an HVS (high Voltage stereo System), and obtaining the multi-scale perception contrast of the infrared image
Figure FDA00035919845300000115
And multi-scale perceptual contrast of visible light images
Figure FDA00035919845300000116
The specific method comprises the following steps:
(1) multi-scale adaptive contrast ratio for calculating infrared and visible light images
Figure FDA0003591984530000021
And
Figure FDA0003591984530000022
Figure FDA0003591984530000023
Figure FDA0003591984530000024
wherein the content of the first and second substances,
Figure FDA0003591984530000025
is the j-th layer low-pass image of the infrared image,
Figure FDA0003591984530000026
is a jth layer bandpass image of the infrared image,
Figure FDA0003591984530000027
is the jth layer low pass image of the visible image,
Figure FDA0003591984530000028
the j-th layer bandpass image of the visible light image, t is an adaptive parameter, and is a set value, preferably, t is 1, α is an adjustment parameter, preferably, α is 0.8,
Figure FDA0003591984530000029
(2) calculating multi-scale perception contrast initial value of infrared and visible light image
Figure FDA00035919845300000210
And
Figure FDA00035919845300000211
when in use
Figure FDA00035919845300000212
When the positive value and the 0 value are taken,
Figure FDA00035919845300000213
when in use
Figure FDA00035919845300000214
When the negative value is taken as the value,
Figure FDA00035919845300000215
when in use
Figure FDA00035919845300000216
When the positive value and the 0 value are taken,
Figure FDA00035919845300000217
when in use
Figure FDA00035919845300000218
When the negative value is taken as the value,
Figure FDA0003591984530000031
wherein h is a threshold value and c is a constant;
(3) for the initial value of the multi-scale perception contrast obtained in the step (2)
Figure FDA0003591984530000032
And
Figure FDA0003591984530000033
carrying out noise and intensity saturation suppression to obtain the final multi-scale perception contrast of the infrared image
Figure FDA0003591984530000034
And multi-scale perceptual contrast of visible light images
Figure FDA0003591984530000035
The noise suppression method is as follows:
Figure FDA0003591984530000036
Figure FDA0003591984530000037
wherein, th is a threshold value,
Figure FDA0003591984530000038
the average gray value of the source image is obtained;
Figure FDA0003591984530000039
the strength was suppressed as follows:
Figure FDA00035919845300000310
Figure FDA00035919845300000311
r is an overexposure suppression parameter, I 0 A source image is represented.
4. The method of claim 3, wherein the method comprises:
in the step 3, the lowest layer low-pass image of the infrared and visible light images is subjected to image processing
Figure FDA00035919845300000312
And
Figure FDA00035919845300000313
self-adaptive adjustment is carried out, fusion weight is determined based on a significance strategy, and an infrared and visible light bottom layer low-pass fusion image is obtained
Figure FDA00035919845300000314
The method comprises the following steps:
(1) for the bottom low-pass image of infrared and visible light images
Figure FDA0003591984530000041
And
Figure FDA0003591984530000042
performing self-adaptive adjustment to obtain an adjusted bottommost layer low-pass image
Figure FDA0003591984530000043
And
Figure FDA0003591984530000044
Figure FDA0003591984530000045
Figure FDA0003591984530000046
wherein l is a threshold;
(2) determining the lowest layer low-pass image of visible light according to a significance fusion strategy
Figure FDA0003591984530000047
Fusion weight w of (c):
Figure FDA0003591984530000048
wherein
Figure FDA0003591984530000049
Which represents a gaussian filtering operation, is shown,
Figure FDA00035919845300000410
and
Figure FDA00035919845300000411
are saliency maps of the lowest low pass image of the visible and infrared images respectively,it is calculated as follows:
Figure FDA00035919845300000412
Figure FDA00035919845300000413
wherein the content of the first and second substances,
Figure FDA00035919845300000414
and
Figure FDA00035919845300000415
respectively representing the gray values of a pixel point n and an adjacent pixel point k in an infrared image bottom layer image area omega,
Figure FDA00035919845300000416
and
Figure FDA00035919845300000417
respectively representing the gray values of a pixel point n and an adjacent pixel point k in a bottom image region omega of the visible light image;
thus, the infrared and visible bottom layer low-pass fusion image
Figure FDA00035919845300000418
Is composed of
Figure FDA00035919845300000419
5. The method of claim 4, wherein the method comprises:
step 4, determining the multi-scale perception contrast of the infrared image by using a bidirectional significance aggregation strategy
Figure FDA00035919845300000420
And multi-scale perceptual contrast of visible light images
Figure FDA00035919845300000421
The method for obtaining the fusion image of the j-th layer perception contrast of the infrared and visible light image comprises the following steps:
(1) will perceive contrast
Figure FDA0003591984530000051
And
Figure FDA0003591984530000052
respectively polymerizing from small scale to large scale to obtain the perception contrast of the j layer
Figure FDA0003591984530000053
Pixel level saliency of
Figure FDA0003591984530000054
And j-th layer perceived contrast
Figure FDA0003591984530000055
Pixel level saliency of
Figure FDA0003591984530000056
Figure FDA0003591984530000057
Figure FDA0003591984530000058
(2) Will perceive contrast
Figure FDA0003591984530000059
And
Figure FDA00035919845300000510
respectively polymerizing from large scale to small scale to obtain the perception contrast of the j layer
Figure FDA00035919845300000511
Structural significance of
Figure FDA00035919845300000512
And j-th layer perceived contrast
Figure FDA00035919845300000513
Structural significance of
Figure FDA00035919845300000514
Figure FDA00035919845300000515
Figure FDA00035919845300000516
Where sf is a structural significance function, which is calculated as follows:
Figure FDA00035919845300000517
gamma is a balance parameter, s 1 、s 2 Is obtained by the following formula:
Figure FDA00035919845300000518
wherein I x (X) and I y (X) respectively represent partial windows W i The gradients of the inner pixel points X in the X and y directions;
(3) computing the j-th layer perceived contrast
Figure FDA00035919845300000519
Overall significance of
Figure FDA00035919845300000520
And j-th layer perceived contrast
Figure FDA00035919845300000521
Overall significance of
Figure FDA00035919845300000522
Figure FDA00035919845300000523
Figure FDA00035919845300000524
Wherein beta is a balance parameter,
Figure FDA00035919845300000525
a saliency map for layer j of an infrared image,
Figure FDA00035919845300000526
the saliency adjustment map of the jth layer of the visible light image has the following values:
Figure FDA00035919845300000527
Figure FDA0003591984530000061
wherein the content of the first and second substances,
Figure FDA0003591984530000062
in the form of a source infrared image,
Figure FDA0003591984530000063
representing the average gray value of the infrared image in the neighborhood omega; sg denotes a sigmoid function which is,
Figure FDA0003591984530000064
u is a control parameter;
(4) overall saliency from perceived contrast of layers of visible and infrared images
Figure FDA0003591984530000065
And
Figure FDA0003591984530000066
the fusion weight of each layer of perceived contrast of the visible light image is as follows:
Figure FDA0003591984530000067
fusion image of j-th layer perception contrast of infrared and visible light image
Figure FDA0003591984530000068
Is composed of
Figure FDA0003591984530000069
6. The method of claim 5, wherein the method comprises:
in the step 5, the method for obtaining the final fusion image through the inverse transformation and reconstruction processes comprises the following steps:
Figure FDA00035919845300000610
wherein the content of the first and second substances,
Figure FDA00035919845300000611
for the bottom-most low-pass fused image,
Figure FDA00035919845300000612
resulting from the inverse transformation process in step 2:
Figure FDA00035919845300000613
Figure FDA00035919845300000614
a contrast image is perceived for the fused layers.
7. The method of claim 6, wherein the method comprises:
h=0.5。
8. the method of claim 6, wherein the method comprises:
c=21.3。
9. the method of claim 6, wherein the method comprises:
in different scale layers, the values of p are different, when j is increased from small to large, when N is 4, the values of p are respectively: 1.40,1.15,1.04,1.15.
10. The method of claim 6, wherein the method comprises:
l=128,β=5,γ=0.1。
CN202210381391.0A 2022-04-12 2022-04-12 Infrared and visible light image perception fusion method based on multi-scale structural decomposition Pending CN114897751A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210381391.0A CN114897751A (en) 2022-04-12 2022-04-12 Infrared and visible light image perception fusion method based on multi-scale structural decomposition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210381391.0A CN114897751A (en) 2022-04-12 2022-04-12 Infrared and visible light image perception fusion method based on multi-scale structural decomposition

Publications (1)

Publication Number Publication Date
CN114897751A true CN114897751A (en) 2022-08-12

Family

ID=82718480

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210381391.0A Pending CN114897751A (en) 2022-04-12 2022-04-12 Infrared and visible light image perception fusion method based on multi-scale structural decomposition

Country Status (1)

Country Link
CN (1) CN114897751A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115797244A (en) * 2023-02-07 2023-03-14 中国科学院长春光学精密机械与物理研究所 Image fusion method based on multi-scale direction co-occurrence filter and intensity transmission

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115797244A (en) * 2023-02-07 2023-03-14 中国科学院长春光学精密机械与物理研究所 Image fusion method based on multi-scale direction co-occurrence filter and intensity transmission

Similar Documents

Publication Publication Date Title
CN111583123A (en) Wavelet transform-based image enhancement algorithm for fusing high-frequency and low-frequency information
CN101770639B (en) Enhancement method of low-illumination image
CN104835130A (en) Multi-exposure image fusion method
CN107274365A (en) A kind of mine image intensification method based on unsharp masking and NSCT algorithms
Song et al. High dynamic range infrared images detail enhancement based on local edge preserving filter
CN108921809B (en) Multispectral and panchromatic image fusion method based on spatial frequency under integral principle
CN113222877B (en) Infrared and visible light image fusion method and application thereof in airborne photoelectric video
CN104537678B (en) A kind of method that cloud and mist is removed in the remote sensing images from single width
Li et al. Underwater image high definition display using the multilayer perceptron and color feature-based SRCNN
Student Study of image fusion-techniques method and applications
Gao et al. Improving the performance of infrared and visible image fusion based on latent low-rank representation nested with rolling guided image filtering
Jian et al. Infrared and visible image fusion based on deep decomposition network and saliency analysis
CN113298147B (en) Image fusion method and device based on regional energy and intuitionistic fuzzy set
Karalı et al. Adaptive image enhancement based on clustering of wavelet coefficients for infrared sea surveillance systems
CN107451986B (en) Single infrared image enhancement method based on fusion technology
CN114897751A (en) Infrared and visible light image perception fusion method based on multi-scale structural decomposition
CN110148083B (en) Image fusion method based on rapid BEMD and deep learning
CN104240208A (en) Uncooled infrared focal plane detector image detail enhancement method
CN111815550B (en) Infrared and visible light image fusion method based on gray level co-occurrence matrix
CN116309233A (en) Infrared and visible light image fusion method based on night vision enhancement
CN114066786A (en) Infrared and visible light image fusion method based on sparsity and filter
Thai et al. Performance evaluation of high dynamic range image tone mapping operators based on separable non-linear multiresolution families
CN116597146A (en) Semantic segmentation method for laser radar sparse point cloud data
CN110992287A (en) Method for clarifying non-uniform illumination video
CN110689510A (en) Sparse representation-based image fusion method introducing dictionary information

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination