CN110930315A - Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM - Google Patents

Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM Download PDF

Info

Publication number
CN110930315A
CN110930315A CN201911009926.6A CN201911009926A CN110930315A CN 110930315 A CN110930315 A CN 110930315A CN 201911009926 A CN201911009926 A CN 201911009926A CN 110930315 A CN110930315 A CN 110930315A
Authority
CN
China
Prior art keywords
layer
path
equation
feature
parameter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911009926.6A
Other languages
Chinese (zh)
Other versions
CN110930315B (en
Inventor
李映
王栋
张号逵
白宗文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN201911009926.6A priority Critical patent/CN110930315B/en
Publication of CN110930315A publication Critical patent/CN110930315A/en
Application granted granted Critical
Publication of CN110930315B publication Critical patent/CN110930315B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/73Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The method comprises two parts of model training and multispectral image panchromatic sharpening, wherein in the stage of model training, original clear multispectral and panchromatic images are subjected to down-sampling to obtain a simulation training image pair; secondly, extracting and fusing the characteristics of the panchromatic image and the multispectral image by using a dual-channel convolution network, and realizing the fusion between convolution characteristics of multiple layers and different depths by combining with a layer CLSTM; then, reconstructing a multispectral image with high spatial resolution from the fused features by using a deconvolution network; finally, adjusting the parameters of the model by using an Adam algorithm; in the multispectral image panchromatic sharpening stage, firstly, the trained dual-channel convolution network and the hierarchical CLSTM are used for extracting and fusing the characteristics of panchromatic images and multispectral images. The convolution network is responsible for extracting the characteristics of the multispectral image and the panchromatic image and fusing the characteristics selected by the convolution network and the CLSTM, and the CLSTM selects and memorizes the characteristics of different depths at multiple levels, so that the fusion of the characteristics of multiple levels and different depths is realized.

Description

Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM
Technical Field
The invention belongs to the field of remote sensing image processing, and particularly relates to a multispectral image panchromatic sharpening method based on a dual-channel convolution network and a multi-level fusion strategy.
Background
Remote sensing images have two important properties-spectral resolution and spatial resolution. The spectral resolution refers to the minimum wavelength range which can be distinguished by the sensor when receiving the spectrum of the target radiation, the narrower the wavelength range is, the higher the spectral resolution is, the stronger the ability of the sensor to distinguish and identify light in each wave band in the spectrum is, the more the number of the generated wave bands is, and the richer the spectral information of the obtained remote sensing image is. The spatial resolution refers to the minimum distance between two adjacent ground objects which can be identified on the remote sensing image, the smaller the minimum distance is, the higher the spatial resolution is, the richer the detail information of the ground objects visible in the remote sensing image is, and the stronger the ability of the remote sensing image to identify the objects is.
Most remote sensing applications require remote sensing images with high spatial and spectral resolution. However, it is difficult to directly acquire such images through a single sensor in consideration of data storage amount and sensor signal-to-noise ratio. Therefore, the remote sensing image obtained by the sensor manufactured by the current technology only has the characteristics of high spatial resolution or high spectral resolution. To alleviate this problem, many optical earth observation satellites carry two optical sensors to simultaneously acquire two images with different but complementary characteristics in the same geographic area. For example, IKONOS, high score No. 2, and WorldView-2 all carry a panchromatic sensor and a multispectral image sensor. The panchromatic sensor acquires a single-band high spatial resolution remote sensing image, and the multispectral sensor acquires a multi-band low spatial resolution image. These two types of images are referred to as panchromatic images and multispectral images, respectively.
In practical applications, the color information in the image and the definition of the target are crucial to the interpretation and analysis of the image, so that multispectral images with high spatial resolution are often required in various occasions. Obviously, the original multispectral image or full-color image is often difficult to meet the needs of the user. Therefore, attempts have been made to combine the unique information of the multispectral image and the panchromatic image organically by using an image fusion technique, and to improve the spatial resolution of the multispectral image by using the spatial detail information in the panchromatic image, so as to obtain the multispectral image having the same spatial resolution as the panchromatic image and having the abundant spectral information of the original multispectral image. This is a multispectral image fusion technique, also called multispectral image panchromatic sharpening. At present, the multispectral image and the panchromatic image are fused, and the method is the only method for obtaining the multispectral image with high spatial resolution. In recent years, commercial products using high-resolution remote sensing images (e.g., Google Earth and Bing Maps) have been increasing, and the demand for fused multispectral image data has been increasing. Furthermore, multispectral image panchromatic sharpening techniques are important pre-processing steps for image enhancement for many remote sensing tasks such as change detection, target recognition, image classification, and the like. Therefore, the multispectral image panchromatic sharpening method is widely concerned by the remote sensing field and the image processing field, and is intensively researched all the time.
In recent years, with the development of artificial intelligence and machine learning, many scholars use the technology to solve the key problem in the multispectral image panchromatic sharpening process. In 2015, Wei Huang et al applied the deep neural network in machine learning to the field of multispectral image panchromatic sharpening for the first time. They believe that the relationship between the high resolution image and the low resolution image is the same for both the panchromatic image and the multispectral image. By studying the mapping relationship between the high-resolution and low-resolution full-color images, the mapping relationship between the high-resolution and low-resolution multispectral images can be obtained. Their models do not outperform the traditional methods. Based on the same principle, Azarang and ghassemia propose a stacked auto-encoder structure to generate high-resolution multispectral images, and achieve better performance than the conventional one. However, the method uses the framework of the traditional method, and only partially uses the convolutional neural network. Influenced by the super-resolution field, Masi Giuseppe et al propose a three-layer convolutional neural network PNN by improving SRCNN. The PNN network can obtain very excellent performance. It is too simple to be constructed of only three convolutional layers and needs further improvement. A multi-scale, multi-depth convolutional neural network (MSDCNN) is a multi-scale, multi-depth, dual-branch convolutional neural network (MSDCNN) proposed by Yuan Qiangqiang et al. Due to the multi-scale and multi-depth structure, the MSDCNN has complex nonlinear mapping capability and can process objects of different scales acquired by a plurality of sensors. However, like PNN, MSDCNN directly splices and fuses the multispectral image and the panchromatic image and inputs them into the network, and if the multispectral image and the panchromatic image are not accurately registered, the result is greatly influenced. Recently, Liu Xiangyu et al proposed a network architecture named TFNet. The network does not directly splice and fuse the full-color image and the multispectral image, but extracts the characteristics of the image firstly and then fuses the images indirectly by fusing the characteristics. Therefore, the whole network consists of three modules of feature extraction, feature fusion and image reconstruction. First, in the feature extraction network, the features of the panchromatic image and the multispectral image are extracted with two sets of convolution layers of three-layer convolution, respectively. And (4) connecting the full-color image and the multispectral features in series and inputting the full-color image and the multispectral features into a feature fusion network. The feature fusion sub-network then fuses the resulting features through three convolutional layers to form a more compact representation. Finally, the image reconstruction sub-network reconstructs a multispectral image with high spatial resolution through the 11 convolutional layers. Features of different depths have different meanings, but these networks only fuse features of a certain depth, and do not fully exploit the fusion between features of multiple levels and different depths. In 2019, zhangyonggjun et al proposed a new end-to-end bidirectional pyramid network (BDPN) for sharpening. BDPN can be described as a bi-directional pyramid that processes multispectral and panchromatic images in two branches, respectively. The method uses a bidirectional network for the first time, and has certain innovation. However, the two paths of the network are not balanced, for example, the branch that processes multispectral images contains only two convolutional layers, which is very different from the branch that processes panchromatic images containing 20 convolutional layers.
Disclosure of Invention
Technical problem to be solved
In order to solve the problem of spatial and spectral distortion caused by single aspects of a convolution network type, fusion feature hierarchy, feature fusion depth and the like in the conventional multispectral and panchromatic image fusion method, the invention provides a multispectral image panchromatic sharpening method based on a dual-channel convolution network and a multi-level fusion strategy. The method provided by the invention processes different types of data (full-color image 2D characteristics and multispectral image 3D characteristics) by using two different convolutional neural networks (2D and 3D), and then realizes a multi-level fusion strategy by using a level CLSTM. Because the full-color image of the single waveband is 2D data, the method adopts a 2D convolution neural network to process spatial information; the 3D multispectral data is processed similarly using a 3D network. And features of different depths are selectively memorized by utilizing a hierarchical CLSTM, and the features automatically selected by the CLSTM are fused into multiple layers of the double-channel, so that the fusion of the features of multiple layers and different depths is realized. And finally, reconstructing a multispectral image with high spatial resolution by using a deconvolution network.
Technical scheme
A multispectral image panchromatic sharpening method based on a dual-channel convolution network and a hierarchical CLSTM is characterized by comprising the following steps:
1. training of fusion models
Inputting: image block set F0(MS, PAN), where the original multispectral MS is of size H × W × S and the PAN is of size 4H × 4W × 1, H, W and S representing the height, width and number of spectra of the original multispectral image, respectively;
(1) constructing a simulated training dataset
Step 11: the original multispectral image MS is down-sampled to obtain a simulated multispectral image block
Figure BDA0002243907510000041
The size of the image block is
Figure BDA0002243907510000042
Step 12: multispectral image to be downsampled
Figure BDA0002243907510000043
Performing bilinear interpolation up-sampling to obtain multispectral image with the same height and width as MS
Figure BDA0002243907510000044
Step 13: downsampling original full-color image PANObtaining full color image
Figure BDA0002243907510000045
Figure BDA0002243907510000046
With simulated upsampled multi-spectral images
Figure BDA0002243907510000047
Are the same in height and width;
(2) constructing a dual-path network
Step 21: constructing a spatial characteristic path; firstly, constructing a stem layer; the full color image is characterized in the stem layer using one 2D convolutional layer and the prerlu active layer as shown in equation (1):
Figure BDA0002243907510000048
wherein the content of the first and second substances,
Figure BDA0002243907510000049
layer 0 features representing spatial feature path extraction, F being an abbreviation for feature; 0 represents a 0 th layer; a is used for representing a spatial feature path; wherein W is an abbreviation for weight, which is a parameter of the convolutional layer; b is a bias term;
then constructing T residual blocks, wherein each residual block comprises two 2D convolution layers and two PReLU activation functions; the 1 st residual block has an input of Fa 0The remaining residual block is input
Figure BDA00022439075100000410
And
Figure BDA00022439075100000411
as shown in equations (2) and (3):
Figure BDA0002243907510000051
Figure BDA0002243907510000052
in equation (2), Wa 1,2Subscripts of (A) and
Figure BDA0002243907510000053
the meanings are the same; wa 1,2The superscript 1,2 of (a) indicates the number of layers, where 1 indicates the 1 st layer of the spatial signature path, i.e., the 1 st residual block, and 2 indicates the 2 nd convolutional layer of the residual block;
Figure BDA0002243907510000054
upper and lower meanings of (A) and Wa 1,2The same; in equation (3), the superscript T denotes the number of layers in the path of the network, the maximum number of layers being denoted by T; wa t,1Upper and lower meaning of (1) and Wa 1,2The same;
Figure BDA0002243907510000055
the upper and lower designations of (A) and (B)
Figure BDA0002243907510000056
The same;
Figure BDA0002243907510000057
the following table CLSTM of (a) indicates that the feature is the output of the t-th layer CLSTM, and a indicates the feature output to the spatial feature path;
step 22: constructing a spectral characteristic path: the spectral characteristic path is similar to the spatial characteristic path, and the path consists of a 3D stem layer and T3D residual blocks; in the 3D stem layer, a 3D convolutional layer and a prerlu active layer are first used to extract the features of the multispectral image, as shown in equation (4):
Figure BDA0002243907510000058
in the equation, Fe 0Layer 0 features representing spectral feature path extraction; 0 represents a 0 th layer; e is used for representing a spectral feature path;
then constructing a 3D residual block: each 3D residual block comprises two 3D convolutional layers and two PReLU activation functions; wherein the input of the 1 st 3D residual block is Fe 0The inputs of the other layers are
Figure BDA0002243907510000059
And
Figure BDA00022439075100000510
as shown in equations (5) and (6):
Figure BDA00022439075100000511
Figure BDA00022439075100000512
in equation (5), We 1,2Subscripts of (A) and
Figure BDA00022439075100000513
the meanings are the same; we 1,2The superscripts 1,2 of (a) indicate the number of layers, where 1 indicates the 1 st layer of the spectral signature path and 2 indicates the 2 nd convolutional layer of that layer;
Figure BDA00022439075100000514
upper and lower meanings of (A) and We 1,2The same is true. In equation (6), the superscript t denotes the number of layers in the path of the network;
Figure BDA00022439075100000515
upper and lower meaning of (1) and We 1,2The same;
Figure BDA00022439075100000516
the upper and lower designations of (A) and (B)
Figure BDA00022439075100000517
The same;
Figure BDA00022439075100000518
the following table CLSTM of (a) indicates that the feature is the output of the t-th layer CLSTM;
(3) building hierarchical CLSTM networks
Step 31: construction of forgetting Gate ftThe door forgets the state information; the CLSTM network has T layers, all CLSTMs share parameters, and the number of layers of the CLSTM network is calculated from 1; three gates in the network are a forgetting gate, an input gate and an output gate respectively; the construction of the forgetting gate is shown in equation (7):
Figure BDA0002243907510000061
wherein t represents the number of layers; ct-1State information representing the previous layer, the initialization of the feature being 0; ht-1Representing the history information of the previous layer, and the initialization of the characteristic is also 0;
Figure BDA0002243907510000062
and
Figure BDA0002243907510000063
features of the t-th layer in the spatial feature path and the spectral feature path, respectively; w represents a weight, and this parameter is shared in the T layer; subscript f indicates that the parameter is a forgetting gate parameter; b is an offset term, and the parameters are shared in the same way;
step 32: build input Gate itThe gate selects an input feature; as shown in equation (8):
Figure BDA0002243907510000064
wherein the subscript i indicates that the parameter is a parameter of the input gate;
step 33: build output gate otAs shown in equation (9):
Figure BDA0002243907510000065
wherein the subscript o indicates that the parameter is a parameter of the output gate;
step 34: the state information is updated as shown in equation (10):
Figure BDA0002243907510000066
wherein the subscript c indicates that the parameter is a parameter to the status information update procedure; ctIs an updated status feature; tanh is an activation function;
step 35: extracting output characteristics H from state information in combination with output gatetAs shown in equation (11):
Figure BDA0002243907510000067
step 36: the output information is passed to the spectral signature path, as shown in equation (12):
Figure BDA0002243907510000068
step 37: converting the output information 3D output information into 2D data used by the spatial signature path, and passing to the spatial signature path, as shown in equation (13):
Figure BDA0002243907510000071
wherein the view () function is to splice the features of the same spatial location;
(4) building reconstruction modules
Step 41: the high spatial resolution multispectral image is finally generated by a reconstruction module; this module consists of a deconvolution layer, as shown in equation (14):
Figure BDA0002243907510000072
whereinTRepresenting a deconvolution;
and (3) outputting: high spatial resolution multiple lightSpectral image MSR
(5) Back propagation tuning parameters
Step 51: the Loss function Loss is constructed as shown in equation (15):
Figure BDA0002243907510000073
wherein S represents the number of pairs of simulated training images; i | · | purple wind1Represents an L1 paradigm; i | · | purple wind2Represents an L2 paradigm; λ is the equilibrium error term | | | MSR,s-MSs||1And the regular term | | W | | ceiling2The parameters of (1); s represents an index of the image pair;
step 52: calculating an optimal panchromatic sharpening network parameter { W, b } by using an Adam optimization algorithm;
and (3) outputting: training the finished panchromatic sharpening network;
2. fusion of full-color and multi-spectral images
Inputting: image block set F0The size of the MS is H × W × S, and the size of the PAN is 4H × 4W × 1, H, W and S respectively represent the height, width and number of channels of the multispectral image;
(1) building a data set
Carrying out bilinear interpolation up-sampling on the multispectral image MS, thereby obtaining the multispectral image with the same height and width as the PAN
Figure BDA0002243907510000074
(2) Constructing a dual-path network
Step 61: constructing a spatial characteristic path; firstly, constructing a stem layer; the full color image is characterized in the stem layer using one 2D convolutional layer and the prerlu active layer as shown in equation (16):
Figure BDA0002243907510000081
then, constructing T residual blocks: the 1 st residual block has an input of Fa 0The remaining residaualblock inputs are
Figure BDA0002243907510000082
And
Figure BDA0002243907510000083
as shown in equations (17) and (18):
Figure BDA0002243907510000084
Figure BDA0002243907510000085
wherein
Figure BDA0002243907510000086
Indicating that the feature is a feature output to the spatial feature path a by the t-th layer CLSTM;
step 62: constructing a spectral characteristic path; the spectral signature path is similar to the spatial signature path, which consists of one 3D stem layer and T3D residual blocks, as shown in equation (19):
Figure BDA0002243907510000087
wherein Fe 0Layer 0 features representing spectral feature path extraction; e is used for representing a spectral feature path; the 1 st 3 Dresideal block has an input of Fe 0The remaining residual block is input
Figure BDA0002243907510000088
And
Figure BDA0002243907510000089
as shown in equations (20) and (21):
Figure BDA00022439075100000810
Figure BDA00022439075100000811
(3) building hierarchical CLSTM networks
Step 71: construction of forgetting Gate ftThe door forgets the state information; the CLSTM network has T layers, all CLSTMs share parameters, and the number of layers of the CLSTM network is calculated from 1; three gates in the network are a forgetting gate, an input gate and an output gate respectively; the construction of the forgetting gate is shown as equation (22):
Figure BDA00022439075100000812
wherein t represents the number of layers; ct-1State information representing the previous layer, the initialization of the feature being 0; ht-1Representing the history information of the previous layer, and the initialization of the characteristic is also 0;
Figure BDA00022439075100000813
and
Figure BDA00022439075100000814
features of the t-th layer in the spatial feature path and the spectral feature path, respectively; w represents a weight, and this parameter is shared in the T layer; subscript f indicates that the parameter is a forgetting gate parameter; b is an offset term, and the parameters are shared in the same way;
step 72: build input Gate itThe gate selects an input feature; as shown in equation (23):
Figure BDA0002243907510000091
wherein the subscript i indicates that the parameter is a parameter of the input gate;
step 73: build output gate otAs shown in equation (24):
Figure BDA0002243907510000092
wherein the subscript o indicates that the parameter is a parameter of the output gate;
step 74: the state information is updated as shown in equation (25):
Figure BDA0002243907510000093
wherein the subscript c indicates that the parameter is a parameter to the status information update procedure; ctIs an updated status feature; tanh is an activation function;
step 75: extracting output characteristics H from state information in combination with output gatetAs shown in equation (26):
Figure BDA0002243907510000094
step 76: the output information is passed to the spectral signature path as shown in equation (27):
Figure BDA0002243907510000095
step 77: converting the output information 3D output information into 2D data used by the spatial signature path, and passing to the spatial signature path, as shown in equation (28):
Figure BDA0002243907510000096
wherein the view () function is to splice the features of the same spatial location;
(4) building reconstruction modules
Step 81: the high spatial resolution multispectral image is finally generated by a reconstruction module; this module consists of a deconvolution layer, as shown in equation (29):
Figure BDA0002243907510000097
(5) forward propagation to obtain multispectral images
Step 91: PAN and multispectral images of full-color images
Figure BDA0002243907510000101
Inputting the result into a panchromatic sharpening network, and acquiring a result MS of forward propagation of the panchromatic sharpening networkR
And (3) outputting: high spatial resolution multispectral image MSR
Advantageous effects
The multispectral image panchromatic sharpening method based on the dual-channel convolution network and the hierarchical CLSTM provided by the invention fully utilizes abundant spectral information in the multispectral image and spatial detail information in the panchromatic image. The multispectral and panchromatic image features extracted by the algorithm are extracted and fused by utilizing 2D and 3D networks respectively, and features of different depths are selected and extracted by combining with hierarchical CLSTM, so that the visual quality of a reconstructed image is effectively improved, and the structural features such as edges, textures and the like of the image are more effectively reconstructed in a spatial domain. 2D spatial information and 3D multispectral information are processed by utilizing 2D and 3D networks, and features of different depths are fused on multiple levels, so that a multispectral image with high spatial resolution can be well reconstructed.
Drawings
FIG. 1 is a flow chart of the present invention.
Fig. 2 shows a specific network structure of the present invention.
Fig. 3 shows a specific structure of CLSTM.
Detailed Description
The invention will now be further described with reference to the following examples and drawings:
the invention provides a multispectral image panchromatic sharpening method based on a dual-channel convolution network and a hierarchical CLSTM. The method comprises two parts of model training and multispectral image panchromatic sharpening. In the model training stage, firstly, the original clear multispectral and panchromatic images are subjected to down-sampling to obtain a simulation training image pair; secondly, extracting and fusing the characteristics of the panchromatic image and the multispectral image by using a dual-channel convolution network, and realizing the fusion between convolution characteristics of multiple layers and different depths by combining with a layer CLSTM; then, reconstructing a multispectral image with high spatial resolution from the fused features by using a deconvolution network; and finally, adjusting the parameters of the model by using an Adam algorithm. In the multispectral image panchromatic sharpening stage, firstly, the trained dual-channel convolution network and the hierarchical CLSTM are used for extracting and fusing the characteristics of panchromatic images and multispectral images. The specific network structure is shown in fig. 2, the convolution network is responsible for extracting the characteristics of the multispectral image and the panchromatic image and fusing the characteristics selected by the convolution network and the characteristics selected by the CLSTM, and the CLSTM selects and memorizes the characteristics at different depths in multiple levels, so that the fusion of the characteristics at multiple levels and different depths is realized.
The specific implementation flow is as follows:
1. training of fusion models
Inputting: image block set F0Where the original multispectral MS is of size H × W × S and the PAN is of size 4H × 4W × 1, H, W and S representing the height, width and number of spectra of the original multispectral image, respectively.
(1) Constructing a simulated training dataset
Step 1: the original multispectral image MS is down-sampled to obtain a simulated multispectral image block
Figure BDA0002243907510000111
The size of the image block is
Figure BDA0002243907510000112
Step 2: multispectral image to be downsampled
Figure BDA0002243907510000113
Performing bilinear interpolation up-sampling to obtain multispectral image with the same height and width as MS
Figure BDA0002243907510000114
And step 3: downsampling original full-color image PAN to obtain full-color image
Figure BDA0002243907510000115
Figure BDA0002243907510000116
With simulated upsampled multi-spectral images
Figure BDA0002243907510000117
Are the same in height and width.
(2) Constructing a dual-path network
Step 1: and constructing a spatial feature passage. The spatial signature path is used to process full color image information and is located in the upper box in the middle of fig. 2. The spatial signature path comprises a stem layer and T2D residual blocks. In the stem layer, a 2D convolutional layer and a prerlu active layer are first used to extract the features of a full color image, as shown in equation (1). In the equation, Fa 0Layer 0 features representing spatial feature path extraction, F is an abbreviation for feature; 0 represents a 0 th layer; a is used for representing a spatial feature path; then, T2D residual blocks are constructed. The 1 st 2D residual block is input
Figure BDA0002243907510000118
The 2 Dresideal block is input as
Figure BDA0002243907510000119
And
Figure BDA00022439075100001110
as shown in equations (2) and (3); in equation (3), the superscript T denotes the number of layers in the path of the network, the maximum number of layers being denoted by T;
Figure BDA00022439075100001111
the subscript CLSTM of (a) indicates that the feature is the output of the t-th layer CLSTM, and a indicates the feature output to the spatial feature path.
Figure BDA0002243907510000121
Figure BDA0002243907510000122
Figure BDA0002243907510000123
Step 2: constructing a spectral characteristic path; the spectral feature path is used for processing multispectral information and is positioned in a lower frame in the middle of the graph 2; the spectral signature path is similar to the spatial signature path, which consists of one 3D stem layer and T3D residalblock. Firstly, extracting the characteristics of the multispectral image by using a 3D convolution layer and a PReLU activation layer in a 3D stem layer, as shown in an equation (4); in the equation, the first and second phases are,
Figure BDA0002243907510000124
layer 0 features representing spectral feature path extraction; 0 represents a 0 th layer; e is used for representing a spectral feature path; then constructing a 3D residual block; wherein the 1 st 3D residual block has the input of
Figure BDA0002243907510000125
The inputs to the remaining layers are
Figure BDA0002243907510000126
And
Figure BDA0002243907510000127
as shown in equations (5) and (6);
Figure BDA0002243907510000128
Figure BDA0002243907510000129
Figure BDA00022439075100001210
(3) building hierarchical CLSTM networks
Step 1: construction of forgetting Gate ftThe door forgets the state information; the hierarchical CLSTM network has a T layer, all CLSTMs share parameters, and the structure of the hierarchical CLSTM network is shown in FIG. 3; the number of layers of the CLSTM network is counted from 1Calculating; three gates in the network are a forgetting gate, an input gate and an output gate respectively; the construction of the forgetting gate is shown in equation (7); wherein t represents the number of layers; ct-1State information representing the previous layer, the initialization of the feature being 0; ht-1Representing the history information of the previous layer, and the initialization of the characteristic is also 0;
Figure BDA00022439075100001211
and
Figure BDA00022439075100001212
features of the t-th layer in the spatial feature path and the spectral feature path, respectively; w represents a weight, and this parameter is shared in the T layer; subscript f indicates that the parameter is a forgetting gate parameter; b is an offset term, and the parameters are shared in the same way;
Figure BDA00022439075100001213
step 2: build input Gate itThe gate selects the input feature. As shown in equation (8), where the subscript i indicates that the parameter is a parameter of the input gate;
Figure BDA0002243907510000131
and step 3: build output gate otAs shown in equation (9); wherein the subscript o indicates that the parameter is a parameter of the output gate;
Figure BDA0002243907510000132
and 4, step 4: updating the state information as shown in equation (10); wherein the subscript c indicates that the parameter is a parameter to the status information update procedure; ctIs an updated status feature; tanh is an activation function;
Figure BDA0002243907510000133
step (ii) of5: extracting output characteristics H from state information in combination with output gatetAs shown in equation (11);
Figure BDA0002243907510000134
step 6: passing the output information to the spectral signature path, as shown in equation (12);
Figure BDA0002243907510000135
and 7: converting the output information 3D output information into 2D data used by a spatial feature path, and transmitting the 2D data to the spatial feature path, as shown in equation (13); the view () function is to splice the features of the same spatial position;
Figure BDA0002243907510000136
(4) building reconstruction modules
Step 1: as shown in fig. 2, the high spatial resolution multispectral image is ultimately generated by the reconstruction module; this module consists of a deconvolution layer, as shown in equation (14).
Figure BDA0002243907510000137
(5) Back propagation tuning parameters
Step 1: constructing a Loss function Loss, as shown in equation (15); s represents the number of pairs of simulated training images; i | · | purple wind1Represents an L1 paradigm; i | · | purple wind2Represents an L2 paradigm; λ is the equilibrium error term | | | MSR,s-MSs||1And the regular term | | W | | ceiling2The parameters of (1); s represents an index of the image pair;
Figure BDA0002243907510000141
step 2: and calculating the optimal network parameters W, b by using an Adam optimization algorithm.
And (3) outputting: a well-learned network.
2. Fusion of full-color and multi-spectral images
Inputting: image block set F0Where the size of the MS is H × W × S and the size of the PAN is 4H × 4W × 1, H, W and S respectively represent the height, width and number of channels of the multispectral image.
(1) Building a data set
Carrying out bilinear interpolation up-sampling on the multispectral image MS, thereby obtaining the multispectral image with the same height and width as the PAN
Figure BDA0002243907510000142
(2) Constructing a dual-path network
Step 1: constructing a spatial characteristic path; firstly, constructing a stem layer; extracting the feature of the full color image using one 2D convolution layer and the prerlu active layer in the stem layer as shown in equation (16); then constructing T residual blocks; the 1 st residual block has an input of Fa 0The remaining residual block is input
Figure BDA0002243907510000143
And
Figure BDA0002243907510000144
as shown in equations (17) and (18);
Figure BDA0002243907510000145
indicating that the feature is a feature output by the t-th layer CLSTM to the spatial feature path a.
Figure BDA0002243907510000146
Figure BDA0002243907510000147
Figure BDA0002243907510000148
Step 2: constructing a spectral characteristic path; the spectral feature path is used for processing multispectral information and is positioned in a part of light cyan shadow in the graph 2; the spectral signature path is similar to the spatial signature path, which consists of one 3D stem layer and T3 dresidic blocks, as shown in equation (19); fe 0Layer 0 features representing spectral feature path extraction; e is used for representing a spectral feature path; the input of the 1 st 3D residual block is Fe 0The remaining residual block is input
Figure BDA0002243907510000151
And
Figure BDA0002243907510000152
as shown in equations (20) and (21).
Figure BDA0002243907510000153
Figure BDA0002243907510000154
Figure BDA0002243907510000155
(3) Building hierarchical CLSTM networks
Step 1: construction of forgetting Gate ftThe door forgets the state information; the hierarchical CLSTM network has a T layer, all CLSTMs share parameters, and the structure of the hierarchical CLSTM network is shown in FIG. 3; the number of layers of the CLSTM network is calculated from 1; three gates in the network are a forgetting gate, an input gate and an output gate respectively; the construction of the forgetting gate is shown in equation (22); wherein t represents the number of layers; ct-1State information representing the previous layer, the initialization of the feature being 0; ht-1Representing the history information of the previous layer, and the initialization of the characteristic is also 0;
Figure BDA0002243907510000156
and
Figure BDA0002243907510000157
features of the t-th layer in the spatial feature path and the spectral feature path, respectively; w represents a weight, and this parameter is shared in the T layer; subscript f indicates that the parameter is a forgetting gate parameter; b is an offset term, and the parameters are shared in the same way;
Figure BDA0002243907510000158
step 2: build input Gate itThe gate selects the input feature. As shown in equation (23), where the subscript i indicates that the parameter is a parameter of the input gate;
Figure BDA0002243907510000159
and step 3: build output gate otAs shown in equation (24); wherein the subscript o indicates that the parameter is a parameter of the output gate;
Figure BDA00022439075100001510
and 4, step 4: updating the state information as shown in equation (25); wherein the subscript c indicates that the parameter is a parameter to the status information update procedure; ctIs an updated status feature; tanh is an activation function;
Figure BDA00022439075100001511
and 5: extracting output characteristics H from state information in combination with output gatetAs shown in equation (26);
Figure BDA0002243907510000161
step 6: passing the output information to the spectral signature path, as shown in equation (27);
Figure BDA0002243907510000162
and 7: converting the output information 3D output information into 2D data used by a spatial feature path, and transmitting the 2D data to the spatial feature path, as shown in equation (28); the view () function is to splice the features of the same spatial position;
Figure BDA0002243907510000163
(4) building reconstruction modules
Step 1: as shown in fig. 2, the high spatial resolution multispectral image is ultimately generated by the reconstruction module; this module consists of a deconvolution layer, as shown in equation (29).
Figure BDA0002243907510000164
(5) Forward propagation to obtain multispectral images
Step 1: PAN and multispectral images of full-color images
Figure BDA0002243907510000165
Inputting into network, obtaining result MS of network forward propagationR
And (3) outputting: high spatial resolution multispectral image MSR

Claims (1)

1. A multispectral image panchromatic sharpening method based on a dual-channel convolution network and a hierarchical CLSTM is characterized by comprising the following steps:
1. training of fusion models
Inputting: image block set F0(MS, PAN), where the original multispectral MS is of size H × W × S and the PAN is of size 4H × 4W × 1, H, W and S representing the height, width and number of spectra of the original multispectral image, respectively;
(1) constructing a simulated training dataset
Step 11: the original multispectral image MS is down-sampled to obtain a simulated multispectral image block
Figure FDA0002243907500000011
The size of the image block is
Figure FDA0002243907500000012
Step 12: multispectral image to be downsampled
Figure FDA0002243907500000013
Performing bilinear interpolation up-sampling to obtain multispectral image with the same height and width as MS
Figure FDA0002243907500000014
Step 13: downsampling original full-color image PAN to obtain full-color image
Figure FDA0002243907500000015
Figure FDA0002243907500000016
With simulated upsampled multi-spectral images
Figure FDA0002243907500000017
Are the same in height and width;
(2) constructing a dual-path network
Step 21: constructing a spatial characteristic path; firstly, constructing a stem layer; the full color image is characterized in the stem layer using one 2D convolutional layer and the prerlu active layer as shown in equation (1):
Figure FDA0002243907500000018
wherein the content of the first and second substances,
Figure FDA0002243907500000019
level 0 features representing spatial feature path extraction, F is an abbreviation for feature(ii) a 0 represents a 0 th layer; a is used for representing a spatial feature path; wherein W is an abbreviation for weight, which is a parameter of the convolutional layer; b is a bias term;
then constructing T residual blocks, wherein each residual block comprises two 2D convolution layers and two PReLU activation functions; the 1 st Residualblock input is
Figure FDA00022439075000000110
The remaining residaualblock inputs are
Figure FDA00022439075000000111
And
Figure FDA00022439075000000112
as shown in equations (2) and (3):
Figure FDA00022439075000000113
Figure FDA00022439075000000114
in the case of the equation (2),
Figure FDA0002243907500000021
subscripts of (A) and
Figure FDA0002243907500000022
the meanings are the same;
Figure FDA0002243907500000023
the superscript 1,2 of (a) indicates the number of layers, where 1 indicates the 1 st layer of the spatial signature path, i.e., the 1 st residual block, and 2 indicates the 2 nd convolutional layer of the residual block;
Figure FDA0002243907500000024
the upper and lower indices of
Figure FDA0002243907500000025
The same; in equation (3), the superscript T denotes the number of layers in the path of the network, the maximum number of layers being denoted by T;
Figure FDA0002243907500000026
the upper and lower designations of (A) and (B)
Figure FDA0002243907500000027
The same;
Figure FDA0002243907500000028
the upper and lower designations of (A) and (B)
Figure FDA0002243907500000029
The same;
Figure FDA00022439075000000210
the following table CLSTM of (a) indicates that the feature is the output of the t-th layer CLSTM, and a indicates the feature output to the spatial feature path;
step 22: constructing a spectral characteristic path: the spectral characteristic path is similar to the spatial characteristic path, and the path consists of a 3D stem layer and T3D residual blocks; in the 3D stem layer, a 3D convolutional layer and a prerlu active layer are first used to extract the features of the multispectral image, as shown in equation (4):
Figure FDA00022439075000000211
in the equation, the first and second phases are,
Figure FDA00022439075000000212
layer 0 features representing spectral feature path extraction; 0 represents a 0 th layer; e is used for representing a spectral feature path;
then constructing a 3D residual block: each 3D residual block comprises two 3D convolutional layers and two PReLU activation functions; wherein the 1 st 3D residual block has the input of
Figure FDA00022439075000000213
The inputs to the remaining layers are
Figure FDA00022439075000000214
And
Figure FDA00022439075000000215
as shown in equations (5) and (6):
Figure FDA00022439075000000216
Figure FDA00022439075000000217
in the case of the equation (5),
Figure FDA00022439075000000218
subscripts of (A) and
Figure FDA00022439075000000219
the meanings are the same;
Figure FDA00022439075000000220
the superscripts 1,2 of (a) indicate the number of layers, where 1 indicates the 1 st layer of the spectral signature path and 2 indicates the 2 nd convolutional layer of that layer;
Figure FDA00022439075000000221
the upper and lower indices of
Figure FDA00022439075000000222
The same is true. In equation (6), the superscript t denotes the number of layers in the path of the network;
Figure FDA00022439075000000223
the upper and lower designations of (A) and (B)
Figure FDA00022439075000000224
The same;
Figure FDA00022439075000000225
the upper and lower designations of (A) and (B)
Figure FDA00022439075000000226
The same;
Figure FDA00022439075000000227
the following table CLSTM of (a) indicates that the feature is the output of the t-th layer CLSTM;
(3) building hierarchical CLSTM networks
Step 31: construction of forgetting Gate ftThe door forgets the state information; the CLSTM network has T layers, all CLSTMs share parameters, and the number of layers of the CLSTM network is calculated from 1; three gates in the network are a forgetting gate, an input gate and an output gate respectively; the construction of the forgetting gate is shown in equation (7):
Figure FDA0002243907500000031
wherein t represents the number of layers; ct-1State information representing the previous layer, the initialization of the feature being 0; ht-1Representing the history information of the previous layer, and the initialization of the characteristic is also 0;
Figure FDA0002243907500000032
and
Figure FDA0002243907500000033
features of the t-th layer in the spatial feature path and the spectral feature path, respectively; w represents a weight, and this parameter is shared in the T layer; subscript f indicates that the parameter is a forgetting gate parameter; b is an offset term, and the parameters are shared in the same way;
step 32: build input Gate itThe gate selects an input feature; as shown in equation (8):
Figure FDA0002243907500000034
wherein the subscript i indicates that the parameter is a parameter of the input gate;
step 33: build output gate otAs shown in equation (9):
Figure FDA0002243907500000035
wherein the subscript o indicates that the parameter is a parameter of the output gate;
step 34: the state information is updated as shown in equation (10):
Figure FDA0002243907500000036
wherein the subscript c indicates that the parameter is a parameter to the status information update procedure; ctIs an updated status feature; tanh is an activation function;
step 35: extracting output characteristics H from state information in combination with output gatetAs shown in equation (11):
Figure FDA0002243907500000037
step 36: the output information is passed to the spectral signature path, as shown in equation (12):
Figure FDA0002243907500000038
step 37: converting the output information 3D output information into 2D data used by the spatial signature path, and passing to the spatial signature path, as shown in equation (13):
Figure FDA0002243907500000039
wherein the view () function is to splice the features of the same spatial location;
(4) building reconstruction modules
Step 41: the high spatial resolution multispectral image is finally generated by a reconstruction module; this module consists of a deconvolution layer, as shown in equation (14):
Figure FDA0002243907500000041
whereinTRepresenting a deconvolution;
and (3) outputting: high spatial resolution multispectral image MSR
(5) Back propagation tuning parameters
Step 51: the Loss function Loss is constructed as shown in equation (15):
Figure FDA0002243907500000042
wherein S represents the number of pairs of simulated training images; i | · | purple wind1Represents an L1 paradigm; i | · | purple wind2Represents an L2 paradigm; λ is the equilibrium error term | | | MSR,s-MSs||1And the regular term | | W | | ceiling2The parameters of (1); s represents an index of the image pair;
step 52: calculating an optimal panchromatic sharpening network parameter { W, b } by using an Adam optimization algorithm;
and (3) outputting: training the finished panchromatic sharpening network;
2. fusion of full-color and multi-spectral images
Inputting: image block set F0The size of the MS is H × W × S, and the size of the PAN is 4H × 4W × 1, H, W and S respectively represent the height, width and number of channels of the multispectral image;
(1) building a data set
Carrying out bilinear interpolation up-sampling on the multispectral image MS, thereby obtaining the multispectral image with the same height and width as the PAN
Figure FDA0002243907500000043
(2) Constructing a dual-path network
Step 61: constructing a spatial characteristic path; firstly, constructing a stem layer; the full color image is characterized in the stem layer using one 2D convolutional layer and the prerlu active layer as shown in equation (16):
Figure FDA0002243907500000044
then, constructing T residual blocks: the 1 st residual block is input
Figure FDA0002243907500000051
The remaining residaualblock inputs are
Figure FDA0002243907500000052
And
Figure FDA0002243907500000053
as shown in equations (17) and (18):
Figure FDA0002243907500000054
Figure FDA0002243907500000055
wherein
Figure FDA00022439075000000516
Indicating that the feature is a feature output to the spatial feature path a by the t-th layer CLSTM;
step 62: constructing a spectral characteristic path; the spectral signature path is similar to the spatial signature path, which consists of one 3D stem layer and T3D residual blocks, as shown in equation (19):
Figure FDA0002243907500000056
wherein
Figure FDA0002243907500000057
Layer 0 features representing spectral feature path extraction; e is used for representing a spectral feature path; the 1 st 3 Dresideal block has as input
Figure FDA0002243907500000058
The remaining residual block is input as
Figure FDA0002243907500000059
And
Figure FDA00022439075000000510
as shown in equations (20) and (21):
Figure FDA00022439075000000511
Figure FDA00022439075000000512
(3) building hierarchical CLSTM networks
Step 71: construction of forgetting Gate ftThe door forgets the state information; the CLSTM network has T layers, all CLSTMs share parameters, and the number of layers of the CLSTM network is calculated from 1; three gates in the network are a forgetting gate, an input gate and an output gate respectively; the construction of the forgetting gate is shown as equation (22):
Figure FDA00022439075000000513
wherein t represents the number of layers; ct-1State information representing the previous layer, the initialization of the feature being 0; ht-1Representing the history information of the previous layer, and the initialization of the characteristic is also 0;
Figure FDA00022439075000000514
and
Figure FDA00022439075000000515
features of the t-th layer in the spatial feature path and the spectral feature path, respectively; w represents a weight, and this parameter is shared in the T layer; subscript f indicates that the parameter is a forgetting gate parameter; b is an offset term, and the parameters are shared in the same way;
step 72: build input Gate itThe gate selects an input feature; as shown in equation (23):
Figure FDA0002243907500000061
wherein the subscript i indicates that the parameter is a parameter of the input gate;
step 73: build output gate otAs shown in equation (24):
Figure FDA0002243907500000062
wherein the subscript o indicates that the parameter is a parameter of the output gate;
step 74: the state information is updated as shown in equation (25):
Figure FDA0002243907500000063
wherein the subscript c indicates that the parameter is a parameter to the status information update procedure; ctIs an updated status feature; tanh is an activation function;
step 75: extracting output characteristics H from state information in combination with output gatetAs shown in equation (26):
Figure FDA0002243907500000064
step 76: the output information is passed to the spectral signature path as shown in equation (27):
Figure FDA0002243907500000065
step 77: converting the output information 3D output information into 2D data used by the spatial signature path, and passing to the spatial signature path, as shown in equation (28):
Figure FDA0002243907500000066
wherein the view () function is to splice the features of the same spatial location;
(4) building reconstruction modules
Step 81: the high spatial resolution multispectral image is finally generated by a reconstruction module; this module consists of a deconvolution layer, as shown in equation (29):
Figure FDA0002243907500000067
(5) forward propagation to obtain multispectral images
Step 91: PAN and multispectral images of full-color images
Figure FDA0002243907500000068
Inputting the result into a panchromatic sharpening network, and acquiring a result MS of forward propagation of the panchromatic sharpening networkR
And (3) outputting: high spatial resolution multispectral image MSR
CN201911009926.6A 2019-10-23 2019-10-23 Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM Active CN110930315B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911009926.6A CN110930315B (en) 2019-10-23 2019-10-23 Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911009926.6A CN110930315B (en) 2019-10-23 2019-10-23 Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM

Publications (2)

Publication Number Publication Date
CN110930315A true CN110930315A (en) 2020-03-27
CN110930315B CN110930315B (en) 2022-02-11

Family

ID=69849322

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911009926.6A Active CN110930315B (en) 2019-10-23 2019-10-23 Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM

Country Status (1)

Country Link
CN (1) CN110930315B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112861774A (en) * 2021-03-04 2021-05-28 山东产研卫星信息技术产业研究院有限公司 Method and system for identifying ship target by using remote sensing image
CN113139902A (en) * 2021-04-23 2021-07-20 深圳大学 Hyperspectral image super-resolution reconstruction method and device and electronic equipment
CN113902650A (en) * 2021-12-07 2022-01-07 南湖实验室 Remote sensing image sharpening method based on parallel deep learning network architecture
CN114581347A (en) * 2022-01-24 2022-06-03 中国科学院空天信息创新研究院 Optical remote sensing space spectrum fusion method, device, equipment and medium without reference image

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170032221A1 (en) * 2015-07-29 2017-02-02 Htc Corporation Method, electronic apparatus, and computer readable medium of constructing classifier for disease detection
CN106611410A (en) * 2016-11-29 2017-05-03 北京空间机电研究所 Pansharpen fusion optimization method based on pyramid model
CN107808131A (en) * 2017-10-23 2018-03-16 华南理工大学 Dynamic gesture identification method based on binary channel depth convolutional neural networks
CN108038501A (en) * 2017-12-08 2018-05-15 桂林电子科技大学 Hyperspectral image classification method based on multi-modal compression bilinearity pond
CN109102469A (en) * 2018-07-04 2018-12-28 华南理工大学 A kind of panchromatic sharpening method of remote sensing images based on convolutional neural networks
CN109146831A (en) * 2018-08-01 2019-01-04 武汉大学 Remote sensing image fusion method and system based on double branch deep learning networks
CN109272010A (en) * 2018-07-27 2019-01-25 吉林大学 Multi-scale Remote Sensing Image fusion method based on convolutional neural networks
CN109767412A (en) * 2018-12-28 2019-05-17 珠海大横琴科技发展有限公司 A kind of remote sensing image fusing method and system based on depth residual error neural network
CN109858540A (en) * 2019-01-24 2019-06-07 青岛中科智康医疗科技有限公司 A kind of medical image recognition system and method based on multi-modal fusion
CN109886870A (en) * 2018-12-29 2019-06-14 西北大学 Remote sensing image fusion method based on binary channels neural network
CN109886929A (en) * 2019-01-24 2019-06-14 江苏大学 A kind of MRI tumour voxel detection method based on convolutional neural networks
CN110189282A (en) * 2019-05-09 2019-08-30 西北工业大学 Based on intensive and jump connection depth convolutional network multispectral and panchromatic image fusion method

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170032221A1 (en) * 2015-07-29 2017-02-02 Htc Corporation Method, electronic apparatus, and computer readable medium of constructing classifier for disease detection
CN106611410A (en) * 2016-11-29 2017-05-03 北京空间机电研究所 Pansharpen fusion optimization method based on pyramid model
CN107808131A (en) * 2017-10-23 2018-03-16 华南理工大学 Dynamic gesture identification method based on binary channel depth convolutional neural networks
CN108038501A (en) * 2017-12-08 2018-05-15 桂林电子科技大学 Hyperspectral image classification method based on multi-modal compression bilinearity pond
CN109102469A (en) * 2018-07-04 2018-12-28 华南理工大学 A kind of panchromatic sharpening method of remote sensing images based on convolutional neural networks
CN109272010A (en) * 2018-07-27 2019-01-25 吉林大学 Multi-scale Remote Sensing Image fusion method based on convolutional neural networks
CN109146831A (en) * 2018-08-01 2019-01-04 武汉大学 Remote sensing image fusion method and system based on double branch deep learning networks
CN109767412A (en) * 2018-12-28 2019-05-17 珠海大横琴科技发展有限公司 A kind of remote sensing image fusing method and system based on depth residual error neural network
CN109886870A (en) * 2018-12-29 2019-06-14 西北大学 Remote sensing image fusion method based on binary channels neural network
CN109858540A (en) * 2019-01-24 2019-06-07 青岛中科智康医疗科技有限公司 A kind of medical image recognition system and method based on multi-modal fusion
CN109886929A (en) * 2019-01-24 2019-06-14 江苏大学 A kind of MRI tumour voxel detection method based on convolutional neural networks
CN110189282A (en) * 2019-05-09 2019-08-30 西北工业大学 Based on intensive and jump connection depth convolutional network multispectral and panchromatic image fusion method

Non-Patent Citations (7)

* Cited by examiner, † Cited by third party
Title
ARIAN AZARANG ET AL: "A new pansharpening method using multi resolution analysis framework and deep neural networks", 《2017 3RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION AND IMAGE ANALYSIS (IPRIA)》 *
CHAO TAN ET AL: "FORECAST-CLSTM: A New Convolutional LSTM Network for Cloudage Nowcasting", 《2018 IEEE VISUAL COMMUNICATIONS AND IMAGE PROCESSING (VCIP)》 *
QIANGQIANG YUAN ET AL: "A Multiscale and Multidepth Convolutional Neural Network for Remote Sensing Imagery Pan-Sharpening", 《 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 》 *
XIANGYU LIU ET AL: "Remote Sensing Image Fusion Based on Two-stream Fusion Network", 《PREPRINT SUBMITTED TO INFORMATION FUSION》 *
XUN WANG ET AL: "Semantic annotation for complex video street views based on 2D-3D multi-feature fusion and aggregated boosting decision forests", 《PREPRINT SUBMITTED TO PATTERN RECOGNITION》 *
宋立飞等: "多尺度输入3D卷积融合双流模型的行为识别方法", 《计算机辅助设计与图形学学报》 *
秦峰等: "融合主题的CLSTM短文本情感分类", 《安徽工业大学学报(自然科学版)》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112861774A (en) * 2021-03-04 2021-05-28 山东产研卫星信息技术产业研究院有限公司 Method and system for identifying ship target by using remote sensing image
CN113139902A (en) * 2021-04-23 2021-07-20 深圳大学 Hyperspectral image super-resolution reconstruction method and device and electronic equipment
CN113902650A (en) * 2021-12-07 2022-01-07 南湖实验室 Remote sensing image sharpening method based on parallel deep learning network architecture
CN114581347A (en) * 2022-01-24 2022-06-03 中国科学院空天信息创新研究院 Optical remote sensing space spectrum fusion method, device, equipment and medium without reference image
CN114581347B (en) * 2022-01-24 2024-03-12 中国科学院空天信息创新研究院 Optical remote sensing spatial spectrum fusion method, device, equipment and medium without reference image

Also Published As

Publication number Publication date
CN110930315B (en) 2022-02-11

Similar Documents

Publication Publication Date Title
CN110930315B (en) Multispectral image panchromatic sharpening method based on dual-channel convolution network and hierarchical CLSTM
CN109272010B (en) Multi-scale remote sensing image fusion method based on convolutional neural network
CN112184554B (en) Remote sensing image fusion method based on residual mixed expansion convolution
Liao et al. Deep learning for fusion of APEX hyperspectral and full-waveform LiDAR remote sensing data for tree species mapping
CN107123089B (en) Remote sensing image super-resolution reconstruction method and system based on depth convolution network
CN104112263B (en) The method of full-colour image and Multispectral Image Fusion based on deep neural network
Gao et al. Cross-scale mixing attention for multisource remote sensing data fusion and classification
CN113469094A (en) Multi-mode remote sensing data depth fusion-based earth surface coverage classification method
CN109087375B (en) Deep learning-based image cavity filling method
CN113129247B (en) Remote sensing image fusion method and medium based on self-adaptive multi-scale residual convolution
CN110189282A (en) Based on intensive and jump connection depth convolutional network multispectral and panchromatic image fusion method
CN112560624B (en) High-resolution remote sensing image semantic segmentation method based on model depth integration
CN114937202A (en) Double-current Swin transform remote sensing scene classification method
CN113744136A (en) Image super-resolution reconstruction method and system based on channel constraint multi-feature fusion
Hu et al. Hyperspectral image super resolution based on multiscale feature fusion and aggregation network with 3-D convolution
CN116309070A (en) Super-resolution reconstruction method and device for hyperspectral remote sensing image and computer equipment
CN113610905A (en) Deep learning remote sensing image registration method based on subimage matching and application
CN112634184A (en) RGB image spectral reflectivity reconstruction method based on fusion convolution neural network
Dang et al. U-shaped deep-learning models for island ecosystem type classification, a case study in Con Dao Island of Vietnam
CN112529828B (en) Reference data non-sensitive remote sensing image space-time fusion model construction method
CN113902646A (en) Remote sensing image pan-sharpening method based on depth layer feature weighted fusion network
CN114092803A (en) Cloud detection method and device based on remote sensing image, electronic device and medium
CN112686830A (en) Super-resolution method of single depth map based on image decomposition
CN116563606A (en) Hyperspectral image classification method based on dual-branch spatial spectrum global feature extraction network
CN116310828A (en) High-resolution remote sensing image change detection method and device combining transducer and CNN

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant