CN113327304A - Hyperspectral image saliency map generation method based on end-to-end neural network - Google Patents

Hyperspectral image saliency map generation method based on end-to-end neural network Download PDF

Info

Publication number
CN113327304A
CN113327304A CN202110593767.XA CN202110593767A CN113327304A CN 113327304 A CN113327304 A CN 113327304A CN 202110593767 A CN202110593767 A CN 202110593767A CN 113327304 A CN113327304 A CN 113327304A
Authority
CN
China
Prior art keywords
neural network
saliency map
convolution
hyperspectral image
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110593767.XA
Other languages
Chinese (zh)
Inventor
许廷发
黄晨
徐畅
樊阿馨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Institute of Technology BIT
Chongqing Innovation Center of Beijing University of Technology
Original Assignee
Beijing Institute of Technology BIT
Chongqing Innovation Center of Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Institute of Technology BIT, Chongqing Innovation Center of Beijing University of Technology filed Critical Beijing Institute of Technology BIT
Priority to CN202110593767.XA priority Critical patent/CN113327304A/en
Publication of CN113327304A publication Critical patent/CN113327304A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T9/00Image coding
    • G06T9/002Image coding using neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10036Multispectral image; Hyperspectral image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a hyperspectral image saliency map generation method based on an end-to-end neural network, which comprises the following steps of S1: preprocessing an image; s2: constructing an end-to-end neural network model and extracting space spectrum characteristics; s3: training a neural network; the end-to-end neural network model structure is a W2 type convolutional neural network, the neural network model comprises a left coding channel, a right coding channel and a middle decoding channel which are respectively a space coding module, a spectrum coding module and a decoding module, and the model further comprises a result prediction module for generating a prediction saliency map. According to the invention, through the constructed end-to-end neural network model, the deep spatial spectrum characteristics of the image can be extracted, and the prediction saliency map is directly generated, so that the consumption of computing resources is saved, and the robustness of the characteristics is improved.

Description

Hyperspectral image saliency map generation method based on end-to-end neural network
Technical Field
The invention relates to the technical field of image processing, in particular to a hyperspectral image saliency map generation method based on an end-to-end neural network.
Background
The hyperspectral image is composed of tens or hundreds of continuous narrow-band images, can capture spatial and spectral dimensional information of a target scene simultaneously, and is called a data cube. With the development of the hyperspectral imaging technology, the hyperspectral imager can acquire hyperspectral data with higher spatial resolution and spectral resolution. Currently, hyperspectral images have been applied and gained in effect in many fields, such as ground object remote sensing, precision agriculture, medical diagnosis, target detection, and the like.
A saliency map is a model that simulates the human visual attention mechanism, describing salient objects or regions of interest to the human eye in a real scene, also referred to as "regions of interest". Saliency detection is the simulation of visual attention mechanisms through algorithms, extraction of salient regions in images and generation of saliency maps. In the conventional method, the local or global contrast is calculated mainly by using primary features such as color and texture of an image, so as to obtain the significance of a region. In recent years, neural network models have been studied intensively in computer vision for extracting deep features of images, based on which saliency maps of better quality can be generated in saliency detection.
The hyperspectral image has abundant space and spectrum information, but the spectrum data has higher dimensionality and correlation and higher processing difficulty, so that most of the existing hyperspectral image saliency map generation methods are based on shallow spectral features and cannot fully utilize the space spectral information of the hyperspectral image; the existing method based on deep features generally adopts a neural network to extract features and then generates a saliency map through saliency detection, and a two-stage process consumes more computing resources, so that the method is complex, not convenient and flexible enough in network training and practicality.
Disclosure of Invention
The invention provides a hyperspectral image saliency map generation method based on an end-to-end neural network, which comprises the steps of preprocessing a hyperspectral image to enable input image normalization to be more suitable for inputting into a model, inputting processed image data into the end-to-end neural network model to extract spatial spectral features, obtaining deep spatial spectral features of the hyperspectral image, directly generating a predicted saliency map, inputting the generated predicted saliency map into a neural network, and training the neural network model by calculating loss between the predicted saliency map and a true saliency map.
The invention provides a hyperspectral image saliency map generation method based on an end-to-end neural network, which comprises the following steps:
step S1: image preprocessing, namely preprocessing the initial hyperspectral image to obtain processed image data and inputting the processed image data into an end-to-end neural network;
step S2: extracting deep layer space spectrum characteristics, constructing an end-to-end neural network model, inputting the preprocessed hyperspectral image data into the neural network model, extracting spatial characteristics and spectral characteristics, fusing the spatial characteristics and the spectral characteristics, predicting the result and fusing the predicted result to obtain a final predicted saliency map;
step S3: training a neural network model, constructing a training data set, performing data expansion on hyperspectral image data, inputting the hyperspectral image data into the network model, calculating and predicting the loss of the saliency map and the truth saliency map by using a loss function, and optimizing a parameter training neural network.
Further, in step S1, the preprocessing of the image is to calculate a mean and a variance of the initial hyperspectral image data obtained by sampling, and normalize the mean and the variance to obtain the processed hyperspectral image data.
Further, in step S2, the end-to-end neural network model includes a spatial coding module, a spectral coding module, a decoding module, and a result prediction, where the spatial coding module and the spectral coding module are respectively connected to the decoding module, the spatial coding module is used to code spatial features, the spectral coding module is used to code spectral features, and the decoding module is used to output a prediction saliency map by fusing input spatial features and spectral features, and input the prediction saliency map into the result prediction module.
Furthermore, the spatial coding module and the decoding module have the same structure of each convolution layer of the convolution block, each convolution layer comprises conv3 × 3 convolution, a batch normalization layer bn and an activation function relu, the upper convolution layers are connected through a down sampling layer, the lower convolution layers are connected through an up sampling layer, and the convolution blocks form a U-shaped structure integrally.
Further, the spatial coding module includes 6 volume blocks, the depth L of the volume blocks is 7, 6, 5, 4, and 4, respectively, and the volume blocks are connected by a maximum pooling layer maxpool.
Further, the decoding module includes 5 convolutional blocks, the depths L of the convolutional blocks are 7, 6, 5, 4, and 4, respectively, and the convolutional blocks are connected by an upsampling layer upsample or a max-pooling layer maxpool.
Furthermore, the spectrum coding module comprises 6 rolling blocks, the depths of the rolling blocks are all 4, and the rolling blocks are connected through an averaging pooling layer avgpool.
Furthermore, each convolution layer of the convolution block in the spectral coding module includes a conv1 × 1 convolution layer and a batch normalization layer, the convolution layers in the convolution block are connected with each other through an activation function relu, and the convolution block integrally forms a U-shaped structure.
Further, the result prediction module receives the output of each convolution block in the decoding module through conv3 × 3 convolution and an activation function sigmoid respectively to obtain a prediction saliency map, and fuses the prediction saliency map through conv1 × 1 convolution and the activation function sigmoid to output a final prediction saliency map.
Further, in step S3, the hyperspectral image size of the data set subjected to neural network training is 1024 × 768, and the data of the training set is subjected to data expansion by horizontal inversion with 50% probability or downsampling with 25% probability, so as to obtain a hyperspectral image with a size of 512 × 384 as an input training set.
Further, in step S3, the loss function is a two-class cross entropy loss function, and the training of the neural network optimizes the model parameters through a back propagation algorithm according to the calculated loss.
The invention has the following beneficial effects:
1. an end-to-end double-branch neural network model is adopted to extract deep features of a hyperspectral image and directly generate a prediction significant map, network training is carried out according to the calculation of the generated prediction significant map and the loss of a true value significant map, a final network model is obtained, computing resources are saved, time consumption is reduced, and flexibility in neural network training and practicability is improved.
2. The constructed end-to-end neural network model structure integrates a spatial coding module, a spectral coding module and a decoding module, fully extracts deep spatial spectral features of a hyperspectral image, improves the robustness of the features, enables the quality of the generated saliency map to be higher, and improves the accuracy of a final result.
3. The input and the output of each convolution block in the decoding module are connected through an upper sampling layer, so that the influence of pixel points of input image data in a parameter range is improved, and the accuracy of result output is ensured.
Drawings
FIG. 1 is a schematic flow diagram of the generation method of the present invention;
FIG. 2 is a schematic diagram of the overall structure of the end-to-end neural network of the present invention;
FIG. 3 is a schematic structural diagram of a 4-layer deep convolution block of the end-to-end neural network of the present invention;
the left side in fig. 3 is a 4-layer depth convolution block structure corresponding to the spatial coding module and the decoding module; the right side is a 4-layer depth convolution block structure corresponding to the spectrum coding module;
S1-S6 respectively represent the predicted saliency maps of each layer corresponding output of the decoding module.
Detailed Description
In the following description, technical solutions in the embodiments of the present invention are clearly and completely described, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1
An embodiment 1 of the present invention provides a hyperspectral image saliency map generation method based on an end-to-end neural network, and as shown in fig. 1, the method includes the following steps:
step S1: image preprocessing, namely preprocessing an initial hyperspectral image to obtain processed image data, and inputting the processed image data into an end-to-end neural network, wherein the specific process is as follows:
in this embodiment, the initial hyperspectral image data is X0∈N512×384×81The size is 512X 384, the spectrum dimension is 81, then the obtained hyperspectral image is preprocessed and artificially marked, the preprocessing calculates the mean value and the variance of the hyperspectral image data, and the whole normalization calculation is carried out to obtain the processed hyperspectral image data X1∈RW×H×LThe continuity of the spectral data in the hyperspectral image is ensured, and the specific formula is as follows:
Figure BDA0003090230610000041
wherein μ (X)0) Is the mean, σ (X), of the raw hyperspectral image data0) Is the variance of the raw hyperspectral image data.
Step S2: constructing an end-to-end neural network model, inputting the hyperspectral image data obtained through preprocessing into the neural network model, extracting spatial features and spectral features, fusing the spatial features and the spectral features, predicting the output result, and fusing the predicted result to obtain a final predicted saliency map;
as shown in fig. 2, the end-to-end neural network model structure in this embodiment is a double branch W2The convolutional neural network comprises a space coding module and a spectrum coding moduleThe device comprises a module, a decoding module and a result prediction module;
the spatial coding module and the spectral coding module are respectively connected with the decoding module to integrally form a U-shaped structure, the spatial coding module is used for coding spatial characteristics, the spectral coding module is used for coding spectral characteristics, and the decoding module is used for fusing input spatial characteristics and spectral characteristics to output a prediction saliency map and inputting the prediction saliency map into the result prediction module;
the spatial coding module comprises 6 volume blocks, the depths L of the volume blocks are respectively 7, 6, 5, 4 and 4, and the volume blocks are connected through a maximum pooling layer maxpool;
the decoding module comprises 5 convolution blocks, the depths L of the convolution blocks are respectively 7, 6, 5, 4 and 4, and the convolution blocks are connected through an upsampling layer or a maximum pooling layer maxpool;
the spectrum coding module comprises 6 rolling blocks, the depths of the rolling blocks are all 4, and the rolling blocks are connected through an average pooling layer avgpool.
And the spatial coding module and each rolling block of the spectral coding module are connected with the input of each layer of rolling block of the decoding module to obtain six prediction results with different degrees.
As shown in fig. 3, a convolution block with a depth of 4 is illustrated;
in this embodiment, the spatial coding module and each convolution layer of the convolution block in the decoding module have the same structure, each convolution layer includes conv3 × 3 convolution, batch normalization layer bn and activation function relu, image data is input to the convolution layer and batch normalization layer bn with convolution kernels of 3 × 3, and then is input to the convolution block with 4 layers of depth, the output of each convolution layer is input to the convolution layer with the same depth, the convolution layers with the front n-1 depth of the upper layer are connected through a down-sampling layer, the convolution layers with the front n-1 depth of the lower layer are connected through a down-sampling layer, n represents the depth of the current convolution block, the convolution blocks form a U-shaped structure as a whole, and finally the outputs of each layer are superimposed and input to the relu activation function;
each convolution layer of the convolution block in the spectrum coding module comprises a conv1 multiplied by 1 convolution layer and a batch normalization layer, input and output of each convolution layer in the convolution block are connected through an activation function relu, similarly, the output of each convolution layer is input into the convolution layer with the same depth, and the convolution layers with different depths are sequentially connected to form a U-shaped structure.
Pre-processed hyperspectral image data X1Inputting end-to-end neural network, extracting deep space spectrum characteristics as F e R through space branching and spectrum branching512×384×64And 64 is the feature dimension, which is then input into the result prediction module.
The result prediction module receives deep-layer space spectrum characteristics output by each convolution block in the decoding module through conv3 x 3 convolution and an activation function sigmoid respectively to generate a prediction saliency map corresponding to each layer, the prediction saliency maps are fused through conv1 x 1 convolution and the activation function sigmoid, and a final prediction saliency map Seeger R is output512×384And the fusion can also output the final result in a mode of superposition and averaging.
Step S3: training a neural network model, constructing a training data set, obtaining a truth-value saliency map by labeling an original hyperspectral image, calculating and predicting the loss of the saliency map and the truth-value saliency map by using a loss function, and optimizing parameters to train the neural network;
in the embodiment, a hyperspectral image with the size of 1024 × 768 is used as a training set, and a hyperspectral image with the size of 512 × 384 is obtained by horizontally turning the image of the training set with 50% probability or sampling the image with 25% probability as input;
calculating the prediction saliency map and the truth saliency map G ∈ N output by the model512×384According to the loss, parameters of the neural network model are optimized through a back propagation algorithm Adam algorithm;
wherein, the truth significance map G epsilon NW×HThe value is 0 or 1, 0 represents background, 1 represents foreground, and the value range of the prediction saliency map is (0.0,1.0), so that the regression problem of the saliency value is regarded as a binary classification problem of the background and the foreground in the neural network training, and the prediction saliency map is subjected to binary classification on a pixel-by-pixel basisAnd the loss function adopts two-classification cross entropy to calculate loss.
In this embodiment, the losses of the predicted saliency map and the true saliency map under six different network depths and the losses of the finally fused saliency map and the true saliency map are calculated respectively, so as to train the network, and the calculation formula of the losses is as follows:
Figure BDA0003090230610000051
where G denotes a true saliency map, S denotes a predicted saliency map, W and H denote the sizes of the saliency map, and x and y denote pixel coordinates.
In this embodiment, 20 epochs are set for training of the neural network model, the batch size is set to 2, the initial learning rate is 0.001, and the attenuation coefficient of each epoch learning rate is 0.95.
The invention is not limited to the foregoing embodiments. The invention extends to any novel feature or any novel combination of features disclosed in this specification and any novel method or process steps or any novel combination of features disclosed.

Claims (10)

1. A hyperspectral image saliency map generation method based on an end-to-end neural network is characterized by comprising the following steps:
step S1: image preprocessing, namely preprocessing the initial hyperspectral image to obtain processed image data and inputting the processed image data into an end-to-end neural network;
step S2: extracting deep layer space spectrum characteristics, constructing an end-to-end neural network model, inputting the preprocessed hyperspectral image data into the neural network model, extracting spatial characteristics and spectral characteristics, fusing the spatial characteristics and the spectral characteristics, predicting the result and fusing the predicted result to obtain a final predicted saliency map;
step S3: training a neural network model, constructing a training data set, performing data expansion on hyperspectral image data, inputting the hyperspectral image data into the network model, calculating and predicting the loss of the saliency map and the truth saliency map by using a loss function, and optimizing a parameter training neural network.
2. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 1, wherein in step S1, the preprocessing of the image is to calculate the mean and variance of the sampled initial hyperspectral image data and normalize the mean and variance to obtain the processed hyperspectral data.
3. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 1, wherein in step S2, the end-to-end neural network model comprises a spatial coding module, a spectral coding module, a decoding module and a result prediction module, the spatial coding module and the spectral coding module are respectively connected with the decoding module, the spatial coding module is used for coding spatial features, the spectral coding module is used for coding spectral features, and the decoding module is used for fusing input spatial features and spectral features to output a predicted saliency map and inputting the predicted saliency map into the result prediction module.
4. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 4, wherein the spatial coding module and the decoding module have the same structure of each convolution layer of a convolution block, each convolution layer comprises conv3 x 3 convolution, a batch normalization layer bn and an activation function relu, the upper convolution layers are connected through a downsampling layer downlink sample, the lower convolution layers are connected through an upsampling layer uplink sample, and the convolution blocks integrally form a U-shaped structure.
5. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 5, wherein the spatial coding module comprises 6 volume blocks, the depth L of each volume block is 7, 6, 5, 4 and 4, and the volume blocks are connected through a max-pooling layer maxpool;
the decoding module comprises 5 convolution blocks, the depths L of the convolution blocks are respectively 7, 6, 5, 4 and 4, and the convolution blocks are connected through an upsampling layer or a maximum pooling layer maxpool.
6. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 4, wherein the spectral coding module comprises 6 volume blocks, the depth of each volume block is 4, and the volume blocks are connected by an average pooling layer avgpool.
7. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 8, wherein each convolution layer of a convolution block in the spectral coding module comprises a conv1 x 1 convolution layer and a batch normalization layer, wherein input ends and output ends of each convolution layer in the convolution block are connected through an activation function relu, and the convolution blocks integrally form a U-shaped structure.
8. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 4 is characterized in that the result prediction module receives the output of each convolution block in the decoding module through conv3 x 3 convolution and an activation function sigmoid respectively to obtain a predicted saliency map, and fuses the predicted saliency map through conv1 x 1 convolution and an activation function sigmoid to output a final predicted saliency map.
9. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 1, wherein in step S3, the hyperspectral image size of the data set subjected to neural network training is 1024 × 768, and the data of the training set is subjected to data expansion by adopting horizontal inversion with 50% probability or downsampling with 25% probability, so as to obtain a hyperspectral image with size 512 × 384 as an input training set.
10. The end-to-end neural network-based hyperspectral image saliency map generation method according to claim 1, wherein in step S3, the loss function is a two-class cross entropy loss function, and model parameters are optimized by a back propagation algorithm according to calculated loss during training of the neural network.
CN202110593767.XA 2021-05-28 2021-05-28 Hyperspectral image saliency map generation method based on end-to-end neural network Pending CN113327304A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110593767.XA CN113327304A (en) 2021-05-28 2021-05-28 Hyperspectral image saliency map generation method based on end-to-end neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110593767.XA CN113327304A (en) 2021-05-28 2021-05-28 Hyperspectral image saliency map generation method based on end-to-end neural network

Publications (1)

Publication Number Publication Date
CN113327304A true CN113327304A (en) 2021-08-31

Family

ID=77422282

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110593767.XA Pending CN113327304A (en) 2021-05-28 2021-05-28 Hyperspectral image saliency map generation method based on end-to-end neural network

Country Status (1)

Country Link
CN (1) CN113327304A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114387258A (en) * 2022-01-14 2022-04-22 北京理工大学重庆创新中心 Hyperspectral image reconstruction method based on regional dynamic depth expansion neural network

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103729848A (en) * 2013-12-28 2014-04-16 北京工业大学 Hyperspectral remote sensing image small target detection method based on spectrum saliency
US20150055824A1 (en) * 2012-04-30 2015-02-26 Nikon Corporation Method of detecting a main subject in an image
CN108090447A (en) * 2017-12-19 2018-05-29 青岛理工大学 Hyperspectral image classification method and device under double branch's deep structures
CN109146831A (en) * 2018-08-01 2019-01-04 武汉大学 Remote sensing image fusion method and system based on double branch deep learning networks
CN109191426A (en) * 2018-07-24 2019-01-11 江南大学 A kind of flat image conspicuousness detection method
CN109871830A (en) * 2019-03-15 2019-06-11 中国人民解放军国防科技大学 Spatial-spectral fusion hyperspectral image classification method based on three-dimensional depth residual error network
CN111160478A (en) * 2019-12-31 2020-05-15 北京理工大学重庆创新中心 Hyperspectral target significance detection method based on deep learning
CN111667489A (en) * 2020-04-30 2020-09-15 华东师范大学 Cancer hyperspectral image segmentation method and system based on double-branch attention deep learning
CN112183360A (en) * 2020-09-29 2021-01-05 上海交通大学 Lightweight semantic segmentation method for high-resolution remote sensing image

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150055824A1 (en) * 2012-04-30 2015-02-26 Nikon Corporation Method of detecting a main subject in an image
CN103729848A (en) * 2013-12-28 2014-04-16 北京工业大学 Hyperspectral remote sensing image small target detection method based on spectrum saliency
CN108090447A (en) * 2017-12-19 2018-05-29 青岛理工大学 Hyperspectral image classification method and device under double branch's deep structures
CN109191426A (en) * 2018-07-24 2019-01-11 江南大学 A kind of flat image conspicuousness detection method
CN109146831A (en) * 2018-08-01 2019-01-04 武汉大学 Remote sensing image fusion method and system based on double branch deep learning networks
CN109871830A (en) * 2019-03-15 2019-06-11 中国人民解放军国防科技大学 Spatial-spectral fusion hyperspectral image classification method based on three-dimensional depth residual error network
CN111160478A (en) * 2019-12-31 2020-05-15 北京理工大学重庆创新中心 Hyperspectral target significance detection method based on deep learning
CN111667489A (en) * 2020-04-30 2020-09-15 华东师范大学 Cancer hyperspectral image segmentation method and system based on double-branch attention deep learning
CN112183360A (en) * 2020-09-29 2021-01-05 上海交通大学 Lightweight semantic segmentation method for high-resolution remote sensing image

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
CHEN HUANG 等: "Salient object detection on hyperspectral images in wireless network using CNN and saliency optimization", 《AD HOC NETWORKS》 *
XIANGYU LIU 等: "Remote sensing image fusion based on two-stream fusion network", 《INFORMATION FUSION》 *
XUEBIN QIN 等: "U2-Net: Going deeper with nested U-structure for salient object detection", 《PATTERN RECOGNITION》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114387258A (en) * 2022-01-14 2022-04-22 北京理工大学重庆创新中心 Hyperspectral image reconstruction method based on regional dynamic depth expansion neural network
CN114387258B (en) * 2022-01-14 2024-03-22 北京理工大学重庆创新中心 Hyperspectral image reconstruction method based on regional dynamic depth expansion neural network

Similar Documents

Publication Publication Date Title
CN111325794B (en) Visual simultaneous localization and map construction method based on depth convolution self-encoder
CN111047548B (en) Attitude transformation data processing method and device, computer equipment and storage medium
CN111259945B (en) Binocular parallax estimation method introducing attention map
Pang et al. Visual haze removal by a unified generative adversarial network
CN111582483B (en) Unsupervised learning optical flow estimation method based on space and channel combined attention mechanism
CN111931787A (en) RGBD significance detection method based on feature polymerization
CN112132023A (en) Crowd counting method based on multi-scale context enhanced network
CN111598998A (en) Three-dimensional virtual model reconstruction method and device, computer equipment and storage medium
CN112419242A (en) No-reference image quality evaluation method based on self-attention mechanism GAN network
Wang et al. VoPiFNet: Voxel-Pixel Fusion Network for Multi-Class 3D Object Detection
CN115359372A (en) Unmanned aerial vehicle video moving object detection method based on optical flow network
CN114049434A (en) 3D modeling method and system based on full convolution neural network
CN115115685A (en) Monocular image depth estimation algorithm based on self-attention neural network
CN110335299A (en) A kind of monocular depth estimating system implementation method based on confrontation network
Zhang et al. Unsupervised depth estimation from monocular videos with hybrid geometric-refined loss and contextual attention
CN117391938B (en) Infrared image super-resolution reconstruction method, system, equipment and terminal
Wang et al. Underwater self-supervised monocular depth estimation and its application in image enhancement
Babu et al. An efficient image dahazing using Googlenet based convolution neural networks
CN113327304A (en) Hyperspectral image saliency map generation method based on end-to-end neural network
Gonzalez-Sabbagh et al. DGD-cGAN: A dual generator for image dewatering and restoration
Liu et al. A video drowning detection device based on underwater computer vision
Huang et al. Underwater image enhancement via LBP‐based attention residual network
Liu et al. SI-SA GAN: A generative adversarial network combined with spatial information and self-attention for removing thin cloud in optical remote sensing images
CN116631064A (en) 3D human body posture estimation method based on complementary enhancement of key points and grid vertexes
CN115170985B (en) Remote sensing image semantic segmentation network and segmentation method based on threshold attention

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210831

RJ01 Rejection of invention patent application after publication