CN110728658A - High-resolution remote sensing image weak target detection method based on deep learning - Google Patents

High-resolution remote sensing image weak target detection method based on deep learning Download PDF

Info

Publication number
CN110728658A
CN110728658A CN201910870991.1A CN201910870991A CN110728658A CN 110728658 A CN110728658 A CN 110728658A CN 201910870991 A CN201910870991 A CN 201910870991A CN 110728658 A CN110728658 A CN 110728658A
Authority
CN
China
Prior art keywords
data set
resolution
image
target detection
remote sensing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201910870991.1A
Other languages
Chinese (zh)
Inventor
王超
张洪艳
张良培
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University (WHU)
Original Assignee
Wuhan University (WHU)
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University (WHU) filed Critical Wuhan University (WHU)
Priority to CN201910870991.1A priority Critical patent/CN110728658A/en
Publication of CN110728658A publication Critical patent/CN110728658A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20092Interactive image processing based on input by user
    • G06T2207/20104Interactive definition of region of interest [ROI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30168Image quality inspection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a method for detecting a weak target of a high-resolution remote sensing image based on deep learning, for the remote sensing image with low resolution, small target size and fuzzy quality, firstly, the resolution is improved by adopting a WGAN-based hyper-resolution reconstruction method, the image with enhanced quality is input into a target detection frame, deep characteristic extraction is carried out on the image by utilizing a residual error network, then the extracted low-level features and high-level features are fused to ensure that the fused multi-level feature map has rich detail information and high-level semantic information, the fused multi-level features are utilized, the region suggestion network carries out region-of-interest coarse extraction on the feature map, and mapping the extracted regions to the same dimension by using a region-of-interest alignment method, and performing subsequent accurate target classification and position refinement to obtain a final target detection result. The method effectively improves the detection precision and recall rate of the weak and small targets under the conditions of low resolution and complex background of the remote sensing image.

Description

High-resolution remote sensing image weak target detection method based on deep learning
Technical Field
The invention relates to the technical field of video image processing and target detection, in particular to a high-resolution remote sensing image weak target detection method based on deep learning.
Background
The remote sensing earth observation technology is improved in image space, spectrum and time resolution, can collect multi-sensor multi-source earth observation data, enables a remote sensing satellite to have high-resolution, all-weather, all-time and large-range earth observation and imaging capabilities through efficient multi-source data fusion, achieves subversive results in multiple fields such as military affairs, agriculture and urban planning, provides strategic support for sustainable development problems such as climate change monitoring and environmental pollution control, and greatly promotes social development and progress. Particularly, the high spatial resolution (high resolution) remote sensing image can reach sub-meter level along with the iterative update of high resolution series satellites of all countries in the world, the obtained image has rich texture and detail information, compared with the low resolution remote sensing image generated by other sensors, the high spatial resolution (high resolution) remote sensing image can finely reflect the state of ground objects, and a rich and reliable data source can be provided for the applications of land utilization investigation, city management, battlefield decision and the like. The high-resolution remote sensing image greatly promotes the development and application of the remote sensing subject. Among the interpretation and processing methods of remote sensing image information, the target detection and identification technology is a popular research direction by virtue of its practicability and wide application field. The target detection consists of a target classification part and a target positioning part, can determine the type of ground objects, simultaneously outputs the position information of the objects, and can reflect the state information of the targets more completely and more finely.
With the improvement of the software and hardware level of a computer, the computing capability and the mass data storage capability of the computer are unprecedentedly developed, so that the intelligent operation and management of the mass data become possible. In this context, deep learning and neural network models, which have previously been slow to develop due to hardware factor limitations, are receiving renewed attention, and more artificial intelligence vision tasks can be excellently performed by deep neural networks, in particular Convolutional Neural Networks (CNNs). Compared with the traditional method, the neural network has strong feature expression capability, nonlinear fitting capability and generalization capability and strong model learning capability, and the model is processed end to end without excessive expert priori knowledge intervention, so that the method is suitable for target classification and regression tasks. And because the network has the characteristics of local perception and parameter sharing when image processing is carried out, the local characteristics of the relation and space between adjacent domains can be kept, and the processing of high-dimensional images is not difficult. By means of good feature extraction and characterization capabilities of the deep neural network, the target detection and identification algorithm based on deep learning develops rapidly, various models with good detection precision and speed are promoted, and compared with the traditional model, the application effect and quality leap is realized.
The inventor of the present application finds that the method of the prior art has at least the following technical problems in the process of implementing the present invention:
although the target detection model based on deep learning is developed vigorously, the application of the target detection model on remote sensing images still has a plurality of difficulties. The remote sensing image has large imaging area and long imaging distance, so that the image space resolution is low and the image background is complex; the top-down imaging mode of the satellite also enables the target to be easily shielded; and some image distortion may occur due to surface reflection, atmospheric refraction, etc. The unique imaging mode of the remote sensing image enables the ground object target in the image to be small in size, low in resolution, lack of sufficient texture information, large in background noise interference and low in contrast, so that detection difficulty is caused, and the application of the remote sensing image in a plurality of fine fields is greatly limited.
Therefore, the method in the prior art has the technical problem of low detection precision.
Disclosure of Invention
The invention aims to provide a high-resolution remote sensing image weak target detection method based on deep learning, and provides a multilayer feature fusion remote sensing image weak target detection algorithm combined with a hyper-resolution reconstruction method aiming at the characteristics of relatively low imaging resolution and small target size of a remote sensing image. In consideration of the influence of image resolution on target detection, the remote sensing image quality is improved by adopting a WGAN-based hyper-resolution reconstruction method, the detail information of the image is enriched, more effective target features can be extracted by a residual network, the low-level features and the high-level features extracted by the residual network are fused, the fused feature map is ensured to have abundant semantic information and detail information, the RPN network is used for carrying out foreground target coarse extraction and interesting region generation, subsequent target accurate detection is carried out, the precision and the recall ratio of target detection are improved, and the target detection precision of the remote sensing image is improved.
In order to achieve the purpose, the invention adopts the technical scheme that: a high-resolution remote sensing image weak target detection method based on deep learning comprises the following steps:
step S1: based on the ground feature type and characteristics in the high-resolution remote sensing image, an initial data set is constructed, the initial data set is divided into a training data set, a verification data set and a test data set, and type information and position information labeling is carried out on ground feature samples in the initial data set;
step S2: taking the training data set and the verification data set as high-resolution samples, performing down-sampling on the high-resolution samples to obtain low-resolution samples, and then performing training and verification evaluation on the WGAN-based image super-resolution reconstruction model by using the low-resolution samples to obtain a trained super-resolution reconstruction model;
step S3: inputting the training data set, the verification data set and the test data set into the trained super-resolution reconstruction model to obtain a reconstructed training data set, verification data set and test data set;
step S4: constructing a remote sensing image target detection model based on multi-layer feature fusion by adopting a target detection basic framework based on region suggestion, and setting parameters of the target detection model according to the characteristics of the reconstructed training data set;
step S5: inputting the reconstructed training data set into a target detection model, training the target detection model, calculating an optimized target function loss value, and performing training and updating on the model weight parameters through a back propagation algorithm;
step S6: verifying the model trained in the step S5 by using the reconstructed verification data set to obtain an optimized target detection model;
step S7: and inputting the reconstructed test data set into an optimized target detection model for target detection and identification, determining the category and position information of the ground object target, and detecting the weak and small target of the remote sensing image.
In one embodiment, when the step S1 labels the feature samples in the initial data set with the category information and the location information, spatial context information is introduced for the weak targets in the initial data set.
In one embodiment, the optimization objective function of the WGAN-based super-resolution image reconstruction model in step S2 is as follows:
Figure BDA0002202799830000031
wherein L represents the total Loss function of WGAN, Loss _ G represents the generator Loss function, Loss _ D represents the discriminator Loss function, fw(x) Representing a continuous mapping function for mapping the samples to have their derivative values within a range; e represents averaging the formula; prRepresenting true distribution, PgThe generation distribution is represented.
In one embodiment, step S3 specifically includes:
performing image reconstruction on remote sensing images contained in a training data set, a verification data set and a test data set by using a trained super-resolution reconstruction model, wherein quality evaluation on the super-resolution images is usually performed by adopting a peak signal-to-noise ratio and an image structure similarity index, wherein the peak signal-to-noise ratio PSNR is usually analyzed by adopting a formula (2), the image Structure Similarity Index (SSIM) is shown in a formula (3), MSE represents a mean square error, M and N represent image length and width values, M and N respectively represent pixel values corresponding to the images before and after reconstruction, s represents a gray level order, and the larger the PSNR value is, the smaller the distortion is;
Figure BDA0002202799830000041
Figure BDA0002202799830000042
wherein c is a constant,
Figure BDA0002202799830000043
μyand σyValue of (d) and mux、σxSimilarly, parameter C1=K1×L,C2=K2xL, L represents the maximum pixel value in a gray-scale image, K1And K2As a constant, SSIM is used to measure the similarity between images from brightness, contrast and texture, with larger values indicating less distortion and higher quality of the image.
In one implementation, the remote sensing image target detection model based on multi-layer feature fusion in step S4 includes a feature extraction and fusion module, a region-of-interest generation module, and a target detection module, where the feature extraction and fusion module is configured to extract feature of a ground object and fuse the extracted low-level features and high-level features, the region-of-interest generation module is configured to extract a region of interest based on the fused feature map, and the target detection module is configured to perform target identification and detection according to the generated region of interest.
In one embodiment, the feature extraction and fusion module includes a down-sampling process and an up-sampling process when fusing the low-level features and the high-level features.
In one embodiment, the formula for optimizing the objective function in step S5 is:
Figure BDA0002202799830000044
in the formula (4), Lcls=-log(pi) Representing the classification error, piRepresenting the probability value of the ground object class output by the classifier;
Figure BDA0002202799830000045
representing the positioning error of the model, where tx=(x-xa)/wa,ty=(y-ya)/ha
Figure BDA0002202799830000046
Data with subscript a is a parameter value of a bounding box output by an anchor box, data with asterisk superscript is a ground truth value, the rest are model prediction output values, lambda is a balance parameter, and N isclsTo train the blocksize value, NlocThe generated Anchor box value.
One or more technical solutions in the embodiments of the present application have at least one or more of the following technical effects:
the invention provides a high-resolution remote sensing image weak target detection method based on deep learning, which is characterized by comprising the steps of firstly, carrying out class information and position information labeling on a ground object sample in an initial data set based on the class and the characteristics of a ground object in a high-resolution remote sensing image, then, taking a training data set and a verification data set as high-resolution samples, carrying out down-sampling on the high-resolution samples to obtain low-resolution samples, then, utilizing the low-resolution samples to train, verify and evaluate an image super-resolution reconstruction model based on WGAN, and obtaining the trained super-resolution reconstruction model; inputting the training data set, the verification data set and the test data set into a trained super-resolution reconstruction model to obtain a reconstructed training data set, a reconstructed verification data set and a reconstructed test data set; then, a target detection basic framework based on region suggestion is adopted, a remote sensing image target detection model based on multi-layer feature fusion is constructed, and parameters of the target detection model are set according to the characteristics of a reconstructed training data set; inputting the reconstructed training data set into a target detection model, training the target detection model, calculating an optimized target function loss value, and performing training and updating on the model weight parameters through a back propagation algorithm; then, verifying the trained model by using the reconstructed verification data set to obtain an optimized target detection model; and finally, inputting the reconstructed test data set into an optimized target detection model for target detection and identification, determining the category and position information of the ground object target, and detecting the remote sensing image weak and small target.
According to the method provided by the invention, the remote sensing image in the training data set, the verification data set and the test data set is reconstructed by adopting the WGAN-based image super-resolution reconstruction model, the quality of the remote sensing image can be improved, the detail information of the image is enriched, so that a remote sensing image target detection model based on multi-layer feature fusion can extract more effective target features, the extracted low-layer features and high-layer features are fused, the fused feature diagram is ensured to have abundant semantic information and detail information, and the fused feature diagram is utilized by an RPN (resilient packet network) to perform foreground target coarse extraction and interesting region generation to perform subsequent target accurate detection, so that the precision and recall ratio of target detection are improved, and the target detection precision of the remote sensing image is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a schematic flow chart of a method for detecting a weak target in a high-resolution remote sensing image based on deep learning according to an embodiment of the present invention;
FIG. 2 is an overall frame diagram of a high-resolution remote sensing image weak target detection method based on deep learning.
Detailed Description
The invention aims to provide a high-resolution remote sensing image weak target detection method based on deep learning, aiming at the technical problem that the detection precision is low in the method in the prior art, the remote sensing image weak target detection and identification are carried out through a multi-layer feature fusion model of a combined hyper-resolution reconstruction method, and the precision and recall rate of high-resolution remote sensing image target detection can be improved through image quality improvement and multi-layer fusion of effective features.
In order to achieve the above purpose, the main concept of the invention is as follows:
for remote sensing images with low resolution, small target size and fuzzy quality, firstly, improving the resolution of the images by adopting a WGAN-based hyper-resolution reconstruction method, and enriching the detail information of weak target ground objects; inputting the image with enhanced quality into a target detection frame, extracting deep features of the image by using a residual error network, and then fusing the extracted low-level features and high-level features, so that the fused multi-level feature map has rich detail information and high-level semantic information, and the accuracy of target positioning and classification is ensured; and performing region-of-interest coarse extraction on the feature map by using the fused multilayer features and the region suggestion network RPN (namely a region-of-interest generation module), mapping the extracted regions to the same dimension by using a region-of-interest alignment method, and performing subsequent accurate target classification and position refinement to obtain a final target detection result. The method effectively improves the detection precision and recall rate of the weak and small targets under the conditions of low resolution and complex background of the remote sensing image.
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention uses the video image acquired by the unmanned aerial vehicle to detect the target. At present, the unmanned aerial vehicle is widely applied in the fields of military affairs, urban management and agricultural production, but is limited by the performance of carrying an image acquisition sensor, the resolution of the acquired image is low, high target detection precision is difficult to obtain, and the high-performance sensor is extremely expensive. Therefore, the present invention will describe a specific process using the unmanned aerial vehicle image as a data source.
The embodiment provides a method for detecting a weak target of a high-resolution remote sensing image based on deep learning, please refer to fig. 1, and the method comprises the following steps:
step S1: based on the ground feature type and characteristics in the high-resolution remote sensing image, an initial data set is constructed, the initial data set is divided into a training data set, a verification data set and a test data set, and type information and position information labeling is carried out on ground feature samples in the initial data set.
Specifically, in the present invention, the weak target mainly refers to a remote sensing image with low resolution, small target size and fuzzy quality. And intercepting the video image shot by the unmanned aerial vehicle through a video frame to obtain image format data, wherein the image data needs to fully cover the multi-mode ground object target, so that the effectiveness of data set manufacturing is ensured. In the specific implementation process, the manufactured data set comprises 800 images, the training set, the verification set and the test set are divided according to the ratio of 2:1:1, effective targets are selected for marking, and in the implementation mode, the acquired video mainly aims at detecting automobile targets.
In one embodiment, when the step S1 labels the feature samples in the initial data set with the category information and the location information, spatial context information is introduced for the weak targets in the initial data set.
Specifically, when the initial data set is labeled, the label format needs to be made to meet the format requirement input by the model, the category and position information of the target object needs to be correctly labeled, wrong detection and missing detection of the target caused by data labeling errors are prevented, and a spatial context information auxiliary model is added for target detection. In the process of constructing the data set, some weak and small targets, fuzzy targets and high-inclination targets are easy to miss the targets and influence the generalization capability of the training model, so that the ground objects need to be labeled in detail, and the multi-scale (size and angle) information of the data set is increased; when targets such as bridges are labeled, spatial context information is introduced, so that the targets can be extracted and identified by an auxiliary model, and the detection precision of the remote sensing targets in a complex scene is improved.
Step S2: and taking the training data set and the verification data set as high-resolution samples, performing down-sampling on the high-resolution samples to obtain low-resolution samples, and performing training, verification and evaluation on the WGAN-based image super-resolution reconstruction model by using the low-resolution samples to obtain a trained super-resolution reconstruction model.
Specifically, the downsampling of the training data set and the validation data set is performed to generate a low resolution data set, which is used as an input sample for training the hyper-differential reconstruction process.
In one embodiment, the optimization objective function of the WGAN-based super-resolution image reconstruction model in step S2 is as follows:
Figure BDA0002202799830000071
wherein L represents the total Loss function of WGAN, Loss _ G represents the generator Loss function, Loss _ D represents the discriminator Loss function, fw(x) Representing a continuous mapping function for mapping the samples to have their derivative values within a range; e represents averaging the formula; prRepresenting true distribution, PgThe generation distribution is represented.
Specifically, the image super-resolution reconstruction model based on the WGAN is used for reconstructing an image, the image super-resolution reconstruction method based on the Wassertein generation countermeasure network (WGAN) is adopted in the invention, and the Wassertein distance can be shown as the following formula (5). Wherein, x and y represent two different samples, formula 5 is used to find the distance between two samples, and the sample similarity is measured by the distance; inf represents a floor function.
Figure BDA0002202799830000081
The WGAN uses the Wasserstein distance to replace the previous K-L divergence and J-S divergence, and effective training of the model under weak correlation of the low-resolution image and the high-resolution real image can be guaranteed.
WGAN-based image super-resolutionIn the optimized objective function of the resolution reconstruction model, fw(x) Representing a continuous mapping function, which is a simple mapping of samples to have their derivative values within a certain range; e represents averaging the formula. By adjusting and optimizing the generator, the Wasserstein distance between the generated sample and the real sample is shortened, and the image quality approaches to the real image. The generation quality of the image is improved through the alternate training of the generator and the discriminator. The WGAN training is stable, the model convergence speed is high, and the quality of the training result can better reflect the quality of the generated image. The quality evaluation of super-resolution images is usually performed by quantitative analysis using peak signal-to-noise ratio (PSNR) and image Structure Similarity Index (SSIM). The PSNR formula is shown below, where MSE represents a mean square error, M and N represent image length and width values, M and N represent pixel values corresponding to images before and after reconstruction, s represents a gray scale order, and generally takes 256, where a larger PSNR value represents a smaller distortion. When image quality evaluation is performed before and after reconstruction, for remote sensing images, generally, PSNR values greater than about 30 can indicate that image quality is effectively improved.
Step S3: and inputting the training data set, the verification data set and the test data set into the trained super-resolution reconstruction model to obtain the reconstructed training data set, the reconstructed verification data set and the reconstructed test data set.
Specifically, the trained super-resolution reconstruction model can be used for performing super-resolution reconstruction on images in the training data set, the verification data set and the test data set, so that the resolution of the remote sensing image is improved, and the detail texture information of the image is enriched.
In one embodiment, step S3 specifically includes:
performing image reconstruction on remote sensing images contained in a training data set, a verification data set and a test data set by using a trained super-resolution reconstruction model, wherein quality evaluation on the super-resolution images is usually performed by adopting a peak signal-to-noise ratio and an image structure similarity index, wherein the peak signal-to-noise ratio PSNR is usually analyzed by adopting a formula (2), the image Structure Similarity Index (SSIM) is shown in a formula (3), MSE represents a mean square error, M and N represent image length and width values, M and N respectively represent pixel values corresponding to the images before and after reconstruction, s represents a gray level order, and the larger the PSNR value is, the smaller the distortion is;
Figure BDA0002202799830000091
wherein c is a constant,μyand σyValue of (d) and mux、σxSimilarly, parameter C1=K1×L,C2=K2xL, L represents the maximum pixel value in a gray-scale image, K1And K2As a constant, SSIM is used to measure the similarity between images from brightness, contrast and texture, with larger values indicating less distortion and higher quality of the image.
In a specific implementation, L is a maximum pixel value in the grayscale image, for example, L is 255, K1And K2Is a small constant, and can be preset by those skilled in the art in specific implementation, and in the embodiment, may be K1=0.01,K2SSIM measures inter-image similarity from brightness, contrast and texture 0.02. Similarly, a larger value indicates a smaller distortion of the image and a higher quality. Similarly, when image quality evaluation is performed before and after reconstruction, the remote sensing image reconstruction effect is considered to be good when the SSIM value in the experiment reaches 0.85 or more.
Step S4: and constructing a remote sensing image target detection model based on multi-layer feature fusion by adopting a target detection basic framework based on region suggestion, and setting parameters of the target detection model according to the characteristics of the reconstructed training data set.
Specifically, the characteristics of the data set mainly include the size, resolution, feature type, and number of images in the data set, and the model parameters include blocksize, IOU threshold, learning rate, and the like.
In one implementation, the remote sensing image target detection model based on multi-layer feature fusion in step S4 includes a feature extraction and fusion module, a region-of-interest generation module, and a target detection module, where the feature extraction and fusion module is configured to extract feature of a ground object and fuse the extracted low-level features and high-level features, the region-of-interest generation module is configured to extract a region of interest based on the fused feature map, and the target detection module is configured to perform target identification and detection according to the generated region of interest.
In a specific implementation process, for an input image, deep feature extraction is performed by using a residual error network, and for extracted features, multi-layer feature fusion is performed by using a feature pyramid network, so that fused features are guaranteed to have high-layer semantic information and low-layer detail texture information. Inputting the fused features into a regional suggestion layer, and extracting all foreground objects (regional suggestion is a process of rough classification and rough positioning, and can distinguish a target ground object from a background, does not distinguish a specific category, determines the approximate position of the target through a window, and roughly extracts the target) and approximate regions (rough positioning) which may exist in the target. Because the regions are different in size, the region alignment method is used, the regions of interest are mapped into feature maps with the same size and input into a subsequent classification regressor, and the specific category identification and the target position refinement are carried out, so that the accurate detection of the target is realized.
More specifically, when the feature of the ground object is extracted, a high-efficiency residual error network is used as a basic network framework (namely, the feature extraction and fusion module adopts a residual error network structure), so that deeper features can be extracted, and the characterization capability of the features is improved; for a multi-level feature mapping graph generated by a depth residual error network, a feature fusion network is used for carrying out low-level and high-level feature fusion, so that the fused feature graph has rich detail information and semantic information, and the classification and positioning accuracy of a model on a weak target is improved; the region of interest generation module may employ a region suggestion network: the RPN is used for inputting the multilayer characteristics into the RPN to perform layer-by-layer processing, generating anchor boxes with different proportions and different scales, providing classification and rough positioning to perform foreground extraction and background separation, performing region-of-interest (ROI) screening through confidence coefficient, and removing redundant regions; and accurately mapping the region of interest extracted based on the fused feature map by adopting a region alignment (ROI Align) method, projecting the ROI to the same size, ensuring the efficiency of the model in reverse propagation, and keeping the rest of the model consistent with a classical target detection framework based on region suggestion.
In one embodiment, the feature extraction and fusion module includes a down-sampling process and an up-sampling process when fusing the low-level features and the high-level features.
Specifically, the image downsampling process is similar to the convolution process of a convolution neural network, and each layer of n layers of features is gradually reduced like each level of a pyramid by setting the size of a convolution kernel; the up-sampling process performs 1 × 1 convolution (to reduce the number of feature channels) on each layer of feature map i generated by down-sampling, the interpolation is to perform superposition fusion on the feature maps with the same size as the i-1 layer, the obtained fusion features are subjected to 3x3 convolution processing (eliminating aliasing effect caused by upsampling) to generate n-1 layer fused feature maps and feature maps (for generating a larger-scale suggested region) after the feature maps of the highest layer are maximally pooled, and then inputting the n-1 layers of fused feature maps and the feature map with the highest layer of feature map maximized and pooled into a subsequent RPN network for layer-by-layer processing to generate an interested region, inputting the fused feature maps (with the pooled feature maps removed) into a region alignment (ROI align) module, and mapping the interested region on each layer of feature map to the same dimension.
Referring to fig. 2, a target detection model is specifically described, first, an input image is input into a WGAN-based image super-resolution reconstruction model to perform image super-resolution reconstruction, and then feature extraction and fusion, region-of-interest generation, and target generation detection are performed through the target detection model.
The target detection model is realized by the following steps:
and substep 1, feature extraction. A skip connection mode is adopted by a residual error network (ResNet), the method is different from the traditional convolution neural network in that input is convoluted to obtain output, and ResNet adds the input characteristics of the layer with the characteristics after convolution and outputs the sum to the next layer, so that the gradient disappearance phenomenon of the network can be avoided along with the deepening of the network layer number when back propagation training is carried out, the extracted detailed characteristic information is enriched, the performance of the model and the training convergence speed are improved, and the deep characteristic extraction and characteristic expression capacity of the model is greatly improved. The basic neural network adopted in the experiment is resnet101, the convolution of 3x3 and 7x7 is mainly used, and the deepening of the layer number can extract deep features of the target
And step 2, performing multi-layer feature fusion on the extracted features. In the multi-layer feature fusion algorithm, the down-sampling process of the image is set to be 5 layers, and each layer of the feature map is gradually decreased like each level of a pyramid; in the up-sampling process, after performing 1x1 convolution on each layer of feature map i generated by down-sampling (in order to reduce the number of feature channels), interpolating feature maps with the same size as the i-1 layer, performing superposition fusion on the feature maps, performing 3x3 convolution processing (eliminating aliasing effect caused by up-sampling) on the obtained fusion features to generate 4 layers of fused feature maps P2-P5 and the feature map P6 after the maximization pooling of the highest layer of feature map (in order to generate a larger-scale suggested region), inputting the feature maps of the P2-P5 layers (without the pooling) into a region alignment (ROI align) module, and mapping the regions of interest on each layer of feature map to the same dimension.
And 3, generating a region of interest. In the RPN, firstly, processing layer by layer is carried out on input multilayer features, 15 candidate frames with different scales and different proportions are generated on each position on a feature map through an Anchor method to serve as candidate target regions, whether the ground feature of each candidate frame belongs to the foreground or the background is obtained through calculation of a ground feature classification layer, and the central point coordinate and the length and width value of the corresponding region are obtained through a window regression layer, so that rough extraction of the region of interest is carried out. The Anchorbox generates bounding boxes with different scales and different length-width ratios on the image, and can detect multi-scale and multi-modal ground object targets. This example produced 15(3x5) different sizes (32,64,128,256,512) and different aspect ratios (1:1,1:2,2:1) of the Anchor box. During model training, the generated Anchor box is roughly classified through a classifier, foreground extraction and background separation are carried out, the Anchor box boundary frame is translated and zoomed, the final rough extraction confidence level is obtained, 3000 Anchor boxes with the confidence levels close to the front are reserved, and the rest redundant boundary frames are deleted.
And substep 4. aligning the regions of interest. When ROI align is used for mapping the interested region, the original length of a boundary frame is reserved, the region is divided into four parts, the value of the central point of each part is obtained by a bilinear interpolation method, and the mapped pixel value is obtained by a maximum pooling or mean pooling method, so that each interested region is mapped to the scale with the same size for subsequent classification regression operation. For a remote sensing image small target, the phenomena of low detection confidence, poor precision and recall caused by positioning deviation can be avoided. And inputting the feature map after aligning the region of interest into a subsequent classifier and a regression device, and performing accurate classification and position refinement on the target to determine the specific category and accurate position information of the target.
And step 5, the initial experimental weight uses the weight parameters trained on the basis of the COCO data set instead of adopting the fully initialized training weight, so that a better migration training effect can be achieved. The Anchor box adopts three ratios (0.5,1 and 2) and five sizes (32,64,128,256 and 512), namely 15 Anchor boxes are generated at each Anchor point, so that the remote sensing image target with multiple scales and multiple angles can be effectively detected; selecting 0.5 as a threshold value of an intersection ratio (IOU), namely, judging that the intersection ratio of a detection result and a real target area is more than 0.5 as effective detection, and avoiding target omission caused by overlarge threshold value for a remote sensing image with a small target size, wherein when the threshold value is set to be 0.75, the target detection precision is greatly reduced; the threshold value of redundant boundary frame removal of the NMS is 0.7, namely, the redundant Bbox of the same target is removed, and the boundary frames of surrounding ground objects are removed as far as possible; the initial learning rate is 0.001, the learning momentum is 0.9, and the weight attenuation is 0.0001; limited by GPU hardware, training the batch size to be 8, namely processing eight images per iteration; in consideration of the complexity and the target diversity of the remote sensing image, the ratio of the training, verifying and testing set is set to be 2:1: 1.
Step S5: inputting the reconstructed training data set into a target detection model, training the target detection model, calculating an optimized target function loss value, and performing training and updating on the model weight parameters through a back propagation algorithm.
In one embodiment, the formula for optimizing the objective function in step S5 is:
Figure BDA0002202799830000121
in the formula (4), Lcls=-log(pi) Representing the classification error, piRepresenting the probability value of the ground object class output by the classifier;
Figure BDA0002202799830000122
representing the positioning error of the model, where tx=(x-xa)/wa,ty=(y-ya)/ha
Figure BDA0002202799830000123
Data with subscript a is a parameter value of a bounding box output by an anchor box, data with asterisk superscript is a ground truth value, the rest are model prediction output values, lambda is a balance parameter, and N isclsTo train the blocksize value, NlocThe generated Anchor box value.
In the specific implementation process, the value of lambda is 10, NclsTo train the blocksize value, NlocTo generate an Anchor box value, this may be 2400.
Step S6: and verifying the model trained in the step S5 by using the reconstructed verification data set to obtain an optimized target detection model.
Specifically, the reconstructed verification data set is used for verifying the model trained in the stage, and training parameters are adjusted according to the actual target detection effect and the batch verification result until the loss functions of the model on the training set and the verification set tend to be stable, the ground object target detection reaches high precision and high recall rate, and the model obtains the optimal solution on the data set.
In the specific implementation process, the experiment adopts a segmented training mode, a training model is output after every 5000 times of training (1000-2000 times in the later period) in the early period, verification is carried out on a verification set, training parameters are adjusted according to the verification result, and training is continued until the training result meets the requirement.
Step S7: and inputting the reconstructed test data set into an optimized target detection model for target detection and identification, determining the category and position information of the ground object target, and detecting the weak and small target of the remote sensing image.
Specifically, when evaluating the target extraction result, it is unlikely that all the features will be correctly extracted due to the limitation of model performance, and this involves the Recall (Recall) concept, which represents the proportion of all the target features that the model can extract to the total number of targets. Similarly, each time the feature extraction is performed, there is a right error and a wrong error in the result output by the model, and the ratio of the number of correct predictions (i.e. the positive samples and the negative samples are both predicted correctly) in the result output by the model to the total prediction amount is the Precision ratio (Precision) of the target detection. AP (Average precision) represents the precision Average value under different recall values, and mAP (mean Average precision) represents the AP Average value of multiple types of ground objects. The AP and mAP values can be used for measuring precision of precision detection of the model, and are the most common target detection measurement indexes, but the recall value is important and cannot be ignored. The method can reflect whether the model is trained sufficiently or not, and can fully and completely detect the multi-modal target ground object without missing detection of the target. In the embodiment, the detection precision of the image weak target detection method constructed by the invention on the automobile is 0.9347, the recall rate is 0.9439, and the detection result of the remote sensing image weak target can be improved.
Generally, the remote sensing image in the training data set, the verification data set and the test data set is reconstructed by adopting a WGAN-based image super-resolution reconstruction model, the quality of the remote sensing image can be improved, the detail information of the image is enriched, so that a remote sensing image target detection model based on multi-layer feature fusion can extract more effective target features, the extracted low-layer features and high-layer features are fused, the semantic information and detail information of a fused feature map are ensured, the fused feature map is used for carrying out foreground target coarse extraction and interesting region generation through an RPN (resilient packet network) to carry out subsequent target accurate detection, the precision and recall ratio of target detection are improved, and the detection precision of the remote sensing image target is improved.
The method has the advantages that as the method is used for constructing the multi-layer characteristic fusion remote sensing image weak target detection algorithm of the combined image super-resolution reconstruction method, the image resolution is improved through the super-resolution reconstruction method aiming at the characteristics of low image imaging resolution and small target size, and the detail texture information of the image is enriched; the extracted low-level features and high-level features can be fused by utilizing a multi-feature fusion network, and the detection precision of the remote sensing image weak target is improved by constructing a target characteristic framework based on multi-level feature fusion.
In the embodiment of the present invention, a remote sensing image is used, but not limited to a remote sensing image. The method has wide universality for other images, such as video images, natural images, medical images and the like, is less limited by objective factors, and has wide application range. Through experimental verification, the multilayer feature fusion target detection algorithm of the combined image hyper-resolution reconstruction method constructed by the patent is proved to be capable of improving the detection precision and the recall rate of weak and small targets in the image.
It is to be noted and understood that various modifications and improvements can be made to the invention described in detail above without departing from the spirit and scope of the invention as claimed in the appended claims. Accordingly, the scope of the claimed subject matter is not limited by any of the specific exemplary teachings provided.

Claims (7)

1. A high-resolution remote sensing image weak target detection method based on deep learning is characterized by comprising the following steps:
step S1: based on the ground feature type and characteristics in the high-resolution remote sensing image, an initial data set is constructed, the initial data set is divided into a training data set, a verification data set and a test data set, and type information and position information labeling is carried out on ground feature samples in the initial data set;
step S2: taking the training data set and the verification data set as high-resolution samples, performing down-sampling on the high-resolution samples to obtain low-resolution samples, and then performing training and verification evaluation on the WGAN-based image super-resolution reconstruction model by using the low-resolution samples to obtain a trained super-resolution reconstruction model;
step S3: inputting the training data set, the verification data set and the test data set into the trained super-resolution reconstruction model to obtain a reconstructed training data set, verification data set and test data set;
step S4: constructing a remote sensing image target detection model based on multi-layer feature fusion by adopting a target detection basic framework based on region suggestion, and setting parameters of the target detection model according to the characteristics of the reconstructed training data set;
step S5: inputting the reconstructed training data set into a target detection model, training the target detection model, calculating an optimized target function loss value, and performing training and updating on the model weight parameters through a back propagation algorithm;
step S6: verifying the model trained in the step S5 by using the reconstructed verification data set to obtain an optimized target detection model;
step S7: and inputting the reconstructed test data set into an optimized target detection model for target detection and identification, determining the category and position information of the ground object target, and detecting the weak and small target of the remote sensing image.
2. The method as claimed in claim 1, wherein the step S1 introduces spatial context information for weak targets in the initial data set when labeling the ground feature samples in the initial data set with category information and location information.
3. The method of claim 1, wherein the optimization objective function of the WGAN-based super resolution reconstruction model in step S2 is:
Figure FDA0002202799820000021
wherein L represents the total Loss function of WGAN, Loss _ G represents the generator Loss function, Loss _ D represents the discriminator Loss function, fw(x) Representing a continuous mapping function for mapping the samples to have their derivative values within a range; e represents averaging the formula; prRepresenting true distribution, PgThe generation distribution is represented.
4. The method according to claim 1, wherein step S3 specifically comprises:
performing image reconstruction on remote sensing images contained in a training data set, a verification data set and a test data set by using a trained super-resolution reconstruction model, wherein quality evaluation on the super-resolution images is usually performed by adopting a peak signal-to-noise ratio and an image structure similarity index, wherein the peak signal-to-noise ratio PSNR is usually analyzed by adopting a formula (2), the image Structure Similarity Index (SSIM) is shown in a formula (3), MSE represents a mean square error, M and N represent image length and width values, M and N respectively represent pixel values corresponding to the images before and after reconstruction, s represents a gray level order, and the larger the PSNR value is, the smaller the distortion is;
Figure FDA0002202799820000022
Figure FDA0002202799820000023
wherein c is a constant,
Figure FDA0002202799820000024
μyand σyValue of (d) and mux、σxSimilarly, parameter C1=K1×L,C2=K2xL, L represents the maximum pixel value in a gray-scale image, K1And K2As a constant, SSIM is used to measure the similarity between images from brightness, contrast and texture, with larger values indicating less distortion and higher quality of the image.
5. The method according to claim 1, wherein the remote sensing image target detection model based on multi-layer feature fusion in step S4 includes a feature extraction and fusion module, a region-of-interest generation module and a target detection module, wherein the feature extraction and fusion module is used for extracting feature of the ground feature and fusing the extracted low-level features and high-level features, the region-of-interest generation module is used for extracting a region of interest based on the fused feature map, and the target detection module is used for performing target identification and detection according to the generated region of interest.
6. The method of claim 5, wherein the feature extraction and fusion module includes a down-sampling process and an up-sampling process when fusing the low-level features and the high-level features.
7. The method of claim 1, wherein the formula of the optimization objective function in step S5 is:
Figure FDA0002202799820000031
in the formula (4), Lcls=-log(pi) Representing the classification error, piRepresenting the probability value of the ground object class output by the classifier;
Figure FDA0002202799820000032
representing the positioning error of the model, where tx=(x-xa)/wa,ty=(y-ya)/ha
Figure FDA0002202799820000033
Data with subscript a is a parameter value of a bounding box output by an anchor box, data with asterisk superscript is a ground truth value, the rest are model prediction output values, lambda is a balance parameter, and N isclsTo train the blocksize value, NlocThe generated Anchor box value.
CN201910870991.1A 2019-09-16 2019-09-16 High-resolution remote sensing image weak target detection method based on deep learning Pending CN110728658A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910870991.1A CN110728658A (en) 2019-09-16 2019-09-16 High-resolution remote sensing image weak target detection method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910870991.1A CN110728658A (en) 2019-09-16 2019-09-16 High-resolution remote sensing image weak target detection method based on deep learning

Publications (1)

Publication Number Publication Date
CN110728658A true CN110728658A (en) 2020-01-24

Family

ID=69219060

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910870991.1A Pending CN110728658A (en) 2019-09-16 2019-09-16 High-resolution remote sensing image weak target detection method based on deep learning

Country Status (1)

Country Link
CN (1) CN110728658A (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111340700A (en) * 2020-02-21 2020-06-26 北京中科虹霸科技有限公司 Model generation method, resolution improvement method, image identification method and device
CN111460060A (en) * 2020-03-04 2020-07-28 华中科技大学 User region-of-interest remote sensing image space indexing method
CN111709307A (en) * 2020-05-22 2020-09-25 哈尔滨工业大学 Resolution enhancement-based remote sensing image small target detection method
CN111709291A (en) * 2020-05-18 2020-09-25 杭州电子科技大学 Takeaway personnel identity identification method based on fusion information
CN111861884A (en) * 2020-07-15 2020-10-30 南京信息工程大学 Satellite cloud image super-resolution reconstruction method based on deep learning
CN111899172A (en) * 2020-07-16 2020-11-06 武汉大学 Vehicle target detection method oriented to remote sensing application scene
CN111968064A (en) * 2020-10-22 2020-11-20 成都睿沿科技有限公司 Image processing method and device, electronic equipment and storage medium
CN112098997A (en) * 2020-09-18 2020-12-18 欧必翼太赫兹科技(北京)有限公司 Three-dimensional holographic imaging security inspection radar image foreign matter detection method
CN112329852A (en) * 2020-11-05 2021-02-05 西安泽塔云科技股份有限公司 Classification method and device for earth surface coverage images and electronic equipment
CN112395958A (en) * 2020-10-29 2021-02-23 中国地质大学(武汉) Remote sensing image small target detection method based on four-scale depth and shallow layer feature fusion
CN112907485A (en) * 2021-03-18 2021-06-04 国家海洋信息中心 Remote sensing image batch color matching method based on l alpha beta space color mapping
CN113033714A (en) * 2021-05-24 2021-06-25 华中师范大学 Object-oriented automatic machine learning method and system for multi-mode multi-granularity remote sensing image
CN113033446A (en) * 2021-04-01 2021-06-25 辽宁工程技术大学 Transmission tower identification and positioning method based on high-resolution remote sensing image
CN113052868A (en) * 2021-03-11 2021-06-29 奥比中光科技集团股份有限公司 Cutout model training and image cutout method and device
CN113111727A (en) * 2021-03-19 2021-07-13 西北工业大学 Method for detecting rotating target in remote sensing scene based on feature alignment
CN113378905A (en) * 2021-06-04 2021-09-10 武汉大学 Small target detection method based on distribution distance
CN113487579A (en) * 2021-07-14 2021-10-08 广州柏视医疗科技有限公司 Multi-mode migration method for automatically sketching model
CN113537214A (en) * 2021-07-16 2021-10-22 浙江工业大学 Automatic phase singularity identification method based on fast R-CNN and SRGAN
CN113538247A (en) * 2021-08-12 2021-10-22 中国科学院空天信息创新研究院 Super-resolution generation and conditional countermeasure network remote sensing image sample generation method
CN113591608A (en) * 2021-07-12 2021-11-02 浙江大学 High-resolution remote sensing image impervious surface extraction method based on deep learning
CN114331849A (en) * 2022-03-15 2022-04-12 之江实验室 Cross-mode nuclear magnetic resonance hyper-resolution network and image super-resolution method
CN114782255A (en) * 2022-06-16 2022-07-22 武汉大学 Semantic-based noctilucent remote sensing image high-resolution reconstruction method
CN114972737A (en) * 2022-06-08 2022-08-30 湖南大学 Remote sensing image target detection system and method based on prototype comparison learning
CN114998242A (en) * 2022-05-26 2022-09-02 电子科技大学 Method for detecting power transmission line pole tower in satellite image
CN115063428A (en) * 2022-08-18 2022-09-16 中国科学院国家空间科学中心 Spatial dim small target detection method based on deep reinforcement learning
CN115457396A (en) * 2022-09-26 2022-12-09 河北省科学院地理科学研究所 Surface target ground object detection method based on remote sensing image
CN116309274A (en) * 2022-12-12 2023-06-23 湖南红普创新科技发展有限公司 Method and device for detecting small target in image, computer equipment and storage medium
CN116862317A (en) * 2023-08-08 2023-10-10 广西壮族自治区自然资源遥感院 Satellite remote sensing monitoring system based on project full life cycle performance evaluation management
CN117095299A (en) * 2023-10-18 2023-11-21 浙江省测绘科学技术研究院 Grain crop extraction method, system, equipment and medium for crushing cultivation area
CN117253155A (en) * 2023-11-17 2023-12-19 山东大学 Human activity detection method and system based on deep learning
WO2024020676A1 (en) * 2022-07-26 2024-02-01 Deeplite Inc. System and method to utilize a reduced image resolution for computer vision applications
CN113111727B (en) * 2021-03-19 2024-05-31 西北工业大学 Feature alignment-based method for detecting rotating target in remote sensing scene

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1831556A (en) * 2006-04-14 2006-09-13 武汉大学 Single satellite remote sensing image small target super resolution ratio reconstruction method
CN105389797A (en) * 2015-10-16 2016-03-09 西安电子科技大学 Unmanned aerial vehicle video small-object detecting method based on super-resolution reconstruction
CN109344821A (en) * 2018-08-30 2019-02-15 西安电子科技大学 Small target detecting method based on Fusion Features and deep learning
CN109816593A (en) * 2019-01-18 2019-05-28 大连海事大学 A kind of super-resolution image reconstruction method of the generation confrontation network based on attention mechanism

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1831556A (en) * 2006-04-14 2006-09-13 武汉大学 Single satellite remote sensing image small target super resolution ratio reconstruction method
CN105389797A (en) * 2015-10-16 2016-03-09 西安电子科技大学 Unmanned aerial vehicle video small-object detecting method based on super-resolution reconstruction
CN109344821A (en) * 2018-08-30 2019-02-15 西安电子科技大学 Small target detecting method based on Fusion Features and deep learning
CN109816593A (en) * 2019-01-18 2019-05-28 大连海事大学 A kind of super-resolution image reconstruction method of the generation confrontation network based on attention mechanism

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
彭小飞 等: "复杂条件下小目标检测算法研究", 《智能计算机与应用》 *
畅青 等: "基于图像超分辨网络的目标检测算法", 《图形图像》 *

Cited By (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111340700A (en) * 2020-02-21 2020-06-26 北京中科虹霸科技有限公司 Model generation method, resolution improvement method, image identification method and device
CN111340700B (en) * 2020-02-21 2023-04-25 北京中科虹霸科技有限公司 Model generation method, resolution improvement method, image recognition method and device
CN111460060A (en) * 2020-03-04 2020-07-28 华中科技大学 User region-of-interest remote sensing image space indexing method
CN111460060B (en) * 2020-03-04 2023-03-28 华中科技大学 User region-of-interest remote sensing image space indexing method
CN111709291B (en) * 2020-05-18 2023-05-26 杭州电子科技大学 Takeaway personnel identity recognition method based on fusion information
CN111709291A (en) * 2020-05-18 2020-09-25 杭州电子科技大学 Takeaway personnel identity identification method based on fusion information
CN111709307B (en) * 2020-05-22 2022-08-30 哈尔滨工业大学 Resolution enhancement-based remote sensing image small target detection method
CN111709307A (en) * 2020-05-22 2020-09-25 哈尔滨工业大学 Resolution enhancement-based remote sensing image small target detection method
CN111861884A (en) * 2020-07-15 2020-10-30 南京信息工程大学 Satellite cloud image super-resolution reconstruction method based on deep learning
CN111861884B (en) * 2020-07-15 2023-05-16 南京信息工程大学 Satellite cloud image super-resolution reconstruction method based on deep learning
CN111899172A (en) * 2020-07-16 2020-11-06 武汉大学 Vehicle target detection method oriented to remote sensing application scene
CN112098997A (en) * 2020-09-18 2020-12-18 欧必翼太赫兹科技(北京)有限公司 Three-dimensional holographic imaging security inspection radar image foreign matter detection method
CN112098997B (en) * 2020-09-18 2021-10-15 欧必翼太赫兹科技(北京)有限公司 Three-dimensional holographic imaging security inspection radar image foreign matter detection method
CN111968064B (en) * 2020-10-22 2021-01-15 成都睿沿科技有限公司 Image processing method and device, electronic equipment and storage medium
CN111968064A (en) * 2020-10-22 2020-11-20 成都睿沿科技有限公司 Image processing method and device, electronic equipment and storage medium
CN112395958A (en) * 2020-10-29 2021-02-23 中国地质大学(武汉) Remote sensing image small target detection method based on four-scale depth and shallow layer feature fusion
CN112329852B (en) * 2020-11-05 2022-04-05 西安锐思数智科技股份有限公司 Classification method and device for earth surface coverage images and electronic equipment
CN112329852A (en) * 2020-11-05 2021-02-05 西安泽塔云科技股份有限公司 Classification method and device for earth surface coverage images and electronic equipment
CN113052868A (en) * 2021-03-11 2021-06-29 奥比中光科技集团股份有限公司 Cutout model training and image cutout method and device
WO2022188886A1 (en) * 2021-03-11 2022-09-15 奥比中光科技集团股份有限公司 Image matting model training method and apparatus, and image matting method and apparatus
CN112907485A (en) * 2021-03-18 2021-06-04 国家海洋信息中心 Remote sensing image batch color matching method based on l alpha beta space color mapping
CN113111727A (en) * 2021-03-19 2021-07-13 西北工业大学 Method for detecting rotating target in remote sensing scene based on feature alignment
CN113111727B (en) * 2021-03-19 2024-05-31 西北工业大学 Feature alignment-based method for detecting rotating target in remote sensing scene
CN113033446A (en) * 2021-04-01 2021-06-25 辽宁工程技术大学 Transmission tower identification and positioning method based on high-resolution remote sensing image
CN113033446B (en) * 2021-04-01 2024-02-02 辽宁工程技术大学 Transmission tower identification and positioning method based on high-resolution remote sensing image
CN113033714A (en) * 2021-05-24 2021-06-25 华中师范大学 Object-oriented automatic machine learning method and system for multi-mode multi-granularity remote sensing image
CN113033714B (en) * 2021-05-24 2021-08-03 华中师范大学 Object-oriented full-automatic machine learning method and system for multi-mode multi-granularity remote sensing image
CN113378905B (en) * 2021-06-04 2022-06-03 武汉大学 Small target detection method based on distribution distance
CN113378905A (en) * 2021-06-04 2021-09-10 武汉大学 Small target detection method based on distribution distance
CN113591608A (en) * 2021-07-12 2021-11-02 浙江大学 High-resolution remote sensing image impervious surface extraction method based on deep learning
CN113487579A (en) * 2021-07-14 2021-10-08 广州柏视医疗科技有限公司 Multi-mode migration method for automatically sketching model
CN113487579B (en) * 2021-07-14 2022-04-01 广州柏视医疗科技有限公司 Multi-mode migration method for automatically sketching model
CN113537214A (en) * 2021-07-16 2021-10-22 浙江工业大学 Automatic phase singularity identification method based on fast R-CNN and SRGAN
CN113538247A (en) * 2021-08-12 2021-10-22 中国科学院空天信息创新研究院 Super-resolution generation and conditional countermeasure network remote sensing image sample generation method
CN114331849A (en) * 2022-03-15 2022-04-12 之江实验室 Cross-mode nuclear magnetic resonance hyper-resolution network and image super-resolution method
CN114998242A (en) * 2022-05-26 2022-09-02 电子科技大学 Method for detecting power transmission line pole tower in satellite image
CN114972737B (en) * 2022-06-08 2024-03-15 湖南大学 Remote sensing image target detection system and method based on prototype contrast learning
CN114972737A (en) * 2022-06-08 2022-08-30 湖南大学 Remote sensing image target detection system and method based on prototype comparison learning
CN114782255A (en) * 2022-06-16 2022-07-22 武汉大学 Semantic-based noctilucent remote sensing image high-resolution reconstruction method
WO2024020676A1 (en) * 2022-07-26 2024-02-01 Deeplite Inc. System and method to utilize a reduced image resolution for computer vision applications
CN115063428A (en) * 2022-08-18 2022-09-16 中国科学院国家空间科学中心 Spatial dim small target detection method based on deep reinforcement learning
CN115457396B (en) * 2022-09-26 2023-06-23 河北省科学院地理科学研究所 Surface target ground object detection method based on remote sensing image
CN115457396A (en) * 2022-09-26 2022-12-09 河北省科学院地理科学研究所 Surface target ground object detection method based on remote sensing image
CN116309274A (en) * 2022-12-12 2023-06-23 湖南红普创新科技发展有限公司 Method and device for detecting small target in image, computer equipment and storage medium
CN116309274B (en) * 2022-12-12 2024-01-30 湖南红普创新科技发展有限公司 Method and device for detecting small target in image, computer equipment and storage medium
CN116862317A (en) * 2023-08-08 2023-10-10 广西壮族自治区自然资源遥感院 Satellite remote sensing monitoring system based on project full life cycle performance evaluation management
CN117095299B (en) * 2023-10-18 2024-01-26 浙江省测绘科学技术研究院 Grain crop extraction method, system, equipment and medium for crushing cultivation area
CN117095299A (en) * 2023-10-18 2023-11-21 浙江省测绘科学技术研究院 Grain crop extraction method, system, equipment and medium for crushing cultivation area
CN117253155A (en) * 2023-11-17 2023-12-19 山东大学 Human activity detection method and system based on deep learning
CN117253155B (en) * 2023-11-17 2024-03-15 山东大学 Human activity detection method and system based on deep learning

Similar Documents

Publication Publication Date Title
CN110728658A (en) High-resolution remote sensing image weak target detection method based on deep learning
CN110738697B (en) Monocular depth estimation method based on deep learning
CN113298818B (en) Remote sensing image building segmentation method based on attention mechanism and multi-scale features
CN110472627B (en) End-to-end SAR image recognition method, device and storage medium
CN110363215B (en) Method for converting SAR image into optical image based on generating type countermeasure network
CN112488210A (en) Three-dimensional point cloud automatic classification method based on graph convolution neural network
CN111161218A (en) High-resolution remote sensing image change detection method based on twin convolutional neural network
CN110111345B (en) Attention network-based 3D point cloud segmentation method
CN112836713A (en) Image anchor-frame-free detection-based mesoscale convection system identification and tracking method
CN110598613B (en) Expressway agglomerate fog monitoring method
Chawan et al. Automatic detection of flood using remote sensing images
CN109635726B (en) Landslide identification method based on combination of symmetric deep network and multi-scale pooling
CN112487900B (en) SAR image ship target detection method based on feature fusion
CN112950780B (en) Intelligent network map generation method and system based on remote sensing image
CN114913498A (en) Parallel multi-scale feature aggregation lane line detection method based on key point estimation
CN116279592A (en) Method for dividing travelable area of unmanned logistics vehicle
CN113610905A (en) Deep learning remote sensing image registration method based on subimage matching and application
CN113469097B (en) Multi-camera real-time detection method for water surface floaters based on SSD network
CN114266947A (en) Classification method and device based on fusion of laser point cloud and visible light image
CN114612315A (en) High-resolution image missing region reconstruction method based on multi-task learning
CN116665153A (en) Road scene segmentation method based on improved deep bv3+ network model
CN115456957B (en) Method for detecting change of remote sensing image by full-scale feature aggregation
CN110910497A (en) Method and system for realizing augmented reality map
CN116129234A (en) Attention-based 4D millimeter wave radar and vision fusion method
CN113762271A (en) SAR image semantic segmentation method and system based on irregular convolution kernel neural network model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200124