CN105243154A - Remote sensing image retrieval method and system based on significant point characteristics and spare self-encodings - Google Patents

Remote sensing image retrieval method and system based on significant point characteristics and spare self-encodings Download PDF

Info

Publication number
CN105243154A
CN105243154A CN201510708598.4A CN201510708598A CN105243154A CN 105243154 A CN105243154 A CN 105243154A CN 201510708598 A CN201510708598 A CN 201510708598A CN 105243154 A CN105243154 A CN 105243154A
Authority
CN
China
Prior art keywords
image
matrix
remarkable
characteristic
sparse
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510708598.4A
Other languages
Chinese (zh)
Other versions
CN105243154B (en
Inventor
邵振峰
周维勋
李从敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Original Assignee
Wuhan University WHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU filed Critical Wuhan University WHU
Priority to CN201510708598.4A priority Critical patent/CN105243154B/en
Publication of CN105243154A publication Critical patent/CN105243154A/en
Application granted granted Critical
Publication of CN105243154B publication Critical patent/CN105243154B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5838Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Library & Information Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

A remote sensing image retrieval method and system based on significant point characteristics and spare self-encodings are disclosed. The method comprises the steps of extracting characteristic points of each image from an image library to obtain a characteristic point matrix, and calculating a salient map of each image based on a visual attention model; performing binaryzation on the salient maps by a self-adaption threshold value method, performing a mask calculation with the characteristic point matrix to obtain filtered significant characteristics points; separately choosing a plurality of significant characteristics points from each training image to configure training samples; training a spare auto-encoder network according to a whitened training sample set to obtain a characteristic extractor; extracting characteristics by the characteristic extractor, performing a sparsification treatment on the extracted image characteristics by a threshold function to obtain a final characteristic vector for retrieval; and performing image retrieval according to a preset similarity measurement criterion based on the extracted characteristic vector. The automatic extraction of the image characteristics is realized through the trained spare auto-encoder network; and in addition, the extracted characteristics are quite high in identification level, so that the retrieval precision ratio is ensured.

Description

Based on remote sensing image retrieval method and the system of the sparse own coding of significant point characteristic sum
Technical field
The invention belongs to technical field of image processing, relate to a kind of remote sensing image retrieval method based on the sparse own coding of significant point characteristic sum and system.
Background technology
Along with the raising of remote sensing earth observation ability, retrievable remotely-sensed data presents the feature of diversification and magnanimity.But, mass remote sensing data for while all kinds of major application demand provides enrich data source, due to current ground data process and analysis ability not enough, " data magnanimity, information are flooded " problem of the large data of remote sensing becomes increasingly conspicuous.How utilizing emerging scientific algorithm technology and means, the interesting target in quick position and intelligent retrieval remote sensing images or region, be the large data processing of remote sensing and analyze facing challenges, is also field of remote sensing image processing problem in science urgently to be resolved hurrily.Remote Sensing Image Retrieval technology is the effective ways solving this bottleneck problem, studies efficient image retrieval technologies and has great importance.
Current remote Sensing Image Retrieval technology mainly by carrying out similarity measurement to the low-level feature of image, and then returns similar image.Compare traditional search method based on key word, content-based retrieval method efficiency and accuracy higher, but to design a kind of character description method that effectively can describe various complicated remote sensing images scene be very difficult.In recent years, degree of depth study became the study hotspot of field of image recognition gradually owing to having good feature learning ability.Compare the feature of engineer, the method based on degree of depth study obtains a feature extractor to realize the automatic extraction of characteristics of image by sample training, is applicable to the remote Sensing Image Retrieval comprising complex scene.Because network design is relative with training simple, sparse own coding has become a kind of conventional degree of depth learning method, and is widely used in image procossing.
Train for sparse autoencoder network, in structure training sample, existing method is normally from the image block structure training sample of random selecting some and size training image, and this sample architecture method has following defect.The first, from the angle of human eye vision theory, what people paid close attention to is specific objective on remote sensing images, and the image block of random selecting may not comprise the specific objective of concern.The second, because the size of training image is fixed, therefore the method for random selecting image block structure training sample may cause lack of training samples.3rd, because training sample is image block, obtain being image block but not the feature of entire image when utilizing the network trained to carry out feature extraction, therefore can not be directly used in image retrieval.In order to obtain the feature of entire image, usually need to adopt the method for convolution, but this process not only counting yield is low but also can introduce other parameters.In activation function selection, existing method adopts sigmoid function as the activation function of network hidden layer neuron usually, and sigmoid function exists the problems such as serious gradient disappearance when network backpropagation, be unfavorable for network training.For sparse autoencoder network feature extraction, existing method normally direct using the activation value of hidden layer as extract feature and not through LS-SVM sparseness, and experiment show that sparse features performance is better.
Summary of the invention
For the deficiency that prior art exists, the invention provides a kind of remote Sensing Image Retrieval technical scheme based on the sparse own coding of significant point characteristic sum.The present invention will extract the significant point feature of remote sensing images as the input of sparse autoencoder network and then train it, finally utilizes the feature extractor of training to extract characteristics of image to realize remote Sensing Image Retrieval.
The technical solution adopted in the present invention is a kind of remote sensing image retrieval method based on the sparse own coding of significant point characteristic sum, comprises the following steps:
Step 1, the unique point extracting each image in image library obtains characteristic point matrix, and utilizes visual attention model to calculate the remarkable figure of each image;
Step 2, for the remarkable figure of image each in image library, respectively adopt Adaptive Thresholding by remarkable figure binaryzation, and the characteristic point matrix corresponding to image carry out mask computing obtain filter after remarkable characteristic; Implementation is as follows,
When adopting Adaptive Thresholding by remarkable figure binaryzation, according to the conspicuousness size of specific image element, determine that the binary-state threshold T of remarkable figure is as follows,
T = 2 w × h Σ x = 1 w Σ y = 1 h I ( x , y )
Wherein, w and h represents that remarkable figure's is wide and high respectively, and I (x, y) represents the saliency value of specific image element (x, y);
If according to binary-state threshold T to remarkable figure binaryzation, obtain binaryzation and significantly scheme, matrix I should be had mutually binaryif P represents the characteristic point matrix of image, P irepresent the notable feature dot matrix after filtering, calculate notable feature dot matrix as follows,
P I = P ⊗ I b i n a r y
Step 3, from image library, get some images as training image, choose some remarkable characteristics structure training sample respectively from each training image, obtain training sample set X, train sparse autoencoder network according to the training sample set X ' after albefaction, obtain feature extractor;
Described sparse autoencoder network input layer, hidden layer and output layer, wherein hidden layer neuron adopts ReLU function as activation function, and output layer neuron adopts softplus function as activation function, and the cost function of sparse autoencoder network is defined as follows,
J ( W , b ) = 1 2 | | X ′ - H W , b | | 2 + λ 2 | | W | | 2
Wherein, Section 1 is square error item, and Section 2 is regular terms, H w,brepresent the network output valve of training sample set X ', W=[W 1, W 2] and b=[b 1, b 2] represent weights W between network input layer and hidden layer respectively 1with biased b 1and the weights W between hidden layer and output layer 2with biased b 2the weight matrix formed, λ represents regularization coefficient;
Step 4, to all images in image library, utilizes step 3 to train the feature extractor of gained to carry out feature extraction, and carries out LS-SVM sparseness with threshold function table to the characteristics of image extracted, and obtains the final proper vector for retrieving; Implementation is as follows,
The characteristics of image Y extracted is expressed as follows,
Y=f 1(W 1P I′+b 1)
Wherein, notable feature dot matrix P i' be according to step 2 gained filter after notable feature dot matrix P iresult after albefaction;
For the characteristics of image Y extracted, carry out following LS-SVM sparseness and obtain sparse features matrix Z,
Z=[Z +,Z -]=[max(0,Y-α),max(0,α-Y)]
Wherein, α represents the threshold value of threshold function table, matrix Z +=max (0, Y-α), Z -=max (0, α-Y);
If the SIFT point number detected from piece image is n, sparse features matrix Z is processed further, obtains proper vector F as follows,
F = 1 n Σ i = 1 n [ Z + i , Z - i ]
Wherein, with representing matrix Z respectively +and Z -i-th column vector.
Step 5, based on the proper vector that step 4 is extracted, the similarity measurement criterion according to presetting carries out image retrieval.
And in step 1, the unique point extracting each image in image library obtains characteristic point matrix, utilizes SIFT operator extraction to realize.
And in step 5, the similarity measurement criterion preset adopts city distance.
The present invention is also corresponding provides a kind of Content-based Remote Sensing Image Retrieval System based on the sparse own coding of significant point characteristic sum, comprises with lower module,
Feature point extraction module, obtains characteristic point matrix for the unique point extracting each image in image library, and utilizes visual attention model to calculate the remarkable figure of each image;
Remarkable characteristic extraction module, for the remarkable figure for image each in image library, respectively adopt Adaptive Thresholding by remarkable figure binaryzation, and the characteristic point matrix corresponding to image carry out mask computing obtain filter after remarkable characteristic; Implementation is as follows,
When adopting Adaptive Thresholding by remarkable figure binaryzation, according to the conspicuousness size of specific image element, determine that the binary-state threshold T of remarkable figure is as follows,
T = 2 w × h Σ x = 1 w Σ y = 1 h I ( x , y )
Wherein, w and h represents that remarkable figure's is wide and high respectively, and I (x, y) represents the saliency value of specific image element (x, y);
If according to binary-state threshold T to remarkable figure binaryzation, obtain binaryzation and significantly scheme, matrix I should be had mutually binaryif P represents the characteristic point matrix of image, P irepresent the notable feature dot matrix after filtering, calculate notable feature dot matrix as follows,
P I = P ⊗ I b i n a r y
Training module, for getting some images as training image from image library, choosing some remarkable characteristics structure training sample respectively from each training image, obtaining training sample set X, train sparse autoencoder network according to the training sample set X ' after albefaction, obtain feature extractor;
Described sparse autoencoder network input layer, hidden layer and output layer, wherein hidden layer neuron adopts ReLU function as activation function, and output layer neuron adopts softplus function as activation function, and the cost function of sparse autoencoder network is defined as follows,
J ( W , b ) = 1 2 | | X ′ - H W , b | | 2 + λ 2 | | W | | 2
Wherein, Section 1 is square error item, and Section 2 is regular terms, H w,brepresent the network output valve of training sample set X ', W=[W 1, W 2] and b=[b 1, b 2] represent weights W between network input layer and hidden layer respectively 1with biased b 1and the weights W between hidden layer and output layer 2with biased b 2the weight matrix formed, λ represents regularization coefficient;
Characteristic extracting module, for all images in image library, utilizes step 3 to train the feature extractor of gained to carry out feature extraction, and carries out LS-SVM sparseness with threshold function table to the characteristics of image extracted, and obtains the final proper vector for retrieving; Implementation is as follows,
The characteristics of image Y extracted is expressed as follows,
Y=f 1(W 1P I′+b 1)
Wherein, notable feature dot matrix P i' be according to step 2 gained filter after notable feature dot matrix P iresult after albefaction;
For the characteristics of image Y extracted, carry out following LS-SVM sparseness and obtain sparse features matrix Z,
Z=[Z +,Z -]=[max(0,Y-α),max(0,α-Y)]
Wherein, α represents the threshold value of threshold function table, matrix Z +=max (0, Y-α), Z -=max (0, α-Y);
If the SIFT point number detected from piece image is n, sparse features matrix Z is processed further, obtains proper vector F as follows,
F = 1 n Σ i = 1 n [ Z + i , Z - i ]
Wherein, with representing matrix Z respectively +and Z -i-th column vector.
Retrieval module, for the proper vector that feature based extraction module extracts, the similarity measurement criterion according to presetting carries out image retrieval.
And in feature point extraction module, the unique point extracting each image in image library obtains characteristic point matrix, utilizes SIFT operator extraction to realize.
And in retrieval module, the similarity measurement criterion preset adopts city distance.
Compared with prior art, the present invention has following features and beneficial effect,
1, adopt the remarkable figure of visual attention model computed image, and remarkable figure binaryzation is filtered the remarkable characteristic obtaining image to the unique point that SIFT extracts, not only meet the vision attention feature of human eye but also the Search Requirement of people can be reflected better.
2, choose the remarkable characteristic structure training sample of image, compensate for the defect of the traditional structure of grab sample on training image training sample.
3, utilize sparse autoencoder network to train the feature extractor obtained to achieve the automatic extraction of characteristics of image, eliminate the characteristic Design process for complicated remote sensing images.
4, favorable expandability, training sample includes but are not limited to remarkable characteristic.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the embodiment of the present invention.
Embodiment
The unique point that first remote sensing image retrieval method based on the sparse own coding of significant point characteristic sum that the present invention proposes extracts image obtains characteristic point matrix, and the remarkable figure of computed image, then adopt adaptive threshold that remarkable figure binaryzation and characteristic point matrix are carried out " mask " computing and obtain remarkable characteristic, then the remarkable characteristic structure training sample choosing some trains sparse autoencoder network, and the proper vector utilizing the feature extractor of training automatically to extract characteristics of image to obtain for retrieving, the method for measuring similarity that last basis is preset carries out image retrieval and returns similar image.
For describing technical solution of the present invention in detail, see Fig. 1, embodiment flow process is provided to be described as follows:
Step 1, the unique point extracting each image in image library obtains characteristic point matrix, and utilizes visual attention model to calculate the remarkable figure of each image.
During concrete enforcement, the image library that can adopt existing image library or be built voluntarily by those skilled in the art.Such as choose the high-resolution remote sensing image that a width comprises multiple atural object classification, adopt Tiles partitioned mode to carry out cutting and build the retrieving images storehouse comprising multiple classification.For the every piece image in image library, first embodiment adopts the unique point (key point) of SIFT (ScaleInvariantFeatureTransform) operator extraction image to obtain characteristic point matrix, then the remarkable figure of GBVS (Graph-BasedVisualSaliency) model computed image is adopted, Tile method of partition, SIFT operator and GBVS model are prior aries, and it will not go into details in the present invention.
Step 2, for the remarkable figure of image each in image library, respectively adopt Adaptive Thresholding by remarkable figure binaryzation, and the characteristic point matrix corresponding to image carry out " mask " computing obtain filter after remarkable characteristic.
Determine the binary-state threshold of remarkable figure in embodiment according to the conspicuousness size of pixel, binaryzation obtains remarkable characteristic after significantly scheming to carry out " mask " computing with characteristic point matrix, realizes as follows:
According to the conspicuousness size of specific image element, determined the binary-state threshold T of remarkable figure by formula (1).
T = 2 w × h Σ x = 1 w Σ y = 1 h I ( x , y ) - - - ( 1 )
Wherein, w and h represents that remarkable figure's is wide and high respectively, and I (x, y) represents the saliency value of significantly figure (x, y) place pixel.
According to binary-state threshold T to remarkable figure binaryzation, obtain binaryzation and significantly scheme, matrix I should be had mutually binary.Utilize the remarkable figure of binaryzation to carry out filtration to the characteristic point matrix of image and obtain remarkable characteristic.If P represents the characteristic point matrix of image, P irepresent the notable feature dot matrix after filtering, then notable feature dot matrix calculates by formula (2).
P I = P ⊗ I b i n a r y - - - ( 2 )
Wherein,
Matrix
Each element representation SIFT key point characteristic of correspondence vector of matrix P, and SIFT key point characteristic of correspondence vector is generally 128 dimensions, the corresponding use 128 of the embodiment of the present invention is tieed up;
Matrix
Wherein, P 128(x, y) representation feature point characteristic of correspondence vector, if (x, y) place pixel does not have unique point, P 128(x, y)=0.I binaryin each element be 0 or 1, I binary(x, y) represents the value of the remarkable figure of binaryzation at (x, y) place.Symbol for scale multiplication symbol.
Step 3, chooses some images as training image from image library, chooses some remarkable characteristics structure training sample respectively, train sparse autoencoder network, obtain feature extractor from each training image.
In embodiment, choose the remarkable characteristic of the training image of some and unconventional image block structure training sample in step 3, during training, select ReLU (RectifiedLinearUnits) function and unconventional sigmoid function as the activation function of sparse autoencoder network hidden layer neuron.Such as, in step 3, each remarkable characteristic is the proper vector of 4 × 4 × 8=128 dimension, and a unique point forms a training sample.During concrete enforcement, in the number of training image, a width training image, the number of remarkable characteristic can be specified voluntarily by those skilled in the art.
Be implemented as follows:
First, choose the remarkable characteristic of image, structure training sample set.
First from image library, the image of random selecting some is as training image for embodiment, and then the remarkable characteristic of the training image of random selecting some constructs training sample set.Training sample set can represent with formula (3):
Wherein, m represents the number of training sample, and a remarkable characteristic is shown in each list of X, i.e. a training sample.Such as, [x 1,1, x 2,1..., x 128,1] be the 1st training sample, [x 1,2, x 2,2..., x 128,2] be the 2nd training sample.
Then, sparse autoencoder network is trained to obtain feature extractor.
There is certain correlativity in the remarkable characteristic extracted due to same width training image, therefore directly training sample set X can not be inputted sparse autoencoder network and train.ZCA (ZeroComponentAnalysis) albefaction is adopted to process the training sample set X ' after obtaining albefaction to training sample before training, and correlation parameter when preserving ZCA albefaction, ZCA albefaction is embodied as prior art, and it will not go into details in the present invention.
Embodiment defines the sparse autoencoder network that comprises input layer, hidden layer and output layer 3 layers, and wherein hidden layer neuron adopts ReLU function f 1(0, x) as activation function, output layer neuron adopts softplus function f to=max 2=ln (1+e x) as activation function.Compare traditional sigmoid function, ReLU function can be alleviated gradient disappearance problem to a certain extent and be more conducive to network training.Given training sample set X ', then the cost function of sparse autoencoder network may be defined as formula (4).
J ( W , b ) = 1 2 | | X ′ - H W , b | | 2 + λ 2 | | W | | 2 - - - ( 4 )
In formula, Section 1 is square error item, and Section 2 is regular terms, H w,brepresent the network output valve of training sample set X ', W=[W 1, W 2] and b=[b 1, b 2] represent weights W between network input layer and hidden layer respectively 1with biased b 1and the weights W between hidden layer and output layer 2with biased b 2the weight matrix formed, λ represents regularization coefficient.During concrete enforcement, the cost function in the method optimized-types (4) such as Gradient Descent during training, can be adopted to obtain weight and bias matrix parameter W and b.
Step 4, to all images in image library, utilizes step 3 to train the feature extractor of gained to carry out feature extraction, and carries out LS-SVM sparseness with threshold function table to the feature extracted, and obtains the final proper vector for retrieving.
In the step 4 of embodiment, the remarkable characteristic input feature vector extraction apparatus of image is carried out mapping and obtain corresponding characteristics of image, recycling threshold function table carries out LS-SVM sparseness to the feature extracted and can obtain the final proper vector for retrieving.
The characteristics of image Y extracted can be expressed as follows with formula (5),
Y=f 1(W 1P I′+b 1)(5)
Wherein, by W 1p i+ b 1reLU function f is substituted into as variable x 1=max (0, x), notable feature dot matrix P used herein i' be notable feature dot matrix after filtering according to step 2 gained, use the ZCA whitening parameters identical with when carrying out albefaction to training sample set X to carry out pretreated result.For the characteristics of image Y extracted, carry out LS-SVM sparseness by formula (6) and obtain sparse features matrix Z.
Z=[Z +,Z -]=[max(0,Y-α),max(0,α-Y)](6)
Wherein, α represents the threshold value of threshold function table f=max (0, x-α) and f=max (0, α-Y), matrix Z +=max (0, Y-α), Z -=max (0, α-Y).
In order to obtain the final proper vector F for retrieving, if the SIFT point number detected from piece image is n, by formula (7), sparse features matrix Z is processed further.
F = 1 n Σ i = 1 n [ Z + i , Z - i ] - - - ( 7 )
Wherein, with representing matrix Z respectively +and Z -i-th column vector.
Step 5, based on the proper vector that step 4 is extracted, the similarity measurement criterion according to presetting carries out image retrieval: when specifically implementing, those skilled in the art can preset similarity measurement criterion voluntarily.Embodiment adopts city distance (L1 norm) to calculate the similarity of query image and other images, and returns associated picture by similarity size.During concrete enforcement, can arbitrary image be query image in image library, obtain the associated picture returned by similarity size, to other images beyond image library, also can adopt and extract proper vector in the same way, and retrieve from image library.
During concrete enforcement, above flow process can adopt computer software mode to realize automatic operational scheme, and modular mode also can be adopted to provide corresponding system.The present invention is also corresponding provides a kind of Content-based Remote Sensing Image Retrieval System based on the sparse own coding of significant point characteristic sum, comprises with lower module,
Feature point extraction module, obtains characteristic point matrix for the unique point extracting each image in image library, and utilizes visual attention model to calculate the remarkable figure of each image;
Remarkable characteristic extraction module, for the remarkable figure for image each in image library, respectively adopt Adaptive Thresholding by remarkable figure binaryzation, and the characteristic point matrix corresponding to image carry out mask computing obtain filter after remarkable characteristic; Implementation is as follows,
When adopting Adaptive Thresholding by remarkable figure binaryzation, according to the conspicuousness size of specific image element, determine that the binary-state threshold T of remarkable figure is as follows,
T = 2 w × h Σ x = 1 w Σ y = 1 h I ( x , y )
Wherein, w and h represents that remarkable figure's is wide and high respectively, and I (x, y) represents the saliency value of specific image element (x, y);
If according to binary-state threshold T to remarkable figure binaryzation, obtain binaryzation and significantly scheme, matrix I should be had mutually binaryif P represents the characteristic point matrix of image, P irepresent the notable feature dot matrix after filtering, calculate notable feature dot matrix as follows,
P I = P ⊗ I b i n a r y
Training module, for getting some images as training image from image library, choosing some remarkable characteristics structure training sample respectively from each training image, obtaining training sample set X, train sparse autoencoder network according to the training sample set X ' after albefaction, obtain feature extractor;
Described sparse autoencoder network input layer, hidden layer and output layer, wherein hidden layer neuron adopts ReLU function as activation function, and output layer neuron adopts softplus function as activation function, and the cost function of sparse autoencoder network is defined as follows,
J ( W , b ) = 1 2 | | X ′ - H W , b | | 2 + λ 2 | | W | | 2
Wherein, Section 1 is square error item, and Section 2 is regular terms, H w,brepresent the network output valve of training sample set X ', W=[W 1, W 2] and b=[b 1, b 2] represent weights W between network input layer and hidden layer respectively 1with biased b 1and the weights W between hidden layer and output layer 2with biased b 2the weight matrix formed, λ represents regularization coefficient;
Query characteristics extraction module, for image to be checked, utilizes step 3 to train the feature extractor of gained to carry out feature extraction, and carries out LS-SVM sparseness with threshold function table to the characteristics of image extracted, and obtains the final proper vector for retrieving; Implementation is as follows,
The characteristics of image Y extracted is expressed as follows,
Y=f 1(W 1P I′+b 1)
Wherein, notable feature dot matrix P i' be according to step 2 gained filter after notable feature dot matrix P iresult after albefaction;
For the characteristics of image Y extracted, carry out following LS-SVM sparseness and obtain sparse features matrix Z,
Z=[Z +,Z -]=[max(0,Y-α),max(0,α-Y)]
Wherein, α represents the threshold value of threshold function table, matrix Z +=max (0, Y-α), Z -=max (0, α-Y);
If the SIFT point number detected from piece image is n, sparse features matrix Z is processed further, obtains proper vector F as follows,
F = 1 n Σ i = 1 n [ Z + i , Z - i ]
Wherein, with representing matrix Z respectively +and Z -i-th column vector.
Retrieval module, for the proper vector extracted based on query characteristics extraction module, the similarity measurement criterion according to presetting carries out image retrieval.
Specific embodiment described herein is only to the present invention's explanation for example.Those skilled in the art can make various amendment or supplement or adopt similar mode to substitute to described specific embodiment, but can't depart from spirit of the present invention or surmount the scope that appended claims defines.

Claims (6)

1. based on a remote sensing image retrieval method for the sparse own coding of significant point characteristic sum, it is characterized in that: comprise the following steps,
Step 1, the unique point extracting each image in image library obtains characteristic point matrix, and utilizes visual attention model to calculate the remarkable figure of each image;
Step 2, for the remarkable figure of image each in image library, respectively adopt Adaptive Thresholding by remarkable figure binaryzation, and the characteristic point matrix corresponding to image carry out mask computing obtain filter after remarkable characteristic; Implementation is as follows,
When adopting Adaptive Thresholding by remarkable figure binaryzation, according to the conspicuousness size of specific image element, determine that the binary-state threshold T of remarkable figure is as follows,
T = 2 w × h Σ x = 1 w Σ y = 1 h I ( x , y )
Wherein, w and h represents that remarkable figure's is wide and high respectively, and I (x, y) represents the saliency value of specific image element (x, y);
If according to binary-state threshold T to remarkable figure binaryzation, obtain binaryzation and significantly scheme, matrix I should be had mutually binaryif P represents the characteristic point matrix of image, P irepresent the notable feature dot matrix after filtering, calculate notable feature dot matrix as follows,
P I = P ⊗ I b i n a r y
Step 3, from image library, get some images as training image, choose some remarkable characteristics structure training sample respectively from each training image, obtain training sample set X, train sparse autoencoder network according to the training sample set X ' after albefaction, obtain feature extractor;
Described sparse autoencoder network input layer, hidden layer and output layer, wherein hidden layer neuron adopts ReLU function as activation function, and output layer neuron adopts softplus function as activation function, and the cost function of sparse autoencoder network is defined as follows,
J ( W , b ) = 1 2 | | X ′ - H W , b | | 2 + λ 2 | | W | | 2
Wherein, Section 1 is square error item, and Section 2 is regular terms, H w,brepresent the network output valve of training sample set X ', W=[W 1, W 2] and b=[b 1, b 2] represent weights W between network input layer and hidden layer respectively 1with biased b 1and the weights W between hidden layer and output layer 2with biased b 2the weight matrix formed, λ represents regularization coefficient;
Step 4, to all images in image library, utilizes step 3 to train the feature extractor of gained to carry out feature extraction, and carries out LS-SVM sparseness with threshold function table to the characteristics of image extracted, and obtains the final proper vector for retrieving; Implementation is as follows,
The characteristics of image Y extracted is expressed as follows,
Y=f 1(W 1P I′+b 1)
Wherein, notable feature dot matrix P i' be according to step 2 gained filter after notable feature dot matrix P iresult after albefaction;
For the characteristics of image Y extracted, carry out following LS-SVM sparseness and obtain sparse features matrix Z,
Z=[Z +,Z -]=[max(0,Y-α),max(0,α-Y)]
Wherein, α represents the threshold value of threshold function table, matrix Z +=max (0, Y-α), Z -=max (0, α-Y);
If the SIFT point number detected from piece image is n, sparse features matrix Z is processed further, obtains proper vector F as follows,
F = 1 n Σ i = 1 n [ Z + i , Z - i ]
Wherein, with representing matrix Z respectively +and Z -i-th column vector.
Step 5, based on the proper vector that step 4 is extracted, the similarity measurement criterion according to presetting carries out image retrieval.
2. according to claim 1 based on the remote sensing image retrieval method of the sparse own coding of significant point characteristic sum, it is characterized in that: in step 1, the unique point extracting each image in image library obtains characteristic point matrix, utilizes SIFT operator extraction to realize.
3. according to claim 1 or 2 based on the remote sensing image retrieval method of the sparse own coding of significant point characteristic sum, it is characterized in that: in step 5, the similarity measurement criterion preset adopts city distance.
4. based on a Content-based Remote Sensing Image Retrieval System for the sparse own coding of significant point characteristic sum, it is characterized in that: comprise with lower module,
Feature point extraction module, obtains characteristic point matrix for the unique point extracting each image in image library, and utilizes visual attention model to calculate the remarkable figure of each image;
Remarkable characteristic extraction module, for the remarkable figure for image each in image library, respectively adopt Adaptive Thresholding by remarkable figure binaryzation, and the characteristic point matrix corresponding to image carry out mask computing obtain filter after remarkable characteristic; Implementation is as follows,
When adopting Adaptive Thresholding by remarkable figure binaryzation, according to the conspicuousness size of specific image element, determine that the binary-state threshold T of remarkable figure is as follows,
T = 2 w × h Σ x = 1 w Σ y = 1 h I ( x , y )
Wherein, w and h represents that remarkable figure's is wide and high respectively, and I (x, y) represents the saliency value of specific image element (x, y);
If according to binary-state threshold T to remarkable figure binaryzation, obtain binaryzation and significantly scheme, matrix I should be had mutually binaryif P represents the characteristic point matrix of image, P irepresent the notable feature dot matrix after filtering, calculate notable feature dot matrix as follows,
P I = P ⊗ I b i n a r y
Training module, for getting some images as training image from image library, choosing some remarkable characteristics structure training sample respectively from each training image, obtaining training sample set X, train sparse autoencoder network according to the training sample set X ' after albefaction, obtain feature extractor;
Described sparse autoencoder network input layer, hidden layer and output layer, wherein hidden layer neuron adopts ReLU function as activation function, and output layer neuron adopts softplus function as activation function, and the cost function of sparse autoencoder network is defined as follows,
J ( W , b ) = 1 2 | | X ′ - H W , b | | 2 + λ 2 | | W | | 2
Wherein, Section 1 is square error item, and Section 2 is regular terms, H w,brepresent the network output valve of training sample set X ', W=[W 1, W 2] and b=[b 1, b 2] represent weights W between network input layer and hidden layer respectively 1with biased b 1and the weights W between hidden layer and output layer 2with biased b 2the weight matrix formed, λ represents regularization coefficient;
Characteristic extracting module, for all images in image library, utilizes step 3 to train the feature extractor of gained to carry out feature extraction, and carries out LS-SVM sparseness with threshold function table to the characteristics of image extracted, and obtains the final proper vector for retrieving; Implementation is as follows,
The characteristics of image Y extracted is expressed as follows,
Y=f 1(W 1P I′+b 1)
Wherein, notable feature dot matrix P i' be according to step 2 gained filter after notable feature dot matrix P iresult after albefaction;
For the characteristics of image Y extracted, carry out following LS-SVM sparseness and obtain sparse features matrix Z,
Z=[Z +,Z -]=[max(0,Y-α),max(0,α-Y)]
Wherein, α represents the threshold value of threshold function table, matrix Z +=max (0, Y-α), Z -=max (0, α-Y);
If the SIFT point number detected from piece image is n, sparse features matrix Z is processed further, obtains proper vector F as follows,
F = 1 n Σ i = 1 n [ Z + i , Z - i ]
Wherein, with representing matrix Z respectively +and Z -i-th column vector.
Retrieval module, for the proper vector extracted based on query characteristics extraction module, the similarity measurement criterion according to presetting carries out image retrieval.
5. according to claim 4 based on the Content-based Remote Sensing Image Retrieval System of the sparse own coding of significant point characteristic sum, it is characterized in that: in feature point extraction module, the unique point extracting each image in image library obtains characteristic point matrix, utilizes SIFT operator extraction to realize.
6. according to claim 4 or 5 based on the Content-based Remote Sensing Image Retrieval System of the sparse own coding of significant point characteristic sum, it is characterized in that: in retrieval module, the similarity measurement criterion preset adopts city distance.
CN201510708598.4A 2015-10-27 2015-10-27 Remote sensing image retrieval method based on notable point feature and sparse own coding and system Active CN105243154B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510708598.4A CN105243154B (en) 2015-10-27 2015-10-27 Remote sensing image retrieval method based on notable point feature and sparse own coding and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510708598.4A CN105243154B (en) 2015-10-27 2015-10-27 Remote sensing image retrieval method based on notable point feature and sparse own coding and system

Publications (2)

Publication Number Publication Date
CN105243154A true CN105243154A (en) 2016-01-13
CN105243154B CN105243154B (en) 2018-08-21

Family

ID=55040802

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510708598.4A Active CN105243154B (en) 2015-10-27 2015-10-27 Remote sensing image retrieval method based on notable point feature and sparse own coding and system

Country Status (1)

Country Link
CN (1) CN105243154B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105718531A (en) * 2016-01-14 2016-06-29 广州市万联信息科技有限公司 Image database building method and image recognition method
CN106228130A (en) * 2016-07-19 2016-12-14 武汉大学 Remote sensing image cloud detection method of optic based on fuzzy autoencoder network
CN106295613A (en) * 2016-08-23 2017-01-04 哈尔滨理工大学 A kind of unmanned plane target localization method and system
CN106909924A (en) * 2017-02-18 2017-06-30 北京工业大学 A kind of remote sensing image method for quickly retrieving based on depth conspicuousness
CN107122809A (en) * 2017-04-24 2017-09-01 北京工业大学 Neural network characteristics learning method based on image own coding
CN107515895A (en) * 2017-07-14 2017-12-26 中国科学院计算技术研究所 A kind of sensation target search method and system based on target detection
CN108830172A (en) * 2018-05-24 2018-11-16 天津大学 Aircraft remote sensing images detection method based on depth residual error network and SV coding
CN109259733A (en) * 2018-10-25 2019-01-25 深圳和而泰智能控制股份有限公司 Apnea detection method, apparatus and detection device in a kind of sleep
CN111144483A (en) * 2019-12-26 2020-05-12 歌尔股份有限公司 Image feature point filtering method and terminal
CN112731410A (en) * 2020-12-25 2021-04-30 上海大学 Underwater target sonar detection method based on CNN

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102073748A (en) * 2011-03-08 2011-05-25 武汉大学 Visual keyword based remote sensing image semantic searching method
CN102867196A (en) * 2012-09-13 2013-01-09 武汉大学 Method for detecting complex sea-surface remote sensing image ships based on Gist characteristic study
CN103309982A (en) * 2013-06-17 2013-09-18 武汉大学 Remote sensing image retrieval method based on vision saliency point characteristics
CN104462494A (en) * 2014-12-22 2015-03-25 武汉大学 Remote sensing image retrieval method and system based on non-supervision characteristic learning

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102073748A (en) * 2011-03-08 2011-05-25 武汉大学 Visual keyword based remote sensing image semantic searching method
CN102867196A (en) * 2012-09-13 2013-01-09 武汉大学 Method for detecting complex sea-surface remote sensing image ships based on Gist characteristic study
CN103309982A (en) * 2013-06-17 2013-09-18 武汉大学 Remote sensing image retrieval method based on vision saliency point characteristics
CN104462494A (en) * 2014-12-22 2015-03-25 武汉大学 Remote sensing image retrieval method and system based on non-supervision characteristic learning

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
周维勋 等: "利用视觉注意模型和局部特征的遥感影像检索方法", 《武汉大学学报.信息科学版》 *
王星 等: "基于视觉显著点特征的遥感影像检索方法", 《测绘科学》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105718531A (en) * 2016-01-14 2016-06-29 广州市万联信息科技有限公司 Image database building method and image recognition method
CN105718531B (en) * 2016-01-14 2019-12-17 广州市万联信息科技有限公司 Image database establishing method and image identification method
CN106228130B (en) * 2016-07-19 2019-09-10 武汉大学 Remote sensing image cloud detection method of optic based on fuzzy autoencoder network
CN106228130A (en) * 2016-07-19 2016-12-14 武汉大学 Remote sensing image cloud detection method of optic based on fuzzy autoencoder network
CN106295613A (en) * 2016-08-23 2017-01-04 哈尔滨理工大学 A kind of unmanned plane target localization method and system
CN106909924A (en) * 2017-02-18 2017-06-30 北京工业大学 A kind of remote sensing image method for quickly retrieving based on depth conspicuousness
CN106909924B (en) * 2017-02-18 2020-08-28 北京工业大学 Remote sensing image rapid retrieval method based on depth significance
CN107122809A (en) * 2017-04-24 2017-09-01 北京工业大学 Neural network characteristics learning method based on image own coding
CN107122809B (en) * 2017-04-24 2020-04-28 北京工业大学 Neural network feature learning method based on image self-coding
CN107515895B (en) * 2017-07-14 2020-06-05 中国科学院计算技术研究所 Visual target retrieval method and system based on target detection
CN107515895A (en) * 2017-07-14 2017-12-26 中国科学院计算技术研究所 A kind of sensation target search method and system based on target detection
CN108830172A (en) * 2018-05-24 2018-11-16 天津大学 Aircraft remote sensing images detection method based on depth residual error network and SV coding
CN109259733A (en) * 2018-10-25 2019-01-25 深圳和而泰智能控制股份有限公司 Apnea detection method, apparatus and detection device in a kind of sleep
CN111144483A (en) * 2019-12-26 2020-05-12 歌尔股份有限公司 Image feature point filtering method and terminal
CN111144483B (en) * 2019-12-26 2023-10-17 歌尔股份有限公司 Image feature point filtering method and terminal
CN112731410A (en) * 2020-12-25 2021-04-30 上海大学 Underwater target sonar detection method based on CNN
CN112731410B (en) * 2020-12-25 2021-11-05 上海大学 Underwater target sonar detection method based on CNN

Also Published As

Publication number Publication date
CN105243154B (en) 2018-08-21

Similar Documents

Publication Publication Date Title
CN105243154A (en) Remote sensing image retrieval method and system based on significant point characteristics and spare self-encodings
Yuan et al. Exploring a fine-grained multiscale method for cross-modal remote sensing image retrieval
CN111259850B (en) Pedestrian re-identification method integrating random batch mask and multi-scale representation learning
CN109919108B (en) Remote sensing image rapid target detection method based on deep hash auxiliary network
Zhang et al. Weakly supervised learning based on coupled convolutional neural networks for aircraft detection
CN108804530B (en) Subtitling areas of an image
CN104462494B (en) A kind of remote sensing image retrieval method and system based on unsupervised feature learning
CN112750140B (en) Information mining-based disguised target image segmentation method
CN109344821A (en) Small target detecting method based on Fusion Features and deep learning
CN111680176A (en) Remote sensing image retrieval method and system based on attention and bidirectional feature fusion
CN110929080B (en) Optical remote sensing image retrieval method based on attention and generation countermeasure network
CN111914107B (en) Instance retrieval method based on multi-channel attention area expansion
CN103745201B (en) A kind of program identification method and device
CN108921850B (en) Image local feature extraction method based on image segmentation technology
CN110222760A (en) A kind of fast image processing method based on winograd algorithm
CN107767416A (en) The recognition methods of pedestrian's direction in a kind of low-resolution image
CN109408655A (en) The freehand sketch retrieval method of incorporate voids convolution and multiple dimensioned sensing network
KR20200010672A (en) Smart merchandise searching method and system using deep learning
CN108446588B (en) Double-temporal remote sensing image change detection method and system
Li et al. An aerial image segmentation approach based on enhanced multi-scale convolutional neural network
CN115410081A (en) Multi-scale aggregated cloud and cloud shadow identification method, system, equipment and storage medium
CN104778699A (en) Adaptive object feature tracking method
Zhang et al. A multiple feature fully convolutional network for road extraction from high-resolution remote sensing image over mountainous areas
CN115222998A (en) Image classification method
Liu et al. Image retrieval using CNN and low-level feature fusion for crime scene investigation image database

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant