CN102867195B - Method for detecting and identifying a plurality of types of objects in remote sensing image - Google Patents

Method for detecting and identifying a plurality of types of objects in remote sensing image Download PDF

Info

Publication number
CN102867195B
CN102867195B CN201210300645.8A CN201210300645A CN102867195B CN 102867195 B CN102867195 B CN 102867195B CN 201210300645 A CN201210300645 A CN 201210300645A CN 102867195 B CN102867195 B CN 102867195B
Authority
CN
China
Prior art keywords
image
matrix
target
class
subimage block
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201210300645.8A
Other languages
Chinese (zh)
Other versions
CN102867195A (en
Inventor
韩军伟
周培诚
王东阳
郭雷
程塨
李晖晖
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Northwestern Polytechnical University
Original Assignee
Northwestern Polytechnical University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Northwestern Polytechnical University filed Critical Northwestern Polytechnical University
Priority to CN201210300645.8A priority Critical patent/CN102867195B/en
Publication of CN102867195A publication Critical patent/CN102867195A/en
Application granted granted Critical
Publication of CN102867195B publication Critical patent/CN102867195B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a method for detecting and identifying a plurality of types of objects in a remote sensing image based on sparse representation dictionary learning. The method is technically characterized by comprising the following steps of: training preprocessed training data to form a dictionary by using a sparse representation dictionary training method; performing sparse coding on the dictionary obtained by training by using an image sub-block in a test image, calculating a sparse representation coefficient to further obtain a reconstruction error of the image sub-block, and determining a candidate target area by thresholding the reconstruction error; and accurately detecting and identifying a plurality of types of objects in the remote sensing image by using post processing. By the method, a plurality of types of objects in the remote sensing image under a complex background can be detected and identified. The method is high in detection and identification accuracy and low in false alarm rate.

Description

A kind of remote sensing images multi-class targets detection and Identification method
Technical field
The present invention relates to a kind of remote sensing images multi-class targets detection and Identification method, can be applied to the polytype Target detection and identification under complex background remote sensing images.
Background technology
As an application of Remote Sensing Image Processing Technology, object detection and recognition under complex background remote sensing images is a gordian technique in the field such as military surveillance and precision strike, also be study hotspot and the difficult point in this field always, there is important military and civilian to be worth, be subject to people and more and more pay close attention to.
Remote Sensing Target detects the two kinds of methods that mainly contain at present.A kind of is that some shape, the geometric properties having by detection target in remote sensing images solves target detection problems, but due to remote sensing images background complexity, exist a large amount of shape, geometric properties similar with target, only rely on these features to detect target and there will be a large amount of undetected, flase drop.Another kind is the thought based on classification, wherein modal is Bag-of-Words(BoW) sorting technique, first the method is that image is extracted to SIFT feature cluster, one group of standard base (image-region of standard) using cluster centre in image space, then can organize standard base with this image is carried out to vector representation, finally obtained vector is classified and thresholding by use svm classifier device, obtain testing result; But BoW method, although the SIFT feature of extracting has yardstick and rotational invariance, has only utilized the statistical nature of characteristic area, and has ignored the spatial information of characteristic area, therefore uses the method verification and measurement ratio of BoW low, and false alarm rate is high; And another sorting technique Linear Spatial Pyramid Matching Using Sparse Coding(ScSPM) although considered the spatial information of characteristic area, the vectorial dimension for classifying obtaining is too high, operand is excessive.In addition, the object detection method of most based on classification also only limits to simple target to detect, and can not detect and identification multiple targets simultaneously.
Summary of the invention
The technical matters solving
For fear of the deficiencies in the prior art part, the present invention proposes a kind of method of the remote sensing images multi-class targets detection and Identification based on rarefaction representation dictionary learning.This method can automatically detect and identify dissimilar target from the remote sensing images of complex background, has higher accuracy of detection and lower false alarm rate.
Technical scheme
A kind of remote sensing images multi-class targets detection and Identification method, is characterized in that step is as follows:
Step 1: use the method training dictionary based on rarefaction representation dictionary learning, concrete steps are as follows:
Step a1 training image early stage processes: first by generic target unification to principal direction in original image, then by the image after unified direction along 0 ° to 360 °, according to step-length rotate to be the image of individual different directions; The original image of different classes of target is all processed according to the method described above, obtained class training image, the wherein different classes of number of targets of p for detecting, for the anglec of rotation, c is the total number of classification of different target different directions image in obtained training image; Wherein: for rounding downwards;
Step b1 data pre-service: adopt method of weighted mean pair tri-components of RGB of class training image are weighted and on average obtain gray level image, then gray level image are carried out to down-sampling processing, obtain the image of n × n size; The image of n × n size is carried out to energy normalized processing and obtain normalized image, then normalized image is converted to n 2the column vector of × 1 dimension, the row using column vector in training data, obtain pretreated training dataset U=[U 1, U 2..., U c], wherein U ithe subdata collection of corresponding i class in training dataset U, i=1,2 ..., c;
Step c1 trains dictionary: the FDDL software package of issuing by Fisher Discrimination Dictionary Learning for Sparse Representation is trained known training dataset U=[U 1, U 2..., U c], obtain dictionary D=[D 1, D 2..., D c], wherein, D ithe sub-dictionary corresponding with i class;
Step 2 sparse coding: the dictionary D=[D obtaining according to training 1, D 2..., D c], the each subimage block in test pattern is carried out to sparse coding, obtain the sparse coefficient that each subimage block is corresponding, concrete treatment step is as follows:
Step a2 test pattern pre-service: first use the method for weighted mean described in step b1 that test pattern is converted into test gray level image, then use the moving window that size is S × S to slide and obtain subimage block with interval steps b along test gray level image; Subimage block down-sampling is processed to the image that size is n × n, then carry out energy normalized processing, then image after treatment energy normalized is converted to a n 2the column vector β of × 1 dimension, represents the pixel grey scale value information of the subimage block obtaining by moving window with column vector β;
Step b2 sparse coding: each subimage block is passed through to Optimized model
α ^ = arg min | | α | | 1 s . t . | | β - Dα | | 2 2 ≤ ϵ
Obtain the sparse coding coefficient of corresponding each subimage block wherein be and sub-dictionary D icorresponding coefficient vector, ε > 0 is allowable error, || || 1for l 1norm, || || 2for l 2norm;
Step c2 asks for reconstructed error: according to sparse coding coefficient calculate the reconstructed error e of each subimage block and each class i, get e=min{e ias the reconstructed error of this subimage block, and record its corresponding classification then judge in this subimage block, whether to comprise target according to the magnitude relationship between reconstructed error e and predefined threshold tau: if e < is τ, illustrates and comprise target, otherwise, illustrate that this subimage block is background;
Step 3 object detection and recognition:
Step a3: will judge the corresponding reconstructed error e of each subimage block that comprises target in step c2, form a reconstructed error matrix E=(e of the same size with test gray level image, to represent candidate target region st) p × Q; Wherein, e stthe value of locating in coordinate points (s, t) for reconstruct error matrix, e st = 0 e &GreaterEqual; &tau; e e < &tau; , P × Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q;
To in step c2, judge the corresponding classification C of each subimage block that comprises target, form a classification matrix L=(C of the same size with test gray level image, to represent candidate target classification st) p × Q; Wherein C stthe value of locating in coordinate points (s, t) for classification matrix, C st = 0 e &GreaterEqual; &tau; C e < &tau; ;
Step b3: big or small G time of window S × S slided in change, repeating step 2 ~ step a3G time, the G obtaining a reconstructed error matrix and G classification matrix, the span of G is 5 ~ 10; By a multiple dimensioned reconstructed error matrix M AP=(e of the G obtaining a reconstructed error matrix composition stg) p × Q × G; Wherein, e stgfor the element in matrix M AP, its value is to change the corresponding e of reconstructed error matrix that sliding window size obtains for the g time st, P × Q × G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G;
The G obtaining a classification matrix formed to a multiple dimensioned classification Matrix C LASS=(C stg) p × Q × G; Wherein, C stgfor the element in Matrix C LASS, its value is to change the corresponding C of classification matrix that sliding window size obtains for the g time st; Obtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP p × Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix p × Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Obtain Scale Matrixes scale=(scale (s, t)) according to multiple dimensioned reconstructed error matrix M AP p × Q, scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t), scale ( s , t ) = 0 e st = 0 arg min g { e stg } e st &NotEqual; 0 ;
Step c3: ask for minimal reconstruction error matrix (map (s, t)) p × Qlocal neighborhood minimal value as the target response value detecting, local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t)) p × Qin corresponding coordinate be the center of target, according to center at (class (s, t)) p × Q(scale (s, t)) p × Qthe corresponding classification of target and scale size are found in the position of middle correspondence.
Described calculated with weighted average method formula is f (x, y)=0.3R (x, y)+0.59G (x, y)+0.11B (x, y), in formula, f (x, y) gray level image obtaining for method of weighted mean is at the gray-scale value of pixel (x, y), R (x, y), G (x, y) and B (x, y) be respectively input training image at tri-component values of the RGB of pixel (x, y).
Described energy normalized computing formula is in formula, f norm(x, y) is the gray-scale value of f (x, y) after energy normalized, and u and v are respectively the row and column size of gray level image.
Described l 1the computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k |
In formula, z is the vector of size for M × 1, ξ kfor the element of vectorial z, k=1,2 ..., M.
Described l 2the computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k | 2
In formula, z is the vector of size for M × 1, ξ kfor the element of vectorial z, k=1,2 ..., M.
Described reconstructed error e icomputing formula be
e i = | | &beta; - D i &alpha; ^ i | | 2 2 + &gamma; | | &alpha; ^ - m i | | 2 2
In formula, γ is predefined weights, and the span of γ is 0 ~ 1, m ito Y iin the element of the every a line mean vector obtaining of averaging; Y ifor U ithe optimum code coefficient obtaining through dictionary D sparse coding.
The described anglec of rotation span is 0 ° to 90 °.
Described FDDL software package parameter lambda 1scope be 0.001 ~ 0.01, λ 2scope be 0.01 ~ 0.1.
The span of described S is the integer between 40 ~ 90, and the span of b is the integer between 1 ~ 15.
The span of described threshold tau is 0 ~ 1.
Beneficial effect
The method of a kind of remote sensing images multi-class targets detection and Identification based on rarefaction representation dictionary learning that the present invention proposes, first use pretreated training data to train redundant dictionary, then the dictionary that uses training to obtain to the subimage block in test pattern carries out sparse coding, obtain its rarefaction representation coefficient, and then obtain the reconstructed error of subimage block by rarefaction representation coefficient, and it is carried out to thresholding processing, determine candidate target region; Realize the accurate detection and Identification to remote sensing images multi-class targets finally by crossing some post-processed.
The present invention automatically detects and identifies the target of plurality of classes in the remote sensing images from complex background.Facts have proved, the method has higher detection and accuracy of identification and lower false alarm rate.
Brief description of the drawings
Fig. 1: the basic flow sheet of the inventive method
Fig. 2: the training data in the inventive method
Fig. 3: the partial detection of the inventive method
(a) Aircraft Targets testing result (red square frame represents Aircraft Targets, and yellow square frame is false-alarm)
(b) Ship Target Detection result (white box represents Ship Target)
(c) oil depot target detection result (blue square frame represents oil depot target)
(d) aircraft, Ship Target Detection result
(e) aircraft, oil depot target detection result
(f) naval vessel, oil depot target detection result
Embodiment
Now in conjunction with embodiment, accompanying drawing, the invention will be further described:
For the hardware environment of implementing be: Intel Pentium 2.93GHz CPU computing machine, 2.0GB internal memory, the software environment of operation is: Matlab R2011a and Windows XP.Choose the remote sensing images that 100 width obtain from Google Earth and carried out multi-class targets test experience, mainly comprised tertiary target: aircraft, naval vessel, oil depot, wherein, totally 200 of Aircraft Targets, totally 120 of Ship Targets, totally 420 of oil depot targets.
The present invention is specifically implemented as follows:
1, training redundant dictionary: use the method training dictionary based on rarefaction representation dictionary learning, detailed process is as follows:
(1.1) training image is processed early stage: concrete processing procedure is: first by generic target unification to principal direction in original image, then by the image after unified direction along 0 ° to 360 °, every 10 ° of rotations once, obtain 36 class training datas, the original image of different classes of target, all according to above method processing, is finally obtained to 55 class training image, i.e. c=55, wherein, aircraft is totally 36 classes, naval vessel 18 classes, oil depot 1 class;
(1.2) data pre-service: adopt method of weighted mean to be weighted and on average to obtain gray level image tri-components of RGB of 55 class training images, then gray level image is carried out to down-sampling processing, obtain the image of 15 × 15 sizes, the image of 15 × 15 sizes is carried out to energy normalized processing and obtain normalized image, again normalized image is converted to the column vector of 255 × 1 dimensions, row using this column vector in training data, obtain pretreated training dataset U=[U 1, U 2..., U c], wherein U ithe subdata collection of corresponding i class in training dataset U, i=1,2 ..., c;
(1.3) the FDDL software package of issuing by Lei Zhang is trained known training dataset U=[U 1, U 2..., U c], obtain dictionary D=[D 1, D 2..., D c], wherein, D ithe sub-dictionary corresponding with i class; Software package parameter lambda 1=0.005, λ 2=0.05;
The FDDL software package of described Lei Zhang is shown in paper: Meng Yang, Lei Zhang, Xiangchu Feng, David Zhang.Fisher Discrimination Dictionary Learning for Sparse Representation[C] .ICCV, 2011
2, sparse coding: the dictionary D=[D obtaining according to training 1, D 2..., D c], the each subimage block in test pattern is carried out to sparse coding, obtain the sparse coefficient that each subimage block is corresponding, concrete treatment step is as follows:
(2.1) test pattern pre-service: first use the method for weighted mean described in (1.1) that test pattern is converted into test gray level image, then use the moving window that size is S × S to slide and obtain subimage block as 5 pixels taking interval steps along test gray level image, S initial value gets 90; To each subimage block that uses moving window to obtain, be down sampled to size and be 15 × 15 image, then carry out energy normalized processing, again image after treatment energy normalized is converted to the column vector β of one 225 × 1 dimension, represents the pixel grey scale value information of the subimage block obtaining by moving window with column vector β;
(2.2) sparse coding: to each subimage block by passing through Optimized model:
&alpha; ^ = arg min | | &alpha; | | 1 s . t . | | &beta; - D&alpha; | | 2 2 &le; &epsiv;
Obtain the sparse coding coefficient vector of corresponding each subimage block wherein be and sub-dictionary D i
Corresponding coefficient vector, allowable error ε=0.15, || || 1for l 1norm, || || 2for l 2norm; ;
(2.3) ask for reconstructed error: according to sparse coding coefficient calculate the reconstructed error of subimage block picture piece and each class weights γ=0.5, gets e=min{e ias the reconstructed error of this subimage block, and record its corresponding classification then judge in this subimage block, whether to comprise target according to the magnitude relationship between reconstructed error e and predefined threshold tau=0.3: if e < is τ, illustrates and comprise target, otherwise, illustrate that this subimage block is background;
3, object detection and recognition:
(3.1) will in (2.3), judge the corresponding reconstructed error e of each subimage block that comprises target, form a reconstructed error matrix E=(e of the same size with test gray level image, to represent candidate target region st) p × Q; Wherein e stthe value of locating in coordinate points (s, t) for reconstruct error matrix, e st = 0 e &GreaterEqual; &tau; e e < &tau; , P × Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q; To in (2.3), judge the corresponding classification of each subimage block that comprises target, form a classification matrix L=(C of the same size with test gray level image, to represent candidate target classification st) p × Q; Wherein C stthe value of locating in coordinate points (s, t) for classification matrix, C st = 0 e &GreaterEqual; &tau; C e < &tau; ;
(3.2) change the size of sliding window S × S, become S=90-10 × j, j=1,2 ... G is number of times for a change, repetition 2, step (3.1) G time altogether, the G obtaining a reconstructed error matrix and G classification matrix; By a multiple dimensioned reconstructed error matrix M AP=(e of the G obtaining a reconstructed error matrix composition stg) p × Q × G; Wherein, e stgfor the element in matrix M AP, its value is to change the corresponding e of reconstructed error matrix that sliding window size obtains for the g time st, P × Q × G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G; The G obtaining a classification matrix formed to a multiple dimensioned classification Matrix C LASS=(C stg) p × Q × G; Wherein, C stgfor the element in Matrix C LASS, its value is to change the corresponding C of classification matrix that sliding window size obtains for the g time st; Obtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP p × Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t), then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix p × Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t), obtain Scale Matrixes according to multiple dimensioned reconstructed error matrix M AP scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t),
scale ( s , t ) = 0 e st = 0 arg min g { e stg } e st &NotEqual; 0 ;
(3.3): ask for minimal reconstruction error matrix (map (s, t)) p × Qlocal neighborhood minimal value as the target response value detecting, local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t)) p × Qin corresponding coordinate be the center of target, just can be at (class (s, t)) according to center p × Q(scale (s, t)) p × Qthe corresponding classification of target and scale size are found in the position of middle correspondence.
Described calculated with weighted average method formula is
f(x,y)=0.3R(x,y)+0.59G(x,y)+0.11B(x,y)
In formula, f (x, y) gray level image obtaining for method of weighted mean is at pixel (x, y) gray-scale value, R (x, y), G (x, y) and B (x, y) training image that is respectively input is at tri-component values of the RGB of pixel (x, y).
Described energy normalized computing formula is
f norm ( x , y ) = f ( x , y ) &Sigma; x = 1 u &Sigma; y = 1 v [ f ( x , y ) ] 2
In formula, f norm(x, y) is the gray-scale value of f (x, y) after energy normalized, and u and v are respectively the row and column size of gray level image, u=15, v=15.
Described l 1the computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k |
In formula, z is the vector of size for M × 1, ξ kfor the element of vectorial z, k=1,2 ..., M.
Described l 2the computing formula of norm is
| | z | | 1 = &Sigma; k = 1 M | &xi; k | 2
Described reconstructed error e icomputing formula be
e i = | | &beta; - D i &alpha; ^ i | | 2 2 + &gamma; | | &alpha; ^ - m i | | 2 2
In formula, γ is predefined weights, γ=0.5, m ito Y iin the element of the every a line mean vector obtaining of averaging; Y ifor U ithe optimum code coefficient obtaining through dictionary D sparse coding;
Select correct verification and measurement ratio and false alarm rate to assess validity of the present invention.Wherein, the ratio of the target number that correct verification and measurement ratio is defined as correct detection and total target number, false alarm rate is defined as false-alarm number and the correct target number detecting and the ratio of false-alarm number sum., the testing result of gained of the present invention and the multi-class targets detection algorithm based on BoW are contrasted, comparing result is as shown in table 1 meanwhile.Correct verification and measurement ratio and false alarm rate have all shown the validity of the inventive method.
Table 1 evaluation

Claims (5)

1. a remote sensing images multi-class targets detection and Identification method, is characterized in that step is as follows:
Step 1: use the method training dictionary based on rarefaction representation dictionary learning, concrete steps are as follows:
Step a1 training image early stage processes: first by generic target unification to principal direction in original image, then by the image after unified direction along 0 ° to 360 °, according to step-length rotate to be the image of individual different directions; The original image of different classes of target is all processed according to the method described above, obtained class training image, the wherein different classes of number of targets of p for detecting, for the anglec of rotation, c is the total number of classification of different target different directions image in obtained training image; Wherein: for rounding downwards;
Step b1 data pre-service: adopt method of weighted mean pair tri-components of RGB of class training image are weighted and on average obtain gray level image, described calculated with weighted average method formula is f (x, y)=0.3R (x, y)+0.59G (x, y)+0.11B (x, y), in formula, the gray level image that f (x, y) obtains for method of weighted mean is at pixel (x, y) gray-scale value, R (x, y), G (x, y) and B (x, y) training image that is respectively input is at tri-component values of the RGB of pixel (x, y); Then gray level image is carried out to down-sampling processing, obtain the image of n × n size; Image to n × n size carries out energy normalized processing, and described energy normalized computing formula is in formula, f norm(x, y) is the gray-scale value of f (x, y) after energy normalized, and u and v are respectively the row and column size of gray level image; Carry out having obtained normalized image after energy normalized processing, then normalized image is converted to n 2the column vector of × 1 dimension, the row using column vector in training data, obtain pretreated training dataset U=[U 1, U 2..., U c], wherein U ithe subdata collection of corresponding i class in training dataset U, i=1,2 ..., c;
Step c1 trains dictionary: the FDDL software package of issuing by Fisher Discrimination Dictionary Learning for Sparse Representation is trained known training dataset U=[U 1, U 2..., U c], obtain dictionary D=[D 1, D 2..., D c], wherein, D ithe sub-dictionary corresponding with i class;
Step 2 sparse coding: the dictionary D=[D obtaining according to training 1, D 2..., D c], the each subimage block in test pattern is carried out to sparse coding, obtain the sparse coefficient that each subimage block is corresponding, concrete treatment step is as follows:
Step a2 test pattern pre-service: first use the method for weighted mean described in step b1 that test pattern is converted into test gray level image, then use the moving window that size is S × S to slide and obtain subimage block with interval steps b along test gray level image; Subimage block down-sampling is processed to the image that size is n × n, then carry out energy normalized processing, then image after treatment energy normalized is converted to a n 2the column vector β of × 1 dimension, represents the pixel grey scale value information of the subimage block obtaining by moving window with column vector β;
Step b2 sparse coding: each subimage block is passed through to Optimized model
&alpha; ^ = arg min | | &alpha; | | 1 s . t . | | &beta; - D&alpha; | | 2 2 &le; &epsiv;
Obtain the sparse coding coefficient of corresponding each subimage block wherein be and sub-dictionary D icorresponding coefficient vector, ε > 0 is allowable error, || || 1for l 1norm, || || 2for l 2norm; Described l lthe computing formula of norm is described l 2the computing formula of norm is in two formula, z is the vector of size for M × 1, ξ kfor the element of vectorial z, k=1,2 ..., M
Step c2 asks for reconstructed error: according to sparse coding coefficient calculate the reconstructed error e of each subimage block and each class i, described reconstructed error e icomputing formula be in formula, γ is predefined weights, m ito Y iin the element of the every a line mean vector obtaining of averaging; Y ifor U ithe optimum code coefficient obtaining through dictionary D sparse coding; Get e=min{e ias the reconstructed error of this subimage block, and record its corresponding classification then judge in this subimage block, whether to comprise target according to the magnitude relationship between reconstructed error e and predefined threshold tau: if e < is τ, illustrates and comprise target, otherwise, illustrate that this subimage block is background;
Step 3 object detection and recognition:
Step a3: will judge the corresponding reconstructed error e of each subimage block that comprises target in step c2, form a reconstructed error matrix E=(e of the same size with test gray level image, to represent candidate target region st) p × Q; Wherein, e stthe value of locating in coordinate points (s, t) for reconstruct error matrix, e st = 0 e &GreaterEqual; &tau; e e < &tau; , P × Q is the size of test pattern, s=1, and 2 ... P, t=1,2 ... Q;
To in step c2, judge the corresponding classification C of each subimage block that comprises target, form a classification matrix L=(C of the same size with test gray level image, to represent candidate target classification st) p × Q; Wherein C stthe value of locating in coordinate points (s, t) for classification matrix, C st = 0 e &GreaterEqual; &tau; C e < &tau; ,
Step b3: big or small G time of window S × S slided in change, repeating step 2~step a3G time, the G obtaining a reconstructed error matrix and G classification matrix, the span of G is 5~10; By a multiple dimensioned reconstructed error matrix M AP=(e of the G obtaining a reconstructed error matrix composition stg) p × Q × G; Wherein, e stgfor the element in matrix M AP, its value is to change the corresponding e of reconstructed error matrix that sliding window size obtains for the g time st, P × Q × G is the size of multiple dimensioned reconstructed error matrix, g=1, and 2 ... G;
The G obtaining a classification matrix formed to a multiple dimensioned classification Matrix C LASS=(C stg) p × Q × G; Wherein, C stgfor the element in Matrix C LASS, its value is to change the corresponding C of classification matrix that sliding window size obtains for the g time st; Obtain a minimal reconstruction error matrix (map (s, t)) according to multiple dimensioned reconstructed error matrix M AP p × Q, wherein map (s, t) is the value that corresponding minimal reconstruction error matrix is located in coordinate points (s, t),
Then obtain the minimum classification matrix (class (s, t)) of corresponding minimal reconstruction error matrix p × Q, wherein class (s, t) is the value that minimum classification matrix is located in coordinate points (s, t),
Obtain Scale Matrixes scale=(scale (s, t)) according to multiple dimensioned reconstructed error matrix M AP p × Q, scale (s, t) is the value that corresponding Scale Matrixes is located in coordinate points (r, t), scale ( s , t ) = 0 e st = 0 arg min g { e stg } e st &NotEqual; 0 ;
Step c3: ask for minimal reconstruction error matrix (map (s, t)) p × Qlocal neighborhood minimal value as the target response value detecting, local neighborhood minimal value is at minimal reconstruction error matrix (map (s, t)) p × Qin corresponding coordinate be the center of target, according to center at (class (s, t)) p × Q(scale (s, t)) p × Qthe corresponding classification of target and scale size are found in the position of middle correspondence.
2. remote sensing images multi-class targets detection and Identification method according to claim 1, is characterized in that: the described anglec of rotation span is 0 ° to 90 °.
3. remote sensing images multi-class targets detection and Identification method according to claim 1, is characterized in that: described FDDL software package parameter lambda 1scope be 0.001~0.01, λ 2scope be 0.01~0.1.
4. remote sensing images multi-class targets detection and Identification method according to claim 1, is characterized in that: the span of described S is the integer between 40~90, the span of b is the integer between 1~15.
5. remote sensing images multi-class targets detection and Identification method according to claim 1, is characterized in that: the span of described threshold tau is 0~1.
CN201210300645.8A 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image Active CN102867195B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210300645.8A CN102867195B (en) 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210300645.8A CN102867195B (en) 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image

Publications (2)

Publication Number Publication Date
CN102867195A CN102867195A (en) 2013-01-09
CN102867195B true CN102867195B (en) 2014-11-26

Family

ID=47446059

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210300645.8A Active CN102867195B (en) 2012-08-22 2012-08-22 Method for detecting and identifying a plurality of types of objects in remote sensing image

Country Status (1)

Country Link
CN (1) CN102867195B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103258210B (en) * 2013-05-27 2016-09-14 中山大学 A kind of high-definition image classification method based on dictionary learning
CN103632164B (en) * 2013-11-25 2017-03-01 西北工业大学 The volume firm state classification recognition methodss of the KNN coil image data based on KAP sample optimization
CN104517121A (en) * 2014-12-10 2015-04-15 中国科学院遥感与数字地球研究所 Spatial big data dictionary learning method based on particle swarm optimization
CN105740422B (en) * 2016-01-29 2019-10-29 北京大学 Pedestrian retrieval method and device
CN106067041B (en) * 2016-06-03 2019-05-31 河海大学 A kind of improved multi-target detection method based on rarefaction representation
CN107451595A (en) * 2017-08-04 2017-12-08 河海大学 Infrared image salient region detection method based on hybrid algorithm
CN109190457B (en) * 2018-07-19 2021-12-03 北京市遥感信息研究所 Oil depot cluster target rapid detection method based on large-format remote sensing image
CN109946076B (en) * 2019-01-25 2020-04-28 西安交通大学 Planetary wheel bearing fault identification method of weighted multi-scale dictionary learning framework
CN110189328B (en) * 2019-06-11 2021-02-23 北华航天工业学院 Satellite remote sensing image processing system and processing method thereof

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102129573A (en) * 2011-03-10 2011-07-20 西安电子科技大学 SAR (Synthetic Aperture Radar) image segmentation method based on dictionary learning and sparse representation
CN102324047A (en) * 2011-09-05 2012-01-18 西安电子科技大学 High spectrum image atural object recognition methods based on sparse nuclear coding SKR

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8374442B2 (en) * 2008-11-19 2013-02-12 Nec Laboratories America, Inc. Linear spatial pyramid matching using sparse coding

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102129573A (en) * 2011-03-10 2011-07-20 西安电子科技大学 SAR (Synthetic Aperture Radar) image segmentation method based on dictionary learning and sparse representation
CN102324047A (en) * 2011-09-05 2012-01-18 西安电子科技大学 High spectrum image atural object recognition methods based on sparse nuclear coding SKR

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
梁天一 等.基于稀疏编码的图像语义分类器模型.《华东理工大学学报(自然科学报)》.2007,第33卷(第6期),827-892. *

Also Published As

Publication number Publication date
CN102867195A (en) 2013-01-09

Similar Documents

Publication Publication Date Title
CN102867195B (en) Method for detecting and identifying a plurality of types of objects in remote sensing image
CN110363182B (en) Deep learning-based lane line detection method
CN110348319B (en) Face anti-counterfeiting method based on face depth information and edge image fusion
CN107346436B (en) Visual saliency detection method fusing image classification
Sirmacek et al. Urban area detection using local feature points and spatial voting
Yuan et al. Large-scale solar panel mapping from aerial images using deep convolutional networks
Cord et al. Automatic road defect detection by textural pattern recognition based on AdaBoost
CN106228125B (en) Method for detecting lane lines based on integrated study cascade classifier
CN103034863B (en) The remote sensing image road acquisition methods of a kind of syncaryon Fisher and multiple dimensioned extraction
CN103116744B (en) Based on the false fingerprint detection method of MRF and SVM-KNN classification
CN102819740B (en) A kind of Single Infrared Image Frame Dim targets detection and localization method
CN102542295B (en) Method for detecting landslip from remotely sensed image by adopting image classification technology
CN110555475A (en) few-sample target detection method based on semantic information fusion
CN109409190A (en) Pedestrian detection method based on histogram of gradients and Canny edge detector
CN106023220A (en) Vehicle exterior part image segmentation method based on deep learning
CN105426889A (en) PCA mixed feature fusion based gas-liquid two-phase flow type identification method
CN104077577A (en) Trademark detection method based on convolutional neural network
CN106295124A (en) Utilize the method that multiple image detecting technique comprehensively analyzes gene polyadenylation signal figure likelihood probability amount
CN103390164A (en) Object detection method based on depth image and implementing device thereof
CN104182985A (en) Remote sensing image change detection method
CN108710909B (en) Counting method for deformable, rotary and invariant boxed objects
Yuan et al. Learning to count buildings in diverse aerial scenes
CN105512622A (en) Visible remote-sensing image sea-land segmentation method based on image segmentation and supervised learning
CN108073940A (en) A kind of method of 3D object instance object detections in unstructured moving grids
CN106446854A (en) High-resolution optical remote sensing image target detection method based on rotation invariant HOG feature

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant