CN103823845A - Method for automatically annotating remote sensing images on basis of deep learning - Google Patents
Method for automatically annotating remote sensing images on basis of deep learning Download PDFInfo
- Publication number
- CN103823845A CN103823845A CN201410039584.3A CN201410039584A CN103823845A CN 103823845 A CN103823845 A CN 103823845A CN 201410039584 A CN201410039584 A CN 201410039584A CN 103823845 A CN103823845 A CN 103823845A
- Authority
- CN
- China
- Prior art keywords
- remote sensing
- layer
- sensing image
- degree
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/5866—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, manually generated location and time information
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a method for automatically annotating remote sensing images on the basis of deep learning. The method for automatically annotating the remote sensing images includes extracting visual feature vectors of the to-be-annotated remote sensing images; inputting the visual feature vectors into a DBM (deep Boltzmann machine) model to automatically annotate the to-be-annotated remote sensing images. The DBM model implemented in the method sequentially comprises a visible layer, a first hidden layer, a second hidden layer and a tag layer from bottom to top, and is acquired by means of training. The method for automatically annotating the remote sensing images has the advantages that the deep Boltzmann machine model implemented in the method comprises the two hidden layers (namely, the first hidden layer and the second hidden layer respectively), accordingly, the problem of 'semantic gaps' in image semantic annotation procedures can be effectively solved by the two hidden layers, and the integral annotation accuracy can be improved.
Description
Technical field
The present invention relates to intelligent classification and the retrieval technique of remote sensing image, particularly a kind of remote sensing image automatic marking method based on degree of depth study.
Background technology
Remote sensing image is one of significant data of spatial information, is widely used in geology and flood monitoring, agricultural and forest inventory investigation, soil utilization and city planning and military field.Along with the development of Chinese Space science and earth observation technology, remote sensing image data present the trend of exponential increase data year, effective management of magnanimity remote sensing image data is become and is become more and more important.
Remote sensing image mark is one of important content of remote sensing images analysis and understanding, it is the bottom visual signature by extracting remote sensing image, learn contacting between these bottom visual signatures and high-level semantic by some machine learning model, thereby realize some semantic label automatic markings to remote sensing image, for example, by remote sensing image automatic marking process, can automatically mark out residential block, farmland, shopping centre, desert, forest etc. in magnanimity remote sensing image.The automatic marking of remote sensing image is the understanding to remote sensing image semanteme, is also the important technical basis that magnanimity remote sensing image is carried out to classification cataloguing and retrieval.
Can regard the automatic classification work of the broad sense to remote sensing image as to the automatic marking work of remote sensing image, before remote sensing image is carried out to automatic marking work, need to first determine and need the corresponding class label of remote sensing image (being text label) of mark which has, then different remote sensing images to be got up from different class label corresponding relationships.
Mark work to traditional images mainly contains 3 class methods: method, the method based on machine learning and the method based on relevant feedback based on object body.Mostly the mark work of traditional images is to come by the lower-level vision feature of image the vision content of analysis and understanding image, but this method exists a problem mostly: " semantic gap "." semantic gap " refers to the high-level semantic that can not only infer image by the bottom visual signature of image, between the bottom visual signature of image and the high-level semantic of image, there is no the suitable abstract bridge that does, thereby the effect of mark is unsatisfactory.
In order to overcome the problem of " semantic gap ", people have slowly been developed certain methods the bottom visual signature of image have been mapped to the high-level semantic of image, wherein typical method comprises probability latent semantic analysis (Probabilistic Latent Semantic Analysis, pLSA) model, potential Di Li Cray distributes (Latent Dirichlet Allocation, LDA) model and author's theme (Author Topic Model, ATM) model etc., but these method majorities are just considered the color and vein characteristic of image, and do not consider the spectral characteristic of remote sensing image.Spectral characteristic is the key character of remote sensing image, also be the key property that remote sensing image is different from general image data, for different characters of ground object, to the spectrum of different wave length, absorption characteristic and reflection characteristic have very big difference, and therefore the spectral characteristic of remote sensing image has very strong identification capability to different characters of ground object.
Summary of the invention
For deficiency of the prior art, the invention provides a kind of " semantic gap " problem that can overcome linguistic indexing of pictures, realize the remote sensing image automatic marking method based on degree of depth study of the semantic tagger of degree of precision.
Based on a remote sensing image automatic marking method for degree of depth study, comprising:
(1) the low-level image feature vector that extracts remote sensing image to be marked builds the visual feature vector that obtains corresponding remote sensing image;
(2) degree of depth Boltzmann machine model input of described visual feature vector being trained carries out automatic marking;
The degree of depth Boltzmann machine model training in described step (2) obtains by following steps:
(S1) create the label dictionary that comprises several text labels;
(S2) select the mark of respective classes to have text label remote sensing image as model training data set according to label dictionary;
(S3) the low-level image feature vector that extracts each remote sensing image builds the visual feature vector that obtains corresponding remote sensing image, and determines the Text eigenvector of each remote sensing image according to label dictionary and text label;
(S4) build degree of depth Boltzmann machine model, described degree of depth Boltzmann machine model comprises visible layer, the first hidden layer, the second hidden layer and label layer from bottom to up successively; Any two nodes in each layer are without connection, the two-way connection of any two nodes between adjacent layer;
(S5) utilize the visual feature vector of all remote sensing images of model training data centralization and Text eigenvector to described degree of depth Boltzmann machine model training, the degree of depth Boltzmann machine model that obtains training.
In remote sensing image automatic marking method based on degree of depth study of the present invention, first extract the low-level image feature of remote sensing image to be marked and build the visual feature vector that obtains remote sensing image according to all low-level image features, then the directly input using visual feature vector as the visible layer of degree of depth Boltzmann machine model, now degree of depth Boltzmann machine model (DBM model, Deep Boltzmann Machine model) the output of label layer as Text eigenvector, then adopt the text label corresponding with Text eigenvector to carry out automatic marking to remote sensing image to be marked.
In DBM model, high-level semantic (label layer) is by abstract the obtaining of low-level image feature (input of visible layer), because low-level image feature can not be transitioned into high-level semantic well, therefore can produce " semantic gap ".The number of plies of hidden layer is more in theory, semantic gap is less, but the data volume of considering remotely-sensed data is large, if it is too much that the number of plies of hidden layer is set, can then train speed very slow, therefore as preferred, degree of depth Boltzmann machine model used in the present invention comprises two hidden layers (being respectively the first hidden layer and the second hidden layer), by being set, two hidden layers improve the abstracting power of the centre of DBM model, just in time can fill up " semantic gap " problem in linguistic indexing of pictures process, improve overall mark accuracy rate.
The text label number comprising in the label dictionary creating in described step (S1) and kind need to set according to application.If need the remote sensing image of mark to be only divided into land and waters, the size of so whole label dictionary is 2, in label dictionary, has two text labels, is respectively land and waters.In actual applications, the size of label dictionary can, than 2 much bigger, be determined according to practical application, is mostly in situation, has " residential block ", " river ", " highway ", " forest ", " desert " etc. text label in label dictionary.
A text label represents a classification, and the text label of the model training data centralization remote sensing image in described step (S2) is generally contained all text labels in label dictionary, can also have in theory the text label not having in label dictionary.
Text eigenvector in described step (S3) is a 0-1 vector (being that in vector, all elements can only be 0 or 1), and described Text eigenvector is determined the Text eigenvector of each remote sensing image according to following steps:
(S31) full null vector of initialization, makes the corresponding text label of every one dimension;
(S32) according to the text label of remote sensing image, be 1 by the element assignment of corresponding dimension, obtain the Text eigenvector of this remote sensing image.
The degree of depth Boltzmann machine model that described step (S4) builds, any two nodes in each layer are without connection, the two-way connection of any two nodes between adjacent layer.
The interconnection network that the second described hidden layer and label layer form is unidirectional BP(Back Propagation) neural network.
Adopt BP neural network to be because can be used for feedback regulation, this is the characteristic of BP neural network, BP neural network is a feed-forward regulation network, it can contrast according to the result of current training and actual result, then carry out suitable feed-forward regulation and come the parameter of correction model according to error, using BP neural network is exactly the learning model that has supervision of having realized a band feedforward in fact.
The training process of the degree of depth Boltzmann machine model in described step (S5) is as follows:
(S51) using visual feature vector as visible layer, using Text eigenvector as label layer;
(S52) using visible layer and the first hidden layer as limited Boltzmann machine, input using visual feature vector as visible layer, uses sdpecific dispersion algorithm is trained to the connection weight that obtains between visible layer and the first hidden layer and the end-state of the first hidden layer to this Boltzmann machine;
(S53) using the first hidden layer and the second hidden layer as limited Boltzmann machine, input using the end-state of the first hidden layer as the first hidden layer, uses sdpecific dispersion algorithm is trained to the connection weight that obtains between the first hidden layer and the second hidden layer and the end-state of the 2nd hidden layer to this Boltzmann machine;
(S54) input using the end-state of the second hidden layer as the second hidden layer, input using Text eigenvector as label layer, adopt the training of BP neural net method to obtain the connection weight of the second hidden layer and label layer, complete the training of degree of depth Boltzmann machine model.
Training process depends on the structure of the degree of depth Boltzmann machine model of step (S4) structure, input take visual feature vector as visible layer when training, the visual feature vector of each remote sensing image of model training data centralization is thought once to train, by a large amount of training, in training process, constantly adjust the coefficient of connection of adjacent two layers, finally obtain the Best link weight of each group, and then the degree of depth Boltzmann machine model that obtains training.
The node number of described visible layer is identical with the dimension of visual feature vector.
In identification and training process, all inputs using visual feature vector as visible layer, therefore each node of visible layer must be mutually corresponding with the element of every one dimension in visual feature vector, and the node number of visible layer is identical with the dimension of visual feature vector.
The node number of described label layer is identical with the dimension of Text eigenvector.
The dimension of described Text eigenvector is identical with the number of label dictionary Chinese version label.
Obtain Text eigenvector according to the output of label layer, then use the text label corresponding with Text eigenvector to mark remote sensing image to be marked.And the input using the text label of each remote sensing image as label layer in training process, for guaranteeing to carry out smoothly, the node number of label layer is identical with the dimension of Text eigenvector.
Further determine that according to Text eigenvector method can know, a text label in Text eigenvector in each element corresponding label dictionary, the dimension of explanatory text proper vector is identical with the text label number in label dictionary.
Node number in described the first hidden layer and the second hidden layer is set based on experience value, is generally 500~1500, in actual application, can adjust according to experiment effect.
Described low-level image feature vector comprise averaged spectrum Reflectivity vector, color layout's description vectors, color structure description vectors, scalable color description vector, homogeneity texture description vector, edge histogram description vectors, GIST proper vector and based on SIFT(Scale-invariant feature transform) the visual word bag vector of feature.
A remote sensing image has a series of low-level image feature vectors, a kind of low-level image feature vector of many employings in existing remote sensing image identification or mask method.The present invention is the identification that increases degree of depth Boltzmann machine model, improve the accuracy rate of mark, low-level image feature vector of the present invention has comprised the texture feature vectors such as color layout's description vectors, color structure description vectors, scalable color description vector, homogeneity texture description vector, edge histogram description vectors, GIST proper vector and the visual word bag vector based on SIFT feature simultaneously, has also introduced spectral characteristic (averaged spectrum Reflectivity vector).Extraction obtains low-level image feature vector and directly as required the splicing of each low-level image feature vector is obtained to visual feature vector according to book afterwards, the dimension of visual feature vector be all low-level image feature vectors dimension and.
As preferably, described averaged spectrum Reflectivity vector comprises the spectrum average reflectance of four kinds of wavelength, and four kinds of described wavelength are respectively 0.44~0.51 micron, 0.53~0.62 micron, 0.63~0.70 micron and 0.74~0.80 micron.These four kinds of wavelength have stronger separating capacity.
Visual word bag vector based on SIFT feature obtains by following steps:
(a) calculate the SIFT proper vector of all remote sensing images of described model training data centralization;
(b) all SIFT proper vectors are carried out to cluster and obtain 500~1000 cluster centres (SIFT proper vector);
(c), using each cluster centre as vision word, add up the occurrence number of each vision word in the SIFT proper vector of every remote sensing image and form the visual word bag vector based on SIFT feature.
A remote sensing image generally has multiple SIFT proper vectors, and the number of the SIFT proper vector that remote sensing image is corresponding depends on the size of remote sensing image and the content comprising, and remote sensing image is larger, and the content comprising is more, and corresponding SIFT proper vector is more.A common remote sensing image comprises 500~2000 SIFT proper vectors.Vision word, for whole model training data set, obtains by cluster.When cluster take the SIFT proper vector of all remote sensing images of whole model training data centralization as object, cluster obtains cluster centre, so-called cluster centre in fact can think to asking of same class SIFT proper vector flat the vector that all arrives, therefore vision word is also a vector, and same class SIFT proper vector adopts identical vision word to represent.The visual word bag vector dimension of every remote sensing image is identical with vision word number, and the each element in visual word bag vector represents respectively the number of times that in this remote sensing image, different vision words occur.
The number of cluster centre can be set as required, and SIFT proper vector numbers general and all remote sensing images of model training data centralization are suitable.As preferably, in the present invention, the number of cluster centre is 500~1000.
According to the feature of degree of depth Boltzmann machine model, the detailed process of described step (2) is as follows:
The visible layer of the degree of depth Boltzmann machine model that the input of described visual feature vector is trained, according to the output of the label layer of degree of depth Boltzmann machine model, determine the Text eigenvector of remote sensing images to be marked, then mark remote sensing images to be marked with text label corresponding to text proper vector.
Because degree of depth Boltzmann machine model is from visible layer, successively calculate the value (all nodal values are Text eigenvector) of each node of top layer (label layer) according to the connection weight between adjacent layer, the Text eigenvector that in some situation, label layer calculates may occur that non-vanishing is not 1 element yet, therefore, the Text eigenvector that need to calculate label layer is normalized, then according to normalized result, the element of military order value maximum is 1, all the other are all zero, the final Text eigenvector that just obtains figure to be marked.
In remote sensing image automatic marking method based on degree of depth study of the present invention, use the DBM model with two hidden layers, visible layer is the low-level image feature vector of corresponding remote sensing image, the superiors are text labels of remote sensing image, two hidden layers in the middle of DBM just in time can be filled up " semantic gap " problem in linguistic indexing of pictures process like this, and overall mark accuracy rate is improved.And in the time that being described, remote sensing image low-level image feature except using color and vein characteristic, introduced spectral characteristic (average reflectance of the spectrum of different wave length), the identification to different earth objects can greatly increase like this remote sensing image mark time simultaneously.
Embodiment
Below in conjunction with specific embodiment, the present invention is described in further detail.
Based on a remote sensing image automatic marking method for degree of depth study, comprising:
(1) the low-level image feature vector that extracts remote sensing image to be marked builds the visual feature vector that obtains corresponding remote sensing image;
In this enforcement, low-level image feature vector comprises averaged spectrum Reflectivity vector, color layout's description vectors, color structure description vectors, scalable color description vector, homogeneity texture description vector, edge histogram description vectors, GIST proper vector and the visual word bag vector based on SIFT feature.
Averaged spectrum Reflectivity vector can directly be obtained from remote sensing image data, and remote sensing image data is different from normal image, has just gathered spectral information when satellite is taken remote sensing image.
In the present embodiment, the proper vector of averaged spectrum reflectivity has 4 dimensions, comprise the spectrum average reflectance of four kinds of wavelength, four kinds of wavelength are respectively 0.44~0.51 micron, 0.53~0.62 micron, 0.63~0.70 micron and 0.74~0.80 micron, the color layout's description vectors obtaining is 192 dimensions, color structure description vectors is 256 dimensions, scalable color description vector is 256 dimensions, scalable color description vector is 43 dimensions, edge histogram description vectors is 150 dimensions, GIST proper vector is 960 dimensions, visual word bag vector based on SIFT feature is 1000 dimensions, all low-level image feature vector splicings obtain the visual feature vector of remote sensing image to be marked, this feature visual feature vector is 2861 dimensions.
Visual word bag vector based on SIFT feature is extracted and is obtained by following steps:
(a) calculate the SIFT proper vector of all remote sensing images of described model training data centralization;
(b) all SIFT proper vectors are carried out to cluster and obtain 1000 cluster centres;
(c) using each cluster centre as vision word, add up the occurrence number of each vision word in the SIFT proper vector of every remote sensing image and form the visual word bag vector based on SIFT feature of corresponding remote sensing image, the dimension of visual word bag vector equals 1000(and equals the number of cluster centre), in visual word bag vector, each element is respectively the number of times that in all SIFT proper vectors of corresponding remote sensing image, different vision words occur.
(2) degree of depth Boltzmann machine model input of the visual feature vector of remote sensing image to be marked being trained carries out automatic marking.
The degree of depth Boltzmann machine model training using in step in the present embodiment (2) obtains by following steps:
(S1) create the label dictionary that comprises several text labels;
The text label number comprising in the label dictionary creating in step (S1) and kind need to set according to application.If need the remote sensing image of mark to be only divided into land and waters, the size of so whole label dictionary is 2, in label dictionary, has two text labels, is respectively land and waters.In actual applications, the size of label dictionary can, than 2 much bigger, be determined according to practical application, is mostly in situation, has " residential block ", " river ", " highway ", " forest ", " desert " etc. text label in label dictionary.In the present embodiment, in label dictionary, contain 21 text labels.
(S2) select the remote sensing image that marks text label of respective classes as model training data set according to label dictionary, the text label of this model training data centralization remote sensing image comprises all text labels in label dictionary, and only includes all text labels in label dictionary.
(S3) the low-level image feature vector that extracts each remote sensing image builds the visual feature vector that obtains corresponding remote sensing image, and determines the Text eigenvector of each remote sensing image according to label dictionary and text label;
In this step, the acquisition methods of the visual feature vector of each remote sensing image is identical with step (1).
Text eigenvector is a 0-1 vector (being that in vector, all elements can only be 0 or 1), and described Text eigenvector is determined the Text eigenvector of each remote sensing image according to following steps:
(S31) full null vector of initialization, makes the corresponding text label of every one dimension;
(S32) according to the text label of remote sensing image, be 1 by the element assignment of corresponding dimension, obtain the Text eigenvector of this remote sensing image.
(S4) build degree of depth Boltzmann machine model, described degree of depth Boltzmann machine model comprises visible layer, the first hidden layer, the second hidden layer and label layer from bottom to up successively, any two nodes in each layer are without connection, the two-way connection of any two nodes between adjacent layer, and the interconnection network of the second hidden layer and label layer formation is unidirectional BP neural network.
The node number of visible layer is identical with the dimension of visual feature vector, is 2861 dimensions.
The number of the node number of label layer, the dimension of Text eigenvector and label dictionary Chinese version label is all identical, is 21 in this enforcement.
Node number in the first hidden layer and the second hidden layer is 1024.
(S5) utilize the visual feature vector of all remote sensing images of model training data set and Text eigenvector to described degree of depth Boltzmann machine model training, the degree of depth Boltzmann machine model that obtains training, concrete training process is as follows:
(S51) using visual feature vector as visible layer, using Text eigenvector as label layer;
(S52) using visible layer and the first hidden layer as limited Boltzmann machine, input using visual feature vector as visible layer, uses sdpecific dispersion algorithm is trained to the connection weight that obtains between visible layer and the first hidden layer and the end-state of the first hidden layer to this Boltzmann machine;
(S53) using the first hidden layer and the second hidden layer as limited Boltzmann machine, input using the end-state of the first hidden layer as the first hidden layer, uses sdpecific dispersion algorithm is trained to the connection weight that obtains between the first hidden layer and the second hidden layer and the end-state of the 2nd hidden layer to this Boltzmann machine;
(S54) using the second hidden layer and label layer as limited Boltzmann machine, input using the end-state of the second hidden layer as the second hidden layer, input using Text eigenvector as label layer, adopt the training of BP neural net method to obtain the connection weight of the second hidden layer and label layer, complete the training of degree of depth Boltzmann machine model.
The above; be only the specific embodiment of the present invention, but protection scope of the present invention is not limited to this, any be familiar with those skilled in the art the present invention disclose technical scope in; the variation that can expect easily or replacement, within all should being encompassed in protection scope of the present invention.
Claims (9)
1. the remote sensing image automatic marking method based on degree of depth study, is characterized in that, comprising:
(1) the low-level image feature vector that extracts remote sensing image to be marked builds the visual feature vector that obtains corresponding remote sensing image;
(2) degree of depth Boltzmann machine model input of described visual feature vector being trained carries out automatic marking;
The degree of depth Boltzmann machine model training in described step (2) obtains by following steps:
(S1) create the label dictionary that comprises several text labels;
(S2) select the mark of respective classes to have text label remote sensing image as model training data set according to label dictionary;
(S3) the low-level image feature vector that extracts each remote sensing image builds the visual feature vector that obtains corresponding remote sensing image, and determines the Text eigenvector of each remote sensing image according to label dictionary and text label;
(S4) build degree of depth Boltzmann machine model, described degree of depth Boltzmann machine model comprises visible layer, the first hidden layer, the second hidden layer and label layer from bottom to up successively, any two nodes in each layer are without connection, the two-way connection of any two nodes between adjacent layer;
(S5) utilize the visual feature vector of all remote sensing images of model training data centralization and Text eigenvector to described degree of depth Boltzmann machine model training, the degree of depth Boltzmann machine model that obtains training.
2. the remote sensing image automatic marking method based on degree of depth study as claimed in claim 1, is characterized in that, the interconnection network that the second described hidden layer and label layer form is unidirectional BP neural network.
3. the remote sensing image automatic marking method based on degree of depth study as claimed in claim 2, is characterized in that, the training process of the degree of depth Boltzmann machine model in described step (S5) is as follows:
(S51) using visual feature vector as visible layer, using Text eigenvector as label layer;
(S52) using visible layer and the first hidden layer as limited Boltzmann machine, input using visual feature vector as visible layer, uses sdpecific dispersion algorithm is trained to the connection weight that obtains between visible layer and the first hidden layer and the end-state of the first hidden layer to this Boltzmann machine;
(S53) using the first hidden layer and the second hidden layer as limited Boltzmann machine, input using the end-state of the first hidden layer as the first hidden layer, uses sdpecific dispersion algorithm is trained to the connection weight that obtains between the first hidden layer and the second hidden layer and the end-state of the 2nd hidden layer to this Boltzmann machine;
(S54) using the second hidden layer and label layer as limited Boltzmann machine, input using the end-state of the second hidden layer as the second hidden layer, input using Text eigenvector as label layer, adopt the training of BP neural net method to obtain the connection weight of the second hidden layer and label layer, complete the training of degree of depth Boltzmann machine model.
4. the remote sensing image automatic marking method based on degree of depth study as claimed in claim 3, is characterized in that, the node number of described visible layer is identical with the dimension of visual feature vector.
5. the remote sensing image automatic marking method based on degree of depth study as claimed in claim 4, is characterized in that, the node number of described label layer is identical with the dimension of Text eigenvector.
6. the remote sensing image automatic marking method based on degree of depth study as claimed in claim 5, is characterized in that, the dimension of described Text eigenvector is identical with the number of label dictionary Chinese version label.
7. the remote sensing image automatic marking method based on degree of depth study as described in arbitrary claim in claim 1~6, it is characterized in that, described low-level image feature vector comprises averaged spectrum Reflectivity vector, color layout's description vectors, color structure description vectors, scalable color description vector, homogeneity texture description vector, edge histogram description vectors, GIST proper vector and the visual word bag vector based on SIFT feature.
8. the remote sensing image automatic marking method based on degree of depth study as claimed in claim 7, it is characterized in that, described averaged spectrum Reflectivity vector comprises the spectrum average reflectance of four kinds of wavelength, and four kinds of described wavelength are respectively 0.44~0.51 micron, 0.53~0.62 micron, 0.63~0.70 micron and 0.74~0.80 micron.
9. the remote sensing image automatic marking method based on degree of depth study as claimed in claim 8, is characterized in that, the visual word bag vector based on SIFT feature obtains by following steps:
(a) calculate the SIFT proper vector of all remote sensing images of described model training data centralization;
(b) all SIFT proper vectors are carried out to cluster and obtain 500~1000 cluster centres;
(c), using each cluster centre as vision word, add up the occurrence number of each vision word in the SIFT proper vector of every remote sensing image and form the visual word bag vector based on SIFT feature.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410039584.3A CN103823845B (en) | 2014-01-28 | 2014-01-28 | Method for automatically annotating remote sensing images on basis of deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410039584.3A CN103823845B (en) | 2014-01-28 | 2014-01-28 | Method for automatically annotating remote sensing images on basis of deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103823845A true CN103823845A (en) | 2014-05-28 |
CN103823845B CN103823845B (en) | 2017-01-18 |
Family
ID=50758909
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410039584.3A Active CN103823845B (en) | 2014-01-28 | 2014-01-28 | Method for automatically annotating remote sensing images on basis of deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103823845B (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104021224A (en) * | 2014-06-25 | 2014-09-03 | 中国科学院自动化研究所 | Image labeling method based on layer-by-layer label fusing deep network |
CN104063720A (en) * | 2014-07-03 | 2014-09-24 | 浙江大学 | Method for detecting images of prohibited commodities of e-commerce websites based on deep Boltzmann machine |
CN104572940A (en) * | 2014-12-30 | 2015-04-29 | 中国人民解放军海军航空工程学院 | Automatic image annotation method based on deep learning and canonical correlation analysis |
CN104700100A (en) * | 2015-04-01 | 2015-06-10 | 哈尔滨工业大学 | Feature extraction method for high spatial resolution remote sensing big data |
CN105205448A (en) * | 2015-08-11 | 2015-12-30 | 中国科学院自动化研究所 | Character recognition model training method based on deep learning and recognition method thereof |
CN105894025A (en) * | 2016-03-30 | 2016-08-24 | 中国科学院自动化研究所 | Natural image aesthetic feeling quality assessment method based on multitask deep learning |
CN106056609A (en) * | 2016-06-02 | 2016-10-26 | 上海海洋大学 | Method based on DBNMI model for realizing automatic annotation of remote sensing image |
CN107044976A (en) * | 2017-05-10 | 2017-08-15 | 中国科学院合肥物质科学研究院 | Heavy metal content in soil analyzing and predicting method based on LIBS Yu stack RBM depth learning technologies |
WO2017166137A1 (en) * | 2016-03-30 | 2017-10-05 | 中国科学院自动化研究所 | Method for multi-task deep learning-based aesthetic quality assessment on natural image |
CN108764263A (en) * | 2018-02-12 | 2018-11-06 | 北京佳格天地科技有限公司 | The atural object annotation equipment and method of remote sensing image |
CN109636838A (en) * | 2018-12-11 | 2019-04-16 | 北京市燃气集团有限责任公司 | A kind of combustion gas Analysis of Potential method and device based on remote sensing image variation detection |
CN109886106A (en) * | 2019-01-15 | 2019-06-14 | 浙江大学 | A kind of remote sensing images building change detecting method based on deep learning |
CN110442721A (en) * | 2018-11-28 | 2019-11-12 | 腾讯科技(深圳)有限公司 | Neural network language model, training method, device and storage medium |
CN110738630A (en) * | 2018-07-02 | 2020-01-31 | 由田新技股份有限公司 | Training method and detection system of recursive deep learning system |
CN110765111A (en) * | 2019-10-28 | 2020-02-07 | 深圳市商汤科技有限公司 | Storage and reading method and device, electronic equipment and storage medium |
US10936905B2 (en) | 2018-07-06 | 2021-03-02 | Tata Consultancy Services Limited | Method and system for automatic object annotation using deep network |
CN113011584A (en) * | 2021-03-18 | 2021-06-22 | 广东南方数码科技股份有限公司 | Coding model training method, coding device and storage medium |
CN116521041A (en) * | 2023-05-10 | 2023-08-01 | 北京天工科仪空间技术有限公司 | Remote sensing data labeling method based on front page, computer equipment and medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101770584A (en) * | 2009-12-30 | 2010-07-07 | 重庆大学 | Extraction method for identification characteristic of high spectrum remote sensing data |
US7890512B2 (en) * | 2008-06-11 | 2011-02-15 | Microsoft Corporation | Automatic image annotation using semantic distance learning |
CN102129571A (en) * | 2011-01-31 | 2011-07-20 | 重庆大学 | Method for classifying multi-spectral remote sensing data land use based on semi-supervisor manifold learning |
CN102902984A (en) * | 2012-09-27 | 2013-01-30 | 西安电子科技大学 | Remote-sensing image semi-supervised projection dimension reducing method based on local consistency |
CN103198333A (en) * | 2013-04-15 | 2013-07-10 | 中国科学院电子学研究所 | Automatic semantic labeling method of high resolution remote sensing image |
-
2014
- 2014-01-28 CN CN201410039584.3A patent/CN103823845B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7890512B2 (en) * | 2008-06-11 | 2011-02-15 | Microsoft Corporation | Automatic image annotation using semantic distance learning |
CN101770584A (en) * | 2009-12-30 | 2010-07-07 | 重庆大学 | Extraction method for identification characteristic of high spectrum remote sensing data |
CN102129571A (en) * | 2011-01-31 | 2011-07-20 | 重庆大学 | Method for classifying multi-spectral remote sensing data land use based on semi-supervisor manifold learning |
CN102902984A (en) * | 2012-09-27 | 2013-01-30 | 西安电子科技大学 | Remote-sensing image semi-supervised projection dimension reducing method based on local consistency |
CN103198333A (en) * | 2013-04-15 | 2013-07-10 | 中国科学院电子学研究所 | Automatic semantic labeling method of high resolution remote sensing image |
Cited By (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104021224A (en) * | 2014-06-25 | 2014-09-03 | 中国科学院自动化研究所 | Image labeling method based on layer-by-layer label fusing deep network |
CN104063720A (en) * | 2014-07-03 | 2014-09-24 | 浙江大学 | Method for detecting images of prohibited commodities of e-commerce websites based on deep Boltzmann machine |
CN104572940B (en) * | 2014-12-30 | 2017-11-21 | 中国人民解放军海军航空工程学院 | A kind of image automatic annotation method based on deep learning and canonical correlation analysis |
CN104572940A (en) * | 2014-12-30 | 2015-04-29 | 中国人民解放军海军航空工程学院 | Automatic image annotation method based on deep learning and canonical correlation analysis |
CN104700100A (en) * | 2015-04-01 | 2015-06-10 | 哈尔滨工业大学 | Feature extraction method for high spatial resolution remote sensing big data |
CN105205448A (en) * | 2015-08-11 | 2015-12-30 | 中国科学院自动化研究所 | Character recognition model training method based on deep learning and recognition method thereof |
CN105205448B (en) * | 2015-08-11 | 2019-03-15 | 中国科学院自动化研究所 | Text region model training method and recognition methods based on deep learning |
US10685434B2 (en) | 2016-03-30 | 2020-06-16 | Institute Of Automation, Chinese Academy Of Sciences | Method for assessing aesthetic quality of natural image based on multi-task deep learning |
WO2017166137A1 (en) * | 2016-03-30 | 2017-10-05 | 中国科学院自动化研究所 | Method for multi-task deep learning-based aesthetic quality assessment on natural image |
CN105894025A (en) * | 2016-03-30 | 2016-08-24 | 中国科学院自动化研究所 | Natural image aesthetic feeling quality assessment method based on multitask deep learning |
CN106056609B (en) * | 2016-06-02 | 2018-11-06 | 上海海洋大学 | Method based on DBNMI model realization remote sensing image automatic markings |
CN106056609A (en) * | 2016-06-02 | 2016-10-26 | 上海海洋大学 | Method based on DBNMI model for realizing automatic annotation of remote sensing image |
CN107044976A (en) * | 2017-05-10 | 2017-08-15 | 中国科学院合肥物质科学研究院 | Heavy metal content in soil analyzing and predicting method based on LIBS Yu stack RBM depth learning technologies |
CN108764263A (en) * | 2018-02-12 | 2018-11-06 | 北京佳格天地科技有限公司 | The atural object annotation equipment and method of remote sensing image |
CN110738630A (en) * | 2018-07-02 | 2020-01-31 | 由田新技股份有限公司 | Training method and detection system of recursive deep learning system |
US10936905B2 (en) | 2018-07-06 | 2021-03-02 | Tata Consultancy Services Limited | Method and system for automatic object annotation using deep network |
CN110442721A (en) * | 2018-11-28 | 2019-11-12 | 腾讯科技(深圳)有限公司 | Neural network language model, training method, device and storage medium |
CN110442721B (en) * | 2018-11-28 | 2023-01-06 | 腾讯科技(深圳)有限公司 | Neural network language model, training method, device and storage medium |
CN109636838A (en) * | 2018-12-11 | 2019-04-16 | 北京市燃气集团有限责任公司 | A kind of combustion gas Analysis of Potential method and device based on remote sensing image variation detection |
CN109886106A (en) * | 2019-01-15 | 2019-06-14 | 浙江大学 | A kind of remote sensing images building change detecting method based on deep learning |
CN110765111A (en) * | 2019-10-28 | 2020-02-07 | 深圳市商汤科技有限公司 | Storage and reading method and device, electronic equipment and storage medium |
CN110765111B (en) * | 2019-10-28 | 2023-03-31 | 深圳市商汤科技有限公司 | Storage and reading method and device, electronic equipment and storage medium |
CN113011584A (en) * | 2021-03-18 | 2021-06-22 | 广东南方数码科技股份有限公司 | Coding model training method, coding device and storage medium |
CN113011584B (en) * | 2021-03-18 | 2024-04-16 | 广东南方数码科技股份有限公司 | Coding model training method, coding device and storage medium |
CN116521041A (en) * | 2023-05-10 | 2023-08-01 | 北京天工科仪空间技术有限公司 | Remote sensing data labeling method based on front page, computer equipment and medium |
Also Published As
Publication number | Publication date |
---|---|
CN103823845B (en) | 2017-01-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103823845A (en) | Method for automatically annotating remote sensing images on basis of deep learning | |
Zhao et al. | Dirichlet-derived multiple topic scene classification model for high spatial resolution remote sensing imagery | |
Zhao et al. | ApLeaf: An efficient android-based plant leaf identification system | |
CN107704877A (en) | A kind of image privacy cognitive method based on deep learning | |
CN103345656B (en) | A kind of data identification method based on multitask deep neural network and device | |
Zheng et al. | Topic modeling of multimodal data: an autoregressive approach | |
CN104834748A (en) | Image retrieval method utilizing deep semantic to rank hash codes | |
CN105279554B (en) | The training method and device of deep neural network based on Hash coding layer | |
CN107644235A (en) | Image automatic annotation method based on semi-supervised learning | |
CN105893349A (en) | Category label matching and mapping method and device | |
CN107067020A (en) | Image identification method and device | |
CN105631479A (en) | Imbalance-learning-based depth convolution network image marking method and apparatus | |
CN109614973A (en) | Rice seedling and Weeds at seedling image, semantic dividing method, system, equipment and medium | |
Wu et al. | Online asymmetric similarity learning for cross-modal retrieval | |
CN107832351A (en) | Cross-module state search method based on depth related network | |
CN109766424A (en) | Filtering method and device for reading understanding model training data | |
CN109522797A (en) | Rice seedling and Weeds at seedling recognition methods and system based on convolutional neural networks | |
CN107463658A (en) | File classification method and device | |
CN104268140B (en) | Image search method based on weight self study hypergraph and multivariate information fusion | |
CN104572940A (en) | Automatic image annotation method based on deep learning and canonical correlation analysis | |
CN103530403B (en) | A kind of structurized Image Description Methods | |
Serrano-Talamantes et al. | Self organizing natural scene image retrieval | |
CN103530405B (en) | A kind of image search method based on hierarchy | |
CN104298974A (en) | Human body behavior recognition method based on depth video sequence | |
CN103324954A (en) | Image classification method based on tree structure and system using same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant |