CN105590108A - Scene identification method used in noisy environment - Google Patents
Scene identification method used in noisy environment Download PDFInfo
- Publication number
- CN105590108A CN105590108A CN201610103825.5A CN201610103825A CN105590108A CN 105590108 A CN105590108 A CN 105590108A CN 201610103825 A CN201610103825 A CN 201610103825A CN 105590108 A CN105590108 A CN 105590108A
- Authority
- CN
- China
- Prior art keywords
- sample
- feature
- scene
- image
- dimensional
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/35—Categorising the entire scene, e.g. birthday party or wedding scene
- G06V20/36—Indoor scenes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
Abstract
The invention relates to a scene identification method used in a noisy environment, and relates to the corresponding technical field. The scene identification method is characterized in that (1) scene images comprising samples of marks can be acquired by adopting a sensor; (2) the characteristic extraction and the characteristic expression of the color images and the depth images of the scenes can be respectively carried out, and the color image characteristics and the depth image characteristics of the same group can be combined together; (3)the characteristic selecting models of the samples of the marks can be acquired by adopting the characteristic selecting algorithm according to the second step; (4) the classification can be carried out by adopting a classifier. The scene identification method is advantageous in that the accurate identification of the scenes can be carried out in the noisy environment, and a certain identification capability can be guaranteed under the condition of the samples mixed with the noises, and therefore the performance can be improved under the condition of the indoor scene data mixed with the noises in a concentrated manner.
Description
Technical field
The invention belongs to a kind of scene Recognition method and technology field, especially belong to a kind of scene Recognition having under the environment of making an uproarMethod and technology field.
Background technology
In general, scene classification can be regarded independently object identification problem of a kind of visual angle as, and a scene is by oneThe entity composition of series. For example, indoor scene can include chair, desk, people and bookshelf, and the ornaments of these things neitherChangeless. Accurate identification to scene contributes to solve a lot of practical applications, such as CBIR, and machinePeople's Path Planning Technique and image labeling, etc. Nowadays, scene Recognition has more and more received researcher's concern.
Research widely shows that the characteristics of image dimension obtaining after feature extraction is very high, is limited to computational resource, higherCharacteristic dimension can affect the scene Recognition of RGB-D (in conjunction with coloured image and depth image (depthimage)) sensor compositionThe practicality of system. Although existing Feature Selection can make high dimensional feature become more succinct and effective, but existingSome feature selection approach have been ignored the problem that is mixed with much noise in sample, but, in actual application, due to systemComplexity issue and device processes precision problem a lot of noise that usually can adulterate, the knowledge of so existing feature selecting algorithmOther effect just has certain limitation.
Through retrieval, disclosed scene Recognition process patent is existing more than hundred sections through retrieving, but under the environment of making an uproarScene Recognition method few in number, applicant is in conjunction with Cauchy's estimation theory, forms stream shape Cauchy learning algorithm, reached and made an uproarAccurate identification to scene under environment.
Summary of the invention
The present invention describes a kind of scene Recognition method and calculation of a kind of new feature selecting having under the environment of making an uproar in detailMethod---stream shape Cauchy learning algorithm.
The present invention adopts following technical scheme to realize.
There is the scene Recognition method under the environment of making an uproar, comprise the following steps: 1) utilize sensor to obtain scene image, inContaining the sample of mark; 2) coloured image to scene and depth image (depthimage) carry out feature extraction and feature respectivelyExpress, merge coloured image feature and depth image (depthimage) feature of same group; 3) select feature selecting algorithm pairThe sample of mark obtains feature selecting model by the feature that the 2nd step obtains; 4) adopt grader to classify.
Step 1 of the present invention) be to utilize Kinect sensor to obtain scene image.
Step 2 of the present invention) be specially, convert all images to gray-scale map, image is carried out to convergent-divergent, then rightIn the localized mass of coloured image and depth image (depthimage), extract special by yardstick invariant features conversion (SIFT) methodLevy, then adopt partial restriction uniform enconding (LLC) algorithm to carry out feature representation.
Step 3 of the present invention) feature selecting algorithm is for flowing shape Cauchy learning algorithm, and concrete steps are as follows:
For a given sample xiBelong to sample set X=[x1,x2,...xn]∈RD×N(N is number of samples here,D is the original dimension of sample, and R represents at real number space), the low-dimensional sample y that it is correspondingiBelong to sample set Y=[y1,y2,...yn]∈Rd×N(d is the dimension after dimensionality reduction here), finds the similar and inhomogeneous sample of K arest neighborsItsIn, there is k1Individually be and xiSimilar sample, remaining k2Individually be and xiInhomogeneous sample, wherein K=k1+k2, use respectivelyWithRepresent this two groups of samples; For whole xiLocalized mass be expressed as:(whereinRepresent D × (k1+k2+ 1) linear space of dimension), rightThe low-dimensional of answering is expressedIn a low-dimensional localized mass newly obtaining,Reach between similar sample enough far away of the enough distance near and between inhomogeneity sample of distance, therefore above majorized function tableShow as follows:
α is scale factor, is used for controlling the impact of sample between the interior sample of class and class;
Define a coefficient vector ωi:
Utilize the coefficient vector of definition, (1) formula will be become following form by abbreviation:
What tr () represented is mark computing, in formula
Introduce selection matrix (S belowi)pq:
Therefore, obtain low-dimensional and express Yi=YSi, object function (2) is rewritten as:
Introduced Cauchy's estimation theory, overcome the impact that noise brings, (4) formula becomes following form:
C is the coefficient for weighing noise;
Owing to there being Y=UTX relation, (7) formula is reduced to:
Distant between the sample out of the space representation in low-dimensional just represents for each sampleFor enough far away of: the distance at the sample of each lower dimensional space and all sample classes center, be expressed as following object function:
Be exactly all sample Lei centers,
For fear of the situation of over-fitting occurs, added two norms, integrate so all above-mentioned situations, just write as withUnder object function:
The C here1And C2It is regularization coefficient;
In order to make (8) have unique solution, the given qualifications of applicant is UTU=I; Projection matrix U is by the method for iterationWith the method for solving solution of characteristic value out.
The invention has the beneficial effects as follows under the environment of making an uproar scene is accurately identified, ensured to sneak into noise at sampleEqually also there is certain identification capability later; Therefore just improve indoor scene data centralization and be mixed with the performance under noise situations.
Below in conjunction with the drawings and specific embodiments, the present invention is further explained.
Brief description of the drawings
Fig. 1 is the logical framework figure of technical solution of the present invention.
Detailed description of the invention
See Fig. 1, the object of the invention is to overcome existing combination coloured image and depth image (depthiamge) and passIn the indoor scene recognition system of sensor, do not consider the problem of influence of noise, proposed a kind of scene Recognition having under the environment of making an uproarMethod, comprises the following steps: 1) utilize Kinect sensor to obtain scene image; 2) coloured image to scene and the degree of depth respectivelyImage (depthimage) carries out feature extraction and feature representation, merges coloured image feature and the depth image of same group(depthimage) feature; 3) adopt stream shape Cauchy learning algorithm to obtain feature to the sample of mark by the feature that the 2nd step obtainsPreference pattern; 4) adopt SVMs (SVM) grader to classify.
The described first step utilizes Kinect sensor to obtain scene coloured image and corresponding depth image (depthimage)。
The feature of the coloured image that described second step obtains the first step and corresponding depth image (depthimage)The process of extraction and feature representation and merging is as follows: 1) convert all images to gray-scale map, and by certain proportion to imageCarry out convergent-divergent, make its size be less than or equal to 300 × 300 pixels. 2) to coloured image and depth image (depthimage)Localized mass on extract feature by yardstick invariant features conversion (SIFT) method, the size of this localized mass is 16 × 16 pixels, phaseBetween adjacent localized mass on image level or vertically have the overlapping region of 8 pixels, the yardstick invariant features extracting in localized massConverting characteristic dimension is 128. 3) adopt LLC algorithm to carry out feature representation. Adopt partial restriction uniform enconding (LLC) algorithm to carry outWhen expression, need to carry out k average (k-means) cluster to the localized mass on all data sets, thereby form a code book (wordAllusion quotation). K means clustering algorithm chooses at random first cluster centre, when cluster centre is in the time changing among a small circle, and the termination of iteration.Suppose that the number of code book is 1024 in embodiments. Applicant carries out maximum polymerization on three sheaf space pyramid models,This three sheaf spaces pyramid model is divided into 1 × 1,2 × 2 and 4 × 4 subregion. To paired coloured image and depth image(depthimage) concerning, the characteristic length of partial restriction uniform enconding (LLC) is all (1+4+16) × 1024=21504.The characteristics of image finally coloured image and depth image (depthimage) being obtained respectively merges and obtains 21504 × 2=43008 dimensional features.
The 3rd described step feature selecting algorithm, flows shape Cauchy learning algorithm concrete steps as follows:
For a given sample xiBelong to sample set X=[x1,x2,...xn]∈RD×N(N is number of samples here,D is the original dimension of sample, and R represents at real number space), the low-dimensional sample y that it is correspondingiBelong to sample set Y=[y1,y2,...yn]∈Rd×N(d is the dimension after dimensionality reduction here), finds the similar and inhomogeneous sample of K arest neighborsItsIn, there is k1Individually be and xiSimilar sample, remaining k2Individually be and xiInhomogeneous sample, wherein K=k1+k2, applicant respectivelyWithWithRepresent this two groups of samples; For whole xiLocalized mass be expressed as:(whereinRepresent D × (k1+k2+ 1) linear space of dimension), rightThe low-dimensional of answering is expressedIn a low-dimensional localized mass newly obtaining,Reach between similar sample enough far away of the enough distance near and between inhomogeneity sample of distance, therefore above saying optimizationFunction representation is as follows:
Here α is scale factor, is used for controlling the impact of sample between the interior sample of class and class.
Here can define a coefficient vector ωi:
Utilize the coefficient vector of definition, (1) formula will be become following form by abbreviation:
What the tr () here represented is mark computing, in formula
Introduce selection matrix (S belowi)pq:
Here
Therefore, obtain low-dimensional and express Yi=YSi, object function (2) is rewritten as:
Introduce Cauchy's estimation theory [1],
[1]M.IvanandC.H.Muller,“Breakdownpointsofcauchyregression-scaleestimators,”Statistics&probabilityletters,vol.57,no.1,pp.79–89,Feb.2002.
Overcome the impact that noise brings, (4) formula becomes following form:
The c is here the coefficient for weighing noise.
Owing to there being Y=UTX relation, (7) formula abbreviation is:
Distant between the sample out of the space representation in low-dimensional just represents for each sampleFor enough far away of: the distance at the sample of each lower dimensional space and all sample classes center, be expressed as following object function:
HereBe exactly all sample Lei centers,
For fear of the situation that over-fitting occurs, so added two norms, integrate so all above-mentioned situations, just writeBecome following object function:
The C here1And C2It is regularization coefficient.
In order to make (8) have unique solution, the given qualifications of applicant is UTU=I. Projection matrix U is by the method for iterationWith the method for solving solution of characteristic value out.
The 4th described step adopts SVMs (SVM) grader to classify.
Claims (4)
1. there is the scene Recognition method under the environment of making an uproar, it is characterized in that, comprise the following steps: 1) utilize sensor to obtain fieldScape image, the sample of intragenic marker; 2) coloured image to scene and depth image depthimage carry out feature extraction respectivelyAnd feature representation, merge coloured image feature and the depth image depthimage feature of same group; 3) select feature selecting to calculateMethod is obtained feature selecting model to the sample of mark by the feature that the 2nd step obtains; 4) adopt grader to classify.
2. a kind of scene Recognition method having under the environment of making an uproar according to claim 1, is characterized in that described step 1)To utilize Kinect sensor to obtain scene image.
3. a kind of scene Recognition method having under the environment of making an uproar according to claim 1, is characterized in that described step 2)Be specially, convert all images to gray-scale map, image is carried out to convergent-divergent, then to coloured image and depth image depthIn the localized mass of image, extract feature by yardstick invariant features conversion method, then adopt partial restriction uniform enconding algorithm to carry outFeature representation.
4. a kind of scene Recognition method having under the environment of making an uproar according to claim 1, is characterized in that described step 3)Feature selecting algorithm is stream shape Cauchy learning algorithm, and concrete steps are as follows:
For a given sample xiBelong to sample set X=[x1,x2,...xn]∈RD×N(N is number of samples here, and D isThe original dimension of sample, R represents at real number space), the low-dimensional sample y that it is correspondingiBelong to sample set Y=[y1,y2,...yn]∈Rd×N(d is the dimension after dimensionality reduction here), finds the similar and inhomogeneous sample of K arest neighborsWherein, there is k1Individually be and xiSimilar sample, remaining k2Individually be and xiInhomogeneous sample, wherein K=k1+k2, use respectivelyWithRepresent this two groups of samples; For whole xiLocalized mass be expressed as:
α is scale factor, is used for controlling the impact of sample between the interior sample of class and class;
Define a coefficient vector ωi:
Utilize the coefficient vector of definition, (1) formula will be become following form by abbreviation:
What tr () represented is mark computing, in formula
Introduce selection matrix (S belowi)pq:
Therefore, obtain low-dimensional and express Yi=YSi, object function (2) is rewritten as:
Introduced Cauchy's estimation theory, overcome the impact that noise brings, (4) formula becomes following form:
C is the coefficient for weighing noise;
Owing to there being Y=UTX relation, (7) formula is reduced to:
Distant between the sample out of the space representation in low-dimensional is just expressed as: every for each sampleEnough far away of the distance at the sample of individual lower dimensional space and all sample classes center, is expressed as following object function:
Be exactly all sample Lei centers,
For fear of the situation that over-fitting occurs, add two norms, integrate so all above-mentioned situations, just write as followingObject function:
The C here1And C2It is regularization coefficient;
In order to make (8) have unique solution, the given qualifications of applicant is UTU=I; Projection matrix U is by method and the feature of iterationThe method for solving solution of value out.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610103825.5A CN105590108B (en) | 2016-02-25 | 2016-02-25 | A kind of scene recognition method under noisy environment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610103825.5A CN105590108B (en) | 2016-02-25 | 2016-02-25 | A kind of scene recognition method under noisy environment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN105590108A true CN105590108A (en) | 2016-05-18 |
CN105590108B CN105590108B (en) | 2018-08-17 |
Family
ID=55929678
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610103825.5A Active CN105590108B (en) | 2016-02-25 | 2016-02-25 | A kind of scene recognition method under noisy environment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105590108B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050031034A1 (en) * | 2003-06-25 | 2005-02-10 | Nejat Kamaci | Cauchy-distribution based coding system and method |
CN102629330A (en) * | 2012-02-29 | 2012-08-08 | 华南理工大学 | Rapid and high-precision matching method of depth image and color image |
CN102867191A (en) * | 2012-09-04 | 2013-01-09 | 广东群兴玩具股份有限公司 | Dimension reducing method based on manifold sub-space study |
CN103500342A (en) * | 2013-09-18 | 2014-01-08 | 华南理工大学 | Human behavior recognition method based on accelerometer |
CN104732209A (en) * | 2015-03-17 | 2015-06-24 | 深圳先进技术研究院 | Indoor scene recognition method and device |
-
2016
- 2016-02-25 CN CN201610103825.5A patent/CN105590108B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050031034A1 (en) * | 2003-06-25 | 2005-02-10 | Nejat Kamaci | Cauchy-distribution based coding system and method |
CN102629330A (en) * | 2012-02-29 | 2012-08-08 | 华南理工大学 | Rapid and high-precision matching method of depth image and color image |
CN102867191A (en) * | 2012-09-04 | 2013-01-09 | 广东群兴玩具股份有限公司 | Dimension reducing method based on manifold sub-space study |
CN103500342A (en) * | 2013-09-18 | 2014-01-08 | 华南理工大学 | Human behavior recognition method based on accelerometer |
CN104732209A (en) * | 2015-03-17 | 2015-06-24 | 深圳先进技术研究院 | Indoor scene recognition method and device |
Also Published As
Publication number | Publication date |
---|---|
CN105590108B (en) | 2018-08-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106096557B (en) | A kind of semi-supervised learning facial expression recognizing method based on fuzzy training sample | |
CN109753885B (en) | Target detection method and device and pedestrian detection method and system | |
CN104599275B (en) | The RGB-D scene understanding methods of imparametrization based on probability graph model | |
CN105956560A (en) | Vehicle model identification method based on pooling multi-scale depth convolution characteristics | |
CN107766890B (en) | Improved method for discriminant graph block learning in fine-grained identification | |
CN106126581A (en) | Cartographical sketching image search method based on degree of depth study | |
CN107808129A (en) | A kind of facial multi-characteristic points localization method based on single convolutional neural networks | |
CN107330355B (en) | Deep pedestrian re-identification method based on positive sample balance constraint | |
CN105488536A (en) | Agricultural pest image recognition method based on multi-feature deep learning technology | |
CN108038435A (en) | A kind of feature extraction and method for tracking target based on convolutional neural networks | |
CN103268363B (en) | A kind of Chinese calligraphy's image search method based on elastic HOG feature and DDTW coupling | |
CN109101981B (en) | Loop detection method based on global image stripe code in streetscape scene | |
CN103699902A (en) | Sorting method of ground-based visible light cloud picture | |
CN102622609B (en) | Method for automatically classifying three-dimensional models based on support vector machine | |
CN108154158B (en) | Building image segmentation method for augmented reality application | |
CN102122353A (en) | Method for segmenting images by using increment dictionary learning and sparse representation | |
CN107944428A (en) | A kind of indoor scene semanteme marking method based on super-pixel collection | |
CN108537145A (en) | Human bodys' response method based on space-time skeleton character and depth belief network | |
CN107203745A (en) | A kind of across visual angle action identification method based on cross-domain study | |
CN110517270B (en) | Indoor scene semantic segmentation method based on super-pixel depth network | |
CN107767416A (en) | The recognition methods of pedestrian's direction in a kind of low-resolution image | |
CN109934095A (en) | A kind of remote sensing images Clean water withdraw method and system based on deep learning | |
CN105320963B (en) | The semi-supervised feature selection approach of large scale towards high score remote sensing images | |
CN101986295A (en) | Image clustering method based on manifold sparse coding | |
CN105574545B (en) | The semantic cutting method of street environment image various visual angles and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20210218 Address after: 650091 Yunnan province Kunming City Lake Road No. 2 Patentee after: YUNNAN University Patentee after: Tao Dapeng Address before: 650091 Yunnan province Kunming City Lake Road No. 2 Patentee before: YUNNAN University |