CN105590108A - Scene identification method used in noisy environment - Google Patents

Scene identification method used in noisy environment Download PDF

Info

Publication number
CN105590108A
CN105590108A CN201610103825.5A CN201610103825A CN105590108A CN 105590108 A CN105590108 A CN 105590108A CN 201610103825 A CN201610103825 A CN 201610103825A CN 105590108 A CN105590108 A CN 105590108A
Authority
CN
China
Prior art keywords
sample
feature
scene
image
dimensional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610103825.5A
Other languages
Chinese (zh)
Other versions
CN105590108B (en
Inventor
陶大鹏
郭亚男
杨喜鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yunnan University YNU
Original Assignee
Yunnan University YNU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Yunnan University YNU filed Critical Yunnan University YNU
Priority to CN201610103825.5A priority Critical patent/CN105590108B/en
Publication of CN105590108A publication Critical patent/CN105590108A/en
Application granted granted Critical
Publication of CN105590108B publication Critical patent/CN105590108B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/35Categorising the entire scene, e.g. birthday party or wedding scene
    • G06V20/36Indoor scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]

Abstract

The invention relates to a scene identification method used in a noisy environment, and relates to the corresponding technical field. The scene identification method is characterized in that (1) scene images comprising samples of marks can be acquired by adopting a sensor; (2) the characteristic extraction and the characteristic expression of the color images and the depth images of the scenes can be respectively carried out, and the color image characteristics and the depth image characteristics of the same group can be combined together; (3)the characteristic selecting models of the samples of the marks can be acquired by adopting the characteristic selecting algorithm according to the second step; (4) the classification can be carried out by adopting a classifier. The scene identification method is advantageous in that the accurate identification of the scenes can be carried out in the noisy environment, and a certain identification capability can be guaranteed under the condition of the samples mixed with the noises, and therefore the performance can be improved under the condition of the indoor scene data mixed with the noises in a concentrated manner.

Description

A kind of scene Recognition method having under the environment of making an uproar
Technical field
The invention belongs to a kind of scene Recognition method and technology field, especially belong to a kind of scene Recognition having under the environment of making an uproarMethod and technology field.
Background technology
In general, scene classification can be regarded independently object identification problem of a kind of visual angle as, and a scene is by oneThe entity composition of series. For example, indoor scene can include chair, desk, people and bookshelf, and the ornaments of these things neitherChangeless. Accurate identification to scene contributes to solve a lot of practical applications, such as CBIR, and machinePeople's Path Planning Technique and image labeling, etc. Nowadays, scene Recognition has more and more received researcher's concern.
Research widely shows that the characteristics of image dimension obtaining after feature extraction is very high, is limited to computational resource, higherCharacteristic dimension can affect the scene Recognition of RGB-D (in conjunction with coloured image and depth image (depthimage)) sensor compositionThe practicality of system. Although existing Feature Selection can make high dimensional feature become more succinct and effective, but existingSome feature selection approach have been ignored the problem that is mixed with much noise in sample, but, in actual application, due to systemComplexity issue and device processes precision problem a lot of noise that usually can adulterate, the knowledge of so existing feature selecting algorithmOther effect just has certain limitation.
Through retrieval, disclosed scene Recognition process patent is existing more than hundred sections through retrieving, but under the environment of making an uproarScene Recognition method few in number, applicant is in conjunction with Cauchy's estimation theory, forms stream shape Cauchy learning algorithm, reached and made an uproarAccurate identification to scene under environment.
Summary of the invention
The present invention describes a kind of scene Recognition method and calculation of a kind of new feature selecting having under the environment of making an uproar in detailMethod---stream shape Cauchy learning algorithm.
The present invention adopts following technical scheme to realize.
There is the scene Recognition method under the environment of making an uproar, comprise the following steps: 1) utilize sensor to obtain scene image, inContaining the sample of mark; 2) coloured image to scene and depth image (depthimage) carry out feature extraction and feature respectivelyExpress, merge coloured image feature and depth image (depthimage) feature of same group; 3) select feature selecting algorithm pairThe sample of mark obtains feature selecting model by the feature that the 2nd step obtains; 4) adopt grader to classify.
Step 1 of the present invention) be to utilize Kinect sensor to obtain scene image.
Step 2 of the present invention) be specially, convert all images to gray-scale map, image is carried out to convergent-divergent, then rightIn the localized mass of coloured image and depth image (depthimage), extract special by yardstick invariant features conversion (SIFT) methodLevy, then adopt partial restriction uniform enconding (LLC) algorithm to carry out feature representation.
Step 3 of the present invention) feature selecting algorithm is for flowing shape Cauchy learning algorithm, and concrete steps are as follows:
For a given sample xiBelong to sample set X=[x1,x2,...xn]∈RD×N(N is number of samples here,D is the original dimension of sample, and R represents at real number space), the low-dimensional sample y that it is correspondingiBelong to sample set Y=[y1,y2,...yn]∈Rd×N(d is the dimension after dimensionality reduction here), finds the similar and inhomogeneous sample of K arest neighborsItsIn, there is k1Individually be and xiSimilar sample, remaining k2Individually be and xiInhomogeneous sample, wherein K=k1+k2, use respectivelyWithRepresent this two groups of samples; For whole xiLocalized mass be expressed as:(whereinRepresent D × (k1+k2+ 1) linear space of dimension), rightThe low-dimensional of answering is expressedIn a low-dimensional localized mass newly obtaining,Reach between similar sample enough far away of the enough distance near and between inhomogeneity sample of distance, therefore above majorized function tableShow as follows:
min y i Σ j = 1 k 1 | | y i - y i j | | 2 - α Σ p = 1 k 2 | | y i - y i p | | 2 , - - - ( 1 )
α is scale factor, is used for controlling the impact of sample between the interior sample of class and class;
Define a coefficient vector ωi
Utilize the coefficient vector of definition, (1) formula will be become following form by abbreviation:
arg min Y i t r ( Y i L i Y i T ) , - - - ( 2 )
What tr () represented is mark computing, in formula L i = Σ j = 1 k 1 + k 2 ( ω i ) j - ω i T - ω i d i a g ( ω i ) .
Introduce selection matrix (S belowi)pq
( S i ) p q = { 1 , I f p = F i { q } ; 0 , e l s e . - - - ( 3 )
S i ∈ R N × ( 1 + k 1 + k 2 ) ;
Therefore, obtain low-dimensional and express Yi=YSi, object function (2) is rewritten as:
arg min Y Σ i = 1 N t r ( YS i L i S i T Y T ) = arg min Y t r ( YLY T ) . - - - ( 4 )
Introduced Cauchy's estimation theory, overcome the impact that noise brings, (4) formula becomes following form:
arg min Y l o g ( 1 + ( t r ( YLY T ) c 2 ) ) - - - ( 5 )
C is the coefficient for weighing noise;
Owing to there being Y=UTX relation, (7) formula is reduced to:
arg min U l o g ( 1 + ( t r ( U T XLX T U ) c 2 ) ) - - - ( 6 )
Distant between the sample out of the space representation in low-dimensional just represents for each sampleFor enough far away of: the distance at the sample of each lower dimensional space and all sample classes center, be expressed as following object function:
arg m a x y i Σ i = 1 N | | y i - y ‾ | | 2 - - - ( 7 )
Be exactly all sample Lei centers,
For fear of the situation of over-fitting occurs, added two norms, integrate so all above-mentioned situations, just write as withUnder object function:
arg m i n U l o g ( 1 + ( t r ( U T XLX T U ) c 2 ) ) - C 1 t r ( U T S U ) + C 2 | | U | | 2 - - - ( 8 )
The C here1And C2It is regularization coefficient;
In order to make (8) have unique solution, the given qualifications of applicant is UTU=I; Projection matrix U is by the method for iterationWith the method for solving solution of characteristic value out.
The invention has the beneficial effects as follows under the environment of making an uproar scene is accurately identified, ensured to sneak into noise at sampleEqually also there is certain identification capability later; Therefore just improve indoor scene data centralization and be mixed with the performance under noise situations.
Below in conjunction with the drawings and specific embodiments, the present invention is further explained.
Brief description of the drawings
Fig. 1 is the logical framework figure of technical solution of the present invention.
Detailed description of the invention
See Fig. 1, the object of the invention is to overcome existing combination coloured image and depth image (depthiamge) and passIn the indoor scene recognition system of sensor, do not consider the problem of influence of noise, proposed a kind of scene Recognition having under the environment of making an uproarMethod, comprises the following steps: 1) utilize Kinect sensor to obtain scene image; 2) coloured image to scene and the degree of depth respectivelyImage (depthimage) carries out feature extraction and feature representation, merges coloured image feature and the depth image of same group(depthimage) feature; 3) adopt stream shape Cauchy learning algorithm to obtain feature to the sample of mark by the feature that the 2nd step obtainsPreference pattern; 4) adopt SVMs (SVM) grader to classify.
The described first step utilizes Kinect sensor to obtain scene coloured image and corresponding depth image (depthimage)。
The feature of the coloured image that described second step obtains the first step and corresponding depth image (depthimage)The process of extraction and feature representation and merging is as follows: 1) convert all images to gray-scale map, and by certain proportion to imageCarry out convergent-divergent, make its size be less than or equal to 300 × 300 pixels. 2) to coloured image and depth image (depthimage)Localized mass on extract feature by yardstick invariant features conversion (SIFT) method, the size of this localized mass is 16 × 16 pixels, phaseBetween adjacent localized mass on image level or vertically have the overlapping region of 8 pixels, the yardstick invariant features extracting in localized massConverting characteristic dimension is 128. 3) adopt LLC algorithm to carry out feature representation. Adopt partial restriction uniform enconding (LLC) algorithm to carry outWhen expression, need to carry out k average (k-means) cluster to the localized mass on all data sets, thereby form a code book (wordAllusion quotation). K means clustering algorithm chooses at random first cluster centre, when cluster centre is in the time changing among a small circle, and the termination of iteration.Suppose that the number of code book is 1024 in embodiments. Applicant carries out maximum polymerization on three sheaf space pyramid models,This three sheaf spaces pyramid model is divided into 1 × 1,2 × 2 and 4 × 4 subregion. To paired coloured image and depth image(depthimage) concerning, the characteristic length of partial restriction uniform enconding (LLC) is all (1+4+16) × 1024=21504.The characteristics of image finally coloured image and depth image (depthimage) being obtained respectively merges and obtains 21504 × 2=43008 dimensional features.
The 3rd described step feature selecting algorithm, flows shape Cauchy learning algorithm concrete steps as follows:
For a given sample xiBelong to sample set X=[x1,x2,...xn]∈RD×N(N is number of samples here,D is the original dimension of sample, and R represents at real number space), the low-dimensional sample y that it is correspondingiBelong to sample set Y=[y1,y2,...yn]∈Rd×N(d is the dimension after dimensionality reduction here), finds the similar and inhomogeneous sample of K arest neighborsItsIn, there is k1Individually be and xiSimilar sample, remaining k2Individually be and xiInhomogeneous sample, wherein K=k1+k2, applicant respectivelyWithWithRepresent this two groups of samples; For whole xiLocalized mass be expressed as:(whereinRepresent D × (k1+k2+ 1) linear space of dimension), rightThe low-dimensional of answering is expressedIn a low-dimensional localized mass newly obtaining,Reach between similar sample enough far away of the enough distance near and between inhomogeneity sample of distance, therefore above saying optimizationFunction representation is as follows:
min y i Σ j = 1 k 1 | | y i - y i j | | 2 - α Σ p = 1 k 2 | | y i - y i p | | 2 , - - - ( 1 )
Here α is scale factor, is used for controlling the impact of sample between the interior sample of class and class.
Here can define a coefficient vector ωi
Utilize the coefficient vector of definition, (1) formula will be become following form by abbreviation:
arg min Y i t r ( Y i L i Y i T ) - - - ( 2 )
What the tr () here represented is mark computing, in formula L i = Σ j = 1 k 1 + k 2 ( ω i ) j - ω i T - ω i d i a g ( ω i ) .
Introduce selection matrix (S belowi)pq
( S i ) p q = { 1 , I f p = F i { q } ; 0 , e l s e . - - - ( 3 )
Here S i ∈ R N × ( 1 + k 1 + k 2 ) .
Therefore, obtain low-dimensional and express Yi=YSi, object function (2) is rewritten as:
arg min Y Σ i = 1 N t r ( YS i L i S i T Y T ) = arg min Y t r ( YLY T ) . - - - ( 4 )
Introduce Cauchy's estimation theory [1],
[1]M.IvanandC.H.Muller,“Breakdownpointsofcauchyregression-scaleestimators,”Statistics&probabilityletters,vol.57,no.1,pp.79–89,Feb.2002.
Overcome the impact that noise brings, (4) formula becomes following form:
arg min Y l o g ( 1 + ( t r ( YLY T ) c 2 ) ) - - - ( 5 )
The c is here the coefficient for weighing noise.
Owing to there being Y=UTX relation, (7) formula abbreviation is:
arg min U l o g ( 1 + ( t r ( U T XLX T U ) c 2 ) ) - - - ( 6 )
Distant between the sample out of the space representation in low-dimensional just represents for each sampleFor enough far away of: the distance at the sample of each lower dimensional space and all sample classes center, be expressed as following object function:
arg m a x y i Σ i = 1 N | | y i - y ‾ | | 2 - - - ( 7 )
HereBe exactly all sample Lei centers,
For fear of the situation that over-fitting occurs, so added two norms, integrate so all above-mentioned situations, just writeBecome following object function:
arg m i n U l o g ( 1 + ( t r ( U T XLX T U ) c 2 ) ) - C 1 t r ( U T S U ) + C 2 | | U | | 2 - - - ( 8 )
The C here1And C2It is regularization coefficient.
In order to make (8) have unique solution, the given qualifications of applicant is UTU=I. Projection matrix U is by the method for iterationWith the method for solving solution of characteristic value out.
The 4th described step adopts SVMs (SVM) grader to classify.

Claims (4)

1. there is the scene Recognition method under the environment of making an uproar, it is characterized in that, comprise the following steps: 1) utilize sensor to obtain fieldScape image, the sample of intragenic marker; 2) coloured image to scene and depth image depthimage carry out feature extraction respectivelyAnd feature representation, merge coloured image feature and the depth image depthimage feature of same group; 3) select feature selecting to calculateMethod is obtained feature selecting model to the sample of mark by the feature that the 2nd step obtains; 4) adopt grader to classify.
2. a kind of scene Recognition method having under the environment of making an uproar according to claim 1, is characterized in that described step 1)To utilize Kinect sensor to obtain scene image.
3. a kind of scene Recognition method having under the environment of making an uproar according to claim 1, is characterized in that described step 2)Be specially, convert all images to gray-scale map, image is carried out to convergent-divergent, then to coloured image and depth image depthIn the localized mass of image, extract feature by yardstick invariant features conversion method, then adopt partial restriction uniform enconding algorithm to carry outFeature representation.
4. a kind of scene Recognition method having under the environment of making an uproar according to claim 1, is characterized in that described step 3)Feature selecting algorithm is stream shape Cauchy learning algorithm, and concrete steps are as follows:
For a given sample xiBelong to sample set X=[x1,x2,...xn]∈RD×N(N is number of samples here, and D isThe original dimension of sample, R represents at real number space), the low-dimensional sample y that it is correspondingiBelong to sample set Y=[y1,y2,...yn]∈Rd×N(d is the dimension after dimensionality reduction here), finds the similar and inhomogeneous sample of K arest neighborsWherein, there is k1Individually be and xiSimilar sample, remaining k2Individually be and xiInhomogeneous sample, wherein K=k1+k2, use respectivelyWithRepresent this two groups of samples; For whole xiLocalized mass be expressed as:
X i = [ x i , x i 1 , ... x i k 1 , x i 1 , ... , x i k 2 ] ∈ R D × ( k 1 + k 2 + 1 ) (whereinRepresent D × (k1+k2+ 1) the linearity sky of dimensionBetween), corresponding low-dimensional is expressed and is Y i = [ y i , y i 1 , ... , y i k 1 , y i 1 , ... , y i k 2 ] ∈ R d × ( k 1 + k 2 + 1 ) ; A low-dimensional part newly obtainingIn piece, reach between similar sample enough far away of the enough distance near and between inhomogeneity sample of distance, therefore above optimizationFunction representation is as follows:
m i n y i Σ j = 1 k 1 | | y i - y i j | | 2 - α Σ p = 1 k 2 | | y i - y i p | | 2 , - - - ( 1 )
α is scale factor, is used for controlling the impact of sample between the interior sample of class and class;
Define a coefficient vector ωi
Utilize the coefficient vector of definition, (1) formula will be become following form by abbreviation:
arg min Y i t r ( Y i L i Y i T ) , - - - ( 2 )
What tr () represented is mark computing, in formula L i = Σ j = 1 k 1 + k 2 ( ω i ) j - ω i T - ω i d i a g ( ω i ) .
Introduce selection matrix (S belowi)pq
( S i ) p q = 1 , I f p = F i { q } ; 0 , e l s e . - - - ( 3 )
S i ∈ R N × ( 1 + k 1 + k 2 ) ;
Therefore, obtain low-dimensional and express Yi=YSi, object function (2) is rewritten as:
arg min Y Σ i = 1 N t r ( YS i L i S i T Y T ) = arg min Y t r ( YLY T ) . - - - ( 4 )
Introduced Cauchy's estimation theory, overcome the impact that noise brings, (4) formula becomes following form:
arg m i n Y l o g ( 1 + ( t r ( YLY T ) c 2 ) ) - - - ( 5 )
C is the coefficient for weighing noise;
Owing to there being Y=UTX relation, (7) formula is reduced to:
arg m i n U l o g ( 1 + ( t r ( U T XLX T U ) c 2 ) ) - - - ( 6 )
Distant between the sample out of the space representation in low-dimensional is just expressed as: every for each sampleEnough far away of the distance at the sample of individual lower dimensional space and all sample classes center, is expressed as following object function:
arg m a x y i Σ i = 1 N | | y i - y ‾ | | 2 - - - ( 7 )
Be exactly all sample Lei centers,
For fear of the situation that over-fitting occurs, add two norms, integrate so all above-mentioned situations, just write as followingObject function:
arg m i n U l o g ( 1 + ( t r ( U T XLX T U ) c 2 ) ) - C 1 t r ( U T S U ) + C 2 | | U | | 2 - - - ( 8 )
The C here1And C2It is regularization coefficient;
In order to make (8) have unique solution, the given qualifications of applicant is UTU=I; Projection matrix U is by method and the feature of iterationThe method for solving solution of value out.
CN201610103825.5A 2016-02-25 2016-02-25 A kind of scene recognition method under noisy environment Active CN105590108B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610103825.5A CN105590108B (en) 2016-02-25 2016-02-25 A kind of scene recognition method under noisy environment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610103825.5A CN105590108B (en) 2016-02-25 2016-02-25 A kind of scene recognition method under noisy environment

Publications (2)

Publication Number Publication Date
CN105590108A true CN105590108A (en) 2016-05-18
CN105590108B CN105590108B (en) 2018-08-17

Family

ID=55929678

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610103825.5A Active CN105590108B (en) 2016-02-25 2016-02-25 A kind of scene recognition method under noisy environment

Country Status (1)

Country Link
CN (1) CN105590108B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050031034A1 (en) * 2003-06-25 2005-02-10 Nejat Kamaci Cauchy-distribution based coding system and method
CN102629330A (en) * 2012-02-29 2012-08-08 华南理工大学 Rapid and high-precision matching method of depth image and color image
CN102867191A (en) * 2012-09-04 2013-01-09 广东群兴玩具股份有限公司 Dimension reducing method based on manifold sub-space study
CN103500342A (en) * 2013-09-18 2014-01-08 华南理工大学 Human behavior recognition method based on accelerometer
CN104732209A (en) * 2015-03-17 2015-06-24 深圳先进技术研究院 Indoor scene recognition method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050031034A1 (en) * 2003-06-25 2005-02-10 Nejat Kamaci Cauchy-distribution based coding system and method
CN102629330A (en) * 2012-02-29 2012-08-08 华南理工大学 Rapid and high-precision matching method of depth image and color image
CN102867191A (en) * 2012-09-04 2013-01-09 广东群兴玩具股份有限公司 Dimension reducing method based on manifold sub-space study
CN103500342A (en) * 2013-09-18 2014-01-08 华南理工大学 Human behavior recognition method based on accelerometer
CN104732209A (en) * 2015-03-17 2015-06-24 深圳先进技术研究院 Indoor scene recognition method and device

Also Published As

Publication number Publication date
CN105590108B (en) 2018-08-17

Similar Documents

Publication Publication Date Title
CN106096557B (en) A kind of semi-supervised learning facial expression recognizing method based on fuzzy training sample
CN109753885B (en) Target detection method and device and pedestrian detection method and system
CN104599275B (en) The RGB-D scene understanding methods of imparametrization based on probability graph model
CN105956560A (en) Vehicle model identification method based on pooling multi-scale depth convolution characteristics
CN107766890B (en) Improved method for discriminant graph block learning in fine-grained identification
CN106126581A (en) Cartographical sketching image search method based on degree of depth study
CN107808129A (en) A kind of facial multi-characteristic points localization method based on single convolutional neural networks
CN107330355B (en) Deep pedestrian re-identification method based on positive sample balance constraint
CN105488536A (en) Agricultural pest image recognition method based on multi-feature deep learning technology
CN108038435A (en) A kind of feature extraction and method for tracking target based on convolutional neural networks
CN103268363B (en) A kind of Chinese calligraphy's image search method based on elastic HOG feature and DDTW coupling
CN109101981B (en) Loop detection method based on global image stripe code in streetscape scene
CN103699902A (en) Sorting method of ground-based visible light cloud picture
CN102622609B (en) Method for automatically classifying three-dimensional models based on support vector machine
CN108154158B (en) Building image segmentation method for augmented reality application
CN102122353A (en) Method for segmenting images by using increment dictionary learning and sparse representation
CN107944428A (en) A kind of indoor scene semanteme marking method based on super-pixel collection
CN108537145A (en) Human bodys' response method based on space-time skeleton character and depth belief network
CN107203745A (en) A kind of across visual angle action identification method based on cross-domain study
CN110517270B (en) Indoor scene semantic segmentation method based on super-pixel depth network
CN107767416A (en) The recognition methods of pedestrian's direction in a kind of low-resolution image
CN109934095A (en) A kind of remote sensing images Clean water withdraw method and system based on deep learning
CN105320963B (en) The semi-supervised feature selection approach of large scale towards high score remote sensing images
CN101986295A (en) Image clustering method based on manifold sparse coding
CN105574545B (en) The semantic cutting method of street environment image various visual angles and device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20210218

Address after: 650091 Yunnan province Kunming City Lake Road No. 2

Patentee after: YUNNAN University

Patentee after: Tao Dapeng

Address before: 650091 Yunnan province Kunming City Lake Road No. 2

Patentee before: YUNNAN University