CN109325434A - A kind of image scene classification method of the probability topic model of multiple features - Google Patents

A kind of image scene classification method of the probability topic model of multiple features Download PDF

Info

Publication number
CN109325434A
CN109325434A CN201811077631.8A CN201811077631A CN109325434A CN 109325434 A CN109325434 A CN 109325434A CN 201811077631 A CN201811077631 A CN 201811077631A CN 109325434 A CN109325434 A CN 109325434A
Authority
CN
China
Prior art keywords
image
distribution
visual
theme
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811077631.8A
Other languages
Chinese (zh)
Inventor
孙雪莹
陈锦言
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201811077631.8A priority Critical patent/CN109325434A/en
Publication of CN109325434A publication Critical patent/CN109325434A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/35Categorising the entire scene, e.g. birthday party or wedding scene
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/42Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
    • G06V10/422Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation for representing the structure of the pattern or shape of an object therefor
    • G06V10/424Syntactic representation, e.g. by using alphabets or grammars

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to a kind of image scene classification methods of the probability topic model of multiple features, comprising: training image is concentrated all image blocks, piecemeal size is 9 × 9, and local color, SIFT and the textural characteristics of image are extracted to each piece;The feature of extraction is indicated with vector, generates visual dictionary with K- mean algorithm;Quantify the visual word in visual dictionary, obtains the characteristic visual word distribution in training image collection;Using LDA model, the distribution and the distribution situation of each training image theme probability θ of each affiliated theme Z of visual word in training sample is obtained;Obtained visual theme is distributed input KNN-SVM classifier, optimizes parameter K, V and Z of KNN-SVM classifier;To the unfiled image that test image is concentrated, corresponding visual theme is obtained with LDA model learning and is distributed;The theme distribution that LDA is handled inputs KNN-SVM classifier.

Description

A kind of image scene classification method of the probability topic model of multiple features
Technical field
The present invention relates to a kind of Fusion of Color features, the image scene classification method of SIFT feature and textural characteristics, especially It is a kind of image scene classification method based on LDA probability topic model.
Background technique
Scene image classification is one of a basic problem and computer vision field that robotics research faces Vital task.In recent years, with the fast development of machine vision technique, numerous scene classification methods and technology are emerged, these It is very extensive that method is related to face.So-called scene image classification, refers to given image, the content for being included by observing it, And then judge the classification of its photographed scene.In robotics research field, in order to estimate robot in locating ring in real time Position and direction in border, it usually needs the system that can be completed at the same time map and positioning is established for it, and scene image divides Class is exactly the key link in the system development.In computer vision field, with the rapid development of internet multimedia technology, Emerge the complicated image data of magnanimity, in order to which effectively these data are analyzed and are managed, need be according to picture material It sticks semantic label, and scene image classification is by chance a kind of important channel for solving the problems, such as such.
Common scene can be roughly divided into 4 classes: natural scene, City scenarios, indoor scene and event scenarios.Due to difference The difference of scene constitution element is larger, classifying quality of the same classification method on different contextual data collection be frequently present of compared with Big difference, and this species diversity is especially pronounced between outdoor scene and indoor scene.In early stage, scene image classification is mainly adopted With the method based on low-level features and the method based on scene image structure;And in the later period, scene image classification mainly uses base In the method for visual vocabulary.Therefore, the research method of scene image classification can substantially be divided into three classes: the side based on low-level features Method, the method for method and view-based access control model vocabulary based on scene image structure.
In the history of scene classification, SIFT (scale-invariant feature transform) is a kind of comparison Popular iamge description.It is to identify the same target occurred in different images, it is for translation, scaling, rotation, light Certain stability can be kept according to situations such as even blocking, there is resolving ability powerful, outstanding.Textural characteristics are also a kind of Global characteristics, it also illustrates the surface nature of scenery corresponding to image or image-region but since texture is a kind of object The characteristic on surface can not reflect the essential attribute of object completely, so being that can not obtain high level just with textural characteristics The of secondary picture material is different from color characteristic, and textural characteristics are not based on the feature of pixel, it needs including multiple pixels Statistics is carried out in the region of point and calculates in pattern match, and this zonal feature has biggish superiority, will not be due to The deviation of part and can not successful match as a kind of statistical nature, textural characteristics are often with there is rotational invariance, and for making an uproar Sound has stronger resistivity
Summary of the invention
The object of the present invention is to provide the probability topic model of a kind of Fusion of Color feature, textural characteristics and SIFT feature into The classification method of row scene image, the present invention can significantly improve the effect of classification.Technical solution is as follows:
A kind of image scene classification method of the probability topic model of multiple features, including the following steps:
Step 1: the image that 2/3rds are randomly choosed in data set is used as training set;Training image is concentrated all Image block, piecemeal size are 9 × 9, and local color, SIFT and the textural characteristics of image are extracted to each piece;
Step 2: the feature of extraction being indicated with vector, generates visual dictionary with K- mean algorithm;
Step 3: the visual word in quantization visual dictionary obtains the characteristic visual word distribution in training image collection;
Step 4: applying LDA model, obtain the distribution of each affiliated theme Z of visual word and each training image master in training sample Inscribe the distribution situation of probability θ;
Step 5: by obtained visual theme be distributed input KNN-SVM classifier, using training image concentrate image into Row experiment, optimizes parameter K, V and Z of KNN-SVM classifier;
Step 6: the unfiled image concentrated to test image repeats (1) to (3) step, obtains the distribution of characteristic visual word, Theme distribution situation of the distribution as word each in test image for using the z that training image obtains in (4), to test image, Corresponding visual theme distribution is obtained with LDA model learning;
Step 7: the theme distribution that LDA is handled inputs KNN-SVM classifier.
Detailed description of the invention
Fig. 1 is hsv color space structure schematic diagram
The mentioned method flow diagram of Fig. 2
Specific embodiment
1. color feature extracted and expression
Using hsv color space, hsv color space is defined according to the visual perception of people, including tone, saturation degree With three color attributes of brightness, these three dimensions are irrelevant.In hsv color space the colour information of luminance component and image without It closes, and tone and saturation degree component meet people to the visual perception of color, more can express face from the vision system of people Color characteristic, this is two distinguishing features of HSV.These features make hsv color space be very suitable to description digital picture.HSV face It is independent mutually between three components of the colour space, different visual characteristics of human eyes is respectively indicated, therefore human eye can also be independently Perceive the variation of each color component, especially tone variations, the vision energy accurate judgement of human eye.In summary feature, in number In the application such as image procossing and image scene classification, hsv color model is more suitable for describing the content of image.
General digital picture is all made of RGB color model expression, therefore with hsv color model extraction and is indicating color spy When sign, needing to convert RGB color value to hsv color value is indicated, the formula mutually converted are as follows:
V=max (R, G, B)
It enables:
Then have:
H=H' × 60
Wherein, H ∈ [0 °, 360 °], S ∈ [0,1], V ∈ [0,1]
2. SIFT feature indicates
SIFT (Scale Invariant Feature Transform) feature, there is good adaptability and robustness, Good invariance can be kept to the scaling of image, rotation, translation and the transformation such as affine.It is the figure based on scale space As local feature representation method.The algorithm steps of SIFT feature are described in detail as follows:
Step 1: constructing and initializing scale space
The Analysis On Multi-scale Features of image data are simulated with Scale-space theory, and the scale space of a width two dimensional image can use height This convolution kernel realizes that change of scale formula is as follows:
L (x, y, σ)=G (x, y, σ) * I (x, y)
D (x, y, σ)=(G (x, y, k σ)-G (x, y, σ)) * I (x, y)=L (x, y, k σ)-L (x, y, σ)
Wherein, G (x, y, σ) indicates changeable scale Gaussian function.The space coordinate of (x, y) expression two dimensional image pixel.σ Indicating scale coordinate, its size embodies the smoothness of image, and scale is smaller, and the minutia of correspondence image is more obvious, Finer, resolution ratio is higher, and image is more smooth.
Step 2: detection DOG scale space extreme point
Each sampled point will be adjacent with it all the points (scale domain and image area) be compared, to find scale space Extreme point.In this layer of scale space all the points corresponding with the minimum that adjacent layer is got or maximum value, all as image at this Characteristic point under scale.Since first layer does not have upper layer field, the last layer does not have lower layer field, the first and last two of each group of image Layer cannot carry out extreme value comparison, and for the two special layers, this chapter finds image Min-max using difference of Gaussian algorithm Approximation characteristic point, to guarantee the continuity of scale space variation.
Step 3: the bad point of effect in removal characteristic point
DOG operator can generate larger fluctuation in edge, remove the characteristic point of low contrast and the biggish edge that floats is rung Ying Dian can obtain more accurate key point position and scale.
Step 4: the direction of distribution key point
The directioin parameter of each key point is calculated and is specified by the gradient direction distribution of key point neighborhood territory pixel, It can guarantee the rotational invariance of characteristic point operator in this way.
Specifically, indicating the gradient direction of the neighborhood territory pixel of key point with statistic histogram.The peak value of histogram represents The principal direction of neighborhood gradient at the key point, the peak value that main peak value 80% energy is approximately equal in histogram can be used as the pass The auxiliary direction of key point.Only one possible principal direction of one key point, it is also possible to have a principal direction and multiple auxiliary directions.Pass through After above 4 step calculates, each key point is there are three dimension, scale, position and direction where respectively indicating.And piece image SIFT feature size determined by the keypoint quantity chosen.In practical application, Lowe suggests (each using 4 × 4 seed points Seed point has 8 direction gradient values) each key point is described, and each key point includes scale, position and direction, most end form The SIFT feature vector tieed up at one 3 × 128.Using normalization and standardization, illumination variation can be eliminated to image Bring influences.So far, SIFT feature vector eliminate scaling, translation, rotation and illumination factor influence.
3. the extraction of textural characteristics
Texture is showed by pixel and its intensity profile of surrounding space neighborhood, i.e. local grain information.In addition, part The repeatability of texture information in varying degrees is exactly global texture information.While textural characteristics embody the property of global characteristics, It also illustrates the surface nature of scenery corresponding to image or image-region.But since texture is a kind of spy of body surface Property, it can not reflect the essential attribute of object completely, so just with textural characteristics can not being obtained in high level diagram picture Hold.Different from color characteristic, textural characteristics are not based on the feature of pixel, it is needed in the region comprising multiple pixels In carry out statistics calculating.In pattern match, this zonal feature have biggish superiority, will not due to part it is inclined Difference and can not successful match.
Using gray level co-occurrence matrixes method.Co-occurrence matrix is defined with the joint probability density of the pixel of two positions, it is not Only reflect the distribution character of brightness, also reflection has same brightness or close to the position distribution characteristic between the pixel of brightness, is The second-order statistics feature of related brightness of image variation.The gray level co-occurrence matrixes of piece image can reflect image grayscale about side To, adjacent spaces, the integrated information of amplitude of variation, it be analyze image local mode and their queueing disciplines basis.
It takes any point (x, y) in image (N × N) and deviates its another point (x+a, y+b), if the gray value of the point pair For (g1,g2).It enables point (x, y) move on entire picture, then can obtain various (g1,g2) value, if the series of gray value is k, then (g1,g2) combination share k2Kind.For entire picture, each (g is counted1,g2) value occur number, be then arranged in one A square matrix, then with (g1,g2) occur total degree by they be normalized to occur probability P (g1,g2), such square matrix is known as Gray level co-occurrence matrixes.
4. hidden Di Li Cray distribution (Latent Dirichlet Allocation, LDA) model scene classification
According to the color of extraction, SIFT and textural characteristics, then Fusion Features are quantized into visual word, being modeled with LDA will Image carries out dimensionality reduction, and the visual theme distribution for obtaining image indicates, uses KNN (K arest neighbors sorting algorithm (k- NearestNeighbor)-SVM classifier, K arest neighbors (kNN, k-NearestNeighbor) sorting algorithm, according to image Visual theme distribution probability carries out scene classification to test image.
Step 1: the image of random selection 2/3rds is as training set in data set.Training image is concentrated into institute There is image block, piecemeal size is 9 × 9, and local color, SIFT and the textural characteristics of image are extracted to each piece.
Step 2: the feature of extraction is indicated with vector, visual dictionary is generated with K- mean algorithm.
Step 3: the visual word in quantization visual dictionary, obtains the characteristic visual word distribution in training image collection.
Step 4: obtaining the distribution of each affiliated theme Z of visual word and each training image in training sample using LDA model The distribution situation of theme probability θ
Step 5: by obtained visual theme be distributed input KNN-SVM classifier, using training image concentrate image into Row experiment, optimizes parameter K, V and Z of KNN-SVM classifier.
Step 6: repeating (1) to (3) step to the unfiled image that test image is concentrated, point of characteristic visual word is obtained Cloth uses theme distribution situation of the distribution for the z that training image obtains in (4) as word each in test image, to test chart Picture obtains corresponding visual theme with LDA model learning and is distributed.
Step 7: the theme distribution that LDA is handled inputs KNN-SVM classifier.

Claims (1)

1. a kind of image scene classification method of the probability topic model of multiple features, including the following steps:
Step 1: the image that 2/3rds are randomly choosed in data set is used as training set;Training image is concentrated into all images Piecemeal, piecemeal size are 9 × 9, and local color, SIFT and the textural characteristics of image are extracted to each piece.
Step 2: the feature of extraction being indicated with vector, generates visual dictionary with K- mean algorithm;
Step 3: the visual word in quantization visual dictionary obtains the characteristic visual word distribution in training image collection;
Step 4: applying LDA model, it is general to obtain the distribution of each affiliated theme Z of visual word and each training image theme in training sample The distribution situation of rate θ;
Step 5: obtained visual theme being distributed input KNN-SVM classifier, is carried out using the image that training image is concentrated real It tests, optimizes parameter K, V and Z of KNN-SVM classifier;
Step 6: the unfiled image concentrated to test image repeats (1) to (3) step, obtains the distribution of characteristic visual word, uses (4) distribution for the z that training image obtains in uses LDA to test image as the theme distribution situation of word each in test image Model learning obtains corresponding visual theme distribution;
Step 7: the theme distribution that LDA is handled inputs KNN-SVM classifier.
CN201811077631.8A 2018-09-15 2018-09-15 A kind of image scene classification method of the probability topic model of multiple features Pending CN109325434A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811077631.8A CN109325434A (en) 2018-09-15 2018-09-15 A kind of image scene classification method of the probability topic model of multiple features

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811077631.8A CN109325434A (en) 2018-09-15 2018-09-15 A kind of image scene classification method of the probability topic model of multiple features

Publications (1)

Publication Number Publication Date
CN109325434A true CN109325434A (en) 2019-02-12

Family

ID=65265647

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811077631.8A Pending CN109325434A (en) 2018-09-15 2018-09-15 A kind of image scene classification method of the probability topic model of multiple features

Country Status (1)

Country Link
CN (1) CN109325434A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111724349A (en) * 2020-05-29 2020-09-29 同济大学 Image smudge recognition method based on HSV and SVM
CN111797875A (en) * 2019-04-09 2020-10-20 Oppo广东移动通信有限公司 Scene modeling method and device, storage medium and electronic equipment
CN113223668A (en) * 2021-04-15 2021-08-06 中南民族大学 Capsule endoscopy image redundant data screening method
CN113447771A (en) * 2021-06-09 2021-09-28 上海交通大学 Partial discharge pattern recognition method based on SIFT-LDA characteristics

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102622607A (en) * 2012-02-24 2012-08-01 河海大学 Remote sensing image classification method based on multi-feature fusion
US20120213426A1 (en) * 2011-02-22 2012-08-23 The Board Of Trustees Of The Leland Stanford Junior University Method for Implementing a High-Level Image Representation for Image Analysis
CN104268546A (en) * 2014-05-28 2015-01-07 苏州大学 Dynamic scene classification method based on topic model
CN106156798A (en) * 2016-07-25 2016-11-23 河海大学 Scene image classification method based on annular space pyramid and Multiple Kernel Learning
CN106250919A (en) * 2016-07-25 2016-12-21 河海大学 The scene image classification method that combination of multiple features based on spatial pyramid model is expressed
CN107273928A (en) * 2017-06-14 2017-10-20 上海海洋大学 A kind of remote sensing images automatic marking method based on weight Fusion Features
CN107644235A (en) * 2017-10-24 2018-01-30 广西师范大学 Image automatic annotation method based on semi-supervised learning

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120213426A1 (en) * 2011-02-22 2012-08-23 The Board Of Trustees Of The Leland Stanford Junior University Method for Implementing a High-Level Image Representation for Image Analysis
CN102622607A (en) * 2012-02-24 2012-08-01 河海大学 Remote sensing image classification method based on multi-feature fusion
CN104268546A (en) * 2014-05-28 2015-01-07 苏州大学 Dynamic scene classification method based on topic model
CN106156798A (en) * 2016-07-25 2016-11-23 河海大学 Scene image classification method based on annular space pyramid and Multiple Kernel Learning
CN106250919A (en) * 2016-07-25 2016-12-21 河海大学 The scene image classification method that combination of multiple features based on spatial pyramid model is expressed
CN107273928A (en) * 2017-06-14 2017-10-20 上海海洋大学 A kind of remote sensing images automatic marking method based on weight Fusion Features
CN107644235A (en) * 2017-10-24 2018-01-30 广西师范大学 Image automatic annotation method based on semi-supervised learning

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
何刚等: "兼顾特征级和决策级融合的场景分类", 《计算机应用》 *
孙伟等: "多特征融合的室内场景分类研究", 《广东工业大学学报》 *
曾培龙: "基于概率主题模型的图像场景分类研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *
陈红娟: "基于概率潜在语义分析的图像场景分类", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111797875A (en) * 2019-04-09 2020-10-20 Oppo广东移动通信有限公司 Scene modeling method and device, storage medium and electronic equipment
CN111797875B (en) * 2019-04-09 2023-12-01 Oppo广东移动通信有限公司 Scene modeling method and device, storage medium and electronic equipment
CN111724349A (en) * 2020-05-29 2020-09-29 同济大学 Image smudge recognition method based on HSV and SVM
CN111724349B (en) * 2020-05-29 2022-09-20 同济大学 Image smudge recognition method based on HSV and SVM
CN113223668A (en) * 2021-04-15 2021-08-06 中南民族大学 Capsule endoscopy image redundant data screening method
CN113447771A (en) * 2021-06-09 2021-09-28 上海交通大学 Partial discharge pattern recognition method based on SIFT-LDA characteristics

Similar Documents

Publication Publication Date Title
Cimpoi et al. Deep filter banks for texture recognition and segmentation
Zhu et al. Learning a discriminative model for the perception of realism in composite images
Cimpoi et al. Deep convolutional filter banks for texture recognition and segmentation
Kuo et al. Data-efficient graph embedding learning for PCB component detection
Narihira et al. Learning lightness from human judgement on relative reflectance
CN112686812B (en) Bank card inclination correction detection method and device, readable storage medium and terminal
CN109325434A (en) A kind of image scene classification method of the probability topic model of multiple features
CN107967482A (en) Icon-based programming method and device
CN109948566B (en) Double-flow face anti-fraud detection method based on weight fusion and feature selection
CN107784284B (en) Face recognition method and system
Bappy et al. Real estate image classification
CN110728238A (en) Personnel re-detection method of fusion type neural network
CN112418262A (en) Vehicle re-identification method, client and system
Smiatacz Normalization of face illumination using basic knowledge and information extracted from a single image
Selinger et al. Improving appearance-based object recognition in cluttered backgrounds
CN113011506B (en) Texture image classification method based on deep fractal spectrum network
CN108960285A (en) A kind of method of generating classification model, tongue body image classification method and device
Abraham Digital image forgery detection approaches: A review and analysis
CN108876849B (en) Deep learning target identification and positioning method based on auxiliary identification
CN111950565B (en) Abstract picture image direction identification method based on feature fusion and naive Bayes
Roy et al. Detection and classification of geometric shape objects for industrial applications
Prasomphan et al. Feature extraction for image matching in wat phra chetuphon wimonmangklararam balcony painting with sift algorithms
HS et al. A novel method to recognize object in Images using Convolution Neural Networks
Blanc-Talon et al. Advanced Concepts for Intelligent Vision Systems: 12th International Conference, ACIVS 2010, Sydney, Australia, December 13-16, 2010, Proceedings, Part I
Alyosef Large scale partial-and near-duplicate image retrieval using spatial information of local features

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20190212