CN104077382A - Method for improving GDM (Global Data Manager) feature selection of audio classifier - Google Patents
Method for improving GDM (Global Data Manager) feature selection of audio classifier Download PDFInfo
- Publication number
- CN104077382A CN104077382A CN201410298526.2A CN201410298526A CN104077382A CN 104077382 A CN104077382 A CN 104077382A CN 201410298526 A CN201410298526 A CN 201410298526A CN 104077382 A CN104077382 A CN 104077382A
- Authority
- CN
- China
- Prior art keywords
- feature
- sigma
- selecting
- degree
- separation
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/68—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/683—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
Landscapes
- Engineering & Computer Science (AREA)
- Library & Information Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses a method for improving GDM (Global Data Manager) feature selection of an audio classifier. The method comprises the following steps: training the Gaussian mixed model Gfc of each feature f specific to each audio type c, wherein c is between 1 and C; selecting a first feature f1, calculating the degree of separation between every two types specific to each feature, and selecting a first feature under which the average degree of separation among all the types is maximum; after the first feature f1 is selected, removing the feature f1 from a feature set to be selected, and finding out a type c1 and a type c2 between which the degree of separation is minimum in each type corresponding to f1; selecting a second feature f2, and selecting a feature f2 under which the degree of separation between the type c1 and the type c2 found in the third step is maximum; after the second feature f2 is selected, removing the feature f2 from the feature set to be selected, and constructing a feature vector by using the selected features f1 and f2 to obtain a final selected feature. Through a selected feature subset, types which are easily mixed up can be distinguished optimally, thereby increasing the overall classification accuracy of the classifier.
Description
Technical field
The invention belongs to audio feature extraction field, relate in particular to a kind of for improving the GDM feature selection approach of audio classifiers.
Background technology
Audio frequency characteristics is another key factor that affects audio classifiers performance.One section of raw audio streams itself is only that a kind of non-semantic symbol represents and non-structured binary stream, and except containing the limited information such as sample frequency, quantified precision and coding method, itself does not comprise clear and definite structural information and semantic information.
People's ear has extremely strong resolution characteristic, and a given section audio stream not only can be told the type of audio frequency immediately, can also tell the sound characteristic of the very difficult descriptions such as speaker's mood in audio frequency and the keynote of music (excited or constrain etc.).Make computing machine can possess the function of as people's ear, audio frequency being classified and being identified, first need from a series of two-value symbol, to change audio stream into the characteristic parameter that can reflect difference between different audio types, i.e. feature extraction.Feature extraction is the basis of various classification problems.
According to the character of particular problem and specific area, select to have the feature of obvious separating capacity, be a part very crucial in design category process.The in the situation that of limited training sample, we wish that the characteristic Design with the least possible has the sorter of good versatility.
The average dispersion degree of the algorithm that traditional characteristic is selected based on making all categories maximum (GMM based Mean Separability Maximization is called for short GMSM) criterion, the performance of this feature selecting algorithm is easy to be subject to easily sub-category impact.And in fact, the performance of multiclass audio classifiers is except being subject to easily sub-category impact, it is more the impact that is subject to easily to obscure classification, improve the performance of sorter, improving the nicety of grading of easily obscuring between classification is the key point of problem, therefore,, during feature selecting, should choose those and make easily to obscure the feature that classification is more easily distinguished.
Summary of the invention
The object of the present invention is to provide a kind ofly for improving the GDM feature selection approach of audio classifiers, be intended to improve and easily obscure the nicety of grading between classification.
The present invention is achieved in that a kind of as follows for improving the concrete steps of GDM feature selection approach of audio classifiers:
Step 1, training pattern, to each audio types c, c ∈ [1, C], trains the gauss hybrid models G of every kind of feature f
fc;
Step 2, select first feature f
1, to each feature, calculate the degree of separation between every two classifications, select the satisfied average divided degree maximum making between all categories of first feature, that is:
Step 3, selecting first feature f
1after, from characteristic set to be selected, remove feature f
1, and find out f
1two class c of degree of separation minimum in each corresponding classification
1and c
2, that is:
Step 4, second feature f of selection
2, selection makes two class c that find out in step 3
1and c
2the feature f of degree of separation maximum
2, that is:
Step 5, selecting second feature f
2after, from characteristic set to be selected, remove feature f
2, and with the feature f selecting
1and f
2form an eigenvector, iterative step three-step 4, when iteration, replaces respectively two formula in step 3 and step 4 with formula below, that is:
Wherein, l represents the number of times of iteration;
Judge whether to meet iteration cut-off condition, if l<L returns to step 3, otherwise, stop iteration, what obtain selecting is characterized as: f
1, f
2..., f
l.
Further, for improving the problem of the GDM feature selection approach of audio classifiers, be described below:
Suppose to have C audio types, F kind feature, therefrom select L subcharacter, first, for each feature f, trains the gauss hybrid models GMM of each classification
fc, c ∈ [1, C], the probability density function that c class Gaussian Mixture distributes is:
Wherein, K represents the number of mixed components, Θ
c=(π
1..., π
k, θ
1..., θ
k), the parameter of expression model, π
i, the weight of i mixed components of expression, meets constraint condition:
θ
i={ μ
i, Σ
i, the parameter of i mixed components of expression;
P (X| θ
i) be each gaussian component, its expression-form is as follows:
Wherein, μ
ifor the mean value vector of D dimension, represent the average of gaussian component;
Σ
icovariance matrix for D * D;
The degree of separation (Separability) defining between two classifications is:
S
f(GMM
fk,GMM
fl)=dis(GMM
fk,GMM
fl)
Dis () represents the distance between two gauss hybrid models, adopts improved symmetry distance tolerance K-L2 distance, and computing formula is as follows:
effect gathers
Of the present inventionly for improving character subset that the GDM feature selection approach of audio classifiers selects, can make the class discrimination the most easily obscured best, can improve the nicety of grading of sorter integral body.
Accompanying drawing explanation
Fig. 1 be the embodiment of the present invention provide for improving the GDM feature selection approach process flow diagram of audio classifiers.
Embodiment
In order to make object of the present invention, technical scheme and advantage clearer, below in conjunction with drawings and Examples, the present invention is further elaborated.Should be appreciated that specific embodiment described herein, only in order to explain the present invention, is not intended to limit the present invention.
Of the present inventionly for improving the problem of the GDM feature selection approach of audio classifiers, be described below:
Suppose to have C audio types, F kind feature, therefrom select L subcharacter, first, for each feature f, trains the gauss hybrid models GMM of each classification
fc, c ∈ [1, C], the probability density function that c class Gaussian Mixture distributes is:
Wherein, K represents the number of mixed components, Θ
c=(π
1..., π
k, θ
1..., θ
k), the parameter of expression model, π
i, the weight of i mixed components of expression, meets constraint condition:
θ
i={ μ
i, Σ
i, the parameter of i mixed components of expression;
P (X| θ
i) be each gaussian component, its expression-form is as follows:
Wherein, μ
ifor the mean value vector of D dimension, represent the average of gaussian component;
Σ
icovariance matrix for D * D, has various ways, can be complete matrix, Block diagonal matrix or diagonal matrix.For ease of computing, conventionally suppose between each dimensional feature separate, i.e. Σ
iit is diagonal matrix.
The degree of separation (Separability) defining between two classifications is:
S
f(GMM
fk,GMM
fl)=dis(GMM
fk,GMM
fl)
Dis () represents the distance between two gauss hybrid models, and its measurement criterion has a variety of.As shown in table 1, A and B represent respectively two gauss hybrid models, and d (i, j) represents the distance between two gaussian component, and its conventional tolerance has Euclidean distance, mahalanobis distance, K-L distance and Bhattachyaryya distance etc.Wherein, between gauss hybrid models, in class divergence range formula, dis (A, B) can be any one in these four kinds of distances.
Table 1 common distance measurement criterion
In above table, K-L distance is asymmetrical, and for ease of calculating, the present embodiment adopts improved symmetry distance tolerance K-L2 distance, and computing formula is as follows:
By formula S
f(GMM
fk, GMM
fl)=dis (GMM
fk, GMM
fl) computing formula of degree of separation can find out, from two gauss hybrid models away from must be more, its degree of separation is larger, corresponding, and two classifications are also more easily distinguished.For easy differentiation classification, its classification accuracy is higher, therefore, when feature selecting, needn't give too many consideration.On the contrary, two its degree of separation of gauss hybrid models close to must be are more less, mutually deserved, and two such classifications are more easily obscured.For the classification of easily obscuring, its classification accuracy is very low, causes declining to a great extent of whole classifier performance.Therefore, during feature selecting, should choose those and can make easily to obscure the feature that classification degree of separation is large, so just can make final sorter improve the classification accuracy of all classifications.In other words, when selecting feature, first to find those classifications of easily obscuring, then select to make these easily to obscure the feature that classification is easily distinguished.
Fig. 1 shows of the present invention for improving the flow process of the GDM feature selection approach of audio classifiers, and as shown in the figure, the present invention is achieved in that a kind of as follows for improving the concrete steps of GDM feature selection approach of audio classifiers:
S101: training pattern, to each audio types c, c ∈ [1, C], trains the gauss hybrid models G of every kind of feature f
fc;
S102: select first feature f
1, to each feature, calculate the degree of separation between every two classifications, select the satisfied average divided degree maximum making between all categories of first feature, that is:
S103: selecting first feature f
1after, from characteristic set to be selected, remove feature f
1, and find out f
1two class c of degree of separation minimum in each corresponding classification
1and c
2, that is:
S104: select second feature f
2, selection makes two class c that find out in step 3
1and c
2the feature f of degree of separation maximum
2, that is:
S105: selecting second feature f
2after, from characteristic set to be selected, remove feature f
2, and with the feature f selecting
1and f
2form an eigenvector, iterative step three-step 4, when iteration, replaces respectively two formula in step 3 and step 4 with formula below, that is:
Wherein, l represents the number of times of iteration;
Judge whether to meet iteration cut-off condition, if l<L returns to step 3, otherwise, stop iteration, what obtain selecting is characterized as: f
1, f
2..., f
l.
Although above-mentioned, by reference to the accompanying drawings the specific embodiment of the present invention is described; but be not limiting the scope of the invention; one of ordinary skill in the art should be understood that; on the basis of technical scheme of the present invention, those skilled in the art do not need to pay various modifications that performing creative labour can make or distortion still within protection scope of the present invention.
Claims (2)
1. for improving a GDM feature selection approach for audio classifiers, it is characterized in that, described as follows for improving the concrete steps of GDM feature selection approach of audio classifiers:
Step 1, training pattern, to each audio types c, c ∈ [1, C], trains the gauss hybrid models G of every kind of feature f
fc;
Step 2, select first feature f1, to each feature, calculate the degree of separation between every two classifications, the average divided degree of selecting first feature to meet to make between all categories is maximum, that is:
Step 3, selecting first feature f
1after, from characteristic set to be selected, remove feature f
1, and find out f
1two class c of degree of separation minimum in each corresponding classification
1and c
2, that is:
Step 4, second feature f of selection
2, selection makes two class c that find out in step 3
1and c
2the feature f of degree of separation maximum
2, that is:
Step 5, selecting second feature f
2after, from characteristic set to be selected, remove feature f
2, and with the feature f selecting
1and f
2form an eigenvector, iterative step three-step 4, when iteration, replaces respectively two formula in step 3 and step 4 with formula below, that is:
Wherein, l represents the number of times of iteration;
Judge whether to meet iteration cut-off condition, if l<L returns to step 3, otherwise, stop iteration, what obtain selecting is characterized as: f
1, f
2..., f
l.
2. as claimed in claim 1ly for improving the GDM feature selection approach of audio classifiers, it is characterized in that, for improving the problem of the GDM feature selection approach of audio classifiers, be described below:
Suppose to have C audio types, F kind feature, therefrom select L subcharacter, first, for each feature f, trains the gauss hybrid models GMM of each classification
fc, c ∈ [1, C], the probability density function that c class Gaussian Mixture distributes is:
Wherein, K represents the number of mixed components, Θ
c=(π
1..., π
k, θ
1..., θ
k), the parameter of expression model, π
i, the weight of i mixed components of expression, meets constraint condition:
θ
i={ μ
i, Σ
i, the parameter of i mixed components of expression;
P (X| θ
i) be each gaussian component, its expression-form is as follows:
Wherein, μ
ifor the mean value vector of D dimension, represent the average of gaussian component;
Σ
icovariance matrix for D * D;
The degree of separation (Separability) defining between two classifications is:
S
f(GMM
fk,GMM
fl)=dis(GMM
fk,GMM
fl)
Dis () represents the distance between two gauss hybrid models, adopts improved symmetry distance tolerance K-L2 distance, and computing formula is as follows:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410298526.2A CN104077382A (en) | 2014-06-27 | 2014-06-27 | Method for improving GDM (Global Data Manager) feature selection of audio classifier |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410298526.2A CN104077382A (en) | 2014-06-27 | 2014-06-27 | Method for improving GDM (Global Data Manager) feature selection of audio classifier |
Publications (1)
Publication Number | Publication Date |
---|---|
CN104077382A true CN104077382A (en) | 2014-10-01 |
Family
ID=51598636
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410298526.2A Pending CN104077382A (en) | 2014-06-27 | 2014-06-27 | Method for improving GDM (Global Data Manager) feature selection of audio classifier |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104077382A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111554273A (en) * | 2020-04-28 | 2020-08-18 | 华南理工大学 | Method for selecting amplified corpora in voice keyword recognition |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020037083A1 (en) * | 2000-07-14 | 2002-03-28 | Weare Christopher B. | System and methods for providing automatic classification of media entities according to tempo properties |
US7065416B2 (en) * | 2001-08-29 | 2006-06-20 | Microsoft Corporation | System and methods for providing automatic classification of media entities according to melodic movement properties |
CN102129456A (en) * | 2011-03-09 | 2011-07-20 | 天津大学 | Method for monitoring and automatically classifying music factions based on decorrelation sparse mapping |
-
2014
- 2014-06-27 CN CN201410298526.2A patent/CN104077382A/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020037083A1 (en) * | 2000-07-14 | 2002-03-28 | Weare Christopher B. | System and methods for providing automatic classification of media entities according to tempo properties |
US7065416B2 (en) * | 2001-08-29 | 2006-06-20 | Microsoft Corporation | System and methods for providing automatic classification of media entities according to melodic movement properties |
CN102129456A (en) * | 2011-03-09 | 2011-07-20 | 天津大学 | Method for monitoring and automatically classifying music factions based on decorrelation sparse mapping |
Non-Patent Citations (1)
Title |
---|
王荣燕: "复杂音频分类中的关键问题研究", 《中国博士学位论文全文数据库信息科技辑》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111554273A (en) * | 2020-04-28 | 2020-08-18 | 华南理工大学 | Method for selecting amplified corpora in voice keyword recognition |
CN111554273B (en) * | 2020-04-28 | 2023-02-10 | 华南理工大学 | Method for selecting amplified corpora in voice keyword recognition |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108932950B (en) | Sound scene identification method based on label amplification and multi-spectral diagram fusion | |
CN106570178B (en) | High-dimensional text data feature selection method based on graph clustering | |
CN101710490B (en) | Method and device for compensating noise for voice assessment | |
CN103229233B (en) | For identifying the modelling apparatus of speaker and method and Speaker Recognition System | |
CN104091321A (en) | Multi-level-point-set characteristic extraction method applicable to ground laser radar point cloud classification | |
CN104795064B (en) | The recognition methods of sound event under low signal-to-noise ratio sound field scape | |
CN104167208A (en) | Speaker recognition method and device | |
CN103605990B (en) | Integrated multi-classifier fusion classification method and integrated multi-classifier fusion classification system based on graph clustering label propagation | |
CN103745200A (en) | Facial image identification method based on word bag model | |
CN105304078A (en) | Target sound data training device and target sound data training method | |
CN103474061A (en) | Automatic distinguishing method based on integration of classifier for Chinese dialects | |
CN103474072A (en) | Rapid anti-noise twitter identification method by utilizing textural features and random forest (RF) | |
CN103035239B (en) | Speaker recognition method based on partial learning | |
CN104077598A (en) | Emotion recognition method based on speech fuzzy clustering | |
CN111091809B (en) | Regional accent recognition method and device based on depth feature fusion | |
CN104008177A (en) | Method and system for rule base structure optimization and generation facing image semantic annotation | |
Moh et al. | Towards domain independent speaker clustering | |
CN113366500A (en) | Classification result verification method and classification result learning method using verification neural network, and computing device executing the methods | |
Marques et al. | Additional evidence that common low-level features of individual audio frames are not representative of music genres | |
Qin et al. | Graph convolution-based deep clustering for speech separation | |
CN111340057A (en) | Classification model training method and device | |
CN104077382A (en) | Method for improving GDM (Global Data Manager) feature selection of audio classifier | |
CN103871413A (en) | Men and women speaking voice classification method based on SVM and HMM mixing model | |
CN107533671A (en) | Pattern recognition device, mode identification method and program | |
CN109753999A (en) | For the fine granularity model recognizing method of visual angle auto graph |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20141001 |