CN107507611A - A kind of method and device of Classification of Speech identification - Google Patents
A kind of method and device of Classification of Speech identification Download PDFInfo
- Publication number
- CN107507611A CN107507611A CN201710774048.1A CN201710774048A CN107507611A CN 107507611 A CN107507611 A CN 107507611A CN 201710774048 A CN201710774048 A CN 201710774048A CN 107507611 A CN107507611 A CN 107507611A
- Authority
- CN
- China
- Prior art keywords
- sample set
- training sample
- laplce
- parameter
- optimal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/083—Recognition networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
- G10L2015/0631—Creating reference templates; Clustering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L2015/085—Methods for reducing search complexity, pruning
Abstract
The invention discloses a kind of Classification of Speech to know method for distinguishing, and speech data sample to be discriminated is inputted into the sorter model being pre-created, the classification results of the speech data sample are worth to according to the output of the sorter model;Wherein, the sorter model is based on using L1 normal forms regular parameter and the constraints of the determination of Laplce's regular parameter and SVMs, obtain supporting vector sample set, so that the sorter model obtained, with stronger openness and interpretation, and there is stronger filter capacity to noise, so as to have stronger robustness to noise, so as to obtain an accurate result to Classification of Speech.The present invention also provides a kind of device of Classification of Speech identification, has above-mentioned beneficial effect.
Description
Technical field
The present invention relates to artificial intelligence application field, more particularly to a kind of method and device of Classification of Speech identification.
Background technology
With the development of artificial intelligence, computer technology is widely used in every field, and wherein speech recognition is
Wherein there is very much one of direction of application value, and the conventional complexity all suitable for the treatment technology of voice and language, to meter
The computing of calculation machine brings certain burden.
Current relatively simple voice processing technology is to create generation model using semi-supervised algorithm, with the life of establishment
Voice is handled into model, but the generation model of currently acquired processing voice creates process complexity, and to noise
Filter capacity is not very strong, lacks stronger robustness.
The content of the invention
It is an object of the invention to provide a kind of Classification of Speech to know method for distinguishing, solves the generation model of processing voice to making an uproar
The problem of sound filter capacity is low, improve the accuracy of Classification of Speech identification.
It is a further object of the present invention to provide a kind of device of Classification of Speech identification.
In order to solve the above technical problems, the present invention, which provides a kind of Classification of Speech, knows method for distinguishing, this method includes:
Speech data sample to be discriminated is inputted into the sorter model being pre-created, according to the sorter model
Output is worth to the classification results of the speech data sample;Wherein, the sorter model establishment process is:
The training sample set of speech data is inputted into Laplce's SVMs;According to the training sample set and just
Determine parameter and kernel function, obtain optimal L1 normal form regular parameters γAWith optimal Laplce's regular parameter γIAnd core
Jacobian matrix;According to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter γI, the core
The constraints of function and SVMs, obtain supporting vector sample set and offset;According to the supporting vector sample
This set and the offset, obtain sorter model.
Wherein, the training sample set of speech data is inputted in the SVMs to Laplce to be included:
Training sample set is inputted into Laplce's SVMs:Wherein xi∈RD, yiIt is xiLabel,
Show xiClassification, work as yiDuring ∈ { -1 ,+1 }, It is the number for having label training sample, works as yiWhen=0,U is the number of no label training sample, and D is the dimension of luv space.
Wherein, it is described according to the training sample set and positive definite parameter and kernel function, obtain optimal L1 normal form canonicals
Parameter γAWith Laplce's regular parameter γIAnd kernel matrix includes:
The training sample set is divided into several pieces, passes through the training sample set pair L1 normal form regular parameters after division
γAWith Laplce's regular parameter γITested and trained in a manner of cross validation, obtain optimal L1 normal forms canonical ginseng
Number γAWith optimal Laplce's regular parameter γI;
The training sample set is mapped in core Hilbert space by kernel function, obtains kernel matrix K, wherein
Kij=k (xi,xj)。
Wherein, it is described according to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter
γI, the kernel function and SVMs constraints, obtaining supporting vector sample set and offset includes:
Under conditions of, solveObtain discrimination model
Coefficient a=a+-a-=[α1,α2,...,αl+u]TWith offset b=β+-β-, wherein,δ is constant coefficient, ξiFor slack variable,L=D-W is Laplacian Matrix,For parameter preset, and
Dii=∑jWij;
According to the coefficient a of the discrimination model, obtain training sample and concentrate supporting vector sample set SVs={ xi|αi≠
0, i=1 ..., N }.
Wherein, described according to the supporting vector sample set and the offset, obtaining sorter model includes:
According to the supporting vector sample set and the offset, the sorter model is determined:Wherein, x is speech data sample to be judged, wherein x ∈ RD, xsv
It is supporting vector, asvIt is the model coefficient of supporting vector, and y value is speech data sample x differentiation result.
The present invention also provides a kind of device of Classification of Speech identification, and the device includes:
Classifier modules, for inputting speech data sample to be discriminated into the sorter model being pre-created, according to
The output of the sorter model is worth to the classification results of the speech data sample;Wherein, the sorter model is point
Class device creation module, which creates, to be obtained, and the grader creation module is used for:
The training sample set of speech data is inputted into Laplce's SVMs;According to the training sample set and just
Determine parameter and kernel function, obtain optimal L1 normal form regular parameters γAWith optimal Laplce's regular parameter γIWith core letter
Matrix number;According to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter γI, the core letter
The constraints of number and SVMs, obtain supporting vector sample set and offset;According to the supporting vector sample
Set and the offset, obtain sorter model.
Wherein, the grader creation module includes:
Input block, for inputting training sample set into Laplce's SVMs:Wherein xi∈RD,
yiIt is xiLabel, show xiClassification, work as yiDuring ∈ { -1 ,+1 }, It is the number for having label training sample, works as yi
When=0,U is the number of no label training sample, and D is the dimension of luv space.
Wherein, the grader creation module includes:
Parameter processing unit, for the training sample set to be divided into several pieces, pass through the training sample set after division
To L1 normal form regular parameters γAWith Laplce's regular parameter γITested and trained in a manner of cross validation, obtained most
Excellent L1 normal form regular parameters γAWith Laplce's regular parameter γI;The training sample set is mapped to core by kernel function
In Hilbert space, kernel matrix K, wherein K are obtainedij=k (xi,xj)。
Wherein, the grader creation module includes:
Arithmetic element, forUnder conditions of, solveObtain discrimination model
Coefficient a=a+-a-=[α1,α2,...,αl+u]TWith offset b=β+-β-, wherein,δ is constant coefficient, ξiFor slack variable,L=D-W is Laplacian Matrix,For parameter preset, and
Dii=∑jWij;According to the coefficient a of the discrimination model, obtain training sample and concentrate supporting vector sample set SVs={ xi|αi
≠ 0, i=1 ..., N }.
Wherein, the grader creation module includes:
Obtaining unit, for according to the supporting vector sample set and the offset, determining the sorter model:Wherein, x is speech data sample to be judged, wherein x ∈ RD, xsv
It is supporting vector, asvIt is the model coefficient of supporting vector, and y value is speech data sample x differentiation result.
A kind of Classification of Speech provided by the present invention knows method for distinguishing, is input to by speech data sample to be discriminated pre-
In the sorter model first created, obtain and differentiate result, the sorter model is to use L1 normal forms regular parameter and Laplce
Regular parameter determines and the constraints of SVMs, obtains supporting vector sample set, so that the classification obtained
Device model, L1 normal forms regular parameter and Laplce's regular parameter cause the sorter model to have openness and interpretation,
In practice, it is possible to required model is just obtained using seldom sample point, this further enhances grader
To the filter capacity of noise, so as to obtain good robustness, the present invention compared to the prior art, classification of the present invention
Device model complexity is low, and with stronger interpretation and openness, on the basis of the discrimination of sorter model is improved,
Also there is stronger robustness to noise so that the result of Classification of Speech is more accurate.
The present invention also provides a kind of device of Classification of Speech identification, has above-mentioned beneficial effect.
Brief description of the drawings
, below will be to embodiment or existing for the clearer explanation embodiment of the present invention or the technical scheme of prior art
The required accompanying drawing used is briefly described in technology description, it should be apparent that, drawings in the following description are only this hair
Some bright embodiments, for those of ordinary skill in the art, on the premise of not paying creative work, can be with root
Other accompanying drawings are obtained according to these accompanying drawings.
Fig. 1 is that Classification of Speech provided by the invention identifies a kind of flow chart of embodiment;
Fig. 2 is the structured flowchart of the device of Classification of Speech provided in an embodiment of the present invention identification.
Embodiment
In order that those skilled in the art more fully understand the present invention program, with reference to the accompanying drawings and detailed description
The present invention is described in further detail.Obviously, described embodiment is only part of the embodiment of the present invention, rather than
Whole embodiments.Based on the embodiment in the present invention, those of ordinary skill in the art are not making creative work premise
Lower obtained every other embodiment, belongs to the scope of protection of the invention.
Classification of Speech provided by the invention identifies a kind of flow chart of embodiment as shown in figure 1, this method can be with
Including:
Step S101:The training sample set of speech data is inputted into Laplce's SVMs.
Step S102:According to the training sample set and positive definite parameter and kernel function, optimal L1 normal form canonicals are obtained
Parameter γAWith optimal Laplce's regular parameter γIAnd kernel matrix.
Step S103:According to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter
γI, the kernel function and SVMs constraints, obtain supporting vector sample set and offset.
Step S104:According to the supporting vector sample set and the offset, sorter model is obtained.
Step S105:Speech data sample to be discriminated is inputted into sorter model, according to the output of sorter model
It is worth to the classification results of speech data sample.
It should be noted that in the present embodiment, step S101 to step S104 is the process for creating sorter model,
Grader in the present invention can create according to step S101 to step S104 in advance, it is only necessary to defeated according to step S105
Enter speech data sample to be judged, you can obtain classification results.
In addition, generation model of the prior art needs to take the particular content of identification voice, it is necessary to existing including noise
Interior voice content is identified, and so as to cause robust type poor, and the sorter model used in the present invention is to be used to identify
The content of the voice whether be we want without go be concerned about the specific content of voice what is.
Cite a plain example, in sorter model is created, choose 150 project personnel's recording, everyone reads two
52 samples are obtained all over alphabet, wherein each sample dimension is 617.Differentiate for voice content and data set is divided:
Preceding 120 personnel are extracted to alphabetical a's and alphabetical b's pronunciation, obtain 480*617 data set, alphabetical a speech datas division
For positive class, word b speech datas are divided into negative class, mark 80% and are used as training set, remainder is as test set.Wherein training set
Part, mark off 10 data has label data the most, and remaining is as without label data.Trained point using training set
Class device model, input test collection obtain the accuracy based on training sample of the sorter model.
In actual application, the voice that we choose a, b does training sample and obtains that voice a, b can be made a distinction
Sorter model, when there is new phonetic entry, as long as judging whether voice is a or b, without concern for voice it is specific in
Hold.
The sorter model that the present invention uses in Classification of Speech identification, when having such demand in some scenarios, with regard to pole
The big technical difficulty that reduces also saves cost.Simplest application is that, when our voice answers, have fixation for some
The topic of answer, data under voice can be pre-processed and be judged to obtain result by the model again.Such as judge
Topic, in advance learner is trained to obtain corresponding model with " right, wrong " two kinds of voices, the speech data that then will be collected
Result is obtained using the model after pretreatment, and then result is exported and is compared with former topic answer.
In summary, the sorter model employed in the present invention is to have used L1 normal form regularizations so that creates classification
The object function of device model has openness and interpretation, also implies that and is obtained with using seldom training sample point
Required sorter model, so as to exclude noise data well so that sorter model has good robustness.
Based on above-described embodiment, in another specific embodiment of the invention, can include:
The training sample set of input speech data is specially in the SVMs to Laplce:
Training sample set is inputted into Laplce's SVMs:Wherein xi∈RD, yiIt is xiLabel,
Show xiClassification, work as yiDuring ∈ { -1 ,+1 }, It is the number for having label training sample, works as yiWhen=0,U is the number of no label training sample, and D is the dimension of luv space.
Based on above-described embodiment, in another specific embodiment of the invention, can include:
It is described according to the training sample set and positive definite parameter and kernel function, obtain optimal L1 normal form regular parameters
γAWith Laplce's regular parameter γIAnd kernel matrix is specially:
The training sample set is divided into several pieces, passes through the training sample set pair L1 normal form regular parameters after division
γAWith Laplce's regular parameter γITested and trained in a manner of cross validation, obtain optimal L1 normal forms canonical ginseng
Number γAWith optimal Laplce's regular parameter γI;The training sample set is mapped to core Hilbert sky by kernel function
Between in, obtain kernel matrix K, wherein Kij=k (xi,xj)。
Specifically, the cross-validation method in the present invention is to use k-fold cross-validation methods, L1 normal forms regular parameter is with drawing
This regular parameter of pula is obtained by cross validation.
For example, training set is divided into five equal portions, chooses a copy of it and be used as test, other are used as training, obtain five
Accuracy rate is averaging, and obtained Average Accuracy is exactly accurate corresponding to corresponding L1 normal forms regular parameter and Laplce's regular parameter
True rate, L1 normal forms regular parameter corresponding to maximum accuracy rate and Laplce's regular parameter are finally chosen as final parameter.
Based on above-described embodiment, in another specific embodiment of the invention, can include:
It is described according to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter γI, it is described
The constraints of kernel function and SVMs, obtains supporting vector sample set and offset is specially:
Under conditions of, solveObtain discrimination model
Coefficient a=a+-a-=[α1,α2,...,αl+u]TWith offset b=β+-β-, wherein,δ is constant coefficient, ξiFor slack variable,L=D-W is Laplacian Matrix,For parameter preset, and
Dii=∑jWij;
According to the coefficient a of the discrimination model, obtain training sample and concentrate supporting vector sample set SVs={ xi|αi≠
0, i=1 ..., N }.
Specifically, δ takes the normal number of a very little to ensure to get unique solution, it is constant coefficient.
Based on above-described embodiment, in another specific embodiment of the invention, can include:
It is described to be specially according to the supporting vector sample set and the offset, acquisition sorter model:
According to the supporting vector sample set and the offset, the sorter model is determined:Wherein, x is speech data sample to be judged, wherein x ∈ RD, xsv
It is supporting vector, asvIt is the model coefficient of supporting vector, and y value is speech data sample x differentiation result.
The device of Classification of Speech provided in an embodiment of the present invention identification is introduced below, Classification of Speech described below
The device of identification knows method for distinguishing with above-described Classification of Speech can be mutually to should refer to.
Fig. 2 is the structured flowchart of the device of Classification of Speech provided in an embodiment of the present invention identification, and the Classification of Speech of reference picture 2 is known
Other device can include:
Classifier modules 100, for inputting speech data sample to be discriminated, root into the sorter model being pre-created
The classification results of the speech data sample are worth to according to the output of the sorter model;Wherein, the sorter model is
Grader creation module 200, which creates, to be obtained, and the grader creation module 200 is used for:
The training sample set of speech data is inputted into Laplce's SVMs;According to the training sample set and just
Determine parameter and kernel function, obtain optimal L1 normal form regular parameters γAWith optimal Laplce's regular parameter γIWith core letter
Matrix number;According to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter γI, the core letter
The constraints of number and SVMs, obtain supporting vector sample set and offset;According to the supporting vector sample
Set and the offset, obtain sorter model.
Alternatively, the grader creation module 200 includes:
Input block, for inputting training sample set into Laplce's SVMs:Wherein xi∈RD,
yiIt is xiLabel, show xiClassification, work as yiDuring ∈ { -1 ,+1 }, It is the number for having label training sample, works as yi
When=0,U is the number of no label training sample, and D is the dimension of luv space.
Optionally, the grader creation module 200 includes:
Parameter processing unit, for the training sample set to be divided into several pieces, pass through the training sample set after division
To L1 normal form regular parameters γAWith Laplce's regular parameter γITested and trained in a manner of cross validation, obtained most
Excellent L1 normal form regular parameters γAWith Laplce's regular parameter γI;The training sample set is mapped to core by kernel function
In Hilbert space, kernel matrix K, wherein K are obtainedij=k (xi,xj)。
Optionally, the grader creation module 200 includes:
Arithmetic element, forUnder conditions of, solveObtain discrimination model
Coefficient a=a+-a-=[α1,α2,...,αl+u]TWith offset b=β+-β-, wherein,δ is constant coefficient, ξiFor slack variable,L=D-W is Laplacian Matrix,For parameter preset, and
Dii=∑jWij;According to the coefficient a of the discrimination model, obtain training sample and concentrate supporting vector sample set SVs={ xi|αi
≠ 0, i=1 ..., N }.
Optionally, the grader creation module 200 includes:
Obtaining unit, for according to the supporting vector sample set and the offset, determining the sorter model:Wherein, x is speech data sample to be judged, wherein x ∈ RD, xsv
It is supporting vector, asvIt is the model coefficient of supporting vector, and y value is speech data sample x differentiation result.
The device of the Classification of Speech identification of the present embodiment is used to realize that foregoing Classification of Speech knows method for distinguishing, therefore voice
The visible Classification of Speech hereinbefore of embodiment in the device of Classification and Identification knows the embodiment part of method for distinguishing, example
Such as, classifier modules 100, for realizing that above-mentioned Classification of Speech knows step S105 in method for distinguishing, grader creation module 200,
For realizing that above-mentioned Classification of Speech knows step S101, S102, S103 and S104 in method for distinguishing, so, its embodiment
The description of corresponding various pieces embodiment is referred to, will not be repeated here.
Each embodiment is described by the way of progressive in this specification, what each embodiment stressed be with it is other
The difference of embodiment, between each embodiment same or similar part mutually referring to.For dress disclosed in embodiment
For putting, because it is corresponded to the method disclosed in Example, so description is fairly simple, related part is referring to method part
Explanation.
Professional further appreciates that, with reference to the unit of each example of the embodiments described herein description
And algorithm steps, can be realized with electronic hardware, computer software or the combination of the two, in order to clearly demonstrate hardware and
The interchangeability of software, the composition and step of each example are generally described according to function in the above description.These
Function is performed with hardware or software mode actually, application-specific and design constraint depending on technical scheme.Specialty
Technical staff can realize described function using distinct methods to each specific application, but this realization should not
Think beyond the scope of this invention.
Directly it can be held with reference to the step of method or algorithm that the embodiments described herein describes with hardware, processor
Capable software module, or the two combination are implemented.Software module can be placed in random access memory (RAM), internal memory, read-only deposit
Reservoir (ROM), electrically programmable ROM, electrically erasable ROM, register, hard disk, moveable magnetic disc, CD-ROM or technology
In any other form of storage medium well known in field.
Method for distinguishing is known to Classification of Speech provided by the present invention above and device is described in detail.Herein should
The principle and embodiment of the present invention are set forth with specific case, the explanation of above example is only intended to help and managed
Solve the method and its core concept of the present invention.It should be pointed out that for those skilled in the art, do not departing from
On the premise of the principle of the invention, some improvement and modification can also be carried out to the present invention, these are improved and modification also falls into this hair
In bright scope of the claims.
Claims (10)
1. a kind of Classification of Speech knows method for distinguishing, it is characterised in that including:
Speech data sample to be discriminated is inputted into the sorter model being pre-created, according to the output of the sorter model
It is worth to the classification results of the speech data sample;
Wherein, the sorter model establishment process is:
The training sample set of speech data is inputted into Laplce's SVMs;
According to the training sample set and positive definite parameter and kernel function, optimal L1 normal form regular parameters γ is obtainedAWith it is optimal
Laplce's regular parameter γIAnd kernel matrix;
According to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter γI, the kernel function with
And the constraints of SVMs, obtain supporting vector sample set and offset;
According to the supporting vector sample set and the offset, the sorter model is obtained.
2. according to the method for claim 1, it is characterised in that voice number is inputted in the SVMs to Laplce
According to training sample set include:
Training sample set is inputted into Laplce's SVMs:Wherein xi∈RD, yiIt is xiLabel, show xi
Classification, work as yiDuring ∈ { -1 ,+1 }, i=1 ..., l, l are the numbers for having label training sample, work as yiWhen=0, i=l+
1 ..., u, u be no label training sample number, D is the dimension of luv space.
3. according to the method for claim 2, it is characterised in that it is described according to the training sample set and positive definite parameter and
Kernel function, obtain optimal L1 normal form regular parameters γAWith Laplce's regular parameter γIAnd kernel matrix includes:
The training sample set is divided into several pieces, passes through the training sample set pair L1 normal form regular parameters γ after divisionAAnd drawing
Pula this regular parameter γITested and trained in a manner of cross validation, obtain optimal L1 normal form regular parameters γAWith
Optimal Laplce's regular parameter γI;
The training sample set is mapped in core Hilbert space by kernel function, obtains kernel matrix K, wherein Kij=
k(xi,xj)。
4. according to the method for claim 3, it is characterised in that described according to the optimal L1 normal form regular parameters γA、
Optimal Laplce's regular parameter γI, the kernel function and SVMs constraints, obtain supporting vector
Sample set and offset include:
Under conditions of, solveObtain discrimination model
Coefficient a=a+-a-=[α1,α2,...,αl+u]TWith offset b=β+-β-, wherein, ξi>=0, i=1 ..., l,
J=1 ..., l+u, δ are constant coefficient, ξiFor slack variable,L=D-W is Laplacian Matrix,T > 0 are parameter preset, and Dii=∑jWij;
According to the coefficient a of the discrimination model, obtain training sample and concentrate supporting vector sample set SVs={ xi|αi≠ 0, i=
1,…,N}。
5. according to the method for claim 1, it is characterised in that described according to the supporting vector sample set and described inclined
Shifting amount, obtaining sorter model includes:
According to the supporting vector sample set and the offset, the sorter model is determined:Wherein, x is speech data sample to be judged, wherein x ∈ RD, xsv
It is supporting vector, asvIt is the model coefficient of supporting vector, and y value is speech data sample x differentiation result.
A kind of 6. device of Classification of Speech identification, it is characterised in that including:
Classifier modules, for inputting speech data sample to be discriminated into the sorter model being pre-created, according to described
The output of sorter model is worth to the classification results of the speech data sample;
Wherein, the sorter model is that grader creation module creates acquisition, and the grader creation module is used for:
The training sample set of speech data is inputted into Laplce's SVMs;
According to the training sample set and positive definite parameter and kernel function, optimal L1 normal form regular parameters γ is obtainedAWith it is optimal
Laplce's regular parameter γIAnd kernel matrix;
According to the optimal L1 normal form regular parameters γA, optimal Laplce's regular parameter γI, the kernel function with
And the constraints of SVMs, obtain supporting vector sample set and offset;
According to the supporting vector sample set and the offset, sorter model is obtained.
7. according to the method for claim 6, it is characterised in that the grader creation module includes:
Input block, for inputting training sample set into Laplce's SVMs:Wherein xi∈RD, yiIt is xi
Label, show xiClassification, work as yiDuring ∈ { -1 ,+1 }, i=1 ..., l, l are the numbers for having label training sample, work as yi=0
When, i=l+1 ..., u, u are the numbers of no label training sample, and D is the dimension of luv space.
8. device according to claim 7, it is characterised in that the grader creation module includes:
Parameter processing unit, for the training sample set to be divided into several pieces, pass through the training sample set pair L1 after division
Normal form regular parameter γAWith Laplce's regular parameter γITested and trained in a manner of cross validation, obtained optimal
L1 normal form regular parameters γAWith Laplce's regular parameter γI;The training sample set is mapped to core Xi Er by kernel function
In Bert space, kernel matrix K, wherein K are obtainedij=k (xi,xj)。
9. device according to claim 8, it is characterised in that the grader creation module includes:
Arithmetic element, forUnder conditions of, solveObtain discrimination model
Coefficient a=a+-a-=[α1,α2,...,αl+u]TWith offset b=β+-β-, wherein, ξi>=0, i=1 ..., l,
J=1 ..., l+u, δ are constant coefficient, ξiFor slack variable,I=1 ..., l, L=D-W La Pula
This matrix,T > 0 are parameter preset, and Dii=∑jWij;According to the coefficient a of the discrimination model, instructed
Practice supporting vector sample set SVs={ x in sample seti|αi≠ 0, i=1 ..., N }.
10. device according to claim 9, it is characterised in that the grader creation module includes:
Obtaining unit, for according to the supporting vector sample set and the offset, determining the sorter model:Wherein, x is speech data sample to be judged, wherein x ∈ RD, xsv
It is supporting vector, asvIt is the model coefficient of supporting vector, and y value is speech data sample x differentiation result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710774048.1A CN107507611B (en) | 2017-08-31 | 2017-08-31 | Voice classification recognition method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710774048.1A CN107507611B (en) | 2017-08-31 | 2017-08-31 | Voice classification recognition method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107507611A true CN107507611A (en) | 2017-12-22 |
CN107507611B CN107507611B (en) | 2021-08-24 |
Family
ID=60693417
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710774048.1A Active CN107507611B (en) | 2017-08-31 | 2017-08-31 | Voice classification recognition method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107507611B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109065027A (en) * | 2018-06-04 | 2018-12-21 | 平安科技(深圳)有限公司 | Speech differentiation model training method, device, computer equipment and storage medium |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1787075A (en) * | 2005-12-13 | 2006-06-14 | 浙江大学 | Method for distinguishing speek speek person by supporting vector machine model basedon inserted GMM core |
CN1975856A (en) * | 2006-10-30 | 2007-06-06 | 邹采荣 | Speech emotion identifying method based on supporting vector machine |
CN101640043A (en) * | 2009-09-01 | 2010-02-03 | 清华大学 | Speaker recognition method based on multi-coordinate sequence kernel and system thereof |
CN103258532A (en) * | 2012-11-28 | 2013-08-21 | 河海大学常州校区 | Method for recognizing Chinese speech emotions based on fuzzy support vector machine |
CN103605711A (en) * | 2013-11-12 | 2014-02-26 | 中国石油大学(北京) | Construction method and device, classification method and device of support vector machine |
US20160071010A1 (en) * | 2014-05-31 | 2016-03-10 | Huawei Technologies Co., Ltd. | Data Category Identification Method and Apparatus Based on Deep Neural Network |
-
2017
- 2017-08-31 CN CN201710774048.1A patent/CN107507611B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1787075A (en) * | 2005-12-13 | 2006-06-14 | 浙江大学 | Method for distinguishing speek speek person by supporting vector machine model basedon inserted GMM core |
CN1975856A (en) * | 2006-10-30 | 2007-06-06 | 邹采荣 | Speech emotion identifying method based on supporting vector machine |
CN101640043A (en) * | 2009-09-01 | 2010-02-03 | 清华大学 | Speaker recognition method based on multi-coordinate sequence kernel and system thereof |
CN103258532A (en) * | 2012-11-28 | 2013-08-21 | 河海大学常州校区 | Method for recognizing Chinese speech emotions based on fuzzy support vector machine |
CN103605711A (en) * | 2013-11-12 | 2014-02-26 | 中国石油大学(北京) | Construction method and device, classification method and device of support vector machine |
US20160071010A1 (en) * | 2014-05-31 | 2016-03-10 | Huawei Technologies Co., Ltd. | Data Category Identification Method and Apparatus Based on Deep Neural Network |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109065027A (en) * | 2018-06-04 | 2018-12-21 | 平安科技(深圳)有限公司 | Speech differentiation model training method, device, computer equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN107507611B (en) | 2021-08-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108182177A (en) | A kind of mathematics knowledge-ID automation mask method and device | |
CN112183099A (en) | Named entity identification method and system based on semi-supervised small sample extension | |
CN112508334A (en) | Personalized paper combining method and system integrating cognitive characteristics and test question text information | |
CN113111968B (en) | Image recognition model training method, device, electronic equipment and readable storage medium | |
EP3742328A1 (en) | Fatigue state detection method based on facial feature point positioning | |
CN114419736A (en) | Experiment scoring method, system, equipment and readable storage medium | |
CN109086794B (en) | Driving behavior pattern recognition method based on T-LDA topic model | |
CN105989341A (en) | Character recognition method and device | |
CN104347071B (en) | Method and system for generating reference answers of spoken language test | |
CN112926332A (en) | Entity relationship joint extraction method and device | |
CN107993636B (en) | Recursive neural network-based music score modeling and generating method | |
CN110458600A (en) | Portrait model training method, device, computer equipment and storage medium | |
CN112445831B (en) | Data labeling method and device | |
CN107507611A (en) | A kind of method and device of Classification of Speech identification | |
CN108229565A (en) | A kind of image understanding method based on cognition | |
Geertzen et al. | Measuring annotator agreement in a complex hierarchical dialogue act annotation scheme | |
CN109977400B (en) | Verification processing method and device, computer storage medium and terminal | |
CN113627159A (en) | Method, device, medium and product for determining training data of error correction model | |
CN111984790A (en) | Entity relation extraction method | |
CN110085260A (en) | A kind of single syllable stress identification bearing calibration, device, equipment and medium | |
CN115935071A (en) | Knowledge point recommendation method and device, storage medium and electronic equipment | |
CN115730038A (en) | Method and device for automatically generating test paper and examining test paper, electronic equipment and medium | |
CN115438645A (en) | Text data enhancement method and system for sequence labeling task | |
CN113918471A (en) | Test case processing method and device and computer readable storage medium | |
CN114239539A (en) | English composition off-topic detection method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |