CN109785863A - A kind of speech-emotion recognition method and system of deepness belief network - Google Patents
A kind of speech-emotion recognition method and system of deepness belief network Download PDFInfo
- Publication number
- CN109785863A CN109785863A CN201910173690.3A CN201910173690A CN109785863A CN 109785863 A CN109785863 A CN 109785863A CN 201910173690 A CN201910173690 A CN 201910173690A CN 109785863 A CN109785863 A CN 109785863A
- Authority
- CN
- China
- Prior art keywords
- speech
- voice signal
- obtains
- belief network
- emotion recognition
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The present invention discloses the speech-emotion recognition method and system of a kind of deepness belief network.The recognition methods includes: acquisition voice signal;The voice signal is pre-processed, pretreatment voice signal is obtained;Unsupervised speech recognition is carried out using deepness belief network to the pretreatment voice signal, obtains phonic signal character;The identification classification that the phonic signal character is carried out to speech emotional using support vector machines, obtains speech emotion recognition result.Using the multi-categorizer model based on deepness belief network and limitation Boltzmann machine, the multi-classifier system of a speech emotion recognition is established, the discrimination of speech emotional is improved.
Description
Technical field
The present invention relates to field of speech recognition, a kind of speech-emotion recognition method more particularly to deepness belief network and
System.
Background technique
With cloud computing, the development of mobile Internet, big data, machine is that the mankind service intelligence further, people and machine
For the dream to be engaged in the dialogue with natural language gradually close to realizing, requirement of the people to machine interaction capabilities is also higher and higher.Simply
The identification of voice content be no longer satisfied the requirement of people, handle, identify and understand the emotion in voice in practical application
In become particularly important.Language emotion recognition has boundless application prospect, it can be applied not only to man-machine friendship
Mutual system, can be also used for speech recognition, enhance the robustness of speech recognition;Or it is used for speaker identification, improve speaker
Discrimination rate.Speech emotion recognition technology intelligent human-machine interaction, human-computer interaction teaching, be widely used.Automatic language feelings
The other research of perception, be not only able to push computer technology further development, it also by substantially increase people work and
Learning efficiency improves people's lives quality.
Extraneous various emotion signals are sampled to identify various emotions, in terms of deep neural network research, for
The accuracy of emotional semantic classification is low, in terms of pattern-recognition, using the feelings in the prior art based in neural network extraction voice
Sense, it is lower for the discrimination of sad, excited, happy and angry emotion, using adaptive neural network to speech emotional state
Discrimination it is relatively low.
Using traditional neural network in training, each layer of network is trained together as a whole, when facing big data
When situation, the training time of network just will increase, the convergence rate of network is made to become slower.Back-propagation algorithm is neural network
Most commonly used method in training trains entire neural network by the method for iteration, and network parameter is using the side being randomized
Formula is initialized, and adjusts net using the difference of the actual value of the output valve and data that currently calculate network top obtained
The parameter of each layer of network, using traditional gradient descent method, the target of undated parameter be so that neural network forecast value and true value more
It is close, still, come initialization network parameter by the way of random initializtion, it will lead to error correction more down when network updates
Signal is weaker, and gradient also becomes more sparse, so that network is easily trapped into local optimum.So leading to the knowledge of speech emotional state
Not rate is low.
Summary of the invention
The object of the present invention is to provide a kind of speech emotionals of deepness belief network that can be improved speech emotion recognition rate
Recognition methods and system.
To achieve the above object, the present invention provides following schemes:
A kind of speech-emotion recognition method of deepness belief network, which is characterized in that the recognition methods includes:
Obtain voice signal;
The voice signal is pre-processed, pretreatment voice signal is obtained;
Unsupervised speech recognition is carried out using deepness belief network to the pretreatment voice signal, is obtained
Phonic signal character;
The identification classification that the phonic signal character is carried out to speech emotional using support vector machines, obtains speech emotional and knows
Other result.
Optionally, described that unsupervised voice signal spy is carried out using deepness belief network to the pretreatment voice signal
Sign is extracted, and is obtained phonic signal character and is specifically included:
Low layer to high-rise N layer limitation Boltzmann machine is stacked, deepness belief network is obtained;
Limitation Boltzmann machine according to the pretreatment voice signal to i-th layer carries out unsupervised training, obtains i-th most
Excellent parameter, the optimized parameter for the limitation Boltzmann machine that i-th optimized parameter is described i-th layer;Wherein, the value of i is successively
It is 1,2 ..., N;
The limitation Boltzmann machine of i+1 layer is carried out according to i-th optimized parameter and the pretreatment voice signal
Unsupervised training obtains i+1 optimized parameter;
The multiple optimized parameter is finely tuned using the method for overall situation training to the deepness belief network and converges to the overall situation
It is optimal, obtain multiple fine tuning optimized parameters;
The phonic signal character of the pretreatment voice signal is extracted according to the fine tuning optimized parameter.
Optionally, described that the phonic signal character is classified using the identification that support vector machines carries out speech emotional, it obtains
Speech emotion recognition result is obtained to specifically include:
The sample point of the phonic signal character is mapped to by high-dimensional feature space using kernel function, obtaining spatial linear can
The sample divided;
The sample that the support vector machines can divide according to the spatial linear carries out logic to the phonic signal character and sentences
It is disconnected, obtain speech emotion recognition result.
A kind of speech emotion recognition system of deepness belief network, the identifying system include:
Voice signal obtains module, for obtaining voice signal;
Speech signal pre-processing module obtains pretreatment voice signal for pre-processing the voice signal;
Characteristic extracting module, for carrying out unsupervised voice using deepness belief network to the pretreatment voice signal
Signal characteristic abstraction obtains phonic signal character;
Emotion recognition module, for the phonic signal character to be carried out to the identification point of speech emotional using support vector machines
Class obtains speech emotion recognition result.
Optionally, the characteristic extracting module specifically includes:
Deepness belief network establishes unit, for stacking low layer to high-rise N layer limitation Boltzmann machine, obtains depth
Belief network;
Supervised training unit carries out nothing for the limitation Boltzmann machine according to the pretreatment voice signal to i-th layer
Supervised training obtains the i-th optimized parameter, the optimal ginseng for the limitation Boltzmann machine that i-th optimized parameter is described i-th layer
Number;Wherein, the value of i is followed successively by 1,2 ..., N;According to i-th optimized parameter and the pretreatment voice signal to
I+1 layers of limitation Boltzmann machine carries out unsupervised training, obtains i+1 optimized parameter;
Small parameter perturbations unit is believed for being finely tuned the multiple optimized parameter using the method for overall situation training to the depth
Network convergence is read to global optimum, obtains multiple fine tuning optimized parameters;
Speech recognition unit, for extracting the pretreatment voice signal according to the fine tuning optimized parameter
Phonic signal character.
Optionally, the emotion recognition module specifically includes:
Kernel function unit, for the sample point of the phonic signal character to be mapped to high dimensional feature sky using kernel function
Between, obtain the sample that spatial linear can divide;
Logic judgment unit believes the voice according to the sample that the spatial linear can divide for the support vector machines
Number feature carries out logic judgment, obtains speech emotion recognition result.
The specific embodiment provided according to the present invention, the invention discloses following technical effects: the invention discloses one kind
The speech-emotion recognition method and system of deepness belief network.The recognition methods includes: acquisition voice signal;Described in pretreatment
Voice signal obtains pretreatment voice signal;Deepness belief network is used to carry out the pretreatment voice signal unsupervised
Speech recognition obtains phonic signal character;The phonic signal character is subjected to voice feelings using support vector machines
The identification of sense is classified, and speech emotion recognition result is obtained.Each limitation Bohr is successively trained hereby using the deepness belief network
The mode of graceful machine entirely trains the entire deepness belief network to reach training, using based on the deepness belief network and institute
The multi-categorizer model for stating limitation Boltzmann machine, establishes the multi-classifier system of a speech emotion recognition, improves language
The discrimination of sound emotion.
Detailed description of the invention
It in order to more clearly explain the embodiment of the invention or the technical proposal in the existing technology, below will be to institute in embodiment
Attached drawing to be used is needed to be briefly described, it should be apparent that, the accompanying drawings in the following description is only some implementations of the invention
Example, for those of ordinary skill in the art, without any creative labor, can also be according to these attached drawings
Obtain other attached drawings.
Fig. 1 is the flow chart of the speech-emotion recognition method of deepness belief network provided by the invention;
Fig. 2 is the structure composition figure of the speech emotion recognition system of deepness belief network provided by the invention;
Fig. 3 is the feelings identifying system block diagram provided by the invention based on support vector machines.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete
Site preparation description, it is clear that described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.It is based on
Embodiment in the present invention, it is obtained by those of ordinary skill in the art without making creative efforts every other
Embodiment shall fall within the protection scope of the present invention.
The object of the present invention is to provide a kind of speech emotionals of deepness belief network that can be improved speech emotion recognition rate
Recognition methods and system.
In order to make the foregoing objectives, features and advantages of the present invention clearer and more comprehensible, with reference to the accompanying drawing and specific real
Applying mode, the present invention is described in further detail.
As shown in Figure 1, a kind of speech-emotion recognition method of deepness belief network, which is characterized in that the recognition methods
Include:
Step 100: obtaining voice signal;
Step 200: pre-processing the voice signal, obtain pretreatment voice signal;
Step 300: unsupervised phonic signal character is carried out using deepness belief network to the pretreatment voice signal
It extracts, obtains phonic signal character;
Step 400: the phonic signal character being classified using the identification that support vector machines carries out speech emotional, obtains language
Sound emotion recognition result.
The step 300: unsupervised voice signal is carried out using deepness belief network to the pretreatment voice signal
Feature extraction obtains phonic signal character and specifically includes:
Low layer to high-rise N layer limitation Boltzmann machine is stacked, deepness belief network is obtained;
Limitation Boltzmann machine according to the pretreatment voice signal to i-th layer carries out unsupervised training, obtains i-th most
Excellent parameter, the optimized parameter for the limitation Boltzmann machine that i-th optimized parameter is described i-th layer;Wherein, the value of i is successively
It is 1,2 ..., N;
The limitation Boltzmann machine of i+1 layer is carried out according to i-th optimized parameter and the pretreatment voice signal
Unsupervised training obtains i+1 optimized parameter;
The multiple optimized parameter is finely tuned using the method for overall situation training to the deepness belief network and converges to the overall situation
It is optimal, obtain multiple fine tuning optimized parameters;
The phonic signal character of the pretreatment voice signal is extracted according to the fine tuning optimized parameter.
The step 400: the phonic signal character is classified using the identification that support vector machines carries out speech emotional, is obtained
Speech emotion recognition result is obtained to specifically include:
The sample point of the phonic signal character is mapped to by high-dimensional feature space using kernel function, obtaining spatial linear can
The sample divided;
The sample that the support vector machines can divide according to the spatial linear carries out logic to the phonic signal character and sentences
It is disconnected, obtain speech emotion recognition result.
As shown in Fig. 2, a kind of speech emotion recognition system of deepness belief network, the identifying system include:
Voice signal obtains module 1, for obtaining voice signal;
Speech signal pre-processing module 2 obtains pretreatment voice signal for pre-processing the voice signal;
Characteristic extracting module 3, for carrying out unsupervised language using deepness belief network to the pretreatment voice signal
Sound signal feature extraction obtains phonic signal character;
Emotion recognition module 4, for the phonic signal character to be carried out to the identification of speech emotional using support vector machines
Classification obtains speech emotion recognition result.
The characteristic extracting module 3 specifically includes:
Deepness belief network establishes unit, for stacking low layer to high-rise N layer limitation Boltzmann machine, obtains depth
Belief network;
Supervised training unit carries out nothing for the limitation Boltzmann machine according to the pretreatment voice signal to i-th layer
Supervised training obtains the i-th optimized parameter, the optimal ginseng for the limitation Boltzmann machine that i-th optimized parameter is described i-th layer
Number;Wherein, the value of i is followed successively by 1,2 ..., N;According to i-th optimized parameter and the pretreatment voice signal to
I+1 layers of limitation Boltzmann machine carries out unsupervised training, obtains i+1 optimized parameter;
Small parameter perturbations unit is believed for being finely tuned the multiple optimized parameter using the method for overall situation training to the depth
Network convergence is read to global optimum, obtains multiple fine tuning optimized parameters;
Speech recognition unit, for extracting the pretreatment voice signal according to the fine tuning optimized parameter
Phonic signal character.
The emotion recognition module 4 specifically includes:
Kernel function unit, for the sample point of the phonic signal character to be mapped to high dimensional feature sky using kernel function
Between, obtain the sample that spatial linear can divide;
Logic judgment unit believes the voice according to the sample that the spatial linear can divide for the support vector machines
Number feature carries out logic judgment, obtains speech emotion recognition result.
After the multidimensional characteristic vectors for extracting the affective characteristics in voice signal by deepness belief network, one is needed to be suitble to
Emotion classifiers.This method is using support vector machines using one-to-one mode to four kinds of emotions (surprised, glad, angry, sad)
Classify.Deepness belief network is extracted to the multidimensional characteristic vectors of the affective characteristics in voice signal as support vector machines
The sample point of input feature vector is mapped to the Nonlinear separability problem of speech emotional using kernel function by the input of classifier
High-dimensional feature space, so that corresponding sample space linear separability.Feelings identifying system block diagram such as Fig. 3 institute based on support vector machines
Show.
" one-to-one " mode is to construct hyperplane to any two kinds of emotions, needs to train k* (k-1)/2 sub-classifier.It is whole
A training process needs altogetherA support vector machines sub-classifier, i.e., 6.Each sub-classifier is by surprised, glad, anger
Any two kinds of training in anger, sad four kinds of affective characteristics form.That is: glad-indignation, glad-sad, glad-surprised, anger
Anger-sadness, indignation-is surprised, sad-surprised.One classifier of training between every two class is carried out when to a unknown speech emotional
When classification, each classifier carries out its classification to judge and for corresponding classification " throwing a upper ticket ", last who gets the most votes's class
It Ji not be as the classification of the unknown emotion.Decision phase uses ballot method, it is understood that there may be the identical situation of the poll of multiple classes, from
And make unknown sample while belonging to multiple classifications, influence nicety of grading.
It is both needed to before support vector machine classifier training and identification as one label of every emotional speech Design of Signal, to
Indicate emotional category belonging to this emotional speech signal.The type of label must be set as dimorphism.During emotion recognition, together
When feature vector is input in all support vector machines, the output of each support vector machines was selected most later by logical decision
Possible emotional category, finally using the emotion of weight highest (poll is most) as the affective state of voice signal to be identified, energy
Access recognition result.
Each embodiment in this specification is described in a progressive manner, the highlights of each of the examples are with other
The difference of embodiment, the same or similar parts in each embodiment may refer to each other.For system disclosed in embodiment
For, since it is corresponded to the methods disclosed in the examples, so being described relatively simple, related place is said referring to method part
It is bright.
Used herein a specific example illustrates the principle and implementation of the invention, and above embodiments are said
It is bright to be merely used to help understand method and its core concept of the invention;At the same time, for those skilled in the art, foundation
Thought of the invention, there will be changes in the specific implementation manner and application range.In conclusion the content of the present specification is not
It is interpreted as limitation of the present invention.
Claims (6)
1. a kind of speech-emotion recognition method of deepness belief network, which is characterized in that the recognition methods includes:
Obtain voice signal;
The voice signal is pre-processed, pretreatment voice signal is obtained;
Unsupervised speech recognition is carried out using deepness belief network to the pretreatment voice signal, obtains voice
Signal characteristic;
The identification classification that the phonic signal character is carried out to speech emotional using support vector machines, obtains speech emotion recognition knot
Fruit.
2. a kind of speech-emotion recognition method of deepness belief network according to claim 1, which is characterized in that described right
The pretreatment voice signal carries out unsupervised speech recognition using deepness belief network, and it is special to obtain voice signal
Sign specifically includes:
Low layer to high-rise N layer limitation Boltzmann machine is stacked, deepness belief network is obtained;
Limitation Boltzmann machine according to the pretreatment voice signal to i-th layer carries out unsupervised training, obtains the i-th optimal ginseng
Number, the optimized parameter for the limitation Boltzmann machine that i-th optimized parameter is described i-th layer;Wherein, the value of i is followed successively by 1,
2 ..., N;
The limitation Boltzmann machine of i+1 layer is carried out without prison according to i-th optimized parameter and the pretreatment voice signal
Supervise and instruct white silk, obtains i+1 optimized parameter;
The multiple optimized parameter is finely tuned using the method for overall situation training to the deepness belief network and converges to global optimum,
Obtain multiple fine tuning optimized parameters;
The phonic signal character of the pretreatment voice signal is extracted according to the fine tuning optimized parameter.
3. a kind of speech-emotion recognition method of deepness belief network according to claim 1, which is characterized in that described to incite somebody to action
The phonic signal character carries out the identification classification of speech emotional using support vector machines, and it is specific to obtain speech emotion recognition result
Include:
The sample point of the phonic signal character is mapped to by high-dimensional feature space using kernel function, obtains what spatial linear can divide
Sample;
The support vector machines carries out logic judgment to the phonic signal character according to the sample that the spatial linear can divide, and obtains
Obtain speech emotion recognition result.
4. a kind of speech emotion recognition system of deepness belief network, which is characterized in that the identifying system includes:
Voice signal obtains module, for obtaining voice signal;
Speech signal pre-processing module obtains pretreatment voice signal for pre-processing the voice signal;
Characteristic extracting module, for carrying out unsupervised voice signal using deepness belief network to the pretreatment voice signal
Feature extraction obtains phonic signal character;
Emotion recognition module, for the phonic signal character to be classified using the identification that support vector machines carries out speech emotional,
Obtain speech emotion recognition result.
5. a kind of speech emotion recognition system of deepness belief network according to claim 4, which is characterized in that the spy
Sign extraction module specifically includes:
Deepness belief network establishes unit, for stacking low layer to high-rise N layer limitation Boltzmann machine, obtains depth conviction
Network;
Supervised training unit, for unsupervised to i-th layer of limitation Boltzmann machine progress according to the pretreatment voice signal
Training obtains the i-th optimized parameter, the optimized parameter for the limitation Boltzmann machine that i-th optimized parameter is described i-th layer;Its
In, the value of i is followed successively by 1,2 ..., N;According to i-th optimized parameter and the pretreatment voice signal to i+1 layer
Limitation Boltzmann machine carry out unsupervised training, obtain i+1 optimized parameter;
Small parameter perturbations unit, for being finely tuned the multiple optimized parameter using the method for overall situation training to the depth conviction net
Network converges to global optimum, obtains multiple fine tuning optimized parameters;
Speech recognition unit, for extracting the voice of the pretreatment voice signal according to the fine tuning optimized parameter
Signal characteristic.
6. a kind of speech emotion recognition system of deepness belief network according to claim 4, which is characterized in that the feelings
Sense identification module specifically includes:
Kernel function unit is obtained for the sample point of the phonic signal character to be mapped to high-dimensional feature space using kernel function
Obtain the sample of space linear separability;
Logic judgment unit, the sample that can be divided for the support vector machines according to the spatial linear are special to the voice signal
Sign carries out logic judgment, obtains speech emotion recognition result.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910173690.3A CN109785863A (en) | 2019-02-28 | 2019-02-28 | A kind of speech-emotion recognition method and system of deepness belief network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910173690.3A CN109785863A (en) | 2019-02-28 | 2019-02-28 | A kind of speech-emotion recognition method and system of deepness belief network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109785863A true CN109785863A (en) | 2019-05-21 |
Family
ID=66486177
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910173690.3A Pending CN109785863A (en) | 2019-02-28 | 2019-02-28 | A kind of speech-emotion recognition method and system of deepness belief network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109785863A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110619893A (en) * | 2019-09-02 | 2019-12-27 | 合肥工业大学 | Time-frequency feature extraction and artificial intelligence emotion monitoring method of voice signal |
CN112687294A (en) * | 2020-12-21 | 2021-04-20 | 重庆科技学院 | Vehicle-mounted noise identification method |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101561651B1 (en) * | 2014-05-23 | 2015-11-02 | 서강대학교산학협력단 | Interest detecting method and apparatus based feature data of voice signal using Deep Belief Network, recording medium recording program of the method |
CN106297825A (en) * | 2016-07-25 | 2017-01-04 | 华南理工大学 | A kind of speech-emotion recognition method based on integrated degree of depth belief network |
CN107092895A (en) * | 2017-05-09 | 2017-08-25 | 重庆邮电大学 | A kind of multi-modal emotion identification method based on depth belief network |
CN108717856A (en) * | 2018-06-16 | 2018-10-30 | 台州学院 | A kind of speech-emotion recognition method based on multiple dimensioned depth convolution loop neural network |
CN109036468A (en) * | 2018-11-06 | 2018-12-18 | 渤海大学 | Speech-emotion recognition method based on deepness belief network and the non-linear PSVM of core |
-
2019
- 2019-02-28 CN CN201910173690.3A patent/CN109785863A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101561651B1 (en) * | 2014-05-23 | 2015-11-02 | 서강대학교산학협력단 | Interest detecting method and apparatus based feature data of voice signal using Deep Belief Network, recording medium recording program of the method |
CN106297825A (en) * | 2016-07-25 | 2017-01-04 | 华南理工大学 | A kind of speech-emotion recognition method based on integrated degree of depth belief network |
CN107092895A (en) * | 2017-05-09 | 2017-08-25 | 重庆邮电大学 | A kind of multi-modal emotion identification method based on depth belief network |
CN108717856A (en) * | 2018-06-16 | 2018-10-30 | 台州学院 | A kind of speech-emotion recognition method based on multiple dimensioned depth convolution loop neural network |
CN109036468A (en) * | 2018-11-06 | 2018-12-18 | 渤海大学 | Speech-emotion recognition method based on deepness belief network and the non-linear PSVM of core |
Non-Patent Citations (2)
Title |
---|
黄晨晨等: "基于深度信念网络的语音情感识别的研究", 《计算机研究与发展》 * |
黄驹斌: "基于深度信念网络的语音情感识别", 《中国优秀硕士学位论文全文数据库》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110619893A (en) * | 2019-09-02 | 2019-12-27 | 合肥工业大学 | Time-frequency feature extraction and artificial intelligence emotion monitoring method of voice signal |
CN112687294A (en) * | 2020-12-21 | 2021-04-20 | 重庆科技学院 | Vehicle-mounted noise identification method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Kim et al. | Towards speech emotion recognition" in the wild" using aggregated corpora and deep multi-task learning | |
CN108984745B (en) | Neural network text classification method fusing multiple knowledge maps | |
CN108597539A (en) | Speech-emotion recognition method based on parameter migration and sound spectrograph | |
CN110136690A (en) | Phoneme synthesizing method, device and computer readable storage medium | |
CN109523994A (en) | A kind of multitask method of speech classification based on capsule neural network | |
CN107577662A (en) | Towards the semantic understanding system and method for Chinese text | |
CN108874782A (en) | A kind of more wheel dialogue management methods of level attention LSTM and knowledge mapping | |
CN108806668A (en) | A kind of audio and video various dimensions mark and model optimization method | |
CN109409296A (en) | The video feeling recognition methods that facial expression recognition and speech emotion recognition are merged | |
CN108711421A (en) | A kind of voice recognition acoustic model method for building up and device and electronic equipment | |
CN108304823A (en) | A kind of expression recognition method based on two-fold product CNN and long memory network in short-term | |
CN110853656B (en) | Audio tampering identification method based on improved neural network | |
CN111581385A (en) | Chinese text type identification system and method for unbalanced data sampling | |
CN104091602A (en) | Speech emotion recognition method based on fuzzy support vector machine | |
CN111899766B (en) | Speech emotion recognition method based on optimization fusion of depth features and acoustic features | |
CN108364662A (en) | Based on the pairs of speech-emotion recognition method and system for differentiating task | |
Ocquaye et al. | Dual exclusive attentive transfer for unsupervised deep convolutional domain adaptation in speech emotion recognition | |
CN110211595A (en) | A kind of speaker clustering system based on deep learning | |
CN110289002A (en) | A kind of speaker clustering method and system end to end | |
Maheswari et al. | A hybrid model of neural network approach for speaker independent word recognition | |
CN110348482A (en) | A kind of speech emotion recognition system based on depth model integrated architecture | |
CN109785863A (en) | A kind of speech-emotion recognition method and system of deepness belief network | |
Sun et al. | A novel convolutional neural network voiceprint recognition method based on improved pooling method and dropout idea | |
Atkar et al. | Speech emotion recognition using dialogue emotion decoder and CNN Classifier | |
Sivaram et al. | Data-driven and feedback based spectro-temporal features for speech recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190521 |