IL108401A - Method and apparatus for indicating the emotional state of a person - Google Patents
Method and apparatus for indicating the emotional state of a personInfo
- Publication number
- IL108401A IL108401A IL10840194A IL10840194A IL108401A IL 108401 A IL108401 A IL 108401A IL 10840194 A IL10840194 A IL 10840194A IL 10840194 A IL10840194 A IL 10840194A IL 108401 A IL108401 A IL 108401A
- Authority
- IL
- Israel
- Prior art keywords
- speech
- person
- circuit
- cross
- emotional state
- Prior art date
Links
- 230000002996 emotional effect Effects 0.000 title claims description 39
- 238000000034 method Methods 0.000 title claims description 18
- 238000005259 measurement Methods 0.000 claims description 18
- 238000001914 filtration Methods 0.000 claims description 8
- 230000010354 integration Effects 0.000 claims description 7
- 238000012545 processing Methods 0.000 claims description 7
- 230000008569 process Effects 0.000 claims description 2
- 238000000605 extraction Methods 0.000 description 13
- 238000001514 detection method Methods 0.000 description 9
- 230000008451 emotion Effects 0.000 description 6
- 230000000694 effects Effects 0.000 description 3
- 230000002596 correlated effect Effects 0.000 description 2
- 230000000875 corresponding effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000012552 review Methods 0.000 description 2
- 238000001228 spectrum Methods 0.000 description 2
- 210000001260 vocal cord Anatomy 0.000 description 2
- 230000001755 vocal effect Effects 0.000 description 2
- 206010049816 Muscle tightness Diseases 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 210000000867 larynx Anatomy 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 210000003205 muscle Anatomy 0.000 description 1
- 230000029058 respiratory gaseous exchange Effects 0.000 description 1
- 230000000638 stimulation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/90—Pitch determination of speech signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/26—Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
- G10L25/09—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being zero crossing rates
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/03—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
- G10L25/24—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being the cepstrum
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Description
.irru "79 'M n M^n ΐτχ ττππτ no'? ETHOO AND APPARATUS FOR INDICATING THE EMOTIONAL STATE OF A PERSON HASHAVSHEVET (MANUFACTURE) (1988) LTD. n"ya (1988) (ΙΙΪ') πικακνπ B:8742 METHOD AND APPARATUS FOR INDICATING THE EMOTIONAL STATE OF A PERSON The present invention relates to a method for indicating the emotional state of a person by a person's speech. The invention also relates to a speech analyzer for analyzing a person's speech in order to indicate the emotional state of the person.
Studies on the effects of emotion on the acoustic characteristics of speech have shown that average values and ranges of fundamental frequency (Fo) differ from one emotion to another. A review of the state of the art in this respect is found in an article by Iain R. Murray and John L . Arnott : Toward the Stimulation of Emotion in Synthetic Speech: A Review of the Literature on Human Vocal Emotion, appearing in J.Acoust Soc Am, 1 93 : 1 097- 1 1 08 , 1 993 . These studies show that the acoustic properties appearing to be among the most sensitive indicators of emotion are attributes that specify the contours of the fundamental frequency Fo throughout an utterance. Respiration is frequently a sensitive indicator in certain emotional situations, and an increase in its rate results in increased subglottal pressure during speech.
The present invention is based on the assumption that the emotional state of the speaker influences the muscle activity in the larynx and the state of the vocal cords more than and other parts of the speech generating system such as the tongue, lips and jaws. Thus, any analysis of the speech signal that reflects vocal cord activity is more likely to be influenced by physiological changes brought about by the emotional state of the speaker .
Such physiological changes as increased subglottal pressure generally give rise to a narrowing of individual glottal pulses, and hence to a change in the spectrum of the pulses .
One speech analysis technique for determining emotional stress is disclosed in USA Patents 4,093,821 and 4,142,067 by J.D. Williamson. These patents relate to a speech analyzer which determines the emotional state of a person by analyzing the real time frequency or pitch components within the first format band of human speech. Because of the characteristics of the first format speech sounds, the system described in these patents analyzes an FM demodulated first format speech signal and produces an output indicative of nulls or "flat" spots therein. By thus analyzing certain first format frequency distribution patterns, a qualitative measurement of variations in speech-related muscle tension is measured, which in turn is correlated to the emotional state of the speaker.
This approach of pitch detection thus assumes that the speech signal can be viewed as an FM modulated signal, thereby obtaining a modulated pitch signal. However, this assumption does not always hold, and therefore the results obtained from the signal analysis with such a system do not always correlate with the subject's emotional state.
An object of the present invention is to provide a novel method of indicating the emotional state of a person by the person's speech. Another object of the invention is to provide a speech analyzer for analyzing a person's speech and for indicating the emotional state of the person.
According to one aspect of the present invention, there is provided a method of indicating the emotional state of a person by the person's speech, comprising: detecting speech waves of the person; subjecting the detected speech waves to a backward and forward inverse filtering operation to obtain residual signals; cross-correlating said residual signals to produce a cross-correlated output; integrating with maximum overlapping the backward and forward residual signal, thereby achieving an estimation of the glottal wave; and processing the above glottal wave estimation to privide an indication of the emotional state of the speaker.
According to further features in the preferred embodiment of the invention described below, the said cross-correlation output is processed to produce a measurement of the pitch perturbations in the detected speech waves, which pitch perturbations are utilized to indicate the emotional state of the person. In the described preferred embodiment, the above-mentioned estimated glottal wave is also processed to produce measurements of the cepstrum coefficients energy level, zero crossing, etc., in the detected speech waves.
These measurements are also utilized to indicate the emotional state of the person.
The invention also provides a novel speech analyzer for indicating the emotional state of a person in accordance with the above method.
Further features and advantages of the invention will be apparent from the description below.
The invention is herein described, by way of example only, with reference to the accompanying drawings, wherein : Fig. 1 is a general block diagram illustrating one form of speech analyzer system constructed in accordance with the present invention; Fig. 2 is a flow chart explaining the operation of the endpoint detector in the system of Fig. 1; Fig. 3 is a block diagram illustrating the feature extraction circuit in the system of Fig. 1; Figs. 4a and 4b, taken together, constitute a flow chart for explaining the operation of the feature extraction circuit of Fig. 3; Fig. 5 illustrates the pre-processed input signal to the feature extraction circuit and to the endpoint detection circuit in Fig. 1 ; Fig. 6 illustrates the output signal from the endpoint detection circuit of Fig. 1 to the feature extraction circuit; Fig. 7 illustrates the input speech signal applied to the feature extraction circuit as illustrated in Fig. 3; Fig. 8 illustrates the input to the LPC (Linear Predictive Code) analyzer circuit of Fig. 3; Figs. 9a and 9b illustrate the outputs of the inverse filter circuit of Fig. 3; Figs. 10a and 10b illustrate the outputs of the integration circuit in the system of Fig. 3; Figs. 11a and 11b illustrate the inputs to the cross-correlation estimation circuit in Fig. 3; and Fig. 12 illustrates the input signal to the cepstrum estimation circuit illustrated in Fig. 3.
Fig. 1 illustrates the overall system for analyzing the speech signals of a person in order to provide an indication of the emotional state of the person. The illustrated system is based on the assumption that emotion is highly correlated with the glottal speech waveform.
Thus, the system illustrated in Fig. 1 includes a detector, such as a microphone, or telephone, SD, which detects the speech waves of a person and feeds them to bandpass filter and analog-to-digital converter circuit 1. When the source of the voice is a telephone line, this circuit passes the frequency bandpass of 60 to 3200 Hz corresponding to the telephone line frequency bandwidth.
The output from circuit 1 is fed to a processor, generally designated PROC. This processor includes: an endpoint detection circuit 2, whose operation is described with reference to Fig. 2; a feature extraction circuit 3, more particularly described in Fig. 3 and in the flow chart of Figs. 4a and 4b; a decision-making circuit 4; and a reference template 5 which is used in the decision-making process. The output of the decision-making circuit 4 is displayed in a display DISP.
The filtered output of circuit 1 as appearing at port 1b is shown in Fig. 5. This filtered output is fed both to input port 2a of the endpoint detection circuit 2, and to input port 3a of the feature extraction circuit 3.
The endpoint detection circuit 2 detects the start and endpoints of significant signals. For this purpose, it operates according to an algorithm based on the energy measured criterion. Thus, if the measured energy is greater than the adaptively estimated noise energy threshold, then it is considered to be a significant signal. Fig. 2 is a flow chart illustrating the operation of the endpoint detector circuit 2. The output of this circuit, as appearing in output port 2b, is illustrated in Fig. 6, and is applied to input port 3b of the feature extraction circuit 3.
It will thus be seen that the feature extraction circuit 3 illustrated in Fig. 1 receives, via its input port 3a, the pre-filtered and digitized signal from circuit 1 , and via its input port 3b, the output of the endpoint detection circuit 2. Circuit 3 extracts certain features as described more particularly below with reference to Figs. 3, 4a and 4b, and produces an output, via its output port 3c, to the input port 4a of the decision making circuit 4, and also to the input port 5a of the reference template circuit 5.
Circuit 5 maintains a number of reference templates regarding the features extracted by circuit 3 and decisions rules determined by circuit 4, and with which the outputs of circuit 3 are compared in order to indicate the emotional state of the person whose speech is being analyzed. Thus, the reference templates could be pre-recorded templates of that person, or of the population in general, correlating the extracted features to emotional levels; alternatively, the reference templates could be based on continuously obtained and updated data from the speech of the same person correlating the features extracted by circuit 3 and the decision rules determined by circuit 4 to the emotional level of that person.
The details of the feature extraction circuit 3 are more particularly illustrated in Fig. 3. This circuit includes a gate 6 which receives via its input port 6a (corresponding to port 3a in Fig. 1) of the filtered and digitized output from circuit 1 ; it also receives via its input port 3b (Fig. 1), the output from the endpoint detection circuit 2.
Fig. 6 illustrates the input at input port 3b from the endpoint detection circuit, and Fig. 7 illustrates the input at input port 3a from the filter and digitizer circuit 1.
The output port 6b of gate 6 is connected to the input port 7a of a pre-emphasis circuit (differentiator) 7, and also to the input port 9a of an inverse filter 9. The output of the pre-emphasis circuit 7, as appearing on port 7b, is fed to the input port 8a of an LPC-analyzer circuit 8. This circuit determines the coefficients of the filter representing the vocal tract. The output of this analyzer appears in output port 8b and is fed to the input port 9b of an inverse filter 9. The inverse filter 9 also receives, via its input port 9a, the output of gate 6.
The computed coefficients of the LPC-analyzer are used to determine the inverse filter characteristics. The inverse filter 9 performs a forward and backward filtering operation, obtaining two signals at its output port 9c.
These two signals are the residual errors of the speech signal obtained from the inverse filter, and are shown in Figs. 9a and 9b. The residual error signals appearing in output port 9c of the inverse filter 9 are fed to the input port 10a of an integrator circuit 10. This circuit performs the integration as shown in Fig. 3, and produces an output at its output port 10b as shown in Figs. 10a and 10b.
The outputs of the integrator 10 are fed to the input port 11a of a trend estimator circuit 11. The function of this circuit is to remove any DC bias and trend that might have been built during the integration. After removal of the trend, the output signals, as shown in Figs. 11a and 11b, provide the two residual errors.
Using a predetermined range, the output signals of the trend estimator circuit 11 are inputted via input port 12a to a cross-correlator circuit 12. This circuit performs a cross-correlation operation on the two input signals and produces a cross-correlation output signal at its output port 12b.
The output of the cross-correlator circuit 12 is applied to the input port 13a of a first-peak circuit 13 for detecting the first peak in the cross-correlation output signal from circuit 12, and is also applied to a second-peak detector circuit 16 for detecting the second peak in the cross-correlation output signal.
The output from the first-peak detector circuit 13, as appearing in output port 13b, is applied to the input port 14b of a residual error circuit 14, which also receives, via its input port 14a, the output of the trend estimator circuit 11. Circuit 14 computes a new residual error, illustrated in Fig. 12, by integration, with , overlapping of the lag of the first peak, of the backward and forward residual errors, and outputs this new residual error via its output port 14c to input port 15a of a cepstrum estimator circuit 15.
The "cepstrum" is the inverse Fourier Transform of the logarithm of the power spectrum of a signal. Cepstrum estimator 15 computes the cepstrum coefficient vector "c" using Fast Fourier Transform (FFT) as follows: wherein e(n) is the new residual error signal, and IFFT is the inverse FFT. These coefficients are outputted via port 15b to the input port 4a of the decision circuit 4 in Fig. 1.
Another output produced by the feature extraction circuit 3 to the input port 4a of the decision circuit 4 includes the perturbations in the pitch of the detected speech waves. These perturbations are derived by processing the output of the cross-correlation from circuit 12 in the second-peak detector circuit 16. The latter circuit receives the cross-correlation output signal via its input port 16b, and also receives the output signal from the first-peak detector circuit 13 via its input port 16a.
Circuit 16 computes the second peak according to the following criteria: (1) the difference between the , - amplitudes of the first and second peaks of the cross- correlation output should not be significant, and (2) the lag in time between the first and second peaks should be physiologically meaningful. Circuit 16 thus determines the second peak in the cross-correlation output signal from circuit 12, and from that, measures the lag between the first and second peaks. This lag is applied via its output port 16c to a pitch analyzer circuit 17.
The pitch analyzer circuit 17 computes the pitch contour characteristics and pitch perturbations, and outputs these signals, via its output port 17b, to the input port 4a (Fig. 1) of the decision circuit 4, with the cepstrum vector signals from the cepstrum estimator circuit 15.
The operation of the feature extraction circuit 3 illustrated in Fig. 3 is more particularly shown in the flow chart of Figs. 4a and 4b which set forth, for purposes of example, an algorithm for each step in the processing of the speech signals.
The decision circuit 4 thus receives the pitch perturbation measurements and the cepstrum coefficients measurements and other measurements such as energy level, zero crossing, etc., from the feature extraction circuit 3 (Fig. 1), and compares these measurements with the reference templates 5, to thereby provide an indication of the emotional state of the person. As indicated earlier, the reference templates 5 may be pre-recorded references , ■ indicating emotional states of the person whose speech is being analyzed, or of the population in general; alternatively, the reference templates could be data continuously obtained and updated in a self-adapted manner from the speech of the same person.
The output of decision circuit 4 is displayed in display DISP.
It will thus be seen that by subjecting the detected speech waves to a backward and forward inverse filtering operation, two residual signals are obtained.
These two residual signals are then cross-correlated to produce a cross-correlation output. Also a new residual signal is computed by integrating with maximum overlapping the forward and backward residual signals, thereby achieving an estimation of the glottal wave, whose following attributes are then extracted: pitch, cepstrum coefficient, energy level, zero crossing (unvoice). Each measurement is calculated for both its mean and perturbation. These parameters (all or in part) provide an indication of the emotional state of the speaker.
While the invention has been described with respect to one preferred embodiment, it will be appreciated that this is set forth merely for purposes of example, and that many other variations, modifications and applications of the invention may be made.
Claims (23)
1. A method of indicating the emotional state of a person by the person's speech, comprising: detecting speech waves of the person; subjecting the detected speech waves to a backward and forward inverse filtering operation to obtain residual signals; cross-correlating said residual signals to produce a cross-correlated output; integrating with maximum overlapping the backward and forward residual signal, thereby achieving an estimation of the glottal wave; and processing the above glottal wave estimation to provide an indication of the emotional state of the speaker .
2. The method according to Claim 1 , wherein said cross-correlated output is processed to produce a measurement of the pitch perturbations in the detected speech waves, which pitch perturbations are utilized to indicate the emotional state of the person.
3. The method according to Claim 2 , wherein said pitch perturbation measurement includes measuring the lag between first and second peaks in said cross-correlated output signal.
4. The method according to either of Claims 2 or 3, wherein said estimated glottal wave is also processed to produce a measurement of the cepstrum coefficients and/or the energy level, and/or the zero crossing in the detected speech waves, which measurement is also utilized to indicate the emotional state of the person.
5. The method according to any one of Claims 1-4, wherein said detected speech waves are subjected to a Linear Predictive Coding Analysis before being subjected to the backward and forward inverse filtering operation.
6. The method according to any one of Claims 1-5, wherein the detected speech waves are filtered by a bandpass filter and digitized, before being subjected to the backward and forward filtering operation.
7. The method according to Claim 6, wherein the end points of the detected signals, after having been filtered and digitized, are detected before being subjected to the backward and forward filtering operation.
8. The method according to any one of Claims 1-7, wherein the results of processing the glottal wave estimation are compared with a pre-recorded reference to provide an indication of the emotional state of the person.
9. The method according to any one of Claims 1-7, wherein the results of processing the glottal wave estimation are compared with a reference which is continuously obtained and updated from the speech of the same person, to provide an indication of the emotional state of the person.
10. A speech analyzer, comprising: a speech detector for detecting speech waves of the person; a backward and forward reverse filter for subjecting the detected speech waves to a backward and forward inverse filtering operation to obtain residual signals; a cross-correlator circuit for cross-correlating the residual signals to produce a cross-correlated output signal; a circuit for integrating with maximum overlapping the backward and forward signals, thereby achieving an estimation of the glottal wave; and a processor for processing the glottal wave estimation to provide an indication of the emotional state of the person.
11. The speech analyzer according to Claim 10 , wherein said processor produces a measurement of the pitch perturbations in the detected waves, which pitch perturbations are utilized to indicate the emotional state of the person.
12. The speech analyzer according to Claim 11 , wherein said processor also produces a measurement of the cepstrum coefficients energy level, zero crossing (both mean and perturbation) in the detected speech waves, which measurement is also utilized to indicate the emotional state of the person.
13. The speech analyzer according to Claim 10, wherein said processor includes : a first-peak measuring circuit for measuring the first peak in said cross-correlated output signal; a second-peak measuring circuit for measuring the second peak in said cross-correlated output signal, and the lag from said first peak; and a pitch analyzer circuit receiving the output of said second-peak measuring circuit for measuring the pitch perturbations in the original speech waves of the person, which pitch perturbations are utilized for indicating the emotional state of the person.
14. The speech analyzer according to Claim 13, wherein said processor further includes a cepstrum coefficient measuring circuit and/or energy measuring circuit, and/or zero crossing measuring circuit which processes the output of said first-peak measuring circuit to produce a measurement of the cepstrum coefficients in the original speech waves of the person, which measurement is also utilized for indicating the emotional state of the person.
15. The speech analyzer according to any one of Claims 10-14, wherein said processor includes a Linear Predictive Coding Analyzer for analyzing the detected speech waves before they are fed to said backward and forward reverse filter.
16. The speech analyzer according to any one of Claims 10-15, wherein said processor includes an integration circuit which integrates said residual signals outputted by the backward and forward inverse filter before they are cross-correlated in said cross-correlator circuit.
17. The speech analyzer according to Claim 16, wherein said processor further includes a trend estimator circuit receiving the output of said integration circuit and removing DC bias that may have been built up during the integration of said residual signals before they are cross-correlated in said cross-correlator circuit.
18. The speech analyzer according to any one of Claims 10-17, wherein said processor includes an end point detector circuit receiving the output of said speech detector and determining the start and end points of significant signals.
19. The speech analyzer according to any one of Claims 10-18, further including a bandpass filter between said speech detector and said background and forward reverse filter.
20. The speech analyzer according to any one of Claims 10-19, further including a digitizer for digitizing the output of said speech detector before being fed to said backward and forward reverse filter.
21. The speech analyzer according to any one of Claims 10-20, wherein said processor includes a pre-recorded reference template against which the processed cross- correlation output signal is compared to provide an indication of the emotional state of the person.
22. The speech analyzer according to any one of Claims 10-20, wherein said processor includes a continuously-updated reference template which is continuously updated from the speech of the same person and with which the processed cross-correlated output signal is compared to provide an indication of the emotional state of the person.
23. The speech analyzer according to any one of Claims 11-22, further including a display for displaying the output of said processor.
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IL10840194A IL108401A (en) | 1994-01-21 | 1994-01-21 | Method and apparatus for indicating the emotional state of a person |
PCT/US1995/000493 WO1995020216A1 (en) | 1994-01-21 | 1995-01-13 | Method and apparatus for indicating the emotional state of a person |
AU15664/95A AU1566495A (en) | 1994-01-21 | 1995-01-13 | Method and apparatus for indicating the emotional state of a person |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IL10840194A IL108401A (en) | 1994-01-21 | 1994-01-21 | Method and apparatus for indicating the emotional state of a person |
Publications (2)
Publication Number | Publication Date |
---|---|
IL108401A0 IL108401A0 (en) | 1994-04-12 |
IL108401A true IL108401A (en) | 1996-12-05 |
Family
ID=11065722
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
IL10840194A IL108401A (en) | 1994-01-21 | 1994-01-21 | Method and apparatus for indicating the emotional state of a person |
Country Status (3)
Country | Link |
---|---|
AU (1) | AU1566495A (en) |
IL (1) | IL108401A (en) |
WO (1) | WO1995020216A1 (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6006188A (en) * | 1997-03-19 | 1999-12-21 | Dendrite, Inc. | Speech signal processing for determining psychological or physiological characteristics using a knowledge base |
IL122632A0 (en) * | 1997-12-16 | 1998-08-16 | Liberman Amir | Apparatus and methods for detecting emotions |
IL129399A (en) * | 1999-04-12 | 2005-03-20 | Liberman Amir | Apparatus and methods for detecting emotions in the human voice |
US6665644B1 (en) * | 1999-08-10 | 2003-12-16 | International Business Machines Corporation | Conversational data mining |
ES2177437B1 (en) * | 2000-12-13 | 2003-09-01 | Neturin S L | ANIMAL ANALYZING DEVICE FOR MAMMALS. |
IL144818A (en) | 2001-08-09 | 2006-08-20 | Voicesense Ltd | Method and apparatus for speech analysis |
US10068588B2 (en) | 2014-07-21 | 2018-09-04 | Microsoft Technology Licensing, Llc | Real-time emotion recognition from audio signals |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3855417A (en) * | 1972-12-01 | 1974-12-17 | F Fuller | Method and apparatus for phonation analysis lending to valid truth/lie decisions by spectral energy region comparison |
US3855418A (en) * | 1972-12-01 | 1974-12-17 | F Fuller | Method and apparatus for phonation analysis leading to valid truth/lie decisions by vibratto component assessment |
US4093821A (en) * | 1977-06-14 | 1978-06-06 | John Decatur Williamson | Speech analyzer for analyzing pitch or frequency perturbations in individual speech pattern to determine the emotional state of the person |
-
1994
- 1994-01-21 IL IL10840194A patent/IL108401A/en not_active IP Right Cessation
-
1995
- 1995-01-13 WO PCT/US1995/000493 patent/WO1995020216A1/en active Application Filing
- 1995-01-13 AU AU15664/95A patent/AU1566495A/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
WO1995020216A1 (en) | 1995-07-27 |
AU1566495A (en) | 1995-08-08 |
IL108401A0 (en) | 1994-04-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP0235181B1 (en) | A parallel processing pitch detector | |
EP0722164B1 (en) | Method and apparatus for characterizing an input signal | |
EP1083541B1 (en) | A method and apparatus for speech detection | |
JPH08505715A (en) | Discrimination between stationary and nonstationary signals | |
JPS62204652A (en) | Audible frequency signal identification system | |
Wendt et al. | Pitch determination and speech segmentation using the discrete wavelet transform | |
IL108401A (en) | Method and apparatus for indicating the emotional state of a person | |
US20010044714A1 (en) | Method of estimating the pitch of a speech signal using an average distance between peaks, use of the method, and a device adapted therefor | |
US6954726B2 (en) | Method and device for estimating the pitch of a speech signal using a binary signal | |
Vieira et al. | Robust F/sub 0/and jitter estimation in pathological voices | |
US20060150805A1 (en) | Method of automatically detecting vibrato in music | |
JP2564821B2 (en) | Voice judgment detector | |
US20010029447A1 (en) | Method of estimating the pitch of a speech signal using previous estimates, use of the method, and a device adapted therefor | |
KR100526110B1 (en) | Method and System for Pith Synchronous Feature Generation of Speaker Recognition System | |
KR100194953B1 (en) | Pitch detection method by frame in voiced sound section | |
JP2648779B2 (en) | Call signal identification device | |
KR19990011286A (en) | Speech Segmentation Method for Speech Recognition in Noisy Environments | |
KR102443221B1 (en) | Apparatus and method for sleep sound analysis | |
US11881200B2 (en) | Mask generation device, mask generation method, and recording medium | |
JPH0114599B2 (en) | ||
JP2557497B2 (en) | How to identify male and female voices | |
KR0171004B1 (en) | Basic frequency using samdf and ratio technique of the first format frequency | |
JP2880683B2 (en) | Noise suppression device | |
CN116229988A (en) | Voiceprint recognition and authentication method, system and device for personnel of power dispatching system | |
JPH0424717B2 (en) |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FF | Patent granted | ||
KB | Patent renewed | ||
RH | Patent void |