EP1269462B1 - Verfahren und vorrichtung zur sprachaktivitätsdetektion - Google Patents

Verfahren und vorrichtung zur sprachaktivitätsdetektion Download PDF

Info

Publication number
EP1269462B1
EP1269462B1 EP01958309A EP01958309A EP1269462B1 EP 1269462 B1 EP1269462 B1 EP 1269462B1 EP 01958309 A EP01958309 A EP 01958309A EP 01958309 A EP01958309 A EP 01958309A EP 1269462 B1 EP1269462 B1 EP 1269462B1
Authority
EP
European Patent Office
Prior art keywords
audio parameter
unit
delay
audio
averaging
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP01958309A
Other languages
English (en)
French (fr)
Other versions
EP1269462A2 (de
Inventor
Mark Shahaf
Yishay Ben-Shimol
Moti Shor-Haham
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Motorola Solutions Israel Ltd
Original Assignee
Motorola Israel Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Motorola Israel Ltd filed Critical Motorola Israel Ltd
Publication of EP1269462A2 publication Critical patent/EP1269462A2/de
Application granted granted Critical
Publication of EP1269462B1 publication Critical patent/EP1269462B1/de
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/78Detection of presence or absence of voice signals

Definitions

  • the present invention relates to voice processing systems in general, and to methods and apparatus for detecting voice activity in a low resource environment, in particular.
  • a voice activity detector operates under the assumption that speech is present only in part of the audio signals while there are many intervals, which exhibit only silence or background noise.
  • a voice activity detector can be used for many purposes such as suppressing overall transmission activity in a transmission system, when there is no speech, thus potentially saving power and channel bandwidth.
  • the VAD detects that speech activity has resumed, then it can reinitiate transmission activity.
  • a voice activity detector can also be used in conjunction with speech storage devices, by differentiating audio portions which include speech from those that are "speechless". The portions including speech are then stored in the storage device and the "speechless" portions are not stored.
  • Conventional methods for detecting voice are based at least in part on methods for detecting and assessing the power of a speech signal.
  • the estimated power is compared to either constant or adaptive threshold, for determining a decision.
  • the main advantage of these methods is their low complexity, which makes them suitable for low resources implementations.
  • the main disadvantage of such methods is that background noise can result in "speech" being detected when none is present or "speech" which is present not being detected because it is obscured and difficult to detect.
  • Some methods for detecting speech activity are directed at noisy mobile environments and are based on adaptive filtering of the speech signal. This reduces the noise content from the signal, prior to the final decision.
  • the frequency spectrum and noise level may vary because the method will be used for different speakers and in different environments.
  • the input filter and thresholds are often adaptive so as to track these variations. Examples of these methods are provided in GSM specification 06.42 Voice Activity Detector (VAD) for half rate, full rate, and enhanced full rate speech traffic channels respectively.
  • VAD Voice Activity Detector
  • Another such method is the "Multi-Boundary Voice Activity Detection Algorithm" as proposed in ITU G.729 annex B. These methods are more accurate in noisy environment but are significantly complex to implement.
  • European Patent application No. 0785419A2 Benyassine et al. is directed to a method for voice activity detection which includes the following steps: extracting a predetermined set of parameters from the incoming speech signal for each frame and making a frame voicing decision of the incoming speech signal for each frame according to a set of difference measures extracted from the predetermined set of parameters.
  • WO-A-0017856 relates to a method and apparatus for detecting voice activity in a speech signal.
  • WO-A-0017856 has a filing date of 27 August 1999, a publication date of 30 March 2000, and claims a priority date of 18 September 1998.
  • the present invention alleviates the disadvantages of the prior art by providing a method which utilizes conventional vocoder output data of a voice related stream for detecting voice activity therein.
  • the method for voice activity detection is based on the analysis of audio parameters such as Line Spectral Frequencies (LSF) parameters.
  • LSF Line Spectral Frequencies
  • the detection is based on a stationarity estimate of spectral characteristics of the incoming speech frames, which is represented by LSF parameters.
  • Apparatus 100 includes two delay arrays 102 and 110, a plurality of distance measure units 106A, 106B, 106C and 106D, an averaging unit 108, a subtraction unit 114 and decision logic unit (DLU) 116.
  • Delay array 102 includes a plurality of delay units 104A, 104B, 104C, 104D, 104E, 104F and 104G, all connected in series, so that each adds a further delay to the previous one.
  • Delay array 110 includes a plurality of delay units, 112A, 112B, 112C and 112D, all connected in series, so that each adds a further delay to the previous one.
  • Apparatus 100 is further connected to a Line Spectral Frequencies (LSF) generation unit 120, which can be a part of the voice encoder (vocoder) apparatus of an audio system.
  • LSF unit 120 produces LSF values for each received audio frame. It is noted that LSF unit 120 is only one example for an audio parameter generation unit.
  • the output of the LSF unit 120 is coupled to the input of delay unit 104A.
  • the input of each of delay units 104A, 104B, 104C and 104D is connected to a respective one of distance measure units 106A, 106B, 106C and 106D.
  • the input of delay unit 104A is connected to distance measure unit 106A.
  • Delay unit 104A has its output connected to distance measure unit 106B.
  • Unit 104B has its output connected to unit 106C.
  • Unit 104C has its output connected 106D.
  • the output of delay units 104D, 104E, 104F and 104G is connected to a respective one of distance measure units 106A, 106B, 106C and 106D.
  • the output of delay unit 104D is connected to distance measure unit 106A.
  • the LSF value L (n) at the input of delay unit 104A is associated with the value L (n-4) at the output of delay unit 104D.
  • each of the LSF values L (n-1), L (n-2) and L (n-3) is associated with a respective one of LSF values L (n-5), L (n-6) and L (n-7), at a respective one of distance measure units 106B, 106C and 106D.
  • the system includes a different number of delay units and can combine more than two LSF values, which are at different distances from each other, such as L ⁇ n + L ⁇ ⁇ n - 4 + L ⁇ ⁇ n - 6 .
  • the distance measure units 106A, 106B, 106C and 106D are all connected to the averaging unit 108.
  • Averaging unit 108 is further connected to delay unit 112A, subtraction unit 114 and to DLU 116.
  • the output of each of delay units 112A, 112B, 112C and 112D is connected to DLU 116.
  • the output of delay unit 112A is further connected to the subtraction unit 114.
  • Figure 2 is an illustration of a method for operating the apparatus 100 of Figure 1 , operative in accordance with another preferred embodiment of the present invention.
  • a plurality of audio parameters are received.
  • Each of the audio parameters is related to a predetermined audio frame.
  • the audio parameters include LSF values, which represent the short-time frequency spectrum characteristics of the signal envelope for each audio frame.
  • LSF parameters are derived from the Linear Prediction Coefficients (LPC's), which are widely used by many modern speech compression and analysis schemes and are discussed in detail in A. M. Kondoz, Digital Speech: Coding for Low Bit Rate Communications Systems, New York: John Wiley & Sons, 1994 .
  • LPC's Linear Prediction Coefficients
  • step 152 the audio parameters are grouped according to a predetermined pattern of audio frames.
  • each audio frame is associated with a voice frame, which is four places ahead of it.
  • the audio parameters of audio frame n are grouped with the audio parameters of audio frame n-4.
  • any other number can be used for the distance between the frames.
  • further combinations can also be used such as combination (n,n-2,n-7) and the like.
  • distance measure unit 106A groups vector L ( n ) of frame n with vector L ( n -4) of frame n-4.
  • Distance measure unit 106B groups vector L ( n -1) of frame n-1 with vector L ( n -5) of frame n-5.
  • Distance measure unit 106C groups vector L ( n - 2) of frame n-2 with vector L ( n -6) of frame n-6.
  • Distance measure unit 106D groups vector L ( n -3) of frame n-3 with vector L ( n -7) of frame n-7.
  • each distance measure unit 106A, 106B, 106C and 106D performs a two-stage operation.
  • the distance measure units 106A, 106B, 106C and 106D provide the D vectors D ⁇ n D ⁇ ⁇ n - 1 D ⁇ ⁇ n - 2 ... D ⁇ ⁇ n - M 1 - 1 2 to averaging unit 108.
  • step 156 an average value is determined for all of the present characteristic values.
  • the measure a(n) is applied to a second M 2 -stage delay line.
  • the delay line includes four delay units 112A, 112B, 112C and 112D.
  • Averaging unit 108 further provides the latest average value a ( n ) to DLU 116 and to subtraction unit 114.
  • Delay unit 112A provides the previous average value a ( n-1 ) to the subtraction unit 114.
  • step 160 a decision is produced according to the values, which are present. Reference is now made to Figures 3 and 4 .
  • the implementation of each of the decision functions can be according to a Boolean expression, which compares the value e(n) and components of the averaging vector A ( n ) with predetermined or variable threshold values.
  • decision logic can vary according to specific performance requirements to a trade-off between "false alarm”, "miss detect” statistics, and the like.
  • the logic can be either constant or adapted to other components such as background noise characteristic estimator, voicing mode if available, periodicity check, and the like.
  • the instantaneous decision result can further be applied to an additional hangover function.
  • step 180 represents the initial stage of the decision phase, wherein the current state of the VAD (speech-on or speech off) is detected. If the current state of the VAD is speech on, then the system 100 proceeds to step 182. Otherwise, the system 100 proceeds to step 186.
  • VAD speech-on or speech off
  • step 182 compliance with a speech-on-to-off transition condition is detected.
  • a condition includes a predetermined combination of a ( n ) and e ( n ) with respect to predetermined values (Note that the threshold can be adaptive in general case).
  • the system proceeds to step 184, which performs a transition in the VAD state to speech-off. Otherwise, this step is repeated until such compliance is detected.
  • DLU 116 detects if the received values comply with the predetermined condition.
  • step 186 compliance with a speech-off-to-on transition condition is detected.
  • a condition includes another predetermined combination of a ( n ) and e(n) with respect to predetermined values.
  • step 188 which performs a transition in the VAD state to speech-on. Otherwise, this step is repeated until such compliance is detected.
  • step 184 After performing a VAD mode transition (either of step 184 and 188), the system proceeds back to step 180.
  • Apparatus 200 includes a multi stage delay unit 202, two delay arrays 204 and 210, a distance measure unit 206, an averaging unit 208, a subtraction unit 214 and decision logic unit (DLU) 216.
  • Delay array 204 includes a plurality of delay units 218A, 218B and 218M 2 , all connected in series, so that each adds a further delay to the previous one.
  • delay array 210 includes a plurality of delay units, 212A, 212B, 212C and 212D, all connected in series, so that each adds a further delay stage to the previous one.
  • System 200 is further connected to a Line Spectral Frequencies (LSF) generation unit 220, which can be a part of the voice encoder (vocoder) apparatus of an audio system.
  • LSF unit 220 produces LSF values for each received audio frame.
  • the input of multi stage delay unit 202 is connected to LSF unit 220.
  • the output of multi stage delay unit 202 is connected distance measure unit 206.
  • the LSF value L(n) at the input of delay unit 218A is associated with an M 1 stage delayed value L(n-M 1 ) at the output of delay unit 202.
  • the output of distance measure unit 206 is connected to averaging unit 208 and to delay array 204.
  • the output of each of the delay units 218A, 218B and 218M 2 is connected to the averaging unit 208 so that each provides a previously delayed distance measure output value to the averaging unit 208.
  • delay unit 218A provides a distance measure value, which is respective of the pair, L(n-1) and L(n-M 1 -1). Accordingly, only the first distance measure value has to be calculated and the rest are stored, delayed and provided to the averaging unit 208 at the appropriate timing.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Geophysics And Detection Of Objects (AREA)
  • Measuring Pulse, Heart Rate, Blood Pressure Or Blood Flow (AREA)
  • Mobile Radio Communication Systems (AREA)
  • Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
  • Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
  • Time-Division Multiplex Systems (AREA)

Claims (14)

  1. Vorrichtung (100) zur Erfassung einer Sprachaktivität, die umfasst:
    einen Audioparametergenerator (120) zum Erzeugen von Audioparametern von empfangenen Funkrahmen;
    eine mehrstufige Verzögerungseinheit (102), die an den Audioparametergenerator angeschlossen ist, zum Erzeugen einer Reihe von Audioparametern, die um verschiedene Beträge verzögert sind;
    ein Abstandserfassungsmittel (106A, 106B) zum Messen des Abstandes zwischen vorbestimmten Gruppen der Audioparameter, die über verschiedene Verzögerungen verfügen;
    eine Mehrzahl von Bestimmungseinheiten (106A, 106B) zum Bestimmen eines charakteristischen Wertes für jede Gruppe der Audioparameter, deren Abstände gemessen worden sind;
    und eine an die Bestimmungseinheiten angeschlossene Mittelungseinheit (108) zum Bestimmen eines Mittelwertes für alle die charakteristischen Werte.
  2. Vorrichtung (100) zur Erfassung einer Sprachaktivität gemäß Anspruch 1, wobei die mehrstufige Verzögerungseinheit eine Mehrzahl von Audioparameterverzögerungseinheiten (104A, 104B) zum Verzögern der Audioparameter umfasst,
    wobei die Audioparameterverzögerungseinheiten in Reihe zusammengeschaltet sind, so dass jede Audioparameterverzögerungseinheit, die einer vorherigen Audioparameterverzögerungseinheit folgt, eine weitere Verzögerung zu der durch die vorherige addierten Verzögerung addiert, wobei die erste der Audioparameterverzögerungseinheiten an den Audioparametergenerator (120) angeschlossen ist;
    wobei die Abstandserfassungsmittel eine Mehrzahl von Abstanderfassungseinheiten (106A, 106B) umfassen, wobei jede an mindestens zwei der Audioparameterverzögerungseinheiten zum Gruppieren der verzögerten Audiowerte angeschlossen ist, die durch die Audioparameterverzögerungseinheiten gemäß dem vorbestimmten Verzögerungsmuster erzeugt werden, und die Abstandserfassungseinheiten außerdem die Bestimmungseinheiten sind.
  3. Vorrichtung zur Erfassung einer Sprachaktivität gemäß Anspruch 2, wobei die erste Audioparameterverzögerungseinheit (104A) arbeitet, um eine Mehrzahl von Audioparameterwerten bezüglich einer vorbestimmten Sprachperiode von dem Audioparametergenerator zu empfangen, jede von dem Rest der Audioparameterverzögerungseinheiten (104A, 104B, 104C) betreibbar ist, um die Audioparameterwerte von einer vorangehenden der Audioparameterverzögerungseinheiten (104A, 104B, 104C) zu empfangen, und jede der Abstandserfassungseinheiten betreibbar ist, um die Audioparameterwerte, die von ausgewählten der an sie angeschlossenen Audioparameterverzögerungseinheiten empfangen werden, zu verarbeiten, wodurch unterschiedliche Werte erzeugt werden, wobei die Mittelungseinheit betreibbar ist, um aus diesen unterschiedlichen Werten einen Mittelwert zu erzeugen.
  4. Vorrichtung zur Erfassung einer Sprachaktivität gemäß Anspruch 1, Anspruch 2 oder Anspruch 3, die weiterhin umfasst:
    eine Mehrzahl von Mittelungsverzögerungseinheiten (112A, 112B), die in Reihe geschaltet sind, wobei die erste der Mittelungsverzögerungseinheiten (112A) weiterhin an den Ausgang der Mittelungseinheit (108) angeschlossen ist; und
    eine digitale Logikeinheit (116), die an die Mittelungsverzögerungseinheiten angeschlossen ist.
  5. Vorrichtung zur Erfassung einer Sprachaktivität gemäß Anspruch 4, wobei die erste Mittelungsverzögerungseinheit (112A) betreibbar ist, um eine Mehrzahl von verarbeiteten Audioparametermittelwerten von der Mittelungseinheit (108) zu empfangen, wobei jede der Verzögerungseinheiten betreibbar ist, um jeden verarbeiteten Audioparametermittelwert zu verzögern, wobei die digitale Logikeinheit betreibbar ist, um eine Mehrzahl aufeinanderfolgende verarbeitete Audioparametermittelwerte zu empfangen, wobei der neueste der aufeinanderfolgenden verarbeiteten Audioparametermittelwerte von der Mittelungseinheit empfangen wird und der Rest der aufeinanderfolgenden verarbeiteten Audioparametermittelwerte von der Mittelungsverzögerungseinheit empfangen wird, wobei die digitale Logikverarbeitung betreibbar ist, damit die aufeinanderfolgenden verarbeiteten Audioparametermittelwerte dadurch eine Sprachanwesenheitsanzeige erzeugen.
  6. Vorrichtung zur Erfassung einer Sprachaktivität gemäß Anspruch 4 oder Anspruch 5, wobei die erste Audioparameterverzögerungseinheit (104A) betreibbar ist, um eine Mehrzahl von Audioparameterwerten von dem Audioparametergenerator (120) zu empfangen, jede von dem Rest der Audioparameterverzögerungseinheiten (104B, 104C, 104D) betreibbar ist, um die Audioparameterwerte von einer vorangehenden der Audioparameterverzögerungseinheiten (104A, 104B, 104C) zu empfangen, und jede der Abstandserfassungseinheiten (106A, 106B) betreibbar ist, um Audioparameterwerte, die von ausgewählten an sie angeschlossene der Audioparameterverzögerungseinheiten empfangen werden, zusammen zu verarbeiten, wodurch verschiedene Werte erzeugt werden, wobei die Mittelungseinheit betreibbar ist, um von jedem Satz der verschiedenen Werte einen verarbeiteten Audioparametermittelwert zu erzeugen, und
    wobei die erste Mittelungsverzögerungseinheit betreibbar ist, um die verarbeiteten Audioparametermittelwerte von der Mittelungseinheit zu empfangen, wobei jede der Verzögerungseinheiten betreibbar ist, um jeden der verarbeiteten Audioparametermittelwerte zu verzögern, die digitale Logikeinheit eine Mehrzahl von aufeinanderfolgenden verarbeiteten Audioparametermittelwerten empfängt, wobei die digitale Logikeinheit betreibbar ist, um eine Mehrzahl von aufeinanderfolgenden verarbeiteten Audioparametermittelwerten zu empfangen, wobei der neueste der aufeinanderfolgenden verarbeiteten Audioparametermittelwerte von der Mittelungseinheit empfangen wird und der Rest der aufeinanderfolgenden verarbeiteten Audioparametermittelwerte von der Mittelungsverzögerungseinheit empfangen wird, wobei die digitale Logikeinheit betreibbar ist, um die aufeinanderfolgenden verarbeiteten Audioparametermittelwerte zu verarbeiten, wodurch eine Sprachanwesenheitsanzeige erzeugt wird.
  7. Vorrichtung zur Erfassung einer Sprachaktivität gemäß Anspruch 1, Anspruch 4 oder Anspruch 5, wobei jede der Bestimmungseinheiten (218A, 218B) betreibbar ist, um der Mittelungseinheit (208) eine zuvor verzögerte Abstandsmaßausgabe zur Verfügung zu stellen, wobei die erste (218A) der Bestimmungseinheiten durch eine Abstandserfassungseinheit (206), die betreibbar ist, um einen Abstand zwischen jedem einer Reihe von durch die mehrstufige Verzögerungseinheit (206) erzeugten unterschiedlich verzögerten Audioparametern und einer unverzögerten Ausgabe des Audioparametergenerators (120) zu messen, an den Audioparametergenerator (120) angeschlossen ist.
  8. Vorrichtung zur Erfassung einer Sprachaktivität gemäß einem der vorangehenden Ansprüche, wobei der Audioparameter Linienspektralfrequenzen umfasst.
  9. Vorrichtung zur Erfassung einer Sprachaktivität gemäß Anspruch 8, wobei der Audioparametergenerator einen Linienspektralfrequenzgenerator umfasst.
  10. Vorrichtung zur Erfassung einer Sprachaktivität gemäß einem der Ansprüche 4 bis 9, die weiterhin umfasst: eine Subtraktionseinheit (114), die zwischen dem Eingang und dem Ausgang der ersten Mittelungsverzögerungseinheit geschaltet ist und weiterhin an die digitale Logikeinheit angeschlossen ist,
    wobei die Subtraktionseinheit betreibbar ist, um Differenzwerte von verarbeiteten Audioparametermittelwerten, die von der Mittelungseinheit empfangen werden, und von verarbeiteten Audioparametermittelwerten, die durch die erste Mittelungsverzögerungseinheit verzögert werden, zu erzeugen, und
    wobei die digitale Logikeinheit betreibbar ist, um die Differenzwerte zusammen mit den aufeinanderfolgenden verarbeiteten Audioparametermittelwerten zu verarbeiten, wodurch eine Sprachanwesenheitsanzeige erzeugt wird.
  11. Verfahren zur Verwendung der Vorrichtung gemäß einem der vorangehenden Ansprüche zur Erfassung einer Sprachaktivität, das die folgenden Schritte umfasst:
    Gruppieren von Audioparametern, die mit einer vorbestimmten Kombination von Audiorahmen verknüpft sind, wodurch eine Mehrzahl von Gruppen erzeugt wird;
    Bestimmen eines charakteristischen Wertes für jede der Gruppen;
    Bestimmen eines Mittelwertes für jede einer Mehrzahl von Auswahlen einer Mehrzahl der charakteristischen Werte; und
    Bestimmen der Anwesenheit einer Sprachaktivität aus ausgewählten der Mittelwerte.
  12. Verfahren gemäß Anspruch 11, das weiterhin umfasst:
    den Schritt eines Erfassens der Energie von Audioabtastwerten, die mit den Audioparametern verknüpft sind, vor dem Schritt eines Bestimmens der Anwesenheit einer Sprachaktivität.
  13. Verfahren gemäß Anspruch 11 oder Anspruch 12, das weiterhin umfasst: den vorläufigen Schritt eines Empfangens der Audioparameter von einem Audiogenerator.
  14. Verfahren gemäß Anspruch 11, Anspruch 12 oder Anspruch 13, das weiterhin umfasst: den vorläufigen Schritt eines Erzeugens der Audioparameter aus einer Mehrzahl von Audioabtastwerten.
EP01958309A 2000-03-15 2001-03-14 Verfahren und vorrichtung zur sprachaktivitätsdetektion Expired - Lifetime EP1269462B1 (de)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
GB0006312 2000-03-15
GB0006312A GB2360428B (en) 2000-03-15 2000-03-15 Voice activity detection apparatus and method
PCT/IB2001/001603 WO2001080220A2 (en) 2000-03-15 2001-03-14 Voice activity detection apparatus and method

Publications (2)

Publication Number Publication Date
EP1269462A2 EP1269462A2 (de) 2003-01-02
EP1269462B1 true EP1269462B1 (de) 2008-05-14

Family

ID=9887716

Family Applications (1)

Application Number Title Priority Date Filing Date
EP01958309A Expired - Lifetime EP1269462B1 (de) 2000-03-15 2001-03-14 Verfahren und vorrichtung zur sprachaktivitätsdetektion

Country Status (6)

Country Link
EP (1) EP1269462B1 (de)
AT (1) ATE395683T1 (de)
AU (1) AU2001280027A1 (de)
DE (1) DE60133998D1 (de)
GB (1) GB2360428B (de)
WO (1) WO2001080220A2 (de)

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2317084B (en) * 1995-04-28 2000-01-19 Northern Telecom Ltd Methods and apparatus for distinguishing speech intervals from noise intervals in audio signals
US5774849A (en) * 1996-01-22 1998-06-30 Rockwell International Corporation Method and apparatus for generating frame voicing decisions of an incoming speech signal
US6385548B2 (en) * 1997-12-12 2002-05-07 Motorola, Inc. Apparatus and method for detecting and characterizing signals in a communication system
US6188981B1 (en) * 1998-09-18 2001-02-13 Conexant Systems, Inc. Method and apparatus for detecting voice activity in a speech signal

Also Published As

Publication number Publication date
ATE395683T1 (de) 2008-05-15
GB2360428B (en) 2002-09-18
GB2360428A (en) 2001-09-19
EP1269462A2 (de) 2003-01-02
WO2001080220A2 (en) 2001-10-25
WO2001080220A3 (en) 2002-05-23
DE60133998D1 (de) 2008-06-26
AU2001280027A1 (en) 2001-10-30
GB0006312D0 (en) 2000-05-03

Similar Documents

Publication Publication Date Title
EP0909442B1 (de) Sprachaktivitätsdetektor
KR100883712B1 (ko) 음원 방향 추정 방법, 및 음원 방향 추정 장치
EP3493205B1 (de) Verfahren und vorrichtung zur adaptiven detektion einer stimmaktivität in einem audioeingangssignal
EP0551803B1 (de) Verfahren zur Synchronisierung und Kanalschätzung in einem TDMA-Radiokommunikationssystem
JP5006279B2 (ja) 音声活性検出装置及び移動局並びに音声活性検出方法
KR100770839B1 (ko) 음성 신호의 하모닉 정보 및 스펙트럼 포락선 정보,유성음화 비율 추정 방법 및 장치
JP3878482B2 (ja) 音声検出装置および音声検出方法
US8818811B2 (en) Method and apparatus for performing voice activity detection
EP0556992A1 (de) Rauschverminderungseinrichtung
US20070239437A1 (en) Apparatus and method for extracting pitch information from speech signal
EP1548703B1 (de) Vorrichtung und Verfahren zur Sprachaktivitätsdetektion
US6876965B2 (en) Reduced complexity voice activity detector
KR20090080777A (ko) 신호 검출 방법 및 장치
JP3418005B2 (ja) 音声ピッチ検出装置
EP1269462B1 (de) Verfahren und vorrichtung zur sprachaktivitätsdetektion
JPH08221097A (ja) 音声成分の検出法
Beritelli et al. A low‐complexity speech‐pause detection algorithm for communication in noisy environments
CA2279264C (en) Speech immunity enhancement in linear prediction based dtmf detector
US5734679A (en) Voice signal transmission system using spectral parameter and voice parameter encoding apparatus and decoding apparatus used for the voice signal transmission system
JPS61184912A (ja) 定数可変型聴感的重み付けフイルタ
US6993478B2 (en) Vector estimation system, method and associated encoder
GB2437868A (en) Estimating noise power spectrum, sorting time frames, calculating the quantile and interpolating values over all remaining frequencies
JPH10304023A (ja) 電話機
WO1988007740A1 (en) Distance measurement control of a multiple detector system
JPH0311479B2 (de)

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR

AX Request for extension of the european patent

Free format text: AL;LT;LV;MK;RO;SI

17P Request for examination filed

Effective date: 20021125

17Q First examination report despatched

Effective date: 20061201

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

Free format text: LANGUAGE OF EP DOCUMENT: FRENCH

REF Corresponds to:

Ref document number: 60133998

Country of ref document: DE

Date of ref document: 20080626

Kind code of ref document: P

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080825

NLV1 Nl: lapsed or annulled due to failure to fulfill the requirements of art. 29p and 29m of the patents act
PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20081014

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080814

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20090217

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090331

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: FR

Ref legal event code: ST

Effective date: 20091130

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090314

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090331

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090331

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20091123

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080815

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090314

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20200327

Year of fee payment: 20

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20200528

Year of fee payment: 20

REG Reference to a national code

Ref country code: DE

Ref legal event code: R071

Ref document number: 60133998

Country of ref document: DE

REG Reference to a national code

Ref country code: GB

Ref legal event code: PE20

Expiry date: 20210313

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION

Effective date: 20210313