EP1269462B1 - Appareil et procede de detection d'activite vocale - Google Patents

Appareil et procede de detection d'activite vocale Download PDF

Info

Publication number
EP1269462B1
EP1269462B1 EP01958309A EP01958309A EP1269462B1 EP 1269462 B1 EP1269462 B1 EP 1269462B1 EP 01958309 A EP01958309 A EP 01958309A EP 01958309 A EP01958309 A EP 01958309A EP 1269462 B1 EP1269462 B1 EP 1269462B1
Authority
EP
European Patent Office
Prior art keywords
audio parameter
unit
delay
audio
averaging
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP01958309A
Other languages
German (de)
English (en)
Other versions
EP1269462A2 (fr
Inventor
Mark Shahaf
Yishay Ben-Shimol
Moti Shor-Haham
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Motorola Solutions Israel Ltd
Original Assignee
Motorola Israel Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Motorola Israel Ltd filed Critical Motorola Israel Ltd
Publication of EP1269462A2 publication Critical patent/EP1269462A2/fr
Application granted granted Critical
Publication of EP1269462B1 publication Critical patent/EP1269462B1/fr
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/78Detection of presence or absence of voice signals

Definitions

  • the present invention relates to voice processing systems in general, and to methods and apparatus for detecting voice activity in a low resource environment, in particular.
  • a voice activity detector operates under the assumption that speech is present only in part of the audio signals while there are many intervals, which exhibit only silence or background noise.
  • a voice activity detector can be used for many purposes such as suppressing overall transmission activity in a transmission system, when there is no speech, thus potentially saving power and channel bandwidth.
  • the VAD detects that speech activity has resumed, then it can reinitiate transmission activity.
  • a voice activity detector can also be used in conjunction with speech storage devices, by differentiating audio portions which include speech from those that are "speechless". The portions including speech are then stored in the storage device and the "speechless" portions are not stored.
  • Conventional methods for detecting voice are based at least in part on methods for detecting and assessing the power of a speech signal.
  • the estimated power is compared to either constant or adaptive threshold, for determining a decision.
  • the main advantage of these methods is their low complexity, which makes them suitable for low resources implementations.
  • the main disadvantage of such methods is that background noise can result in "speech" being detected when none is present or "speech" which is present not being detected because it is obscured and difficult to detect.
  • Some methods for detecting speech activity are directed at noisy mobile environments and are based on adaptive filtering of the speech signal. This reduces the noise content from the signal, prior to the final decision.
  • the frequency spectrum and noise level may vary because the method will be used for different speakers and in different environments.
  • the input filter and thresholds are often adaptive so as to track these variations. Examples of these methods are provided in GSM specification 06.42 Voice Activity Detector (VAD) for half rate, full rate, and enhanced full rate speech traffic channels respectively.
  • VAD Voice Activity Detector
  • Another such method is the "Multi-Boundary Voice Activity Detection Algorithm" as proposed in ITU G.729 annex B. These methods are more accurate in noisy environment but are significantly complex to implement.
  • European Patent application No. 0785419A2 Benyassine et al. is directed to a method for voice activity detection which includes the following steps: extracting a predetermined set of parameters from the incoming speech signal for each frame and making a frame voicing decision of the incoming speech signal for each frame according to a set of difference measures extracted from the predetermined set of parameters.
  • WO-A-0017856 relates to a method and apparatus for detecting voice activity in a speech signal.
  • WO-A-0017856 has a filing date of 27 August 1999, a publication date of 30 March 2000, and claims a priority date of 18 September 1998.
  • the present invention alleviates the disadvantages of the prior art by providing a method which utilizes conventional vocoder output data of a voice related stream for detecting voice activity therein.
  • the method for voice activity detection is based on the analysis of audio parameters such as Line Spectral Frequencies (LSF) parameters.
  • LSF Line Spectral Frequencies
  • the detection is based on a stationarity estimate of spectral characteristics of the incoming speech frames, which is represented by LSF parameters.
  • Apparatus 100 includes two delay arrays 102 and 110, a plurality of distance measure units 106A, 106B, 106C and 106D, an averaging unit 108, a subtraction unit 114 and decision logic unit (DLU) 116.
  • Delay array 102 includes a plurality of delay units 104A, 104B, 104C, 104D, 104E, 104F and 104G, all connected in series, so that each adds a further delay to the previous one.
  • Delay array 110 includes a plurality of delay units, 112A, 112B, 112C and 112D, all connected in series, so that each adds a further delay to the previous one.
  • Apparatus 100 is further connected to a Line Spectral Frequencies (LSF) generation unit 120, which can be a part of the voice encoder (vocoder) apparatus of an audio system.
  • LSF unit 120 produces LSF values for each received audio frame. It is noted that LSF unit 120 is only one example for an audio parameter generation unit.
  • the output of the LSF unit 120 is coupled to the input of delay unit 104A.
  • the input of each of delay units 104A, 104B, 104C and 104D is connected to a respective one of distance measure units 106A, 106B, 106C and 106D.
  • the input of delay unit 104A is connected to distance measure unit 106A.
  • Delay unit 104A has its output connected to distance measure unit 106B.
  • Unit 104B has its output connected to unit 106C.
  • Unit 104C has its output connected 106D.
  • the output of delay units 104D, 104E, 104F and 104G is connected to a respective one of distance measure units 106A, 106B, 106C and 106D.
  • the output of delay unit 104D is connected to distance measure unit 106A.
  • the LSF value L (n) at the input of delay unit 104A is associated with the value L (n-4) at the output of delay unit 104D.
  • each of the LSF values L (n-1), L (n-2) and L (n-3) is associated with a respective one of LSF values L (n-5), L (n-6) and L (n-7), at a respective one of distance measure units 106B, 106C and 106D.
  • the system includes a different number of delay units and can combine more than two LSF values, which are at different distances from each other, such as L ⁇ n + L ⁇ ⁇ n - 4 + L ⁇ ⁇ n - 6 .
  • the distance measure units 106A, 106B, 106C and 106D are all connected to the averaging unit 108.
  • Averaging unit 108 is further connected to delay unit 112A, subtraction unit 114 and to DLU 116.
  • the output of each of delay units 112A, 112B, 112C and 112D is connected to DLU 116.
  • the output of delay unit 112A is further connected to the subtraction unit 114.
  • Figure 2 is an illustration of a method for operating the apparatus 100 of Figure 1 , operative in accordance with another preferred embodiment of the present invention.
  • a plurality of audio parameters are received.
  • Each of the audio parameters is related to a predetermined audio frame.
  • the audio parameters include LSF values, which represent the short-time frequency spectrum characteristics of the signal envelope for each audio frame.
  • LSF parameters are derived from the Linear Prediction Coefficients (LPC's), which are widely used by many modern speech compression and analysis schemes and are discussed in detail in A. M. Kondoz, Digital Speech: Coding for Low Bit Rate Communications Systems, New York: John Wiley & Sons, 1994 .
  • LPC's Linear Prediction Coefficients
  • step 152 the audio parameters are grouped according to a predetermined pattern of audio frames.
  • each audio frame is associated with a voice frame, which is four places ahead of it.
  • the audio parameters of audio frame n are grouped with the audio parameters of audio frame n-4.
  • any other number can be used for the distance between the frames.
  • further combinations can also be used such as combination (n,n-2,n-7) and the like.
  • distance measure unit 106A groups vector L ( n ) of frame n with vector L ( n -4) of frame n-4.
  • Distance measure unit 106B groups vector L ( n -1) of frame n-1 with vector L ( n -5) of frame n-5.
  • Distance measure unit 106C groups vector L ( n - 2) of frame n-2 with vector L ( n -6) of frame n-6.
  • Distance measure unit 106D groups vector L ( n -3) of frame n-3 with vector L ( n -7) of frame n-7.
  • each distance measure unit 106A, 106B, 106C and 106D performs a two-stage operation.
  • the distance measure units 106A, 106B, 106C and 106D provide the D vectors D ⁇ n D ⁇ ⁇ n - 1 D ⁇ ⁇ n - 2 ... D ⁇ ⁇ n - M 1 - 1 2 to averaging unit 108.
  • step 156 an average value is determined for all of the present characteristic values.
  • the measure a(n) is applied to a second M 2 -stage delay line.
  • the delay line includes four delay units 112A, 112B, 112C and 112D.
  • Averaging unit 108 further provides the latest average value a ( n ) to DLU 116 and to subtraction unit 114.
  • Delay unit 112A provides the previous average value a ( n-1 ) to the subtraction unit 114.
  • step 160 a decision is produced according to the values, which are present. Reference is now made to Figures 3 and 4 .
  • the implementation of each of the decision functions can be according to a Boolean expression, which compares the value e(n) and components of the averaging vector A ( n ) with predetermined or variable threshold values.
  • decision logic can vary according to specific performance requirements to a trade-off between "false alarm”, "miss detect” statistics, and the like.
  • the logic can be either constant or adapted to other components such as background noise characteristic estimator, voicing mode if available, periodicity check, and the like.
  • the instantaneous decision result can further be applied to an additional hangover function.
  • step 180 represents the initial stage of the decision phase, wherein the current state of the VAD (speech-on or speech off) is detected. If the current state of the VAD is speech on, then the system 100 proceeds to step 182. Otherwise, the system 100 proceeds to step 186.
  • VAD speech-on or speech off
  • step 182 compliance with a speech-on-to-off transition condition is detected.
  • a condition includes a predetermined combination of a ( n ) and e ( n ) with respect to predetermined values (Note that the threshold can be adaptive in general case).
  • the system proceeds to step 184, which performs a transition in the VAD state to speech-off. Otherwise, this step is repeated until such compliance is detected.
  • DLU 116 detects if the received values comply with the predetermined condition.
  • step 186 compliance with a speech-off-to-on transition condition is detected.
  • a condition includes another predetermined combination of a ( n ) and e(n) with respect to predetermined values.
  • step 188 which performs a transition in the VAD state to speech-on. Otherwise, this step is repeated until such compliance is detected.
  • step 184 After performing a VAD mode transition (either of step 184 and 188), the system proceeds back to step 180.
  • Apparatus 200 includes a multi stage delay unit 202, two delay arrays 204 and 210, a distance measure unit 206, an averaging unit 208, a subtraction unit 214 and decision logic unit (DLU) 216.
  • Delay array 204 includes a plurality of delay units 218A, 218B and 218M 2 , all connected in series, so that each adds a further delay to the previous one.
  • delay array 210 includes a plurality of delay units, 212A, 212B, 212C and 212D, all connected in series, so that each adds a further delay stage to the previous one.
  • System 200 is further connected to a Line Spectral Frequencies (LSF) generation unit 220, which can be a part of the voice encoder (vocoder) apparatus of an audio system.
  • LSF unit 220 produces LSF values for each received audio frame.
  • the input of multi stage delay unit 202 is connected to LSF unit 220.
  • the output of multi stage delay unit 202 is connected distance measure unit 206.
  • the LSF value L(n) at the input of delay unit 218A is associated with an M 1 stage delayed value L(n-M 1 ) at the output of delay unit 202.
  • the output of distance measure unit 206 is connected to averaging unit 208 and to delay array 204.
  • the output of each of the delay units 218A, 218B and 218M 2 is connected to the averaging unit 208 so that each provides a previously delayed distance measure output value to the averaging unit 208.
  • delay unit 218A provides a distance measure value, which is respective of the pair, L(n-1) and L(n-M 1 -1). Accordingly, only the first distance measure value has to be calculated and the rest are stored, delayed and provided to the averaging unit 208 at the appropriate timing.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Acoustics & Sound (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Computational Linguistics (AREA)
  • Geophysics And Detection Of Objects (AREA)
  • Measuring Pulse, Heart Rate, Blood Pressure Or Blood Flow (AREA)
  • Mobile Radio Communication Systems (AREA)
  • Time-Division Multiplex Systems (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
  • Measuring Or Testing Involving Enzymes Or Micro-Organisms (AREA)
  • Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)

Claims (14)

  1. Appareil de détection d'activité vocale (100) comprenant :
    un générateur de paramètres audio (120) pour produire des paramètres audio à partir de trames audio reçues ;
    une unité de retard multi-étages (102) connectée au générateur de paramètres audio pour produire une série des paramètres audio retardés de quantités différentes ;
    des moyens de mesure de distance (1 06A, 106B) pour mesurer la distance entre des groupes prédéterminés de paramètres audio ayant des retards différents ;
    une pluralité d'unité de détermination (106A, 106B) pour déterminer une valeur caractéristique pour chaque groupe des paramètres audio dont les distances ont été mesurées ;
    et, connectée aux unités de détermination, une unité de calcul de moyenne (108) pour déterminer une valeur moyenne de toutes les valeurs caractéristiques.
  2. Appareil de détection d'activité vocale (100) selon la revendication 1, dans lequel l'unité de retard multi-étages comprend une pluralité d'unités de retard de paramètres audio (104A, 104B) pour retarder les paramètres audio, les unités de retard de paramètres audio étant connectées en série les unes aux autres, de telle sorte que chaque unité de retard de paramètres audio qui suit une unité de retard de paramètres audio précédente ajoute un retard supplémentaire au retard ajouté par la précédente, la première desdites unités de retard de paramètres audio étant connectée au générateur de paramètres audio (120) ;
    dans lequel les moyens de mesure de distance comprennent une pluralité d'unités de mesure de distance (106A, 106B) qui sont chacune connectées à au moins deux desdites unités de retard de paramètres audio, pour regrouper les valeurs audio retardées produites par les unités de retard de paramètres audio en fonction du modèle de retard prédéterminé, et les unités de mesure de distance sont également les unités de détermination.
  3. Appareil de détection d'activité vocale selon la revendication 2, dans lequel ladite première unité de retard de paramètres audio (104A) fonctionne de façon à recevoir une pluralité de valeurs de paramètres audio, concernant une période de parole prédéterminée, en provenance dudit générateur de paramètres audio, chacune desdites unités de retard de paramètres audio restantes (104A, 104B, 104C) peut fonctionner de façon à recevoir lesdites valeurs de paramètres audio en provenance d'une précédente desdites unités de retard de paramètres audio (104A, 104B, 104C), chacune desdites unités de mesure de distance peut fonctionner de façon à traiter les valeurs de paramètres audio reçues en provenance de certaines sélectionnées desdites unités de retard de paramètres audio qui lui sont connectées, pour ainsi produire des valeurs différentielles, l'unité de calcul de moyenne pouvant fonctionner de façon à produire une valeur moyenne à partir desdites valeurs différentielles.
  4. Appareil de détection d'activité vocale selon la revendication 1, la revendication 2 ou la revendication 3, et comprenant, en outre :
    une pluralité d'unités de retard de valeurs moyennes (112A, 112B) connectées en série les unes aux autres, la première desdites unités de retard de valeurs moyennes (112A) étant, en outre, connectée à la sortie de ladite unité de calcul de moyenne (108) ; et
    une unité logique numérique (116) connectée auxdites unités de retard de valeurs moyennes.
  5. Appareil de détection d'activité vocale selon la revendication 4, dans lequel ladite première unité de retard de valeurs moyennes (112A) peut fonctionner de façon à recevoir une pluralité de valeurs moyennes de paramètres audio traités en provenance de ladite unité de calcul de moyenne (108), chacune desdites unités de retard pouvant fonctionner de façon à retarder chacune desdites valeurs moyennes de paramètres audio traités, ladite unité logique numérique pouvant fonctionner de façon à recevoir une pluralité de valeurs moyennes de paramètres audio traités successives, la dernière desdites valeurs moyennes de paramètres audio traités successives étant reçue en provenance de ladite unité de calcul de moyenne et le restant desdites valeurs moyennes de paramètres audio traités successives étant reçu en provenance de ladite unité de retard de valeurs moyennes, ladite unité logique numérique pouvant fonctionner de façon à traiter lesdites valeurs moyennes de paramètres audio traités successives de façon à produire une indication de présence de parole.
  6. Appareil de détection d'activité vocale selon la revendication 4 ou la revendication 5, dans lequel ladite première unité de retard de paramètres audio (104A) peut fonctionner de façon à recevoir une pluralité de valeurs de paramètres audio en provenance dudit générateur de paramètres audio (120), chacune desdites unités de retard de paramètres audio restantes (104B, 104C, 104D) peut fonctionner de façon à recevoir lesdites valeurs de paramètres audio en provenance d'une précédente desdites unités de retard de paramètres audio (104A, 104B, 104C), et chacune desdites unités de mesure de distance (106A, 106B) peut fonctionner de façon à traiter ensemble les valeurs de paramètres audio reçues en provenance de certaines sélectionnées desdites unités de retard de paramètres audio, qui lui sont connectées, pour ainsi produire des valeurs différentielles, ladite unité de calcul de moyenne pouvant fonctionner de façon à produire une valeur moyenne de paramètres audio traités à partir de chaque ensemble desdites valeurs différentielles, et
    dans lequel ladite première unité de retard de valeurs moyennes peut fonctionner de façon à recevoir lesdites valeurs moyennes de paramètres audio traités en provenance de ladite unité de calcul de moyenne, chacun des unités de retard pouvant fonctionner de façon à retarder chacune des valeurs moyennes de paramètres audio traités, ladite unité logique numérique reçoit une pluralité de valeurs moyennes de paramètres audio traités successives, l'unité logique numérique pouvant fonctionner de façon à recevoir une pluralité de valeurs moyennes de paramètres audio traités successives, la dernière desdites valeurs moyennes de paramètres audio traités successives étant reçue en provenance de ladite unité de calcul de moyenne et le restant desdites valeurs moyennes de paramètres audio traités successives étant reçu en provenance de ladite unité de retard de valeurs moyennes, ladite unité logique numérique pouvant fonctionner de façon à traiter lesdites valeurs moyennes de paramètres audio traités successives de façon à produire une indication de présence de parole.
  7. Appareil de détection d'activité vocale selon la revendication 1, la revendication 4 ou la revendication 5, dans lequel chacune des unités de détermination (218A, 218B) peut fonctionner de façon à envoyer une sortie de mesure de distance préalablement retardée à l'unité de calcul de moyenne (208), la première (218A) desdites unités de détermination étant connectée au générateur de paramètres audio (120) par l'intermédiaire d'une unité de mesure de distance (206) qui peut fonctionner de façon à mesurer une distance entre chacun d'une série de paramètres audio différemment retardés produite par l'unité de retard multi-étages (206) et une sortie non retardée du générateur de paramètres audio (120).
  8. Appareil de détection d'activité vocale selon l'une quelconque des revendications précédentes, dans lequel ledit paramètre audio comprend des fréquences de raies spectrales.
  9. Appareil de détection d'activité vocale selon la revendication 8, dans lequel ledit générateur de paramètres audio comprend un générateur de fréquences de raies spectrales.
  10. Appareil de détection d'activité vocale selon l'une quelconque des revendications 4 à 9, et comprenant, en outre, une unité de soustraction (114) connectée entre l'entrée et la sortie de ladite première unité de retard de valeurs moyennes et en plus de ladite unité logique numérique,
    dans lequel ladite unité de soustraction peut fonctionner de façon à produire des valeurs de différence à partir des valeurs moyennes de paramètres audio traités reçues en provenance de ladite unité de calcul de moyenne et à partir des valeurs moyennes de paramètres audio traités retardées par la première unité de retard de valeurs moyennes, et
    dans lequel ladite unité logique numérique peut fonctionner de façon à traiter lesdites valeurs de différence en même temps que lesdites valeurs moyennes de paramètres audio traités successives, pour produire une indication de présence de parole.
  11. Procédé d'utilisation de l'appareil selon l'une quelconque des revendications précédentes pour la détection d'une activité de parole, comprenant les étapes consistant à :
    regrouper des paramètres audio, qui sont associés à une combinaison prédéterminée de trames audio, pour ainsi produire une pluralité de groupes ;
    déterminer une valeur caractéristique pour chacun desdits groupes ;
    déterminer une valeur moyenne pour chacune d'une pluralité de sélections d'une pluralité desdites valeurs caractéristiques ; et
    déterminer la présence d'une activité de parole à partir de certaines sélectionnées desdites valeurs moyennes.
  12. Procédé selon la revendication 11, comprenant, en outre, l'étape consistant à détecter l'énergie d'échantillons audio associés auxdits paramètres audio, avant ladite étape de détermination de la présence d'une activité de parole.
  13. Procédé selon la revendication 11 ou la revendication 12, et comprenant, en outre, l'étape préliminaire consistant à recevoir lesdits paramètres audio en provenance d'un générateur audio.
  14. Procédé selon la revendication 11, la revendication 12 ou la revendication 13, et comprenant, en outre, l'étape préliminaire consistant à produire lesdits paramètres audio à partir d'une pluralité d'échantillons audio.
EP01958309A 2000-03-15 2001-03-14 Appareil et procede de detection d'activite vocale Expired - Lifetime EP1269462B1 (fr)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
GB0006312 2000-03-15
GB0006312A GB2360428B (en) 2000-03-15 2000-03-15 Voice activity detection apparatus and method
PCT/IB2001/001603 WO2001080220A2 (fr) 2000-03-15 2001-03-14 Appareil et procede de detection d'activite vocale

Publications (2)

Publication Number Publication Date
EP1269462A2 EP1269462A2 (fr) 2003-01-02
EP1269462B1 true EP1269462B1 (fr) 2008-05-14

Family

ID=9887716

Family Applications (1)

Application Number Title Priority Date Filing Date
EP01958309A Expired - Lifetime EP1269462B1 (fr) 2000-03-15 2001-03-14 Appareil et procede de detection d'activite vocale

Country Status (6)

Country Link
EP (1) EP1269462B1 (fr)
AT (1) ATE395683T1 (fr)
AU (1) AU2001280027A1 (fr)
DE (1) DE60133998D1 (fr)
GB (1) GB2360428B (fr)
WO (1) WO2001080220A2 (fr)

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2317084B (en) * 1995-04-28 2000-01-19 Northern Telecom Ltd Methods and apparatus for distinguishing speech intervals from noise intervals in audio signals
US5774849A (en) * 1996-01-22 1998-06-30 Rockwell International Corporation Method and apparatus for generating frame voicing decisions of an incoming speech signal
US6385548B2 (en) * 1997-12-12 2002-05-07 Motorola, Inc. Apparatus and method for detecting and characterizing signals in a communication system
US6188981B1 (en) * 1998-09-18 2001-02-13 Conexant Systems, Inc. Method and apparatus for detecting voice activity in a speech signal

Also Published As

Publication number Publication date
GB0006312D0 (en) 2000-05-03
WO2001080220A3 (fr) 2002-05-23
ATE395683T1 (de) 2008-05-15
GB2360428B (en) 2002-09-18
AU2001280027A1 (en) 2001-10-30
WO2001080220A2 (fr) 2001-10-25
DE60133998D1 (de) 2008-06-26
GB2360428A (en) 2001-09-19
EP1269462A2 (fr) 2003-01-02

Similar Documents

Publication Publication Date Title
KR100883712B1 (ko) 음원 방향 추정 방법, 및 음원 방향 추정 장치
US6427134B1 (en) Voice activity detector for calculating spectral irregularity measure on the basis of spectral difference measurements
EP3493205B1 (fr) Procédé et appareil permettant de détecter de façon adaptative une activité vocale dans un signal audio d'entrée
EP0551803B1 (fr) Méthode de synchronisation et d'estimation de canal dans un système de radiocommunication à accès multiple et à division de temps
JP5006279B2 (ja) 音声活性検出装置及び移動局並びに音声活性検出方法
KR100770839B1 (ko) 음성 신호의 하모닉 정보 및 스펙트럼 포락선 정보,유성음화 비율 추정 방법 및 장치
JP3878482B2 (ja) 音声検出装置および音声検出方法
US8818811B2 (en) Method and apparatus for performing voice activity detection
US20030097257A1 (en) Sound signal process method, sound signal processing apparatus and speech recognizer
KR100735343B1 (ko) 음성신호의 피치 정보 추출장치 및 방법
US6876965B2 (en) Reduced complexity voice activity detector
EP1548703B1 (fr) Dispositif et procédé de détection de l'activité vocale
JP3418005B2 (ja) 音声ピッチ検出装置
EP1269462B1 (fr) Appareil et procede de detection d'activite vocale
KR20090080777A (ko) 신호 검출 방법 및 장치
JPH08221097A (ja) 音声成分の検出法
Beritelli et al. A low‐complexity speech‐pause detection algorithm for communication in noisy environments
CA2279264C (fr) Amelioration de l'insensibilite aux signaux vocaux dans un detecteur dtmf fonde sur la prediction lineaire
US5734679A (en) Voice signal transmission system using spectral parameter and voice parameter encoding apparatus and decoding apparatus used for the voice signal transmission system
JPS61184912A (ja) 定数可変型聴感的重み付けフイルタ
US6993478B2 (en) Vector estimation system, method and associated encoder
JPH0832527A (ja) フェージングピッチ推定装置
GB2437868A (en) Estimating noise power spectrum, sorting time frames, calculating the quantile and interpolating values over all remaining frequencies
JPH10304023A (ja) 電話機
JPH0311479B2 (fr)

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR

AX Request for extension of the european patent

Free format text: AL;LT;LV;MK;RO;SI

17P Request for examination filed

Effective date: 20021125

17Q First examination report despatched

Effective date: 20061201

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

Free format text: LANGUAGE OF EP DOCUMENT: FRENCH

REF Corresponds to:

Ref document number: 60133998

Country of ref document: DE

Date of ref document: 20080626

Kind code of ref document: P

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080825

NLV1 Nl: lapsed or annulled due to failure to fulfill the requirements of art. 29p and 29m of the patents act
PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20081014

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080814

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20090217

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090331

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: FR

Ref legal event code: ST

Effective date: 20091130

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090314

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090331

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090331

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20091123

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080815

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20090314

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20080514

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20200327

Year of fee payment: 20

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20200528

Year of fee payment: 20

REG Reference to a national code

Ref country code: DE

Ref legal event code: R071

Ref document number: 60133998

Country of ref document: DE

REG Reference to a national code

Ref country code: GB

Ref legal event code: PE20

Expiry date: 20210313

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION

Effective date: 20210313