EP1058235B1 - Procédé de reproduction pour systèmes contrôlés par la voix avec synthèse de la parole basée sur texte - Google Patents

Procédé de reproduction pour systèmes contrôlés par la voix avec synthèse de la parole basée sur texte Download PDF

Info

Publication number
EP1058235B1
EP1058235B1 EP00108486A EP00108486A EP1058235B1 EP 1058235 B1 EP1058235 B1 EP 1058235B1 EP 00108486 A EP00108486 A EP 00108486A EP 00108486 A EP00108486 A EP 00108486A EP 1058235 B1 EP1058235 B1 EP 1058235B1
Authority
EP
European Patent Office
Prior art keywords
characters
train
variation
converted
speech input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP00108486A
Other languages
German (de)
English (en)
Other versions
EP1058235A3 (fr
EP1058235A2 (fr
Inventor
Peter Buth
Frank Dufhues
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nokia Oyj
Original Assignee
Nokia Oyj
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nokia Oyj filed Critical Nokia Oyj
Publication of EP1058235A2 publication Critical patent/EP1058235A2/fr
Publication of EP1058235A3 publication Critical patent/EP1058235A3/fr
Application granted granted Critical
Publication of EP1058235B1 publication Critical patent/EP1058235B1/fr
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • G10L13/04Details of speech synthesis systems, e.g. synthesiser structure or memory management

Definitions

  • the invention is concerned with the improvement of voice-controlled systems text-based speech synthesis, especially with the improvement of synthetic Playback of saved but determined during pronunciation Strings subject to peculiarities.
  • the subject of speech synthesis is the mechanical transformation of the symbolic Representation of an utterance in an acoustic signal emitted by one human speaker recognized as sufficiently similar to human language becomes.
  • a speech synthesis technique is a technique that involves building a speech synthesizer allowed.
  • Examples of speech synthesis techniques are direct synthesis, the Synthesis using a model and simulation of the vocal tract.
  • either parts of the speech signal are started from stored signal pieces (e.g. one per phoneme) to the corresponding ones Words put together or the transfer function of the vocal tract, which at People used for speech generation by energy of a signal in simulated certain frequency ranges.
  • stored signal pieces e.g. one per phoneme
  • Words put together e.g. one per phoneme
  • the transfer function of the vocal tract which at People used for speech generation by energy of a signal in simulated certain frequency ranges.
  • Voiced sounds through a quasi-periodic excitation of a certain frequency represents.
  • the phoneme mentioned above is the smallest meaning-distinguishing, but itself not meaningful unity of language.
  • Two words different Meaning that only differ by a phoneme e.g. fish - table; forest - Wild
  • form a minimal pair e.g. fish - table; forest - Wild
  • the number of phonemes in a language is proportional small (between 20 and 60).
  • Diphones are mostly used in direct synthesis.
  • Phonemes or sequences of phonemes are created with the help of the International Phonetic alphabet (IPA) noted.
  • IPA International Phonetic alphabet
  • the implementation of a text in a sequence Signs of the Phonetic Alphabet is called Phonetic Transcription designated.
  • a production model is formed, which mostly on minimizing the difference between a digitized human Speech signal (original signal) and a predicted signal based.
  • Another method consists in the simulation of the vocal tract, in which its Shape and position of the individual articulation organs (tongue, jaw, lips) is reproduced.
  • a mathematical model of the flow conditions is used generated in such a defined vocal tract and the speech signal with the help of this model.
  • the phonemes or diphones used in direct synthesis must first can be obtained from natural language by segmentation. Here you can two approaches can be distinguished:
  • features For segmentation, features must first be extracted from the speech signal, on the basis of which it is possible to distinguish between the segments. These characteristics are then classified into classes.
  • Features for feature extraction include spectral analysis, filter bank analysis or the method of linear prediction.
  • Hidden Markov models for the classification, for example, Hidden Markov models, artificial ones neural networks or dynamic time warping (a method for time normalization) to be used.
  • HMM Hidden Markov Model
  • a common approach is the classification of voiced / unvoiced / silent - according to the different forms of stimulation in the generation of language in Vocal tract.
  • the special treatment of certain words of a language is extremely complex, it has been used in speech-controlled arrangements to form the announcement, which an arrangement is to indicate, from a mix of spoken and synthesized language.
  • the desired destination is recorded for a route finder, for example, which has special features in terms of pronunciation compared to the other words of the corresponding language and which is specified by a user in the case of voice-controlled arrangements, and copied into the corresponding destination announcement.
  • a route finder for example, which has special features in terms of pronunciation compared to the other words of the corresponding language and which is specified by a user in the case of voice-controlled arrangements, and copied into the corresponding destination announcement.
  • the procedure is simplified if, according to claim 4 Segmentation of the speech input and the converted character string or the variants formed therefrom. This segmentation allows segments in where no differences or differences below the threshold are found, of exclude further treatment.
  • segmentation approaches can also be used become. This is especially true when looking at the original voice input Advantages because the segmentation contains those contained in the speech signal and information can only be used in a very complex step must, while the segmentation of strings very simply the well-known Number of phonemes contained in the utterance can be used.
  • a particularly simple procedure is achieved if according to claim 9 at least one replacement phoneme similar to this phoneme is linked to each phoneme or stored in a list.
  • the computing work is further reduced if according to claim 10 at a variant of a character string that is determined to be reproducible, the special features, associated with rendering the string, along with the String can be saved. In this case there is the special pronunciation the respective character string if you use it later without much effort available in the memory.
  • strings can be one
  • route finders are street or place names.
  • a mailbox application can do this like in a phone book the names of subscribers his. So that the memory easily with the appropriate information loaded or the stored information can be easily updated, the respective strings are available as text.
  • a memory is designated by 10.
  • This memory 10 which for the Representation of the invention, which should contain German city names, belongs to one Route finder 11.
  • this route finder 11 comprises an arrangement 12 with which natural voice inputs are recorded and temporarily saved can. In the present case, this is realized in such a way that the respective voice input is from one Microphone 13 is detected and stored in a voice memory 14. Now becomes a The user of the route finder 11 is asked to enter his destination, the each destination spoken by the user z. B. "Bochum” or "Itzehoe” from Microphone 13 detected and forwarded to the voice memory 14.
  • the route finder 11 Because the route finder 11 has either been informed of his current location or has still been given it knows, he is first based on the desired destination and the current one Determine the corresponding route to the destination. If the route finder 11 not only show the corresponding route graphically, but spoken ones Deliver announcement, the textual strings of the respective announcement Described phonetically according to general rules and then for speech converted into a purely synthetic form. In that shown in Fig. 1 Exemplary embodiment is the phonetic description of the stored character strings in the converter 15 and the synthesizing in the subsequent ' Speech synthesizing arrangement 16.
  • the respective character string if it has passed through the converter 15 and the speech synthesizer assembly 16 as a word corresponding to the phonetic conditions of the respective language a speaker 17 given to the environment and of this as such be understood.
  • the Play route finder 11 after entering the destination approximately the following sentence: "You have Berlin chosen as the destination. If this does not meet your expectations, give it now set a new goal. "Even though this information follows general rules Problems can be reproduced phonetically, if that The goal should not be Berlin, but Laboe. If the string that the Textual representation of the destination Laboe in the converter 15 according to general rules written phonetically and then in the speech synthesizer 16 Output through speaker 17 like the rest of the information above in brought a synthetic form, that would be given over the speaker 17 Correct result only if, according to general rules, the ending "oe” is generally reproduced as "ö".
  • a comparison arrangement 18 becomes the destination actually spoken by the user and the the character string corresponding to the destination after the converter 15 and the Speech synthesizer 16 has passed, fed and then compared. If the synthesized string shows a high - above a threshold lying - coincidence with the originally spoken destination, is for the Playback uses the synthesized string. Can this match are not ascertained, a variant becomes in the speech synthesis arrangement 16 of the original character string and again in the comparator 18 Comparison between the originally spoken destination and the variant formed carried out.
  • the route finder 11 designed such that as soon as a character string or a Variant has the required agreement with the original, whose Playback via the loudspeaker 17 takes place, further variant formations stopped immediately.
  • the route finder 11 can also be modified such that a A plurality of variants are formed and then one from the variants Variant is selected that most closely matches the original shows.
  • FIG. 2a there is a speech signal in the time domain of actually represented by Itzehoe spoken by a user.
  • Fig. 2b also shows a speech signal in the time domain of the word Itzehoe, but in in Fig. 2b the word Itzehoe from a corresponding present Character string first in the converter 15 phonetically according to general rules described and then subsequently in the speech synthesis arrangement 16 in a synthetic form.
  • the illustration according to FIG. 2b is clear it can be seen that, when applying the general rules, the ending "oe" des Word Itzehoe is reproduced as "ö". However, this incorrect playback exclude the spoken and the synthesized form in one Comparator 18 compared with each other.
  • the converter 15 'in another - not shown - embodiment of the converter 15th can be formed.
  • the process sequence can also be modified. It is found that a Deviation between the spoken and the original synthetic form is given, and there are a plurality of replacement phonemes in the memory 21 stored list, a plurality of variants can also be formed at the same time and be compared to the actual spoken word. Is played then the variant that most closely matches what is spoken Word shows.
  • the additional memory 22 not just on the inclusion of information on the correct pronunciation of stored strings is limited.
  • a comparison in Comparator 18 that between the spoken and the synthesized form of a Word no deviation or below a threshold can be stored in the additional memory 22 for this word, which in the future use of this word an elaborate comparison in Comparator 18 excludes.
  • segment 20.1 has a greater width compared to segment 19.1, while the segment 20.2 compared to the corresponding segment 19.2. essential is narrower.
  • the comparison arrangement 18 is designed so that different lengths Speaking times of a phoneme do not yet indicate a mutual deviation.
  • segment 19, 20 when using different segmentation methods for the spoken and the synthesized Format also a different number of segments 19, 20 can be calculated can. If this occurs, then a certain segment 19, 20 should not only have one correspond to segment 19, 20, but also with the Predecessor and successor of the corresponding segment 19, 20. So it is also possible to replace one phoneme with two other phonemes. This procedure is also possible in the opposite direction. There is no match for a segment 19, 20, this can be excluded, or by two better fitting ones be replaced.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Machine Translation (AREA)
  • Document Processing Apparatus (AREA)
  • Input Circuits Of Receivers And Coupling Of Receivers And Audio Equipment (AREA)

Claims (10)

  1. Procédé de reproduction pour systèmes contrôlés par la voix, avec synthèse de la parole basée sur texte,
    caractérisé en ce que,
    en cas de présence d'une introduction vocale, ayant été effectivement énoncée et correspondant à une chaíne de caractères mémorisée, avant une reproduction de la chaíne de caractères, décrite phonétiquement d'après des règles générales et convertie en une forme purement synthétique, la chaíne de caractères convertie est comparée à l'introduction vocale,
    en ce que, en cas de constatation d'un écart, supérieur à un seuil, de la chaíne de caractères convertie, par rapport à l'introduction vocale, au moins une variante de la chaíne de caractères convertie est formée, et
    en ce que l'une des variantes formées, dans la mesure où celle-ci présente un écart inférieur à un seuil, lors d'une comparaison faite avec l'introduction vocale, est éditée au lieu de la chaíne de caractères convertie.
  2. Procédé de reproduction selon la revendication 1,
    caractérisé en ce qu'une seule variante est toujours formée à l'étape deux, et
    en ce que, dans la mesure où, à l'étape trois, une comparaison entre la variante et l'introduction vocale montre toujours un écart supérieur au seuil, l'étape deux est exécutée encore une fois, dans le but de former une nouvelle variante.
  3. Procédé de reproduction selon la revendication 1,
    caractérisé en ce qu'au moins deux variantes sont formées à l'étape deux, et
    en ce que, en cas de présence de variantes, ayant chaque fois un écart inférieur au seuil par rapport à l'introduction vocale, on reproduit toujours la variante qui présente l'écart minimal par rapport à l'introduction vocale.
  4. Procédé selon l'une des revendications 1 à 3,
    caractérisé en ce que,
    avant de procéder à une comparaison entre l'introduction vocale et la chaíne de caractères convertie, respectivement la ou les variantes formées à partir de celle-ci, on effectue une segmentation de l'introduction vocale et de la chaíne de caractères convertie ou de la ou des variantes formées.
  5. Procédé de reproduction selon la revendication 4,
    caractérisé en ce qu'
    on utilise un échantillon de segmentation identique, pour segmenter tant l'introduction vocale que la chaíne de caractères convertie, respectivement la ou les variantes en étant dérivées.
  6. Procédé de reproduction selon la revendication 4,
    caractérisé en ce que,
    tant pour la segmentation de l'introduction vocale que de la chaíne de caractères convertie, respectivement de la ou des variantes en étant dérivées, on utilise chaque fois un échantillon de segmentation différent.
  7. Procédé de reproduction selon la revendication 4,
    caractérisé en ce que,
    pour segmenter la chaíne de caractères convertie, respectivement la ou les variantes en étant dérivées, on utilise un échantillon de segmentation explicite et, pour effectuer la segmentation de l'introduction vocale, on utilise un échantillon de segmentation implicite.
  8. Procédé de reproduction selon l'une des revendications 4 à 7,
    caractérisé en ce que
    la chaíne de caractères convertie, se présentant sous forme segmentée, et l'introduction de caractères segmentés, dans les segments correspondants, sont examinées pour trouver des éléments communs, et
    en ce que, si dans deux segments correspondants, on se trouve en présence d'un écart supérieur à une valeur de seuil, le phonème se présentant dans le segment de la chaíne de caractères convertie est remplacé par un phonème de substitution.
  9. Procédé de reproduction selon la revendication 8,
    caractérisé en ce qu'
    au moins un phonème de substitution analogue à ce phonème est combiné à chaque phonème.
  10. Procédé de reproduction selon l'une des revendications 1 à 9,
    caractérisé en ce que,
    dès qu'une variante d'une chaíne de caractères a été déterminée comme étant digne de reproduction, les particularités liées à la reproduction de la chaíne de caractères sont mémorisées, en liaison avec la chaíne de caractères.
EP00108486A 1999-05-05 2000-04-19 Procédé de reproduction pour systèmes contrôlés par la voix avec synthèse de la parole basée sur texte Expired - Lifetime EP1058235B1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
DE19920501 1999-05-05
DE19920501A DE19920501A1 (de) 1999-05-05 1999-05-05 Wiedergabeverfahren für sprachgesteuerte Systeme mit textbasierter Sprachsynthese

Publications (3)

Publication Number Publication Date
EP1058235A2 EP1058235A2 (fr) 2000-12-06
EP1058235A3 EP1058235A3 (fr) 2003-02-05
EP1058235B1 true EP1058235B1 (fr) 2003-11-05

Family

ID=7906935

Family Applications (1)

Application Number Title Priority Date Filing Date
EP00108486A Expired - Lifetime EP1058235B1 (fr) 1999-05-05 2000-04-19 Procédé de reproduction pour systèmes contrôlés par la voix avec synthèse de la parole basée sur texte

Country Status (5)

Country Link
US (1) US6546369B1 (fr)
EP (1) EP1058235B1 (fr)
JP (1) JP4602511B2 (fr)
AT (1) ATE253762T1 (fr)
DE (2) DE19920501A1 (fr)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4759827B2 (ja) * 2001-03-28 2011-08-31 日本電気株式会社 音声セグメンテーション装置及びその方法並びにその制御プログラム
US7107215B2 (en) * 2001-04-16 2006-09-12 Sakhr Software Company Determining a compact model to transcribe the arabic language acoustically in a well defined basic phonetic study
AT6920U1 (de) 2002-02-14 2004-05-25 Sail Labs Technology Ag Verfahren zur erzeugung natürlicher sprache in computer-dialogsystemen
DE10253786B4 (de) * 2002-11-19 2009-08-06 Anwaltssozietät BOEHMERT & BOEHMERT GbR (vertretungsberechtigter Gesellschafter: Dr. Carl-Richard Haarmann, 28209 Bremen) Verfahren zur rechnergestützten Ermittlung einer Ähnlichkeit eines elektronisch erfassten ersten Kennzeichens zu mindestens einem elektronisch erfassten zweiten Kennzeichen sowie Vorrichtung und Computerprogramm zur Durchführung desselben
EP1475611B1 (fr) * 2003-05-07 2007-07-11 Harman/Becker Automotive Systems GmbH Procédé et appareil d'application de sortie vocale, support de donées comprenant des donées de parole
JP4714694B2 (ja) * 2003-11-05 2011-06-29 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ 音声−テキストトランスクリプションシステムの誤り検出
JP2006047866A (ja) * 2004-08-06 2006-02-16 Canon Inc 電子辞書装置およびその制御方法
US20060136195A1 (en) * 2004-12-22 2006-06-22 International Business Machines Corporation Text grouping for disambiguation in a speech application
JP4385949B2 (ja) * 2005-01-11 2009-12-16 トヨタ自動車株式会社 車載チャットシステム
US20070016421A1 (en) * 2005-07-12 2007-01-18 Nokia Corporation Correcting a pronunciation of a synthetically generated speech object
US20070129945A1 (en) * 2005-12-06 2007-06-07 Ma Changxue C Voice quality control for high quality speech reconstruction
US8504365B2 (en) * 2008-04-11 2013-08-06 At&T Intellectual Property I, L.P. System and method for detecting synthetic speaker verification
US8380503B2 (en) 2008-06-23 2013-02-19 John Nicholas and Kristin Gross Trust System and method for generating challenge items for CAPTCHAs
US8752141B2 (en) 2008-06-27 2014-06-10 John Nicholas Methods for presenting and determining the efficacy of progressive pictorial and motion-based CAPTCHAs
US9564120B2 (en) * 2010-05-14 2017-02-07 General Motors Llc Speech adaptation in speech synthesis
KR20170044849A (ko) * 2015-10-16 2017-04-26 삼성전자주식회사 전자 장치 및 다국어/다화자의 공통 음향 데이터 셋을 활용하는 tts 변환 방법

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE2435654C2 (de) * 1974-07-24 1983-11-17 Gretag AG, 8105 Regensdorf, Zürich Verfahren und Vorrichtung zur Analyse und Synthese von menschlicher Sprache
NL8302985A (nl) * 1983-08-26 1985-03-18 Philips Nv Multipulse excitatie lineair predictieve spraakcodeerder.
US5029200A (en) * 1989-05-02 1991-07-02 At&T Bell Laboratories Voice message system using synthetic speech
US5293449A (en) * 1990-11-23 1994-03-08 Comsat Corporation Analysis-by-synthesis 2,4 kbps linear predictive speech codec
GB9223066D0 (en) * 1992-11-04 1992-12-16 Secr Defence Children's speech training aid
FI98163C (fi) * 1994-02-08 1997-04-25 Nokia Mobile Phones Ltd Koodausjärjestelmä parametriseen puheenkoodaukseen
US6005549A (en) * 1995-07-24 1999-12-21 Forest; Donald K. User interface method and apparatus
US5913193A (en) * 1996-04-30 1999-06-15 Microsoft Corporation Method and system of runtime acoustic unit selection for speech synthesis
JPH10153998A (ja) * 1996-09-24 1998-06-09 Nippon Telegr & Teleph Corp <Ntt> 補助情報利用型音声合成方法、この方法を実施する手順を記録した記録媒体、およびこの方法を実施する装置
US6163769A (en) * 1997-10-02 2000-12-19 Microsoft Corporation Text-to-speech using clustered context-dependent phoneme-based units
US6081780A (en) * 1998-04-28 2000-06-27 International Business Machines Corporation TTS and prosody based authoring system
US6173263B1 (en) * 1998-08-31 2001-01-09 At&T Corp. Method and system for performing concatenative speech synthesis using half-phonemes
US6266638B1 (en) * 1999-03-30 2001-07-24 At&T Corp Voice quality compensation system for speech synthesis based on unit-selection speech database

Also Published As

Publication number Publication date
DE19920501A1 (de) 2000-11-09
US6546369B1 (en) 2003-04-08
JP2000347681A (ja) 2000-12-15
EP1058235A3 (fr) 2003-02-05
EP1058235A2 (fr) 2000-12-06
ATE253762T1 (de) 2003-11-15
DE50004296D1 (de) 2003-12-11
JP4602511B2 (ja) 2010-12-22

Similar Documents

Publication Publication Date Title
DE60203705T2 (de) Umschreibung und anzeige eines eingegebenen sprachsignals
DE602004012909T2 (de) Verfahren und Vorrichtung zur Modellierung eines Spracherkennungssystems und zur Schätzung einer Wort-Fehlerrate basierend auf einem Text
EP1058235B1 (fr) Procédé de reproduction pour systèmes contrôlés par la voix avec synthèse de la parole basée sur texte
EP1466317B1 (fr) Procede d&#39;exploitation d&#39;un systeme de reconnaissance vocale automatique pour la reconnaissance vocale multilocuteur de mots de differentes langues et systeme de reconnaissance vocale automatique
DE602005002706T2 (de) Verfahren und System für die Umsetzung von Text-zu-Sprache
DE69937176T2 (de) Segmentierungsverfahren zur Erweiterung des aktiven Vokabulars von Spracherkennern
DE69832393T2 (de) Spracherkennungssystem für die erkennung von kontinuierlicher und isolierter sprache
DE60111329T2 (de) Anpassung des phonetischen Kontextes zur Verbesserung der Spracherkennung
DE19610019C2 (de) Digitales Sprachsyntheseverfahren
EP0925578B1 (fr) Systeme et procede de traitement de la parole
DE19825205C2 (de) Verfahren, Vorrichtung und Erzeugnis zum Generieren von postlexikalischen Aussprachen aus lexikalischen Aussprachen mit einem neuronalen Netz
DE19847419A1 (de) Verfahren zur automatischen Erkennung einer buchstabierten sprachlichen Äußerung
EP1273003B1 (fr) Procede et dispositif de determination de marquages prosodiques
EP3010014B1 (fr) Procede d&#39;interpretation de reconnaissance vocale automatique
DE19942178C1 (de) Verfahren zum Aufbereiten einer Datenbank für die automatische Sprachverarbeitung
DE112006000322T5 (de) Audioerkennungssystem zur Erzeugung von Antwort-Audio unter Verwendung extrahierter Audiodaten
EP1264301B1 (fr) Procede pour reconnaitre les enonces verbaux de locuteurs non natifs dans un systeme de traitement de la parole
DE60108104T2 (de) Verfahren zur Sprecheridentifikation
DE60018696T2 (de) Robuste sprachverarbeitung von verrauschten sprachmodellen
EP1282897B1 (fr) Procede pour produire une banque de donnees vocales pour un lexique cible pour l&#39;apprentissage d&#39;un systeme de reconnaissance vocale
DE10040063A1 (de) Verfahren zur Zuordnung von Phonemen
DE102010040553A1 (de) Spracherkennungsverfahren
EP1435087B1 (fr) Procede de production de segments de reference decrivant des blocs vocaux et procede de modelisation d&#39;unites vocales d&#39;un modele de test parle
EP0058130B1 (fr) Procédé pour la synthèse de la parole avec un vocabulaire illimité et dispositif pour la mise en oeuvre dudit procédé
DE4111781A1 (de) Computersystem zur spracherkennung

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Free format text: AL;LT;LV;MK;RO;SI

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: NOKIA CORPORATION

PUAL Search report despatched

Free format text: ORIGINAL CODE: 0009013

AK Designated contracting states

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AX Request for extension of the european patent

Extension state: AL LT LV MK RO SI

17P Request for examination filed

Effective date: 20030120

GRAH Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOS IGRA

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AKX Designation fees paid

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20031105

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT;WARNING: LAPSES OF ITALIAN PATENTS WITH EFFECTIVE DATE BEFORE 2007 MAY HAVE OCCURRED AT ANY TIME BEFORE 2007. THE CORRECT EFFECTIVE DATE MAY BE DIFFERENT FROM THE ONE RECORDED.

Effective date: 20031105

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20031105

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20031105

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

Free format text: NOT ENGLISH

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

REF Corresponds to:

Ref document number: 50004296

Country of ref document: DE

Date of ref document: 20031211

Kind code of ref document: P

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

Free format text: GERMAN

GBT Gb: translation of ep patent filed (gb section 77(6)(a)/1977)

Effective date: 20031224

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20040205

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20040205

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20040205

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20040216

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20040419

Ref country code: AT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20040419

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20040430

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20040430

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20040430

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20040430

REG Reference to a national code

Ref country code: IE

Ref legal event code: FD4D

ET Fr: translation filed
PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20040806

BERE Be: lapsed

Owner name: *NOKIA CORP.

Effective date: 20040430

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: PT

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20040405

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20120425

Year of fee payment: 13

Ref country code: NL

Payment date: 20120413

Year of fee payment: 13

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20120504

Year of fee payment: 13

Ref country code: GB

Payment date: 20120418

Year of fee payment: 13

REG Reference to a national code

Ref country code: NL

Ref legal event code: V1

Effective date: 20131101

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20130419

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20131101

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20130419

REG Reference to a national code

Ref country code: FR

Ref legal event code: ST

Effective date: 20131231

REG Reference to a national code

Ref country code: DE

Ref legal event code: R119

Ref document number: 50004296

Country of ref document: DE

Effective date: 20131101

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: FR

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20130430

Ref country code: NL

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20131101