US7970610B2 - Speech recognition - Google Patents
Speech recognition Download PDFInfo
- Publication number
- US7970610B2 US7970610B2 US10/472,897 US47289703A US7970610B2 US 7970610 B2 US7970610 B2 US 7970610B2 US 47289703 A US47289703 A US 47289703A US 7970610 B2 US7970610 B2 US 7970610B2
- Authority
- US
- United States
- Prior art keywords
- data items
- category
- data
- stored
- uncommon
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000000034 method Methods 0.000 claims description 25
- 230000008901 benefit Effects 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 3
- 230000004044 response Effects 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 230000014616 translation Effects 0.000 description 2
- 238000004833 X-ray photoelectron spectroscopy Methods 0.000 description 1
- 230000000135 prohibitive effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/487—Arrangements for providing information services, e.g. recorded voice services or time announcements
- H04M3/493—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
- H04M3/4931—Directory assistance systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L2015/085—Methods for reducing search complexity, pruning
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/226—Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
- G10L2015/228—Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics of application context
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M2201/00—Electronic components, circuits, software, systems or apparatus used in telephone systems
- H04M2201/40—Electronic components, circuits, software, systems or apparatus used in telephone systems using speech recognition
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/487—Arrangements for providing information services, e.g. recorded voice services or time announcements
- H04M3/493—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
- H04M3/4936—Speech interaction details
Definitions
- the present invention is concerned with speech recognition, particularly although not exclusively for use in automated voice-interactive services for use over a telephone network.
- a typical application is an enquiry service where a user is asked a number of questions in order to elicit replies which, after recognition by a speech recogniser, permit access to one or more desired entries in an information bank.
- An example of this is a directory enquiry system in which a user, requiring the telephone number of a customer, is asked to give the town name and road name of the subscriber's address, and the customer's surname.
- a disadvantage of such a system is that if the correct town is not identified as being one of the closest matches then the enquiry is bound to result in failure.
- a device having corresponding apparatus features to the method features of any of claims 1 to 5 .
- a carrier medium as set out in claim 10 .
- FIG. 1 illustrates an architecture for a directory enquiries system
- FIG. 2 is a flow chart illustrating the operation of the directory enquiries system of FIG. 1 using the method according to the present invention
- FIG. 3 is a second flowchart illustrating the operation of the directory enquiries system of FIG. 1 in using a second embodiment of a method according to the present invention
- FIG. 4 is a flow chart illustrating a method of generating association between surnames which do not have an audio representation stored in the store 8 of FIG. 1 and surnames which do have an audio representation stored in the store 8 .
- FIG. 5 is a flow chart illustrating a second method of generating association between surnames which do not have an audio representation stored in the store 8 of FIG. 1 and surnames which do have an audio representation stored in the store 8 .
- a speech synthesiser 1 is provided for providing announcements to a user via a telephone line interface 2 , by reference to stored, fixed messages in a message data store 3 , or from variable information supplied to it by a main control unit 4 .
- Incoming speech signals from the telephone line interface 2 are conducted to a speech recogniser 5 which is able to recognise spoken words by reference to, respectively, town name, road name or surname recognition data in recognition data stores of 6 , 7 , 8 .
- a main directory database 9 contains, for each telephone customer in the area covered by the directory enquiry service, an entry containing the name, address and telephone number of that customer, in text form.
- the town name recognition data store 6 contains, in text form, the names of all the towns included in the directory database 9 , along with stored data to enable the speech recogniser 5 to recognise those town names in the speech signal received from the telephone line interface 2 .
- the recogniser 5 operates by recognising distinct phonemes in the input speech, which are decoded by reference to stored audio representations in the store 6 representing a tree structure constructed in advance from phonetic translations of the town names stored in the store 6 , decoded by means of a Viterbi algorithm.
- the stores 7 , 8 for road name recognition data and surname recognition data are organised in the same manner.
- the audio representation may equally well be stored in a separate store which is referenced via data in stores 6 , 7 and 8 .
- the audio representation of each phoneme referenced by the stores 6 , 7 and 8 needs only to be stored once in said separate store
- Each entry in the town data store 6 contains, as mentioned above, text corresponding to each of the town names appearing in the database 9 , to act as a label to link the entry in the store 6 to entries in the database 9 (though other kinds of label may be used if preferred).
- the store 6 may contain an entry for every town name that the user might use to refer to geographical locations covered by the database, whether or not all these names are actually present in the database. Noting that some town names are not unique (there are four towns in the UK called Southend), and that some town names carry the same significance (e.g.
- a vocabulary equivalence store 39 is also provided, containing such equivalents, which can be consulted following each recognition of a town name, to return additional possibilities to the set of town names considered to be recognised. For example if “Hammersmith” is recognised, London is added to the set; if “Southend” is recognised, then Southend-on-Sea, Southend (Campbeltown), Southend (Swansea) and Southend (Reading) are added.
- the equivalence data store 39 could, if desired, contain similar information for roads and surnames, or first names if these are used; for example Dave and David are considered to represent the same name.
- the vocabulary equivalence data store 39 may act as a translation between labels used in the name stores 6 , 7 , 8 and the labels used in the database (whether or not the labels are names in text form).
- each leaf in the tree may have one or more textual labels attached to it.
- Attaching several textual labels to a particular leaf in the tree is a known technique for dealing with equivalent ways of referring to the same item of data in a database as described above.
- the technique may also be used for dealing with homophones (words which are pronounced in the same way but spelled differently) for example, “Smith” and “Smyth”.
- the recognition data store 8 contains audio representations of about 50 thousand surnames which correspond to the surnames of about 90% of the population of the UK.
- Several textual labels are associated with a particular audio representation by attaching textual labels to a particular leaf in a tree. These textual labels represent surnames which sound similar to said particular audio representation. Therefore a list of surnames are provided which sound similar to the surname which is represented by a particular audio representation, but which are not themselves represented by audio data in the store 8 . Therefore a greater number of surnames are represented by a smaller data structure, thus reducing the amount of memory required. Furthermore the amount of processing power required is much less and it is possible to perform the speech recognition in real time, using a less powerful processor.
- the operation of the directory enquiry system of FIG. 1 is illustrated in the flow chart of FIG. 2 .
- the process starts ( 10 ) upon receipt of an incoming telephone call signalled to the control unit 4 by the telephone line interface 2 ; the control unit responds by instructing the speech synthesiser 1 to play ( 11 ) a message stored in the message store 3 requesting the caller to give the required surname.
- the caller's response is received ( 12 ) by the recogniser.
- the recogniser 3 then performs its recognition process ( 13 ) with reference to the audio representations stored in the store 8 .
- For common surnames which meet a prescribed threshold of similarity with the received reply any associated uncommon surnames are determined ( 14 ) by reference to the town recognition data store 6 . All of the common surnames which meet a prescribed threshold of similarity with the received reply, together with any uncommon surnames which are associated with the audio representations of these common surnames are then communicated to the control unit 4 .
- the control unit 4 then instructs the speech synthesiser to play ( 15 ) a further message from the message data store 3 requesting the required street name.
- a further response, relating to the street name, is received ( 17 ) from the caller and is processed by the recogniser 3 utilising the data store 7 and the recogniser then communicates to the control unit 4 a set of all of the road names which meet a prescribed threshold of similarity with the received reply.
- the control unit 4 retrieves ( 20 ) from the database 9 a list of all customers having any of the surnames in the set of surnames received by the control unit at step 14 and residing in any of the street names received by the control unit at step 18 .
- the speech signal received at step 12 is an utterance of the uncommon surname ‘Dobson’.
- the set of words which meet the prescribed threshold of similarity with the received reply includes the common surname ‘Robson’.
- ‘Robson’ is associated with similar sounding surnames ‘Hobson, Dobson and Fobson’.
- the speech signal received at step 17 is an utterance of the street name ‘Dove Street’.
- the set of words which meet the prescribed threshold of similarity with the received reply includes the street name ‘Dove Street’.
- the database retrieval at step 22 retrieves the details for customer ‘Dobson’ in ‘Dove Street’ even though the name recognition data store 8 does not contain an audio representation for the name ‘Dobson’.
- the directory enquiries system would operate as illustrated in FIG. 3 , where further information relating to the town name is requested from the caller at step 19 .
- a further response, relating to the town name is received ( 20 ) from the caller and is processed ( 21 ) by the recogniser 3 utilising the data store 6 and the recogniser then communicates to the control unit 4 a set of all of the town names which meet a prescribed threshold of similarity with the received reply.
- This set of town name data is then used, along with street name and surname data in the database retrieval step 22 . If data relating to more than one customer is retrieved from the database then further information may be elicited from the user (steps not shown).
- the speech recogniser 5 provides a score as to how well each utterance matches each audio representation. This score is used to decide which customer data is more likely in the case where data relating to more than one customer is retrieved from the database. In the case of associated uncommon surname the score used can be weighted according to statistics relating to that surname such that the more uncommon a surname is the smaller the weighting factor applied to the score from the recogniser 5 .
- FIG. 4 is a flow chart illustrating a method of generating associations between uncommon surnames and common surnames for use in this invention.
- a speech utterance of a known uncommon surname is received by a speech recogniser, which may be any type of speech recogniser including a phoneme based speech recogniser as described earlier.
- the received speech utterance is compared with audio representations of the common surnames at step 31 , and at step 32 an association is made between the known uncommon surname and the common surname to which the speech recogniser determines that the unknown surname is most similar.
- FIG. 5 illustrates an alternative method of generating associations between uncommon and common surnames for use in the invention.
- a textual representation of an uncommon surname is received.
- this textual representation is converted into a phoneme sequence.
- Such a conversion my be done using a large database associating text to phoneme sequences. The conversion also may be done using letter to sound rules for example as described in Klatt D, ‘Review of text-to-speech conversion for English’, J acoustic Soc Am 82, No. 3 pp 737-793. September 1987.
- the phoneme sequence representing the uncommon surname is then compared to all the phoneme sequences for common surnames for example using a dynamic programming technique such as that described in “Predictive Assessment for Speaker Independent Isolated Word Recognisers” Alison Simons, ESCA EUROSPEECH 95 Madrid 1995 pp 1465-1467. Then at step 43 the uncommon surname is associated with the common surname for which the phonemic sequences are found to be most similar.
- association may be recorded by associating a label representing the known uncommon surname to a leaf in the common surname recognition tree, if a tree based phoneme recogniser is to be used in the directory enquiries system, or by use of a vocabulary equivalence store as discussed previously.
- An advantage of the second technique is that it is not necessary to collect speech data relating to all of the possible uncommon surnames in the database, which is a time consuming exercise. Instead all that is needed is a textual representation of such uncommon surnames.
- a phoneme confusion matrix which records the likelihood of a particular recogniser confusing each phoneme with every other phoneme. Such a matrix is used in the comparison step 42 as described in the above referenced paper.
Landscapes
- Engineering & Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Machine Translation (AREA)
- Telephonic Communication Services (AREA)
Abstract
Description
-
- (i) the user speaks the name of a town;
- (ii) a speech recogniser, by reference to stored town data identifies several towns as having the closest matches to the spoken town name, and produces a “score” or probability indicating the closeness of the match;
- (iii) a list is compiled of all road names occurring in the identified towns;
- (iv) the user speaks the name of a road;
- (v) the speech recogniser identifies several road names, of the ones in the list, having the closest matches to the spoken road name, again with scores;
- (vi) the road scores are each weighted accordingly to the score obtained for the town the road is located in, and the most likely “road” result considered to be the one with the best weighted score.
Claims (10)
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP01303598 | 2001-04-19 | ||
EP01303598.5 | 2001-04-19 | ||
EP01303598 | 2001-04-19 | ||
PCT/GB2002/001748 WO2002086863A1 (en) | 2001-04-19 | 2002-04-15 | Speech recognition |
Publications (2)
Publication Number | Publication Date |
---|---|
US20040117182A1 US20040117182A1 (en) | 2004-06-17 |
US7970610B2 true US7970610B2 (en) | 2011-06-28 |
Family
ID=8181903
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/472,897 Expired - Fee Related US7970610B2 (en) | 2001-04-19 | 2002-04-15 | Speech recognition |
Country Status (5)
Country | Link |
---|---|
US (1) | US7970610B2 (en) |
EP (1) | EP1397797B1 (en) |
CA (1) | CA2440463C (en) |
DE (1) | DE60222413T2 (en) |
WO (1) | WO2002086863A1 (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090248415A1 (en) * | 2008-03-31 | 2009-10-01 | Yap, Inc. | Use of metadata to post process speech recognition output |
US9583107B2 (en) | 2006-04-05 | 2017-02-28 | Amazon Technologies, Inc. | Continuous speech transcription performance indication |
US9973450B2 (en) | 2007-09-17 | 2018-05-15 | Amazon Technologies, Inc. | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2002086863A1 (en) | 2001-04-19 | 2002-10-31 | British Telecommunications Public Limited Company | Speech recognition |
EP2158540A4 (en) * | 2007-06-18 | 2010-10-20 | Geographic Services Inc | Geographic feature name search system |
US9484025B2 (en) | 2013-10-15 | 2016-11-01 | Toyota Jidosha Kabushiki Kaisha | Configuring dynamic custom vocabulary for personalized speech recognition |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5258909A (en) * | 1989-08-31 | 1993-11-02 | International Business Machines Corporation | Method and apparatus for "wrong word" spelling error detection and correction |
US5488652A (en) * | 1994-04-14 | 1996-01-30 | Northern Telecom Limited | Method and apparatus for training speech recognition algorithms for directory assistance applications |
WO1996013030A2 (en) | 1994-10-25 | 1996-05-02 | British Telecommunications Public Limited Company | Voice-operated services |
US5805772A (en) * | 1994-12-30 | 1998-09-08 | Lucent Technologies Inc. | Systems, methods and articles of manufacture for performing high resolution N-best string hypothesization |
US5999902A (en) * | 1995-03-07 | 1999-12-07 | British Telecommunications Public Limited Company | Speech recognition incorporating a priori probability weighting factors |
US6108631A (en) * | 1997-09-24 | 2000-08-22 | U.S. Philips Corporation | Input system for at least location and/or street names |
US6112174A (en) * | 1996-11-13 | 2000-08-29 | Hitachi, Ltd. | Recognition dictionary system structure and changeover method of speech recognition system for car navigation |
US6192337B1 (en) * | 1998-08-14 | 2001-02-20 | International Business Machines Corporation | Apparatus and methods for rejecting confusible words during training associated with a speech recognition system |
US6208965B1 (en) * | 1997-11-20 | 2001-03-27 | At&T Corp. | Method and apparatus for performing a name acquisition based on speech recognition |
US20020049588A1 (en) * | 1993-03-24 | 2002-04-25 | Engate Incorporated | Computer-aided transcription system using pronounceable substitute text with a common cross-reference library |
US6405172B1 (en) * | 2000-09-09 | 2002-06-11 | Mailcode Inc. | Voice-enabled directory look-up based on recognized spoken initial characters |
US20020107689A1 (en) * | 2001-02-08 | 2002-08-08 | Meng-Hsien Liu | Method for voice and speech recognition |
WO2002086863A1 (en) | 2001-04-19 | 2002-10-31 | British Telecommunications Public Limited Company | Speech recognition |
US6483896B1 (en) * | 1998-02-05 | 2002-11-19 | At&T Corp. | Speech recognition using telephone call parameters |
US6937982B2 (en) * | 2000-07-21 | 2005-08-30 | Denso Corporation | Speech recognition apparatus and method using two opposite words |
US6983244B2 (en) * | 2003-08-29 | 2006-01-03 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for improved speech recognition with supplementary information |
-
2002
- 2002-04-15 WO PCT/GB2002/001748 patent/WO2002086863A1/en active IP Right Grant
- 2002-04-15 EP EP02722433A patent/EP1397797B1/en not_active Expired - Lifetime
- 2002-04-15 CA CA002440463A patent/CA2440463C/en not_active Expired - Fee Related
- 2002-04-15 DE DE60222413T patent/DE60222413T2/en not_active Expired - Lifetime
- 2002-04-15 US US10/472,897 patent/US7970610B2/en not_active Expired - Fee Related
Patent Citations (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5258909A (en) * | 1989-08-31 | 1993-11-02 | International Business Machines Corporation | Method and apparatus for "wrong word" spelling error detection and correction |
US20020049588A1 (en) * | 1993-03-24 | 2002-04-25 | Engate Incorporated | Computer-aided transcription system using pronounceable substitute text with a common cross-reference library |
US5488652A (en) * | 1994-04-14 | 1996-01-30 | Northern Telecom Limited | Method and apparatus for training speech recognition algorithms for directory assistance applications |
WO1996013030A2 (en) | 1994-10-25 | 1996-05-02 | British Telecommunications Public Limited Company | Voice-operated services |
US5940793A (en) * | 1994-10-25 | 1999-08-17 | British Telecommunications Public Limited Company | Voice-operated services |
US5805772A (en) * | 1994-12-30 | 1998-09-08 | Lucent Technologies Inc. | Systems, methods and articles of manufacture for performing high resolution N-best string hypothesization |
US5999902A (en) * | 1995-03-07 | 1999-12-07 | British Telecommunications Public Limited Company | Speech recognition incorporating a priori probability weighting factors |
US6112174A (en) * | 1996-11-13 | 2000-08-29 | Hitachi, Ltd. | Recognition dictionary system structure and changeover method of speech recognition system for car navigation |
US6108631A (en) * | 1997-09-24 | 2000-08-22 | U.S. Philips Corporation | Input system for at least location and/or street names |
US6208965B1 (en) * | 1997-11-20 | 2001-03-27 | At&T Corp. | Method and apparatus for performing a name acquisition based on speech recognition |
US6483896B1 (en) * | 1998-02-05 | 2002-11-19 | At&T Corp. | Speech recognition using telephone call parameters |
US6192337B1 (en) * | 1998-08-14 | 2001-02-20 | International Business Machines Corporation | Apparatus and methods for rejecting confusible words during training associated with a speech recognition system |
US6937982B2 (en) * | 2000-07-21 | 2005-08-30 | Denso Corporation | Speech recognition apparatus and method using two opposite words |
US6405172B1 (en) * | 2000-09-09 | 2002-06-11 | Mailcode Inc. | Voice-enabled directory look-up based on recognized spoken initial characters |
US20020107689A1 (en) * | 2001-02-08 | 2002-08-08 | Meng-Hsien Liu | Method for voice and speech recognition |
WO2002086863A1 (en) | 2001-04-19 | 2002-10-31 | British Telecommunications Public Limited Company | Speech recognition |
US6983244B2 (en) * | 2003-08-29 | 2006-01-03 | Matsushita Electric Industrial Co., Ltd. | Method and apparatus for improved speech recognition with supplementary information |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9583107B2 (en) | 2006-04-05 | 2017-02-28 | Amazon Technologies, Inc. | Continuous speech transcription performance indication |
US9973450B2 (en) | 2007-09-17 | 2018-05-15 | Amazon Technologies, Inc. | Methods and systems for dynamically updating web service profile information by parsing transcribed message strings |
US20090248415A1 (en) * | 2008-03-31 | 2009-10-01 | Yap, Inc. | Use of metadata to post process speech recognition output |
US8676577B2 (en) * | 2008-03-31 | 2014-03-18 | Canyon IP Holdings, LLC | Use of metadata to post process speech recognition output |
Also Published As
Publication number | Publication date |
---|---|
DE60222413T2 (en) | 2008-06-12 |
CA2440463A1 (en) | 2002-10-31 |
WO2002086863A1 (en) | 2002-10-31 |
CA2440463C (en) | 2010-02-02 |
EP1397797B1 (en) | 2007-09-12 |
DE60222413D1 (en) | 2007-10-25 |
EP1397797A1 (en) | 2004-03-17 |
US20040117182A1 (en) | 2004-06-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR100383352B1 (en) | Voice-operated service | |
KR100574768B1 (en) | An automated hotel attendant using speech recognition | |
US6570964B1 (en) | Technique for recognizing telephone numbers and other spoken information embedded in voice messages stored in a voice messaging system | |
US6208964B1 (en) | Method and apparatus for providing unsupervised adaptation of transcriptions | |
US6243680B1 (en) | Method and apparatus for obtaining a transcription of phrases through text and spoken utterances | |
US20040153306A1 (en) | Recognition of proper nouns using native-language pronunciation | |
US9286887B2 (en) | Concise dynamic grammars using N-best selection | |
US20030191643A1 (en) | Automatic multi-language phonetic transcribing system | |
US20050004799A1 (en) | System and method for a spoken language interface to a large database of changing records | |
Lamel et al. | Identifying non-linguistic speech features. | |
Kamm et al. | Speech recognition issues for directory assistance applications | |
US7970610B2 (en) | Speech recognition | |
Imperl et al. | Clustering of triphones using phoneme similarity estimation for the definition of a multilingual set of triphones | |
KR20000005278A (en) | Automatic speech recognition | |
EP1158491A2 (en) | Personal data spoken input and retrieval | |
Georgila et al. | A speech-based human-computer interaction system for automating directory assistance services | |
JP2002532763A (en) | Automatic inquiry system operated by voice | |
Nouza | A large Czech vocabulary recognition system for real-time applications | |
Langmann et al. | FRESCO: the French telephone speech data collection-part of the European Speechdat (M) project | |
Petek | Identification of Regional Variants in the Standard Slovenian Speech | |
EP1103954A1 (en) | Digital speech acquisition, transmission, storage and search system and method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: BRITISH TELECOMMUNICATIONS PUBLIC LIMITED COMPANY, Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:DOWNEY, SIMON N.;REEL/FRAME:015070/0658 Effective date: 20020424 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20230628 |