WO1998035491A1 - Interface voix-donnees - Google Patents
Interface voix-donnees Download PDFInfo
- Publication number
- WO1998035491A1 WO1998035491A1 PCT/GB1998/000194 GB9800194W WO9835491A1 WO 1998035491 A1 WO1998035491 A1 WO 1998035491A1 GB 9800194 W GB9800194 W GB 9800194W WO 9835491 A1 WO9835491 A1 WO 9835491A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- words
- coded signals
- speech
- signals
- link
- Prior art date
Links
- 230000005236 sound signal Effects 0.000 claims abstract description 11
- 230000002452 interceptive effect Effects 0.000 claims description 6
- 238000000034 method Methods 0.000 claims description 5
- 239000000284 extract Substances 0.000 abstract 1
- 230000006870 function Effects 0.000 description 6
- 238000007781 pre-processing Methods 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 238000007796 conventional method Methods 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/42—Systems providing special services or facilities to subscribers
- H04M3/487—Arrangements for providing information services, e.g. recorded voice services or time announcements
- H04M3/493—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals
- H04M3/4938—Interactive information services, e.g. directory enquiries ; Arrangements therefor, e.g. interactive voice response [IVR] systems or voice portals comprising a voice browser which renders and interprets, e.g. VoiceXML
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
- G09B5/065—Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/08—Electrically-operated educational appliances providing for individual presentation of information to a plurality of student stations
- G09B5/14—Electrically-operated educational appliances providing for individual presentation of information to a plurality of student stations with provision for individual teacher-student communication
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/221—Announcement of recognition results
Definitions
- an interface for a voice interactive service comprising: a speech synthesiser to receive coded signals representing sequences of words and to generate audio signals corresponding thereto for output; speech recognition means connected to receive the said coded signals and operable upon receipt of a speech signal to be recognised to identify that part of the word sequence represented by the coded signals which most resemble the speech signal to be recognised.
- the invention provides a method of operating a voice interactive service comprising (a) receiving coded signals representing a sequence of words and synthesising audio signals corresponding thereto for output;
- an apparatus 1 for providing a voice-interactive service is shown and in this example it is intended to allow a user to access a text-based information service by voice only, using a telephone 2.
- the apparatus 1 could be located at the user's premises or at the location of the text-based information service, in this example it is located at a telephone exchange or other central location where it can be accessed by many users (at different times or - with duplication of its functions - simultaneously) via a telecommunications link such as a PSTN dialled connection 3.
- the information service is provided by a remote database server 4 which contains (or forms a gateway offering access to) stored pages of textual information - though the database could if desired be incorporated into the apparatus 1 .
- the server is part of a network accessible via a telecommunications link 5, such as the Internet, and responds to addresses transmitted to it by sending a document identified by that address.
- Documents provided by the Internet are commonly formatted according to the hypertext markup language (HTML) which is itself a particular example of the standard generalised markup language according to international standard ISO 8879.
- HTML hypertext markup language
- an HTML document also contains formatting information suggesting the appearance of the document when displayed on a screen (or printed) such as position, font size, italics and so forth. The precise details of these are not important for present purposes; one thing that is of significance however is that these documents also have provision for flagging words or phrases as associated with the address of another document.
- Figure 2a Part of such a document is illustrated in Figure 2a with its displayed appearance shown in Figure 2b. It is seen that this format and control information is enclosed with chevrons " ⁇ > " as delimiters, not being intended for display.
- the text "Patent Office Sites” is to be shown in bold type as indicated by the start and finish codes ⁇ b > and ⁇ /b > .
- the text "US Patent and Trademark Office” is flanked by ⁇ a > and ⁇ /a > delimiters which normally cause the text to be displayed in a distinctive manner - a special colour or underlined, for example - to identify this phrase a representing a link.
- ⁇ a > code contains an associated address "http://www.uspto.gov" which is the address of the Internet page of the US Patent and Trademark Office.
- a user with a visual display terminal receives such a document and wishes to select the USPTO page, he uses a pointing device such as a mouse to point to the underlined phrase, causing the terminal to extract the associated address and transmit it for selection of a new document.
- FIG. 3 shows the apparatus 1 in more detail. It contains a network interface 10 which comprises a modem for connection to the link 5, and a processor programmed with software to transmit addresses via the modem to the server and receive documents from the server.
- This software differs from conventional browser software such as Netscape only in that (a) it receives addresses via a connection 1 1 rather than having them typed in at a keyboard or selected using a mouse and (b) it outputs the received text directly to a file or buffer 1 2 which can be accessed via a connection 1 3.
- a document has been received by the interface 10 and is stored in the buffer 1 2.
- a first portion of text is read out and a correspondingly coded signal is output on the line 13.
- the actual amount of the text output could rely on punctuation characters included in the text, for example up to the first (or second etc.) full stop, or up to the first paragraph mark.
- This is received by a text pre-processing unit 14 which serves to delete unwanted control information, and forward it to a conventional text-to-speech synthesiser 1 5.
- link addresses are stored in the buffer 1 6, they are removed by further text processing 1 7 before forwarding the text to a recognition network generator 1 8 which is connected to a speech recogniser 1 9.
- the recogniser 1 9 is connected to receive audio signals from the telephone line 3, so that responses from the user at 2 may be recognised.
- the recogniser may have permanent programming to enable it to recognise some standard command words for control of the system; however its primary purpose is to match the user's response to the source text which has just been spoken by the synthesiser 1 5; more particularly to identify that part of the source text present in the buffer 1 6 which most closely resembles the user's response.
- the function of the recognition network generator 1 8 is to derive, from the text input to it, parameters for the recogniser defining a vocabulary and grammar corresponding to this task.
- the output of the recognser is a text string corresponding to the matched portion of text (or command word) .
- This output representing the user's response is taken to be a request for a further document information, and the next task is to identify this by locating the text string in the buffer 1 6 and returning the link address contained within in; or if there is none, returning the nearest link address stored in the buffer.
- This function (to be discussed in more detail below) is performed by a link resolve unit 20 which outputs the link address to the interface 10, which transmits it to the database server 4 as a request for a further document. If however the link represents a position in the current document, then this is recognised and a command issued to the buffer 1 2 to read text from a specified point. Control functions - for example if the user wishes to move on to the next
- control words e.g. More, Back, Home, Quit
- control unit 21 which, upon receiving one of these words along, then issues appropriate instructions to the buffer 1 2 and/or interface 10.
- the buffer 1 2 is set up to output one paragraph at a time; suppose further that the user has already heard the title and asked for "More", the buffer 1 2 outputs the next paragraph ""Welcome forests") to the text preprocessor as shown in Figure 4C.
- the recogniser 1 9 matches the speech signal and outputs the text string "Amazon basin", whereupon the link resolver 20 searches in the buffer 1 6 for this text string, finds that it is attached to the link address http://www/amazon. basin", read out this address and forwards it to the interface 1 0 which transmits it to the database server 4 to call up another page.
- the link resolver operates according to the flowchart shown in Figure 5.
- a first test 30 it is determined whether the matched source text is, or contains a link. "Amazon basin", "birds in the Amazon basin” or even “basin many of” would pass this test. In this case, the link address in question is chosen at 31 . Otherwise a second test 32 is performed to establish whether the matched source text lies in a sentence which contains a link; "one thousand species" for example would fall into this category. In this case the address in that sentence (or, if more than one, the one nearest to the matched source text) is chosen.
- the nearest link to the matched source text is chosen, for example by counting the number of words (or the number of characters) from the matched text to the next link above and below it in the buffer, and choosing the link with the lower count.
- a more complex algorithm could examine the nearest links above and below the matched text for the degree of semantic similarity to the matched text and choose the more similar. In a refinement, one could weight this choice to take account of punctuation, for example by increasing by (e.g.) 10 words the count when crossing a paragraph boundary.
- the HTML language also permits links to other parts of the current document - as shown in Figure 4A for the British Wildlife Society.
- the address "#3224" would be recognised by the link resolver as an internal address and forwarded not to the interface 10 but to the buffer 1 2 to cause readout of a paragraph from a point in the document specified by the address.
- the operation of the recognition network generator 1 8 may now be discussed further. There are essentially two components to the setting up of a recogniser for a given function. First, defining its vocabulary, and second, defining its grammar.
- the vocabulary is a question of ensuring that the recogniser has a set of models or templates, typically one for each of the words to be recognised - that is, one for each of the words (other than link addresses) present in the buffer 1 6.
- Vocabulary generation for this purpose may use any of the conventional methods. Typically this is done by using a recogniser preprogrammed with a set of sub-word models (e.g. one per phoneme) and processing each word delivered from the buffer, in similar manner to the operation of a text-to-speech synthesiser, to generate a word template by concatenation of the appropriate sub-word models.
- the recogniser may have a standard store of word models which can be retrieved when the corresponding words are received from the buffer 1 6, though to accommodate proper names and other words not in the standard set the sub-word concatenation method would usually be employed as well.
- the grammar of a recogniser is a set of stored parameters which define what word sequences are permissible; for example, considering the buffer contents shown in Figure 4A whilst “Amazon basin” is a word sequence which is useful to recognise “basin Amazon” is not.
- One possibility is to allow (as sequences for matching against the user's utterance) any number of words from 1 upwards, but only in the sequence in which they appear in the buffer.
- Figure 6 shows this represented graphically (for a portion only of the text) where 40 represents a start node of a recognition "tree", 41 represents an end node, 42 represents word models and the lines 43 represent allowable paths so. It would be possible to include a network of 'carrier phrases' as shown in
- FIG. 7 so that the user could say sentences such as "Tell me more about the Amazon Basin please”.
- a garbage or sink model (Fig. 8) could be included at the beginning and end of the network to allow any speech to surround the echoed phrase.
- the recogniser could simply allow any of the words on the page to be uttered in any order as shown in Figure 9. The accuracy of such a recogniser would not be as high as those shown in Figures ⁇ to 8, but if statistical constraints based on the contents of the HTML page were incorporated in the recognition process a working system could be created.
- the recogniser returns, as a "label" representing its recognition result, the relevant part of the actual text string supplied to the recognition network generator 1 8 by the buffer 1 6, and the link resolver 20 matches this string against the buffer contents to locate the desired links. Whilst this may be convenient to permit use of a conventional unit for the recogniser 1 6, a way of speeding up the operation of the link resolver would be to set up the recogniser to return some parameter enabling faster access to the buffer, for example pointer values giving the addresses in the buffer 1 6 of the first and last characters of the matched source text string.
- This embodiment presupposes that the source text carries hyperlink addresses; however it is also possible to operate this system without embedded addressed of this kind. For example one could transmit to the database server coordinates to identify the point in a (or range of) the source text at which the match occurred. In the case of connectionless service such as the Internet, it would be necessary to concatenate this information with the address of the server before transmitting it.
- the text preprocessor 14 could be arranged to pass certain markings through to the synthesiser 1 5 to allow bold type to be emphasised. Similarly, it would be possible for the preprocessor to pass the hyperlink markings ⁇ a > ... ⁇ /a > (albeit without the addresses) and arrange the synthesiser to respond to these by applying an emphasis, or even switching to a different voice (for example a male instead of female voice) from that used for the remainder of the text. With this expedient, in an alternative embodiment, one can simplify the speech recogniser vocabulary to include only the link words, though it is still preferred to operate the recogniser as described above, against the possibility that the user may not always accurately recollect which words were spoken with the emphasis (or different voice).
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Business, Economics & Management (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Educational Technology (AREA)
- Educational Administration (AREA)
- Multimedia (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Acoustics & Sound (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Computer And Data Communications (AREA)
Abstract
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/GB1998/000194 WO1998035491A1 (fr) | 1997-02-05 | 1998-01-22 | Interface voix-donnees |
AU56743/98A AU5674398A (en) | 1997-02-05 | 1998-01-22 | Voice-data interface |
EP98900943A EP0958692A1 (fr) | 1997-02-05 | 1998-01-22 | Interface voix-donnees |
JP53397198A JP2001510660A (ja) | 1997-02-05 | 1998-01-22 | 音声データインターフェイス |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP97300741.2 | 1997-02-05 | ||
PCT/GB1998/000194 WO1998035491A1 (fr) | 1997-02-05 | 1998-01-22 | Interface voix-donnees |
Publications (1)
Publication Number | Publication Date |
---|---|
WO1998035491A1 true WO1998035491A1 (fr) | 1998-08-13 |
Family
ID=10824876
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/GB1998/000194 WO1998035491A1 (fr) | 1997-02-05 | 1998-01-22 | Interface voix-donnees |
Country Status (1)
Country | Link |
---|---|
WO (1) | WO1998035491A1 (fr) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000014728A1 (fr) * | 1998-09-09 | 2000-03-16 | One Voice Technologies, Inc. | Interface utilisateur interactive de reseau a reconnaissance vocale et a traitement de langage naturel |
EP0992980A2 (fr) * | 1998-10-06 | 2000-04-12 | Lucent Technologies Inc. | Menu interactif à réponse vocale basé sur le web |
WO2000052914A1 (fr) * | 1999-02-27 | 2000-09-08 | Khan Emdadur R | Systeme et procede pour la recherche audio sur internet utilisant un telephone standard |
WO2001016936A1 (fr) * | 1999-08-31 | 2001-03-08 | Accenture Llp | Reconnaissance vocale pour navigation internet |
WO2001043388A2 (fr) * | 1999-12-10 | 2001-06-14 | Deutsche Telekom Ag | Systeme de communication et procede pour permettre un acces internet par l'intermediaire d'un telephone |
EP1134948A2 (fr) * | 2000-03-15 | 2001-09-19 | Nec Corporation | Système de recherche d'information utilisant un terminal radio portable |
EP1168799A2 (fr) * | 2000-06-30 | 2002-01-02 | Fujitsu Limited | Système de traitement de données avec le mécanisme de vocalisation |
JP2002091756A (ja) * | 2000-06-15 | 2002-03-29 | Internatl Business Mach Corp <Ibm> | 多数の音響情報源を同時に提供するためのシステム及び方法 |
DE10201623C1 (de) * | 2002-01-16 | 2003-09-11 | Mediabeam Gmbh | Verfahren zur Datenerfassung von auf einer Internet-Seite bereitgestellten Daten und Verfahren zur Datenübermittlung an eine Internet-Seite |
US6662163B1 (en) * | 2000-03-30 | 2003-12-09 | Voxware, Inc. | System and method for programming portable devices from a remote computer system |
DE102010001564A1 (de) | 2010-02-03 | 2011-08-04 | Bayar, Seher, 51063 | Verfahren und Computerprogrammprodukt zur automatisierten konfigurierbaren akustischen Wiedergabe und Bearbeitung von Internetseiteninhalten |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0735736A2 (fr) * | 1995-03-30 | 1996-10-02 | AT&T IPM Corp. | Méthode pour la reconnaissance automatique de la parole de mots parlés arbitraires |
US5572625A (en) * | 1993-10-22 | 1996-11-05 | Cornell Research Foundation, Inc. | Method for generating audio renderings of digitized works having highly technical content |
GB2307619A (en) * | 1995-11-21 | 1997-05-28 | Alexander James Pollitt | Internet information access system |
WO1997023973A1 (fr) * | 1995-12-22 | 1997-07-03 | Rutgers University | Procede et systeme d'acces audio aux informations d'un reseau informatique grande distance |
WO1997040611A1 (fr) * | 1996-04-22 | 1997-10-30 | At & T Corp. | Procede et appareil de recherche de l'information a l'aide d'une interface audiofrequence |
-
1998
- 1998-01-22 WO PCT/GB1998/000194 patent/WO1998035491A1/fr active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5572625A (en) * | 1993-10-22 | 1996-11-05 | Cornell Research Foundation, Inc. | Method for generating audio renderings of digitized works having highly technical content |
EP0735736A2 (fr) * | 1995-03-30 | 1996-10-02 | AT&T IPM Corp. | Méthode pour la reconnaissance automatique de la parole de mots parlés arbitraires |
GB2307619A (en) * | 1995-11-21 | 1997-05-28 | Alexander James Pollitt | Internet information access system |
WO1997023973A1 (fr) * | 1995-12-22 | 1997-07-03 | Rutgers University | Procede et systeme d'acces audio aux informations d'un reseau informatique grande distance |
WO1997040611A1 (fr) * | 1996-04-22 | 1997-10-30 | At & T Corp. | Procede et appareil de recherche de l'information a l'aide d'une interface audiofrequence |
Non-Patent Citations (5)
Title |
---|
ATKINS D L , BALL T , BARAN T R , BENEDICT M A , COX K C , LADD D A , MATAGA P A, PUCHOL C , RAMMING J C , REHOR K G , TUCKEY C.: "INTEGRATED WEB AND TELEPHONE SERVICE CREATION", BELL LABS TECHNICAL JOURNAL, vol. 2, no. I, 1 January 1997 (1997-01-01), USA, pages 19 - 35, XP002036350 * |
PAGE J H ET AL: "THE LAUREATE TEXT-TO-SPEECH SYSTEM - ARCHITECTURE AND APPLICATIONS", BT TECHNOLOGY JOURNAL, vol. 14, no. 1, 1 January 1996 (1996-01-01), pages 57 - 67, XP000554639 * |
RABINER L R: "The impact of voice processing on modern telecommunications", SPEECH COMMUNICATION, vol. 17, no. 3-4, November 1995 (1995-11-01), pages 217 - 226, XP000641894 * |
RICCIO A ET AL: "VOICE BASED REMOTE DATA BASE ACCESS", PROCEEDINGS OF THE EUROPEAN CONFERENCE ON SPEECH COMMUNICATION AND TECHNOLOGY (EUROSPEECH), PARIS, SEPT. 26 - 28, 1989, vol. 1, 26 September 1989 (1989-09-26), TUBACH J P;MARIANI J J, pages 561 - 564, XP000209922 * |
TAKAHASHI J ET AL: "INTERACTIVE VOICE TECHNOLOGY DEVELOPMENT FOR TELECOMMUNICATIONS APPLICATIONS", SPEECH COMMUNICATION, vol. 17, no. 3/04, November 1995 (1995-11-01), pages 287 - 301, XP000641897 * |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2000014728A1 (fr) * | 1998-09-09 | 2000-03-16 | One Voice Technologies, Inc. | Interface utilisateur interactive de reseau a reconnaissance vocale et a traitement de langage naturel |
EP0992980A2 (fr) * | 1998-10-06 | 2000-04-12 | Lucent Technologies Inc. | Menu interactif à réponse vocale basé sur le web |
JP2000137596A (ja) * | 1998-10-06 | 2000-05-16 | Lucent Technol Inc | 対話型音声応答システム |
EP0992980A3 (fr) * | 1998-10-06 | 2001-05-23 | Lucent Technologies Inc. | Menu interactif à réponse vocale basé sur le web |
US6587822B2 (en) | 1998-10-06 | 2003-07-01 | Lucent Technologies Inc. | Web-based platform for interactive voice response (IVR) |
WO2000052914A1 (fr) * | 1999-02-27 | 2000-09-08 | Khan Emdadur R | Systeme et procede pour la recherche audio sur internet utilisant un telephone standard |
CN100393073C (zh) * | 1999-02-27 | 2008-06-04 | E·R·汗 | 使用标准电话用于网际网络声频浏览的系统及方法 |
US6606611B1 (en) * | 1999-02-27 | 2003-08-12 | Emdadur Khan | System and method for audio-only internet browsing using a standard telephone |
WO2001016936A1 (fr) * | 1999-08-31 | 2001-03-08 | Accenture Llp | Reconnaissance vocale pour navigation internet |
US7590538B2 (en) | 1999-08-31 | 2009-09-15 | Accenture Llp | Voice recognition system for navigating on the internet |
WO2001043388A3 (fr) * | 1999-12-10 | 2002-04-04 | Deutsche Telekom Ag | Systeme de communication et procede pour permettre un acces internet par l'intermediaire d'un telephone |
WO2001043388A2 (fr) * | 1999-12-10 | 2001-06-14 | Deutsche Telekom Ag | Systeme de communication et procede pour permettre un acces internet par l'intermediaire d'un telephone |
EP1134948A3 (fr) * | 2000-03-15 | 2003-04-23 | Nec Corporation | Système de recherche d'information utilisant un terminal radio portable |
EP1134948A2 (fr) * | 2000-03-15 | 2001-09-19 | Nec Corporation | Système de recherche d'information utilisant un terminal radio portable |
US7805145B2 (en) | 2000-03-15 | 2010-09-28 | Nec Corporation | Information search system using radio portable terminal |
US6662163B1 (en) * | 2000-03-30 | 2003-12-09 | Voxware, Inc. | System and method for programming portable devices from a remote computer system |
JP2002091756A (ja) * | 2000-06-15 | 2002-03-29 | Internatl Business Mach Corp <Ibm> | 多数の音響情報源を同時に提供するためのシステム及び方法 |
EP1168799A2 (fr) * | 2000-06-30 | 2002-01-02 | Fujitsu Limited | Système de traitement de données avec le mécanisme de vocalisation |
EP1168799A3 (fr) * | 2000-06-30 | 2005-12-14 | Fujitsu Limited | Système de traitement de données avec le mécanisme de vocalisation |
DE10201623C1 (de) * | 2002-01-16 | 2003-09-11 | Mediabeam Gmbh | Verfahren zur Datenerfassung von auf einer Internet-Seite bereitgestellten Daten und Verfahren zur Datenübermittlung an eine Internet-Seite |
US6741681B2 (en) | 2002-01-16 | 2004-05-25 | Mediabeam Gmbh | Method for acquisition of data provided on an internet site and for data communication to an internet site |
DE102010001564A1 (de) | 2010-02-03 | 2011-08-04 | Bayar, Seher, 51063 | Verfahren und Computerprogrammprodukt zur automatisierten konfigurierbaren akustischen Wiedergabe und Bearbeitung von Internetseiteninhalten |
WO2011095457A2 (fr) | 2010-02-03 | 2011-08-11 | Bayar, Seher | Procédé et produit programme informatique pour la reproduction acoustique et le traitement, configurables de manière automatisée, de contenus de pages internet |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US6532444B1 (en) | Network interactive user interface using speech recognition and natural language processing | |
US6937986B2 (en) | Automatic dynamic speech recognition vocabulary based on external sources of information | |
US6604075B1 (en) | Web-based voice dialog interface | |
KR100661687B1 (ko) | 대화형 음성 응답 애플리케이션 구현 장치 및 방법, 머신 판독가능 매체 및 대화형 음성 응답 시스템 | |
US6282511B1 (en) | Voiced interface with hyperlinked information | |
US20180007201A1 (en) | Personal Voice-Based Information Retrieval System | |
US6188985B1 (en) | Wireless voice-activated device for control of a processor-based host system | |
JP3432076B2 (ja) | 音声対話型ビデオスクリーン表示システム | |
US5884262A (en) | Computer network audio access and conversion system | |
US20020077823A1 (en) | Software development systems and methods | |
US20020010715A1 (en) | System and method for browsing using a limited display device | |
US20060235694A1 (en) | Integrating conversational speech into Web browsers | |
WO1997032427A9 (fr) | Procede et appareil servant a acceder par telephone a internet et a naviguer dans ce dernier | |
EP1221161A1 (fr) | Interface utilisateur interactive a reconnaissance vocale et traitement de langage naturel | |
WO1997032427A1 (fr) | Procede et appareil servant a acceder par telephone a internet et a naviguer dans ce dernier | |
GB2407657A (en) | Automatic grammar generator comprising phase chunking and morphological variation | |
AU2001251354A1 (en) | Natural language and dialogue generation processing | |
WO1998035491A1 (fr) | Interface voix-donnees | |
JP2009187349A (ja) | 文章修正支援システム、文章修正支援方法、および文章修正支援用プログラム | |
Brown et al. | Web page analysis for voice browsing | |
EP0958692A1 (fr) | Interface voix-donnees | |
KR100399574B1 (ko) | 외국인을 위한 전화안내 자동 통역시스템 및 방법 | |
US20030091176A1 (en) | Communication system and method for establishing an internet connection by means of a telephone | |
US7054813B2 (en) | Automatic generation of efficient grammar for heading selection | |
GB2342530A (en) | Gathering user inputs by speech recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 09043165 Country of ref document: US |
|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AL AM AT AU AZ BA BB BG BR BY CA CH CN CU CZ DE DK EE ES FI GB GE GH GM GW HU ID IL IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT UA UG US UZ VN YU ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GH GM KE LS MW SD SZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN ML MR NE SN TD TG |
|
DFPE | Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101) | ||
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 1998 533971 Kind code of ref document: A Format of ref document f/p: F |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 1998900943 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 1998900943 Country of ref document: EP |
|
REG | Reference to national code |
Ref country code: DE Ref legal event code: 8642 |