SE9601812L - Enhancements to, or with respect to, speech-to-speech conversion - Google Patents
Enhancements to, or with respect to, speech-to-speech conversionInfo
- Publication number
- SE9601812L SE9601812L SE9601812A SE9601812A SE9601812L SE 9601812 L SE9601812 L SE 9601812L SE 9601812 A SE9601812 A SE 9601812A SE 9601812 A SE9601812 A SE 9601812A SE 9601812 L SE9601812 L SE 9601812L
- Authority
- SE
- Sweden
- Prior art keywords
- speech
- inputs
- conversion
- enhancements
- respect
- Prior art date
Links
- 238000006243 chemical reaction Methods 0.000 title abstract 2
- 238000009472 formulation Methods 0.000 abstract 1
- 238000000034 method Methods 0.000 abstract 1
- 239000000203 mixture Substances 0.000 abstract 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/263—Language identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/55—Rule-based translation
- G06F40/56—Natural language generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/58—Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/04—Details of speech synthesis systems, e.g. synthesiser structure or memory management
- G10L13/047—Architecture of speech synthesisers
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/005—Language recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Artificial Intelligence (AREA)
- Machine Translation (AREA)
- Use Of Switch Circuits For Exchanges And Methods Of Control Of Multiplex Exchanges (AREA)
Abstract
A system and method for speech-to-speech conversion for providing spoken responses to speech inputs in at least two natural languages wherein speech inputs are recognised and interpreted in said at least two languages. The recognised speech inputs are evaluated to determine the language of the speech inputs, and a dialogue is undertaken with a database containing speech information data, in said at least two natural languages, to obtain data for the formulation of spoken responses to the speech inputs. The speech information data, obtained from the database, is then converted into spoken responses which exhibit the language characteristics of the respective speech inputs.
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SE9601812A SE519273C2 (en) | 1996-05-13 | 1996-05-13 | Improvements to, or with respect to, speech-to-speech conversion |
PCT/SE1997/000584 WO1997043707A1 (en) | 1996-05-13 | 1997-04-08 | Improvements in, or relating to, speech-to-speech conversion |
EP97919841A EP0976026A1 (en) | 1996-05-13 | 1997-04-08 | Improvements in, or relating to, speech-to-speech conversion |
NO19985178A NO318112B1 (en) | 1996-05-13 | 1998-11-06 | Speech-to-speech conversion system and method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SE9601812A SE519273C2 (en) | 1996-05-13 | 1996-05-13 | Improvements to, or with respect to, speech-to-speech conversion |
Publications (3)
Publication Number | Publication Date |
---|---|
SE9601812D0 SE9601812D0 (en) | 1996-05-13 |
SE9601812L true SE9601812L (en) | 1997-11-14 |
SE519273C2 SE519273C2 (en) | 2003-02-11 |
Family
ID=20402544
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
SE9601812A SE519273C2 (en) | 1996-05-13 | 1996-05-13 | Improvements to, or with respect to, speech-to-speech conversion |
Country Status (4)
Country | Link |
---|---|
EP (1) | EP0976026A1 (en) |
NO (1) | NO318112B1 (en) |
SE (1) | SE519273C2 (en) |
WO (1) | WO1997043707A1 (en) |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA2284304A1 (en) * | 1998-12-22 | 2000-06-22 | Nortel Networks Corporation | Communication systems and methods employing automatic language indentification |
US7590538B2 (en) * | 1999-08-31 | 2009-09-15 | Accenture Llp | Voice recognition system for navigating on the internet |
JP2001101187A (en) * | 1999-09-30 | 2001-04-13 | Sony Corp | Device and method for translation and recording medium |
GB2366940B (en) | 2000-09-06 | 2004-08-11 | Ericsson Telefon Ab L M | Text language detection |
DE10063148A1 (en) | 2000-12-18 | 2002-06-20 | Deutsche Telekom Ag | Dialogue system for human-machine interaction with cooperating dialogue facilities |
GB0111012D0 (en) | 2001-05-04 | 2001-06-27 | Nokia Corp | A communication terminal having a predictive text editor application |
DE102006057159A1 (en) | 2006-12-01 | 2008-06-05 | Deutsche Telekom Ag | Method for classifying spoken language in speech dialogue systems |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2165969B (en) * | 1984-10-19 | 1988-07-06 | British Telecomm | Dialogue system |
DE3732849A1 (en) * | 1987-09-29 | 1989-04-20 | Siemens Ag | SYSTEM ARCHITECTURE FOR AN ACOUSTIC HUMAN / MACHINE DIALOG SYSTEM |
EP0543329B1 (en) * | 1991-11-18 | 2002-02-06 | Kabushiki Kaisha Toshiba | Speech dialogue system for facilitating human-computer interaction |
SE500277C2 (en) * | 1993-05-10 | 1994-05-24 | Televerket | Device for increasing speech comprehension when translating speech from a first language to a second language |
SE504177C2 (en) * | 1994-06-29 | 1996-12-02 | Telia Ab | Method and apparatus for adapting a speech recognition equipment for dialectal variations in a language |
-
1996
- 1996-05-13 SE SE9601812A patent/SE519273C2/en unknown
-
1997
- 1997-04-08 WO PCT/SE1997/000584 patent/WO1997043707A1/en active Application Filing
- 1997-04-08 EP EP97919841A patent/EP0976026A1/en not_active Withdrawn
-
1998
- 1998-11-06 NO NO19985178A patent/NO318112B1/en unknown
Also Published As
Publication number | Publication date |
---|---|
NO985178L (en) | 1998-11-11 |
WO1997043707A1 (en) | 1997-11-20 |
NO318112B1 (en) | 2005-01-31 |
EP0976026A1 (en) | 2000-02-02 |
SE519273C2 (en) | 2003-02-11 |
SE9601812D0 (en) | 1996-05-13 |
NO985178D0 (en) | 1998-11-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
BR9815258A (en) | System and method for auditing sgml data pages | |
EP0831460A3 (en) | Speech synthesis method utilizing auxiliary information | |
DE69712216D1 (en) | METHOD AND DEVICE FOR TRANSLATING FROM ONE SPARK TO ANOTHER | |
WO2000033211A3 (en) | Automatic segmentation of a text | |
Klabbers et al. | Speech synthesis development made easy: the bonn open synthesis system. | |
SE9601812L (en) | Enhancements to, or with respect to, speech-to-speech conversion | |
Hongladarom | Rgyalthang Tibetan of Yunnan: A preliminary report | |
SE9303623L (en) | Method and apparatus for automatically extracting prosodic information | |
Post | French tonal structures | |
SE9600959D0 (en) | Speech-to-speech translation method and apparatus | |
SE9601811D0 (en) | A speech-to-speech conversion system | |
WO1997007499A3 (en) | A method and device for preparing and using diphones for multilingual text-to-speech generating | |
Kaiki et al. | Prosodic characteristics of Japanese conversational speech | |
Lee | Machine-to-man communication by speech Part 1: Generation of segmental phonemes from text | |
Vicsi et al. | Text independent neural network/rule based hybrid continuous speech recognition | |
Klabbers et al. | Creation of speech corpora for the multilingual Bonn Open Synthesis System. | |
Smith | On the nonergativity and intransitivity of relative clauses in Labrador Inuttut | |
ATE241196T1 (en) | EXPANSION OF A LANGUAGE RECOGNITION VOCABULARY USING DERIVED WORDS | |
JPS63221475A (en) | Analyzing method for syntax | |
Epitropakis et al. | High quality intonation algorithm for the Greek TTS-system | |
Bruce et al. | Preliminary report from the project," Prosodic Segmentation and Structuring of Dialogue" | |
KR970060042A (en) | Speech synthesis method | |
JPS62293472A (en) | Voice translator | |
KR100194814B1 (en) | Text-to-speech converter using multilevel input information and its method | |
Berthelin et al. | Processing non-expected language. |