US20110153309A1 - Automatic interpretation apparatus and method using utterance similarity measure - Google Patents
Automatic interpretation apparatus and method using utterance similarity measure Download PDFInfo
- Publication number
- US20110153309A1 US20110153309A1 US12/969,017 US96901710A US2011153309A1 US 20110153309 A1 US20110153309 A1 US 20110153309A1 US 96901710 A US96901710 A US 96901710A US 2011153309 A1 US2011153309 A1 US 2011153309A1
- Authority
- US
- United States
- Prior art keywords
- sentence
- language
- language sentence
- voice
- similarity
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/51—Translation evaluation
Definitions
- the following disclosure relates to an automatic interpretation apparatus and method, and in particular, to an automatic interpretation apparatus and method using an inter-sentence utterance similarity measure.
- automatic interpretation devices may fail to perform correct sentence translation in the event of erroneous voice recognition. Also, there may be an error in translation even in the event of errorless voice recognition. Thus, if translated sentences are converted into voice signals prior to output, there may be an error in interpretation.
- related art techniques convert voice recognition results into sentences within a limited range, translate the sentences, and convert the translated sentences into voice signals prior to output. However, if a desired sentence of a user is not within the limited range, sentence translation is limited, thus degrading the interpretation performance.
- an automatic interpretation apparatus includes: a voice recognizing unit receiving a first-language voice and generating a first-language sentence through a voice recognition operation; a language processing unit extracting elements included in the first-language sentence; a similarity calculating unit comparing the extracted elements with elements included in a translated sentence stored in a translated sentence database and calculating the similarity between the first-language sentence and the translated sentence on the basis of the comparison result; a sentence translating unit translating the first-language sentence into a second-language sentence with reference to the translated sentence database according to the calculated similarity; and a voice synthesizing unit detecting voice data corresponding to the second-language sentence and synthesizing the detected voice data to output an analog voice signal corresponding to the second-language sentence.
- an automatic interpretation method includes: receiving a first-language voice and generating a first-language sentence through a voice recognition operation; extracting elements included in the first-language sentence; comparing the extracted elements with elements included in a translated sentence stored in a translated sentence database and calculating the similarity between the first-language sentence and the translated sentence on the basis of the comparison result; receiving the first-language sentence according to the calculated similarity and translating the first-language sentence into a second-language sentence with reference to the translated sentence database; and detecting voice data corresponding to the second-language sentence and synthesizing the detected voice data to output an analog voice signal corresponding to the second-language sentence.
- FIG. 1 is a block diagram of an automatic interpretation apparatus according to an exemplary embodiment.
- FIG. 2 is a flow chart illustrating an automatic interpretation method using the automatic interpretation apparatus illustrated in FIG. 1 .
- FIG. 1 is a block diagram of an automatic interpretation apparatus according to an exemplary embodiment.
- an automatic interpretation apparatus may be applicable to various apparatuses that perform an interpretation from first language to second language.
- the automatic interpretation apparatus recognizes user's voices and determines the similarity between the recognition results and translated sentences including pairs of prepared first-language sentences and second-language sentences.
- the automatic interpretation apparatus uses the determination result to output sentences desired by the user. Accordingly, sentences desired by the user can be displayed to the user even without the use of a complex translator.
- the automatic interpretation apparatus can display an example sentence thereof by using the translated sentences containing the keywords.
- the automatic interpretation apparatus inputs interpretation-target sentences or keywords not only through voice recognition but also through an input unit (e.g., a keypad) to display a list of the most similar candidate sentences (among the translated sentences) on a display screen, thereby enabling the user to select a desired sentence among the displayed sentences.
- an input unit e.g., a keypad
- the automatic interpretation apparatus includes the following units to perform the above-described operations.
- the automatic interpretation apparatus includes a voice recognizing unit 100 , a language processing unit 110 , a similarity calculating unit 120 , a sentence translating unit 130 , a voice synthesizing unit 140 , and a translated sentence database (DB) 150 .
- the voice recognizing unit 100 receives a first-language voice from the user and converts the first-language voice into a first-language sentence through a voice recognition operation. Also, the voice recognizing unit 100 outputs a confidence score for each word of the first-language sentence. The outputted confidence score may be used by the similarity calculating unit 120 . Herein, the confidence score means the matching rate between the first-language voice and the first-language sentence.
- the automatic interpretation apparatus may receive a first-language sentence (instead of the first-language voice) through a character input unit such as a keypad. In this case, the voice recognizing unit 100 may be omitted from the automatic interpretation apparatus.
- the language processing unit 110 receives the first-language sentence from the voice recognizing unit 100 and extracts various elements for similarity calculation from the first-language sentence.
- the various elements include word, word segmentation, morpheme/speech part, sentence pattern, tense, affirmation/negation, modality information, and speech act representing the flow of conversation.
- the language processing unit 110 extracts higher semantic information (class information) together with respect to words such as person name, place name, money amount, date, and numeral. Also, the language processing unit 110 may also extract similar words similar to the word and hetero-form words for the word, through the hetero-form extension and the extension of similar words.
- Examples of the similar words include (Korean)’ and (Korean)’ that are different words with similar meanings.
- Examples of the hetero-form words include adopted words such as (Korean)’ and (Korean)’ that have different forms but have the same meaning.
- the similarity calculating unit 120 considers the confidence score for each word processed by the voice recognizing unit 100 and compares the various elements extracted by the language processing unit 110 with various elements stored in the translated sentence DB 150 to calculate the similarity therebetween.
- the similarity calculation operation is performed by a similarity calculation algorithm expressed as Equation (1).
- S 1 denotes an input sentence
- S 2 denotes a candidate sentence
- f i (e 1,i ) denotes the i th element of the input sentence
- f i (e 2,i ) denotes a similarity function for the i th element of the candidate sentence
- w i denotes a weight for f i .
- the similarity calculation result by Equation (1) is expressed in the form of probability.
- a threshold value is set and it is determined whether the calculated similarity is higher than the threshold value. If the calculated similarity is higher than the threshold value, class information of the second-language sentence corresponding to the first-language sentence selected from the translated sentence DB 150 is translated and the translated result is transferred to the voice synthesizing unit 140 without passing through the sentence translating unit 130 . On the other hand, if the calculated similarity is lower than the threshold value, user selection is requested or the first-language sentence (i.e., the voice recognition result) is transferred to the sentence translating unit 130 .
- the translated sentence DB 150 includes pairs of first-language sentences and second-language sentences. For example, when the first-language sentence is 2 (Korean)’, the second-language sentence is ‘2 tickets to Seoul, please (English)’.
- the sentence translating unit 130 receives the first-language sentence through the similarity calculating unit 120 and translates the first-language sentence with reference to the translated sentence DB 150 .
- the translation result is transferred as a second-language sentence to the voice synthesizing unit 140 .
- the voice synthesizing unit 140 receives the second-language sentence from the similarity calculating unit 120 or the second-language sentence from the sentence translating unit 130 , synthesizes the prestored voice data mapping to the received second-language sentence, and outputs the synthesized voice data in the form of analog signals.
- FIG. 2 is a flow chart illustrating an automatic interpretation method using the automatic interpretation apparatus illustrated in FIG. 1 .
- the voice recognizing unit 100 converts a first-language voice, inputted from a user, into a first-language sentence through a voice recognition operation (S 210 ).
- a confidence score for each word included in the first-language sentence is generated together with the first-language sentence.
- the confidence score is used by the similarity calculating unit 120 .
- an operation of selecting a voice recognition region by the user may be added before the conversion of the first-language voice into the first-language sentence (i.e., before the user voice recognition). For example, if user voice recognition is performed in an airplane or a hotel, an operation of selecting a region of an airplane or a hotel may be added. Thus, the success rate of voice recognition can be increased because a voice recognition operation is performed within the category of the region. If the user does not select a voice recognition region, an operation of classifying the region for the voice recognition result may be added.
- the language processing unit 110 extracts elements for similarity calculation from the first-language sentence (S 220 ).
- the extracted elements include word, word segmentation, morpheme/speech part, sentence pattern, tense, affirmation/negation, modality information, and speech act representing the flow of conversation.
- the similarity calculating unit 120 performs a similarity calculation operation.
- the similarity calculation operation makes it possible to minimize a conversion error that may occur during the conversion of the first-language voice into the first-language sentence through the voice recognition operation.
- the similarity calculating unit 120 compares the elements extracted by the language processing unit 110 with elements included in pairs of first-language sentences and second-language sentences stored in the translated sentence DB 150 to calculate the similarity therebetween.
- the similarity is calculated by Equation (1). If the calculated similarity is higher than the threshold value, class information of the second-language sentence corresponding to the first-language sentence selected from the translated sentence DB 150 is translated. On the other hand, if the calculated similarity is lower than the threshold value, user selection is requested or the first-language sentence is translated (e.g., machine-translated) (S 240 ).
- voice data corresponding to the second-language sentence are searched and the searched voice data are synthesized to output analog voice signals (S 250 ).
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Machine Translation (AREA)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR10-2009-0127709 | 2009-12-21 | ||
KR1020090127709A KR101377459B1 (ko) | 2009-12-21 | 2009-12-21 | 자동 통역 장치 및 그 방법 |
Publications (1)
Publication Number | Publication Date |
---|---|
US20110153309A1 true US20110153309A1 (en) | 2011-06-23 |
Family
ID=44152332
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/969,017 Abandoned US20110153309A1 (en) | 2009-12-21 | 2010-12-15 | Automatic interpretation apparatus and method using utterance similarity measure |
Country Status (2)
Country | Link |
---|---|
US (1) | US20110153309A1 (ko) |
KR (1) | KR101377459B1 (ko) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103902523A (zh) * | 2012-12-28 | 2014-07-02 | 新疆电力信息通信有限责任公司 | 维吾尔语句子相似度计算方法 |
US20140258322A1 (en) * | 2013-03-06 | 2014-09-11 | Electronics And Telecommunications Research Institute | Semantic-based search system and search method thereof |
US8843369B1 (en) * | 2013-12-27 | 2014-09-23 | Google Inc. | Speech endpointing based on voice profile |
US9189476B2 (en) | 2012-04-04 | 2015-11-17 | Electronics And Telecommunications Research Institute | Translation apparatus and method thereof for helping a user to more easily input a sentence to be translated |
US9607613B2 (en) | 2014-04-23 | 2017-03-28 | Google Inc. | Speech endpointing based on word comparisons |
CN107291939A (zh) * | 2017-07-06 | 2017-10-24 | 携程计算机技术(上海)有限公司 | 酒店信息的聚类匹配方法及系统 |
JP2018055622A (ja) * | 2016-09-30 | 2018-04-05 | 大日本印刷株式会社 | 情報処理装置及びプログラム |
US10269341B2 (en) | 2015-10-19 | 2019-04-23 | Google Llc | Speech endpointing |
US10593352B2 (en) | 2017-06-06 | 2020-03-17 | Google Llc | End of query detection |
CN111048073A (zh) * | 2019-12-16 | 2020-04-21 | 北京明略软件系统有限公司 | 一种音频处理方法、装置、电子设备及可读存储介质 |
CN111292814A (zh) * | 2019-12-26 | 2020-06-16 | 北京亚信数据有限公司 | 一种医疗数据标准化的方法及装置 |
US10929754B2 (en) | 2017-06-06 | 2021-02-23 | Google Llc | Unified endpointer using multitask and multidomain learning |
US11062696B2 (en) | 2015-10-19 | 2021-07-13 | Google Llc | Speech endpointing |
US11144732B2 (en) | 2017-10-12 | 2021-10-12 | Electronics And Telecommunications Research Institute | Apparatus and method for user-customized interpretation and translation |
US11887579B1 (en) | 2022-09-28 | 2024-01-30 | Intuit Inc. | Synthetic utterance generation |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101626109B1 (ko) * | 2012-04-04 | 2016-06-13 | 한국전자통신연구원 | 통역 장치 및 방법 |
KR101694286B1 (ko) * | 2012-05-02 | 2017-01-09 | 한국전자통신연구원 | 양방향 자동 통역 및 번역 서비스 제공 장치 및 그 방법 |
KR101827773B1 (ko) * | 2016-08-02 | 2018-02-09 | 주식회사 하이퍼커넥트 | 통역 장치 및 방법 |
KR101854714B1 (ko) * | 2017-12-28 | 2018-05-08 | 주식회사 트위그팜 | 번역 문서 관리 시스템 및 방법 |
Citations (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4507750A (en) * | 1982-05-13 | 1985-03-26 | Texas Instruments Incorporated | Electronic apparatus from a host language |
US5293584A (en) * | 1992-05-21 | 1994-03-08 | International Business Machines Corporation | Speech recognition system for natural language translation |
US5477451A (en) * | 1991-07-25 | 1995-12-19 | International Business Machines Corp. | Method and system for natural language translation |
US5510981A (en) * | 1993-10-28 | 1996-04-23 | International Business Machines Corporation | Language translation apparatus and method using context-based translation models |
US5689616A (en) * | 1993-11-19 | 1997-11-18 | Itt Corporation | Automatic language identification/verification system |
US6233545B1 (en) * | 1997-05-01 | 2001-05-15 | William E. Datig | Universal machine translator of arbitrary languages utilizing epistemic moments |
US20020135618A1 (en) * | 2001-02-05 | 2002-09-26 | International Business Machines Corporation | System and method for multi-modal focus detection, referential ambiguity resolution and mood classification using multi-modal input |
US6763331B2 (en) * | 2001-02-01 | 2004-07-13 | Matsushita Electric Industrial Co., Ltd. | Sentence recognition apparatus, sentence recognition method, program, and medium |
US20040243392A1 (en) * | 2003-05-27 | 2004-12-02 | Kabushiki Kaisha Toshiba | Communication support apparatus, method and program |
US20060004560A1 (en) * | 2004-06-24 | 2006-01-05 | Sharp Kabushiki Kaisha | Method and apparatus for translation based on a repository of existing translations |
US20070016401A1 (en) * | 2004-08-12 | 2007-01-18 | Farzad Ehsani | Speech-to-speech translation system with user-modifiable paraphrasing grammars |
US20070208569A1 (en) * | 2006-03-03 | 2007-09-06 | Balan Subramanian | Communicating across voice and text channels with emotion preservation |
US20080077392A1 (en) * | 2006-09-26 | 2008-03-27 | Kabushiki Kaisha Toshiba | Method, apparatus, system, and computer program product for machine translation |
US20080077391A1 (en) * | 2006-09-22 | 2008-03-27 | Kabushiki Kaisha Toshiba | Method, apparatus, and computer program product for machine translation |
US20080133245A1 (en) * | 2006-12-04 | 2008-06-05 | Sehda, Inc. | Methods for speech-to-speech translation |
US7467085B2 (en) * | 2000-10-17 | 2008-12-16 | Hitachi, Ltd. | Method and apparatus for language translation using registered databases |
US7539619B1 (en) * | 2003-09-05 | 2009-05-26 | Spoken Translation Ind. | Speech-enabled language translation system and method enabling interactive user supervision of translation and speech recognition accuracy |
US20090234636A1 (en) * | 2008-03-14 | 2009-09-17 | Jay Rylander | Hand held language translation and learning device |
US20090299724A1 (en) * | 2008-05-28 | 2009-12-03 | Yonggang Deng | System and method for applying bridging models for robust and efficient speech to speech translation |
US20100008582A1 (en) * | 2008-07-10 | 2010-01-14 | Samsung Electronics Co., Ltd. | Method for recognizing and translating characters in camera-based image |
US20110093254A1 (en) * | 2008-06-09 | 2011-04-21 | Roland Kuhn | Method and System for Using Alignment Means in Matching Translation |
US7983912B2 (en) * | 2005-09-27 | 2011-07-19 | Kabushiki Kaisha Toshiba | Apparatus, method, and computer program product for correcting a misrecognized utterance using a whole or a partial re-utterance |
US20110264439A1 (en) * | 2008-02-29 | 2011-10-27 | Ichiko Sata | Information processing device, method and program |
US20130103382A1 (en) * | 2011-10-19 | 2013-04-25 | Electronics And Telecommunications Research Institute | Method and apparatus for searching similar sentences |
US8527522B2 (en) * | 2008-09-05 | 2013-09-03 | Ramp Holdings, Inc. | Confidence links between name entities in disparate documents |
US8583418B2 (en) * | 2008-09-29 | 2013-11-12 | Apple Inc. | Systems and methods of detecting language and natural language strings for text to speech synthesis |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2006012179A (ja) | 2005-07-19 | 2006-01-12 | Sony Corp | 自然言語処理装置および自然言語処理方法 |
JP5002271B2 (ja) | 2007-01-18 | 2012-08-15 | 株式会社東芝 | 入力された原言語文を目的言語に機械翻訳する装置、方法およびプログラム |
-
2009
- 2009-12-21 KR KR1020090127709A patent/KR101377459B1/ko not_active IP Right Cessation
-
2010
- 2010-12-15 US US12/969,017 patent/US20110153309A1/en not_active Abandoned
Patent Citations (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4507750A (en) * | 1982-05-13 | 1985-03-26 | Texas Instruments Incorporated | Electronic apparatus from a host language |
US5477451A (en) * | 1991-07-25 | 1995-12-19 | International Business Machines Corp. | Method and system for natural language translation |
US5293584A (en) * | 1992-05-21 | 1994-03-08 | International Business Machines Corporation | Speech recognition system for natural language translation |
US5510981A (en) * | 1993-10-28 | 1996-04-23 | International Business Machines Corporation | Language translation apparatus and method using context-based translation models |
US5689616A (en) * | 1993-11-19 | 1997-11-18 | Itt Corporation | Automatic language identification/verification system |
US6233545B1 (en) * | 1997-05-01 | 2001-05-15 | William E. Datig | Universal machine translator of arbitrary languages utilizing epistemic moments |
US7467085B2 (en) * | 2000-10-17 | 2008-12-16 | Hitachi, Ltd. | Method and apparatus for language translation using registered databases |
US6763331B2 (en) * | 2001-02-01 | 2004-07-13 | Matsushita Electric Industrial Co., Ltd. | Sentence recognition apparatus, sentence recognition method, program, and medium |
US20020135618A1 (en) * | 2001-02-05 | 2002-09-26 | International Business Machines Corporation | System and method for multi-modal focus detection, referential ambiguity resolution and mood classification using multi-modal input |
US20040243392A1 (en) * | 2003-05-27 | 2004-12-02 | Kabushiki Kaisha Toshiba | Communication support apparatus, method and program |
US7539619B1 (en) * | 2003-09-05 | 2009-05-26 | Spoken Translation Ind. | Speech-enabled language translation system and method enabling interactive user supervision of translation and speech recognition accuracy |
US20060004560A1 (en) * | 2004-06-24 | 2006-01-05 | Sharp Kabushiki Kaisha | Method and apparatus for translation based on a repository of existing translations |
US20070016401A1 (en) * | 2004-08-12 | 2007-01-18 | Farzad Ehsani | Speech-to-speech translation system with user-modifiable paraphrasing grammars |
US7983912B2 (en) * | 2005-09-27 | 2011-07-19 | Kabushiki Kaisha Toshiba | Apparatus, method, and computer program product for correcting a misrecognized utterance using a whole or a partial re-utterance |
US20070208569A1 (en) * | 2006-03-03 | 2007-09-06 | Balan Subramanian | Communicating across voice and text channels with emotion preservation |
US20080077391A1 (en) * | 2006-09-22 | 2008-03-27 | Kabushiki Kaisha Toshiba | Method, apparatus, and computer program product for machine translation |
US20080077392A1 (en) * | 2006-09-26 | 2008-03-27 | Kabushiki Kaisha Toshiba | Method, apparatus, system, and computer program product for machine translation |
US20080133245A1 (en) * | 2006-12-04 | 2008-06-05 | Sehda, Inc. | Methods for speech-to-speech translation |
US20110264439A1 (en) * | 2008-02-29 | 2011-10-27 | Ichiko Sata | Information processing device, method and program |
US20090234636A1 (en) * | 2008-03-14 | 2009-09-17 | Jay Rylander | Hand held language translation and learning device |
US20090299724A1 (en) * | 2008-05-28 | 2009-12-03 | Yonggang Deng | System and method for applying bridging models for robust and efficient speech to speech translation |
US20110093254A1 (en) * | 2008-06-09 | 2011-04-21 | Roland Kuhn | Method and System for Using Alignment Means in Matching Translation |
US20100008582A1 (en) * | 2008-07-10 | 2010-01-14 | Samsung Electronics Co., Ltd. | Method for recognizing and translating characters in camera-based image |
US8625899B2 (en) * | 2008-07-10 | 2014-01-07 | Samsung Electronics Co., Ltd. | Method for recognizing and translating characters in camera-based image |
US8527522B2 (en) * | 2008-09-05 | 2013-09-03 | Ramp Holdings, Inc. | Confidence links between name entities in disparate documents |
US8583418B2 (en) * | 2008-09-29 | 2013-11-12 | Apple Inc. | Systems and methods of detecting language and natural language strings for text to speech synthesis |
US20130103382A1 (en) * | 2011-10-19 | 2013-04-25 | Electronics And Telecommunications Research Institute | Method and apparatus for searching similar sentences |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9189476B2 (en) | 2012-04-04 | 2015-11-17 | Electronics And Telecommunications Research Institute | Translation apparatus and method thereof for helping a user to more easily input a sentence to be translated |
CN103902523A (zh) * | 2012-12-28 | 2014-07-02 | 新疆电力信息通信有限责任公司 | 维吾尔语句子相似度计算方法 |
US20140258322A1 (en) * | 2013-03-06 | 2014-09-11 | Electronics And Telecommunications Research Institute | Semantic-based search system and search method thereof |
US9268767B2 (en) * | 2013-03-06 | 2016-02-23 | Electronics And Telecommunications Research Institute | Semantic-based search system and search method thereof |
US8843369B1 (en) * | 2013-12-27 | 2014-09-23 | Google Inc. | Speech endpointing based on voice profile |
US10546576B2 (en) | 2014-04-23 | 2020-01-28 | Google Llc | Speech endpointing based on word comparisons |
US11004441B2 (en) | 2014-04-23 | 2021-05-11 | Google Llc | Speech endpointing based on word comparisons |
US10140975B2 (en) | 2014-04-23 | 2018-11-27 | Google Llc | Speech endpointing based on word comparisons |
US9607613B2 (en) | 2014-04-23 | 2017-03-28 | Google Inc. | Speech endpointing based on word comparisons |
US11636846B2 (en) | 2014-04-23 | 2023-04-25 | Google Llc | Speech endpointing based on word comparisons |
US10269341B2 (en) | 2015-10-19 | 2019-04-23 | Google Llc | Speech endpointing |
US11710477B2 (en) | 2015-10-19 | 2023-07-25 | Google Llc | Speech endpointing |
US11062696B2 (en) | 2015-10-19 | 2021-07-13 | Google Llc | Speech endpointing |
JP2018055622A (ja) * | 2016-09-30 | 2018-04-05 | 大日本印刷株式会社 | 情報処理装置及びプログラム |
US10929754B2 (en) | 2017-06-06 | 2021-02-23 | Google Llc | Unified endpointer using multitask and multidomain learning |
US11551709B2 (en) | 2017-06-06 | 2023-01-10 | Google Llc | End of query detection |
US11676625B2 (en) | 2017-06-06 | 2023-06-13 | Google Llc | Unified endpointer using multitask and multidomain learning |
US10593352B2 (en) | 2017-06-06 | 2020-03-17 | Google Llc | End of query detection |
CN107291939A (zh) * | 2017-07-06 | 2017-10-24 | 携程计算机技术(上海)有限公司 | 酒店信息的聚类匹配方法及系统 |
US11144732B2 (en) | 2017-10-12 | 2021-10-12 | Electronics And Telecommunications Research Institute | Apparatus and method for user-customized interpretation and translation |
CN111048073A (zh) * | 2019-12-16 | 2020-04-21 | 北京明略软件系统有限公司 | 一种音频处理方法、装置、电子设备及可读存储介质 |
CN111292814A (zh) * | 2019-12-26 | 2020-06-16 | 北京亚信数据有限公司 | 一种医疗数据标准化的方法及装置 |
US11887579B1 (en) | 2022-09-28 | 2024-01-30 | Intuit Inc. | Synthetic utterance generation |
Also Published As
Publication number | Publication date |
---|---|
KR20110071205A (ko) | 2011-06-29 |
KR101377459B1 (ko) | 2014-03-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20110153309A1 (en) | Automatic interpretation apparatus and method using utterance similarity measure | |
US10796105B2 (en) | Device and method for converting dialect into standard language | |
US9710463B2 (en) | Active error detection and resolution for linguistic translation | |
US8977536B2 (en) | Method and system for translating information with a higher probability of a correct translation | |
US8131536B2 (en) | Extraction-empowered machine translation | |
US8566076B2 (en) | System and method for applying bridging models for robust and efficient speech to speech translation | |
US8370130B2 (en) | Speech understanding system using an example-based semantic representation pattern | |
US20060282255A1 (en) | Collocation translation from monolingual and available bilingual corpora | |
US20080133245A1 (en) | Methods for speech-to-speech translation | |
US9189476B2 (en) | Translation apparatus and method thereof for helping a user to more easily input a sentence to be translated | |
US20150199340A1 (en) | System for translating a language based on user's reaction and method thereof | |
WO2010046782A2 (en) | Hybrid machine translation | |
WO2018055983A1 (ja) | 翻訳装置、翻訳システム、および評価サーバ | |
US9082404B2 (en) | Recognizing device, computer-readable recording medium, recognizing method, generating device, and generating method | |
CN104850543A (zh) | 语音对话支持装置和语音对话支持方法 | |
US20120010873A1 (en) | Sentence translation apparatus and method | |
US10818283B2 (en) | Speech recognition system, terminal device, and dictionary management method | |
US20050060150A1 (en) | Unsupervised training for overlapping ambiguity resolution in word segmentation | |
US20180089176A1 (en) | Method of translating speech signal and electronic device employing the same | |
Sakti et al. | A-STAR: Toward translating Asian spoken languages | |
US8527270B2 (en) | Method and apparatus for conducting an interactive dialogue | |
Ablimit et al. | Stem-affix based Uyghur morphological analyzer | |
CN113268981B (zh) | 一种信息处理方法、装置及电子设备 | |
Huang et al. | Error detection and correction based on Chinese phonemic alphabet in Chinese text | |
Chao et al. | Automatic spelling correction for asr corpus in traditional chinese language using seq2seq models |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ELECTRONICS AND TELECOMMUNICATIONS RESEARCH INSTIT Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KIM, JEONG SE;KIM, SANG HUN;YUN, SEUNG;AND OTHERS;REEL/FRAME:025506/0183 Effective date: 20101117 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |