WO2006001204A1 - Automatic translation device and automatic translation method - Google Patents

Automatic translation device and automatic translation method Download PDF

Info

Publication number
WO2006001204A1
WO2006001204A1 PCT/JP2005/010946 JP2005010946W WO2006001204A1 WO 2006001204 A1 WO2006001204 A1 WO 2006001204A1 JP 2005010946 W JP2005010946 W JP 2005010946W WO 2006001204 A1 WO2006001204 A1 WO 2006001204A1
Authority
WO
WIPO (PCT)
Prior art keywords
language
speech
automatic translation
intermediate language
module
Prior art date
Application number
PCT/JP2005/010946
Other languages
French (fr)
Japanese (ja)
Inventor
Hiromichi Ishibashi
Original Assignee
Matsushita Electric Industrial Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Matsushita Electric Industrial Co., Ltd. filed Critical Matsushita Electric Industrial Co., Ltd.
Publication of WO2006001204A1 publication Critical patent/WO2006001204A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/58Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation

Definitions

  • the present invention relates to an automatic translation apparatus and an automatic translation method for executing interpretation processing between two languages in real time and in a user-friendly manner.
  • Patent Document 1 recognizes the voice of a speaker in the first language, translates it into a second language, converts it into a further language, and converts it to a listener in the second language, and vice versa. Is described. Further, Patent Document 2 describes an invention in which the translated second language is output by a dedicated speaker provided in a mobile phone. Furthermore, Patent Document 3 describes a technique for performing translation between two languages using a display display that does not use voice. Furthermore, Patent Documents 4 to 6 describe configurations suitable for a portable type. Further, Patent Document 7 describes a technique in which both voice and display are used and a voice input confirmation display is displayed on the display. In particular, in Patent Document 7, in order to allow a certain amount of “listening mistakes” in view of the fact that today's technology still does not fully recognize spoken language when performing automatic translation of speech input, It is stated that the first and second languages are displayed simultaneously.
  • Patent Document 1 Japanese Patent Laid-Open No. 9 292971
  • Patent Document 2 JP-A-8-265445
  • Patent Document 3 Japanese Patent Laid-Open No. 6-119372
  • Patent Document 4 JP-A-6-325072
  • Patent Document 5 JP-A-8-77176
  • Patent Document 6 Japanese Patent Laid-Open No. 9 160881
  • Patent Document 7 JP-A-8-278972
  • An object of the present invention is to provide an automatic translation apparatus that executes interpretation processing between two languages in real time and in a user-friendly manner.
  • An automatic translation apparatus comprises: a first speech analysis module that analyzes speech spoken in a first language and generates a first intermediate language;
  • a first translation module for converting the first intermediate language into a second language
  • a character synthesis module for displaying the converted second language on a display.
  • the first intermediate language may further include a first speech synthesis module that reconverts the first intermediate language into a first national language and generates the reconverted speech signal of the first national language. Furthermore, before converted A second speech synthesis module that generates a speech signal in the second language may be further provided.
  • a confirmation button for confirming the first intermediate language is further provided.
  • a correction button may be further provided that makes the first intermediate language indeterminate and prompts the user to re-input the voice in the first language.
  • a second speech analysis module that analyzes speech spoken in the second language and generates a second intermediate language may be further provided.
  • a second translation module for converting the second intermediate language into the first language may be further provided.
  • the character synthesis module may re-convert the second intermediate language into a second language and display it on the display.
  • a confirmation button for confirming the second intermediate language may be further provided.
  • a correction button may be further provided that makes the second intermediate language uncertain and prompts the user to re-input the voice in the second language.
  • a first selection unit that switches an input source of the character synthesis module between the first translation module and the second speech analysis module may be further provided.
  • a second selection unit that switches an input source of the first speech synthesis module between the first speech analysis module and the second translation module may be further provided.
  • An automatic translation method comprises analyzing a speech spoken in a first language and generating a first intermediate language
  • the method may further include a step of reconverting the first intermediate language into a first language and generating a reconverted speech signal of the first language. Furthermore, the method may further comprise the step of generating a converted speech signal of the second language.
  • the method may further include a step of confirming the first intermediate language by listening to the voice signal of the first language.
  • the method may further include a step of listening to the first national language speech signal to make the first intermediate language uncertain and prompting re-input of the first national language speech.
  • the step of analyzing the speech spoken in the second language and generating the second intermediate language is further performed. May be included. Furthermore, the method may further include a step of converting the second intermediate language into the first language. Further, the method may further include a step of re-converting the second intermediate language into the second language and displaying it on the display!
  • the method may further include the step of confirming the second intermediate language by viewing the second national language displayed on the display.
  • the first displayed on the display may further include the step of confirming the second intermediate language by viewing the second national language displayed on the display.
  • It may further include a step of looking at the bilingual language and making the second intermediate language uncertain and prompting the user to re-enter the speech of the second national language.
  • the automatic translation method may be configured as a computer-executable program. Furthermore, the above program may be stored in a computer-readable recording medium.
  • the conversation information is transmitted to the oral language for the first language (Japanese) speaker wearing the device, and visually to the second language (English) speaker.
  • the speaker can recognize the contents spoken by each speaker.
  • FIG. 1 is a block diagram of an automatic translation apparatus according to Embodiment 1 of the present invention.
  • FIG. 2 is a partial external view of the automatic translation apparatus according to Embodiment 1 of the present invention.
  • FIG. 3 is a conceptual diagram for explaining the operation of the automatic translation apparatus according to Embodiment 1 of the present invention.
  • FIG. 4 is a conceptual diagram for explaining the operation of the automatic translation apparatus according to Embodiment 1 of the present invention.
  • FIG. 5 is a block diagram showing an internal configuration of a first translation module.
  • FIG. 6 is a flowchart of an automatic translation method according to the first embodiment of the present invention.
  • FIG. 7 is a flowchart of an automatic translation method according to the first embodiment of the present invention.
  • FIG. 8 is a block diagram of an automatic translation apparatus according to Embodiment 2 of the present invention.
  • FIG. 9 is a partial external view of an automatic translation apparatus according to Embodiment 2 of the present invention. Explanation of symbols
  • FIG. 1 is a block diagram showing the configuration of the automatic translation apparatus according to Embodiment 1 of the present invention.
  • FIG. 2 is a partial external view.
  • This automatic translation apparatus includes a first microphone 1, a first speech analysis module 2, a first speech synthesis module 3, a first translation module 4, a character synthesis module 5, a display 6, a second microphone 7, a second speech analysis module 8 A second translation module 10, a speaker 11, a first selector 9, a second selector 12, a first dictionary 20, and a second dictionary 80.
  • the first microphone 1 converts the first language spoken by the first language speaker into an electrical signal.
  • the first speech analysis module 2 converts the first language speech converted to the electrical signal into the first intermediate language.
  • the first speech synthesis module 3 generates a first language speech signal from the first intermediate language, and also generates 11 speakers.
  • the first translation module 4 converts the first intermediate language into the second Convert to Japanese.
  • the character synthesis module 5 displays the second language on the display 6.
  • the second microphone 7 converts the second language spoken by the second language speaker into an electrical signal.
  • the second speech analysis module 8 converts the second language speech converted into the electrical signal into the second intermediate language.
  • the second translation module 10 converts the second intermediate language into the first language.
  • the first selector 9 switches the input source to the character synthesis module 5 between the first translation module 4 and the second speech analysis module 8.
  • the second selector 12 switches the input source to the first speech synthesis module 3 between the first speech analysis module 2 and the second translation module 10.
  • the first dictionary 20 is used in the first speech analysis module 2 and the first speech synthesis module 3.
  • the second dictionary 80 is used in the second character synthesis module 5 and the second speech analysis module 8.
  • this automatic translation apparatus has, for example, the first microphone 1, the first speech analysis module 2, and the first translation as the most basic configuration in which not all of the above-mentioned members are essential members. It only needs to have module 4 and character synthesis module 5.
  • the display 6 for display may be a display of a mobile phone or another mobile terminal.
  • the first language spoken by the first language speaker is converted into the second language and displayed on the display, so that the second language speaker can easily understand visually.
  • the first speech synthesis module 3 and the speaker 11 may be provided in addition to the basic configuration described above.
  • the first language spoken by the first language speaker is once converted to the first intermediate language, and then the first language speaker himself hears the re-converted first language speech. It is possible to confirm whether or not the Japanese language voice analysis has been performed correctly.
  • the second speech analysis module 8 and the second translation module 10 can be further provided. With this more applied configuration, only the first language speaker can install a two-way conversation between the first language speaker and the second language speaker by installing this automatic translation device. In this case, the first language speaker can orally recognize the translation result, and the second language speaker can visually recognize the translation result.
  • FIG. 5 is a block diagram showing a detailed configuration of the first speech analysis module 2.
  • the first speech analysis module 2 includes a word analysis unit 21, a grammar analysis unit 22, and a semantic analysis unit 23.
  • the word analysis unit 21 recognizes a word from the voice signal.
  • the grammar analyzer 22 The semantic analysis unit 23 analyzes the grammatical relationship between the word strings, extracts the meaning of each word, and outputs the first intermediate language corresponding to the extracted meaning.
  • “intermediate language” means a data string in which the meaning of words and syntactic attributes (ie, classification and connection of verbs, nouns, and adjectives) become clear.
  • the second speech analysis module 8 is different from the first speech analysis module 2 in that it converts the second language speech signal into the second intermediate language, but can be realized by a configuration similar to the configuration shown in FIG.
  • FIG. 2 is a schematic diagram showing a partial appearance of an automatic translation apparatus used in a conversation between a first language speaker and a second language speaker.
  • the right force also shows the front and side views of the first language speaker and the side view of the second language speaker.
  • This automatic translation apparatus is worn by a first language speaker.
  • FIG. 2 shows a microphone 1, a speaker 11, a display 6, and a microphone 7.
  • Fig. 2 shows an example in which “Ookini” spoken by the first language speaker is converted to the second language “Thank youj” and displayed on the display 6.
  • Fig. 3 shows the first language speaker.
  • FIG. 6 is a diagram conceptually showing a process that is displayed on the display 6.
  • Figure 4 shows that the speech spoken by a second language speaker (English “You are welcome”) is converted to an intermediate language code and then converted back to English “You are welcome”. It is a figure which shows notionally the process in which it is converted into 1 national language (Japanese "youtashimashi”) and is output by voice.
  • the voice of a first language (hereinafter, Japanese) speaker is converted into an electrical signal by the microphone 1 and supplied to the first voice analysis module 2.
  • the first speech analysis module 2 recognizes the speech signal and generates the first intermediate language C1. For example, as shown in FIG. 2, when a Japanese speaker utters “Ookini” as he or she urges microphone 1, this electric signal string is stored in (Japanese) dictionary 20 as shown in FIG. Based on the result of the analysis by the speech analysis module 2, a code (“CF5A88Dh" in Fig. 3) is generated as the first intermediate language meaning thanks.
  • the first intermediate language C1 is converted into the second language (English) by the first translation module 4, and simultaneously supplied to the first speech synthesis module 3. Converted.
  • a Japanese speaker can hear the reconverted speech from the speaker 11 attached close to the ear.
  • the first speech synthesis module 3 converts the first intermediate language (code) having the meaning of gratitude into Japanese using the dictionary 20.
  • the spoken word and the intermediate language do not necessarily correspond to 1: 1.
  • “Thank you”, “Ookini”, and “Dandan” (if registered in dictionary 20) may be converted to the same intermediate language “CF5A88Dh” that expresses gratitude.
  • the question is which Japanese to select. The meaning is understood by the speaker.
  • the second national language (English) converted by the first translation module 4 is documented by the character synthesis module 5 and displayed on the display 6.
  • processing is started (S01).
  • the instruction to start the process is not particularly shown, but there is a method of pressing a button provided on the apparatus main body or the microphone, for example.
  • the input speech is converted to the first intermediate language C1 by the first speech analysis module 2 (S03).
  • speech analysis recognition
  • specific speaker recognition technology can be used.
  • the first language speaker can dictate a predetermined basic sentence to the apparatus in advance, and can learn habits and unique passages in the apparatus.
  • the speech recognition is completed, the first intermediate language C1 is generated.
  • the first intermediate language C1 is an information stream whose meaning is fixed including attributes such as nouns and verbs for the first language.
  • the first intermediate language C1 is supplied to the first speech synthesis module 3, converted into the first language, and output through the speaker 11 (S04).
  • the speaker 11 broadly includes a device that converts an electrical signal into sound, for example, a headphone and an earphone.
  • the first language speaker listens to the speaker voice and the content of his speech is accurate. (S05). If it is determined that the recognition has been made in error, the first language speaker presses a correction button (not shown). On the other hand, if it is determined that the speaker is correctly recognized, the speaker presses a confirmation button (not shown).
  • the correction button is not particularly shown in the figure, but for example, a touch sensor provided on the display 6, a push button provided on the apparatus housing, and the like are conceivable.
  • an acceleration sensor is installed there, and the same processing as pressing the correction button is performed when the head is shaken sideways, that is, when horizontal acceleration is detected. Can be done.
  • the confirmation button that is pressed when it is determined that the dictation content has been correctly recognized can have the same mode as the correction button.
  • the acceleration sensor when the vertical acceleration is detected by swinging the head vertically, the same processing as when the enter button is pressed can be performed. .
  • the apparatus determines whether or not the correction button Z confirmation button has been pressed (S06).
  • the voice spoken by the first language speaker is converted into the second language and displayed on the display 6, so that the second language speaker can be visually recognized and is therefore affected by ambient noise. Without understanding the translation results.
  • the first intermediate language C1 is translated by the first translation module 4 into the second intermediate language C2 of the second language (English). Since Japanese and English have different grammatical systems, intermediate language structures (modifications, connection relationships, etc.) must be converted based on the prescribed rules. .
  • the translated second intermediate language C2 is documented by the character synthesis module 5 based on the dictionary 80 and displayed on the display 6. In the case of this embodiment, it is translated into “T hank you” representing gratitude. As shown in Fig. 2, display 6 should be worn around the chest of the first language (Japanese) speaker so that the second language (English) speaker can read with a natural perspective. It is good to do. In the arrangement shown in Fig.
  • a low power type reflective liquid crystal element, an electrophoretic element, or the like may be used. Also, it may use a liquid crystal display part of a mopile AV device such as a mobile phone or PDA, a video movie camera or a digital still camera.
  • processing is started (Sl l).
  • the instruction to start the process is not particularly shown, but there is a method of pressing a button provided on the apparatus main body or the microphone, for example.
  • the case where the acceleration sensor as described above is used is also conceivable. That is, the processing may be started when the head is swung vertically, that is, when acceleration in the vertical direction is detected. Note that this start process has a common purpose in the sense that the power process, which is the same as step 8 in FIG. 6, is advanced, and the first language speaker should not be confused. Conceivable.
  • the second language speaker (English speaker) speaks to the microphone 7 in the second language, and is converted into a voice signal by the microphone 7 (S12). In the example in Figure 4, it is “You are welcome”
  • the second language speech signal is subjected to speech analysis (8) based on the dictionary 80 by the second speech analysis module 8, as shown in FIG. 4, and the second intermediate language C2 ("FAC653BOh" ) Is raw (S13).
  • speech analysis speech recognition
  • speakers of the second language unlike in the case of the first language, speakers of the second language often change from time to time, so use specific speaker recognition technology that can improve recognition efficiency. I can't.
  • the vocabulary that the second language speaker speaks is often “reply” to the first language speaker, the recognition of unspecified speakers is considered to be sufficient.
  • the second intermediate language C 2 is an information stream whose meaning is fixed including attributes such as nouns and verbs for the second language.
  • the second intermediate language C2 is converted into a second national language, documented by the character synthesis module 5, and displayed on the display 6 (S14).
  • the second language speaker can confirm whether or not the word spoken by the second language speaker is correctly recognized (S15). Here, if it is determined that the recognition has been mistaken, the second language speaker presses a correction button (not shown). Alternatively, the second language speaker instructs the first language speaker to press the correction button by a gesture or the like. On the other hand, if it is determined that the speaker has been correctly recognized, the speaker presses a confirmation button (not shown).
  • this correction button is not particularly shown, but can be realized by a touch sensor provided on the display 6, a push button provided on the apparatus housing, or the like.
  • an acceleration sensor is provided there, and the above correction is made when a first language speaker shakes his / her head, that is, when a horizontal acceleration is detected.
  • the same processing as when the normal button is pressed may be performed.
  • the same processing as pressing the correction switch may be performed.
  • the above confirmation button can be realized in the same form as the above correction button.
  • the same processing as when the confirmation button is pressed may be performed when the head is swung vertically, that is, when acceleration in the vertical direction is detected. Good.
  • the apparatus determines whether or not the correction button Z confirmation button has been pressed (S 16). As described above, when it is determined that the dictation content of the second language speaker has been mistakenly recognized, the correction button is pressed or a corresponding means is taken, and the above processing is performed again after returning to step S12. Executed.
  • the speech spoken by the second language speaker is converted into the first language and output from the speaker 11 provided in the vicinity of the ear of the first language speaker. Translation results can be recognized orally without being affected by noise.
  • automatic translation between two languages can be executed in a more natural manner with a configuration suitable for mopile.
  • a Japanese speaker can hear the recognized self-language and the English speaker's language translated into Japanese from the speaker 11 in real time.
  • the “speaker” may be a so-called earphone.
  • the English speaker can see the recognized self-word and the Japanese speaker's word translated into English on the display 6. This is possible even in the open air.
  • selectors 9 and 12 are used to switch between passing the own voice information and passing the other party's voice information. Specifically, it is controlled by a microprocessor (not shown), and for example, the audio input from microphone 1 and the audio input from microphone 2 are input according to the amplitude of each audio signal. The source may be switched.
  • FIG. 8 is a block diagram showing the configuration of the automatic translation apparatus according to Embodiment 2 of the present invention.
  • Fig. 9 is a partial external view.
  • This automatic translation apparatus is different from the automatic translation apparatus according to Embodiment 1 in that it includes a second speech synthesis module 51 and a second speaker 61.
  • the second speech synthesis module 51 converts only English converted by the first translation module 4 into a speech signal.
  • this audio signal is converted into audio by a second speaker 61 provided near the display 6.
  • the English speaker can also visually confirm the translation of the spoken language of the Japanese speaker through the display 6, and in a relatively quiet environment, the second speaker 61 It is also possible to listen to the voice as a voice, and at the same time, it is possible to realize a conversation with higher accuracy.
  • the second speaker 61 is separately provided in the vicinity of the display 6 as shown in FIG. 9.
  • the present invention is not limited to these, and for example, a transparent piezoelectric speaker is used. May be pasted on the display surface of the display 6.
  • the automatic translation apparatus has a function of transmitting conversation information to an oral speaker for a Japanese speaker and visually to an English speaker, and as an electronic interpreter for mopile, etc. Especially useful when used during overseas travel. It can also be used for international conferences.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Machine Translation (AREA)

Abstract

An automatic translation device comprises a first speech analysis module for analyzing a speech spoken in a first language and generating a speech in a first intermediate language, a speech synthesis module for reconverting the speech in the first intermediate language into a speech in the first language and generating a speech signal, a first translation module for converting the speech in the first intermediate language into a speech in a second language, and a character generating module for displaying the speech on a display. Conversation information is aurally conveyed to the person speaking in the first language and wearing this automatic translation device and visually conveyed to the person speaking in the second language. In each conveying form, the content that one of the persons has spoken can be ascertained by the person by oneself.

Description

明 細 書  Specification
自動翻訳装置及び自動翻訳方法  Automatic translation apparatus and automatic translation method
技術分野  Technical field
[0001] 本発明は、リアルタイムで、しかもユーザーフレンドリーに、 2国語間の通訳処理を 実行する自動翻訳装置及び自動翻訳方法に関する。  [0001] The present invention relates to an automatic translation apparatus and an automatic translation method for executing interpretation processing between two languages in real time and in a user-friendly manner.
背景技術  Background art
[0002] 従来の自動翻訳装置としては、以下に列記するものがある。まず、特許文献 1には 第 1国語の話者の音声を認識した後、第 2国語に翻訳'変換して、さらに音声変換し て、第 2国語の聞き手に伝える、またその逆を行うものが記載されている。また、翻訳 された第 2国語を携帯電話に設けられた専用スピーカ一力 音声出力させる形態の 発明が特許文献 2に記載されている。さらに、音声ではなぐディスプレイ表示で二国 語間の翻訳を実行する技術が特許文献 3に記載されている。またさらに、ポータブル タイプに適した構成が特許文献 4〜6に記載されている。また、音声と表示の双方を 用い、ディスプレイに音声入力の確認表示がされるといった技術が特許文献 7に記 載されている。特に、特許文献 7においては、音声入力の自動翻訳を実行するにあ たって、今日の技術では未だ完全に音声言語を認識することはできないことに鑑み、 ある程度の"聞き取りミス"を許容するため、第 1国語と第 2国語を同時表示させる旨が 記載されている。  [0002] Conventional automatic translation apparatuses include those listed below. First, Patent Document 1 recognizes the voice of a speaker in the first language, translates it into a second language, converts it into a further language, and converts it to a listener in the second language, and vice versa. Is described. Further, Patent Document 2 describes an invention in which the translated second language is output by a dedicated speaker provided in a mobile phone. Furthermore, Patent Document 3 describes a technique for performing translation between two languages using a display display that does not use voice. Furthermore, Patent Documents 4 to 6 describe configurations suitable for a portable type. Further, Patent Document 7 describes a technique in which both voice and display are used and a voice input confirmation display is displayed on the display. In particular, in Patent Document 7, in order to allow a certain amount of “listening mistakes” in view of the fact that today's technology still does not fully recognize spoken language when performing automatic translation of speech input, It is stated that the first and second languages are displayed simultaneously.
[0003] 特許文献 1 :特開平 9 292971号公報  Patent Document 1: Japanese Patent Laid-Open No. 9 292971
特許文献 2:特開平 8 - 265445号公報  Patent Document 2: JP-A-8-265445
特許文献 3:特開平 6 - 119372号公報  Patent Document 3: Japanese Patent Laid-Open No. 6-119372
特許文献 4:特開平 6— 325072号公報  Patent Document 4: JP-A-6-325072
特許文献 5:特開平 8 - 77176号公報  Patent Document 5: JP-A-8-77176
特許文献 6:特開平 9 160881号公報  Patent Document 6: Japanese Patent Laid-Open No. 9 160881
特許文献 7:特開平 8 - 278972号公報  Patent Document 7: JP-A-8-278972
発明の開示  Disclosure of the invention
発明が解決しょうとする課題 [0004] しかし、前記列記した従来の構成では、実使用上で問題となるところがある。まず、 特許文献 1に記載されている音声一音声間の翻訳実行方式の課題について説明す る。この方式は、一見、同時通訳者を介したかのように二力国語間の会話が実現でき るといった点で優れている力 モパイルにしかも野外で用いる場合に問題がある。す なわち、音声でもって相手に情報を伝達しょうとすれば十分な音量で音声を発する 必要がある。特に、野外で使用する場合は周囲の雑音に十分うち勝つだけの音量が 必要である。そのため、出力の十分大きいスピーカーを用いる方法がある力 この場 合には機器の形状が大きくなり、し力も消費電力が増えるといったモノィル機器にと つて致命的な問題が生じる。次に、特許文献 2に記載されているような携帯電話の形 式をとれば耳とスピーカーを接近させることができ、サイズ'電力の問題は発生しない 。しかし、自分のみならず、相手方も同等の機器を装着しなければならず煩雑であつ た。 Problems to be solved by the invention [0004] However, the conventional configurations listed above have problems in practical use. First, the problem of the translation-to-speech translation method described in Patent Document 1 will be described. At first glance, this method has a problem when it is used in the field even in the field of mopile power, which is excellent in that conversation between two languages can be realized as if through a simultaneous interpreter. In other words, if you want to convey information to the other party by voice, it is necessary to emit the voice at a sufficient volume. In particular, when using outdoors, it is necessary to have a volume that can overcome the surrounding noise. For this reason, there is a method that uses a speaker with a sufficiently large output. In this case, the shape of the device becomes large, and the power consumption also increases fatal problems. Next, if the cellular phone format described in Patent Document 2 is adopted, the ear and the speaker can be brought close to each other, and the problem of size and power does not occur. However, it was cumbersome not only for me but also for the other party to wear equivalent devices.
[0005] 音声対話よりも、むしろディスプレイ表示、つまり文字によって相手と会話をする方 式 (例えば、特許文献 3から 7参照。)がモパイルでの使用下に適していると考えられ る。しかし、もともど'筆談"というのは日常会話からすれば自然な形態ではなぐ会話 中は相手よりもその翻訳機の表示を見る方に注意が行くことから、コミュニケーション がとりにくくなるのと、やはり安全上の問題 (例えば、筆談に気を取られているすきにピ ッキングに遭う、等)が発生するおそれがある。  [0005] Rather than voice conversation, a display method, that is, a method of talking with the other party using characters (see, for example, Patent Documents 3 to 7) is considered suitable for use in mopile. However, it is difficult to communicate because it is more difficult to communicate with the person who is looking at the display of the translator than the other party during conversation that is not natural in terms of daily conversation. There may still be a safety problem (for example, picking a spot that is distracted by writing).
[0006] 本発明の目的は、リアルタイムで、し力もユーザーフレンドリーに、 2国語間の通訳 処理を実行する自動翻訳装置を提供することである。  [0006] An object of the present invention is to provide an automatic translation apparatus that executes interpretation processing between two languages in real time and in a user-friendly manner.
課題を解決するための手段  Means for solving the problem
[0007] 本発明に係る自動翻訳装置は、第 1国語で話された音声を分析し、第 1中間言語 を生成する第 1音声分析モジュールと、 [0007] An automatic translation apparatus according to the present invention comprises: a first speech analysis module that analyzes speech spoken in a first language and generates a first intermediate language;
前記第 1中間言語を第 2国語に変換する第 1翻訳モジュールと、  A first translation module for converting the first intermediate language into a second language;
変換された前記第 2国語をディスプレイに表示する文字合成モジュールと を備えたことを特徴とする。  And a character synthesis module for displaying the converted second language on a display.
[0008] また、前記第 1中間言語を第 1国語に再変換し、再変換された前記第 1国語の音声 信号を生成する第 1音声合成モジュールをさらに備えてもよい。さらに、変換された前 記第 2国語の音声信号を生成する第 2音声合成モジュールをさらに備えてもよい。 [0008] The first intermediate language may further include a first speech synthesis module that reconverts the first intermediate language into a first national language and generates the reconverted speech signal of the first national language. Furthermore, before converted A second speech synthesis module that generates a speech signal in the second language may be further provided.
[0009] またさらに、前記第 1中間言語を確定する確定ボタンをさらに備えることを特徴とす る。また、前記第 1中間言語を不確定として、第 1国語の音声の再入力を促す訂正ボ タンをさらに備えてもよい。 [0009] Still further, a confirmation button for confirming the first intermediate language is further provided. In addition, a correction button may be further provided that makes the first intermediate language indeterminate and prompts the user to re-input the voice in the first language.
[0010] さらに、第 2国語で話された音声を分析し、第 2中間言語を生成する第 2音声分析 モジュールをさらに備えてもよい。また、前記第 2中間言語を第 1国語に変換する第 2 翻訳モジュールをさらに備えてもよ 、。 [0010] Furthermore, a second speech analysis module that analyzes speech spoken in the second language and generates a second intermediate language may be further provided. Further, a second translation module for converting the second intermediate language into the first language may be further provided.
[0011] なお、前記文字合成モジュールは、前記第 2中間言語を第 2国語に再変換し、ディ スプレイに表示してもよ 、。 [0011] Note that the character synthesis module may re-convert the second intermediate language into a second language and display it on the display.
[0012] また、前記第 2中間言語を確定する確定ボタンをさらに備えてもよい。さらに、前記 第 2中間言語を不確定として、第 2国語の音声の再入力を促す訂正ボタンをさらに備 えてもよい。 [0012] Further, a confirmation button for confirming the second intermediate language may be further provided. Further, a correction button may be further provided that makes the second intermediate language uncertain and prompts the user to re-input the voice in the second language.
[0013] さらに、前記文字合成モジュールの入力元を第 1翻訳モジュールと前記第 2音声分 析モジュールとの間で切り替える第 1選択部をさらに備えてもよい。また、前記第 1音 声合成モジュールの入力元を第 1音声分析モジュールと前記第 2翻訳モジュールと の間で切り替える第 2選択部をさらに備えてもよい。  [0013] Furthermore, a first selection unit that switches an input source of the character synthesis module between the first translation module and the second speech analysis module may be further provided. Further, a second selection unit that switches an input source of the first speech synthesis module between the first speech analysis module and the second translation module may be further provided.
[0014] 本発明に係る自動翻訳方法は、第 1国語で話された音声を分析し、第 1中間言語 を生成するステップと、  [0014] An automatic translation method according to the present invention comprises analyzing a speech spoken in a first language and generating a first intermediate language;
前記第 1中間言語を第 2国語に変換するステップと、  Converting the first intermediate language into a second language;
変換された前記第 2国語をディスプレイに表示するステップと  Displaying the converted second language on a display;
を含むことを特徴とする。  It is characterized by including.
[0015] また、前記第 1中間言語を第 1国語に再変換し、再変換された前記第 1国語の音声 信号を生成するステップをさらに含んでもよい。さらに、変換された前記第 2国語の音 声信号を生成するステップをさらに含んでもょ 、。  [0015] Further, the method may further include a step of reconverting the first intermediate language into a first language and generating a reconverted speech signal of the first language. Furthermore, the method may further comprise the step of generating a converted speech signal of the second language.
[0016] またさらに、前記第 1国語の音声信号を聞いて前記第 1中間言語を確定するステツ プをさらに含んでもよい。また、前記第 1国語の音声信号を聞いて前記第 1中間言語 を不確定として、第 1国語の音声の再入力を促すステップをさらに含んでもよい。  [0016] Further, the method may further include a step of confirming the first intermediate language by listening to the voice signal of the first language. In addition, the method may further include a step of listening to the first national language speech signal to make the first intermediate language uncertain and prompting re-input of the first national language speech.
[0017] さらに、第 2国語で話された音声を分析し、第 2中間言語を生成するステップをさら に含んでもよい。またさらに、前記第 2中間言語を第 1国語に変換するステップをさら に含んでもよい。また、前記第 2中間言語を第 2国語に再変換し、ディスプレイに表示 するステップをさらに含んでもよ!、。 [0017] Further, the step of analyzing the speech spoken in the second language and generating the second intermediate language is further performed. May be included. Furthermore, the method may further include a step of converting the second intermediate language into the first language. Further, the method may further include a step of re-converting the second intermediate language into the second language and displaying it on the display!
[0018] さらに、前記ディスプレイに表示された前記第 2国語を見て、前記第 2中間言語を確 定させるステップをさらに含んでもよい。また、前記ディスプレイに表示された前記第[0018] Further, the method may further include the step of confirming the second intermediate language by viewing the second national language displayed on the display. In addition, the first displayed on the display
2国語を見て、前記第 2中間言語を不確定として、第 2国語の音声の再入力を促すス テツプをさらに含んでもょ 、。 It may further include a step of looking at the bilingual language and making the second intermediate language uncertain and prompting the user to re-enter the speech of the second national language.
[0019] なお、上記自動翻訳方法は、コンピュータで実行可能なプログラムとして構成しても よい。さらに、上記プログラムはコンピュータ読み取り可能な記録媒体に格納してもよ い。 [0019] The automatic translation method may be configured as a computer-executable program. Furthermore, the above program may be stored in a computer-readable recording medium.
発明の効果  The invention's effect
[0020] 本構成によって、当該装置を装着した第一国語(日本語)話者に対してはオーラル に、対する第二国語 (英語)話者に対してはビジュアルに会話情報が伝達され、しか もそれぞれの伝達形態にお!、て、それぞれの話者が話した内容を当該話者自身が ½認することができる。  [0020] With this configuration, the conversation information is transmitted to the oral language for the first language (Japanese) speaker wearing the device, and visually to the second language (English) speaker. In each transmission mode, the speaker can recognize the contents spoken by each speaker.
[0021] 本発明に係る自動翻訳装置によれば、モパイルに適した、しかもより自然な形で、 二国語間の会話を実現することができる。  [0021] According to the automatic translation apparatus of the present invention, conversation between two languages can be realized in a more natural form suitable for mopile.
図面の簡単な説明  Brief Description of Drawings
[0022] [図 1]本発明の実施の形態 1に係る自動翻訳装置のブロック図である。 FIG. 1 is a block diagram of an automatic translation apparatus according to Embodiment 1 of the present invention.
[図 2]本発明の実施の形態 1に係る自動翻訳装置の一部外観図である。  FIG. 2 is a partial external view of the automatic translation apparatus according to Embodiment 1 of the present invention.
[図 3]本発明の実施の形態 1に係る自動翻訳装置の動作説明のための概念図である  FIG. 3 is a conceptual diagram for explaining the operation of the automatic translation apparatus according to Embodiment 1 of the present invention.
[図 4]本発明の実施の形態 1に係る自動翻訳装置の動作説明のための概念図である FIG. 4 is a conceptual diagram for explaining the operation of the automatic translation apparatus according to Embodiment 1 of the present invention.
[図 5]第 1翻訳モジュールの内部構成を示すブロック図である。 FIG. 5 is a block diagram showing an internal configuration of a first translation module.
[図 6]本発明の実施の形態 1に係る自動翻訳方法のフローチャートである。  FIG. 6 is a flowchart of an automatic translation method according to the first embodiment of the present invention.
[図 7]本発明の実施の形態 1に係る自動翻訳方法のフローチャートである。  FIG. 7 is a flowchart of an automatic translation method according to the first embodiment of the present invention.
[図 8]本発明の実施の形態 2に係る自動翻訳装置のブロック図である。 [図 9]本発明の実施の形態 2に係る自動翻訳装置の一部外観図である。 符号の説明 FIG. 8 is a block diagram of an automatic translation apparatus according to Embodiment 2 of the present invention. FIG. 9 is a partial external view of an automatic translation apparatus according to Embodiment 2 of the present invention. Explanation of symbols
[0023] 1、 7 マイク [0023] 1, 7 microphone
2、 8 音声分析モジュール  2, 8 Voice analysis module
3、 51 音声合成モジュール  3, 51 Speech synthesis module
4、 10 翻訳モジュール  4, 10 translation module
5 文字合成モジュール  5 Character synthesis module
6 ディスプレイ  6 display
9、 12 セレクタ  9, 12 selector
11、 61 スピーカー  11, 61 Speaker
20, 80 辞書  20, 80 dictionaries
21 単語解析部  21 Word Analysis Department
22 文法解析部  22 Grammar Analysis Department
23 意味解析部  23 Semantic analysis part
発明を実施するための最良の形態  BEST MODE FOR CARRYING OUT THE INVENTION
[0024] 以下、本発明の実施の形態に係る自動翻訳装置及び自動翻訳方法について、図 面を参照しながら説明する。なお、図面において実質的に同一の部材には同一の符 号を付している。  Hereinafter, an automatic translation apparatus and an automatic translation method according to an embodiment of the present invention will be described with reference to the drawings. In the drawings, substantially the same members are denoted by the same reference numerals.
[0025] (実施の形態 1)  [Embodiment 1]
図 1は、本発明の実施の形態 1における自動翻訳装置の構成を示すブロック図であ る。また、図 2は、その一部の外観図である。この自動翻訳装置は、第 1マイク 1、第 1 音声分析モジュール 2、第 1音声合成モジュール 3、第 1翻訳モジュール 4、文字合成 モジュール 5、ディスプレイ 6、第 2マイク 7、第 2音声分析モジュール 8、第 2翻訳モジ ユール 10、スピーカー 11、第 1セレクタ 9、第 2セレクタ 12、第 1辞書 20、第 2辞書 80 を備える。第 1マイク 1は、第 1国語話者の話す第 1国語の音声を電気信号に変換す る。第 1音声分析モジュール 2は、電気信号に変換された第 1国語の音声を第 1中間 言語に変換する。第 1音声合成モジュール 3は、第 1中間言語から第 1国語の音声信 号を生成し、スピーカー 11力も発する。第 1翻訳モジュール 4は、第 1中間言語を第 2 国語に変換する。文字合成モジュール 5は、第 2国語をディスプレイ 6に表示する。第 2マイク 7は、第 2国語話者の話す第 2国語の音声を電気信号に変換する。第 2音声 分析モジュール 8は、電気信号に変換された第 2国語の音声を第 2中間言語に変換 する。第 2翻訳モジュール 10は、第 2中間言語を第 1国語に変換する。第 1セレクタ 9 は、文字合成モジュール 5への入力元を、第 1翻訳モジュール 4と第 2音声分析モジ ユール 8との間で切り替える。第 2セレクタ 12は、第 1音声合成モジュール 3への入力 元を、第 1音声分析モジュール 2と第 2翻訳モジュール 10との間で切り替える。なお、 第 1辞書 20は、第 1音声分析モジュール 2及び第 1音声合成モジュール 3で利用され る。第 2辞書 80は、第 2文字合成モジュール 5及び第 2音声分析モジュール 8で利用 される。 FIG. 1 is a block diagram showing the configuration of the automatic translation apparatus according to Embodiment 1 of the present invention. FIG. 2 is a partial external view. This automatic translation apparatus includes a first microphone 1, a first speech analysis module 2, a first speech synthesis module 3, a first translation module 4, a character synthesis module 5, a display 6, a second microphone 7, a second speech analysis module 8 A second translation module 10, a speaker 11, a first selector 9, a second selector 12, a first dictionary 20, and a second dictionary 80. The first microphone 1 converts the first language spoken by the first language speaker into an electrical signal. The first speech analysis module 2 converts the first language speech converted to the electrical signal into the first intermediate language. The first speech synthesis module 3 generates a first language speech signal from the first intermediate language, and also generates 11 speakers. The first translation module 4 converts the first intermediate language into the second Convert to Japanese. The character synthesis module 5 displays the second language on the display 6. The second microphone 7 converts the second language spoken by the second language speaker into an electrical signal. The second speech analysis module 8 converts the second language speech converted into the electrical signal into the second intermediate language. The second translation module 10 converts the second intermediate language into the first language. The first selector 9 switches the input source to the character synthesis module 5 between the first translation module 4 and the second speech analysis module 8. The second selector 12 switches the input source to the first speech synthesis module 3 between the first speech analysis module 2 and the second translation module 10. The first dictionary 20 is used in the first speech analysis module 2 and the first speech synthesis module 3. The second dictionary 80 is used in the second character synthesis module 5 and the second speech analysis module 8.
[0026] なお、この自動翻訳装置は、上記部材の全てを必須の部材とするわけではなぐ最 も基本的な構成としては、例えば、第 1マイク 1、第 1音声分析モジュール 2、第 1翻訳 モジュール 4、文字合成モジュール 5を備えていればよい。この場合、表示用のディス プレイ 6は、携帯電話や他の携帯端末のディスプレイを利用すればよい。この基本構 成によって、第 1国語話者の話す第 1国語の音声が、第 2国語に変換されてディスプ レイに表示されるので、第 2国語話者が視覚的に容易に理解できる。さらに、この自 動翻訳装置の応用的な構成としては、上記基本構成に加えて第 1音声合成モジユー ル 3とスピーカー 11とを備えればよい。この応用的構成によって、第 1国語話者が話 した第 1国語がいったん第 1中間言語に変換され、その後、再変換された第 1国語の 音声を第 1国語話者自身が聞いて第 1国語の音声分析が正しく行われた力否かを確 認することができる。またさらに、より応用的には、さらに第 2音声分析モジュール 8、 及び、第 2翻訳モジュール 10を備える構成とすることができる。このより応用的な構成 により、第 1国語話者のみがこの自動翻訳装置を装着するだけで、第 1国語話者と第 2国語話者との双方向の会話を実現できる。この場合、第 1国語話者は、オーラルに 翻訳結果を認識でき、第 2国語話者は、ビジュアルに翻訳結果を認識できる。  [0026] It should be noted that this automatic translation apparatus has, for example, the first microphone 1, the first speech analysis module 2, and the first translation as the most basic configuration in which not all of the above-mentioned members are essential members. It only needs to have module 4 and character synthesis module 5. In this case, the display 6 for display may be a display of a mobile phone or another mobile terminal. With this basic configuration, the first language spoken by the first language speaker is converted into the second language and displayed on the display, so that the second language speaker can easily understand visually. Further, as an applied configuration of the automatic translation apparatus, the first speech synthesis module 3 and the speaker 11 may be provided in addition to the basic configuration described above. With this applied configuration, the first language spoken by the first language speaker is once converted to the first intermediate language, and then the first language speaker himself hears the re-converted first language speech. It is possible to confirm whether or not the Japanese language voice analysis has been performed correctly. Furthermore, in a more applied manner, the second speech analysis module 8 and the second translation module 10 can be further provided. With this more applied configuration, only the first language speaker can install a two-way conversation between the first language speaker and the second language speaker by installing this automatic translation device. In this case, the first language speaker can orally recognize the translation result, and the second language speaker can visually recognize the translation result.
[0027] 図 5は、第 1音声分析モジュール 2の詳細な構成を示すブロック図である。この第 1 音声分析モジュール2は、単語解析部 21、文法解析部 22、意味解析部 23を備える 。単語解析部 21では、音声信号カゝら単語を認識する。文法解析部 22では、複数の 単語列の文法的関係を解析し、意味解析部 23では、上記各単語の意味を抽出し、 抽出した意味に対応する第 1中間言語を出力する。ここで「中間言語」とは、単語の 意味と構文上の属性 (すなわち動詞、名詞、形容詞の分類および接続関係)が明ら 力になったデータ列を意味する。なお、第 2音声分析モジュール 8は、第 2国語の音 声信号を第 2中間言語に変換する点で第 1音声分析モジュール 2と異なるが、図 5に 示す構成と同様の構成によって実現できる。 FIG. 5 is a block diagram showing a detailed configuration of the first speech analysis module 2. The first speech analysis module 2 includes a word analysis unit 21, a grammar analysis unit 22, and a semantic analysis unit 23. The word analysis unit 21 recognizes a word from the voice signal. The grammar analyzer 22 The semantic analysis unit 23 analyzes the grammatical relationship between the word strings, extracts the meaning of each word, and outputs the first intermediate language corresponding to the extracted meaning. Here, “intermediate language” means a data string in which the meaning of words and syntactic attributes (ie, classification and connection of verbs, nouns, and adjectives) become clear. The second speech analysis module 8 is different from the first speech analysis module 2 in that it converts the second language speech signal into the second intermediate language, but can be realized by a configuration similar to the configuration shown in FIG.
[0028] 次に、この自動翻訳装置における動作について、図 1から図 4を用いて説明する。  Next, the operation of this automatic translation apparatus will be described with reference to FIGS. 1 to 4.
図 2は、第 1国語話者と第 2国語話者の会話において用いる自動翻訳装置の一部外 観を示す概略図である。図 2では、右力も第 1国語話者の正面図及び側面図、第 2国 語話者の側面図を示している。この自動翻訳装置は、第 1国語話者が装着しており、 図 2では、マイク 1、スピーカー 11、ディスプレイ 6及びマイク 7を示している。また、図 2では、第 1国語話者の話した「おおきに」が第 2国語「Thank youjに変換されてデ イスプレイ 6に表示される例を示している。図 3は、第 1国語話者の話した音声(日本 語「おおきに」)が、中間言語の符号に変換された後、 日本語「ありがとう」に再変換さ れて確認できると共に、第 2国語 (英語「Thank you」)に変換されてディスプレイ 6 に表示される過程を概念的に示す図である。図 4は、第 2国語話者の話した音声 (英 語「You are welcome」)が、中間言語の符号に変換された後、英語「You are welcome]に再変換されて確認できると共に、第 1国語(日本語「どういたしまして」) に変換されて音声出力される過程を概念的に示す図である。  FIG. 2 is a schematic diagram showing a partial appearance of an automatic translation apparatus used in a conversation between a first language speaker and a second language speaker. In Fig. 2, the right force also shows the front and side views of the first language speaker and the side view of the second language speaker. This automatic translation apparatus is worn by a first language speaker. FIG. 2 shows a microphone 1, a speaker 11, a display 6, and a microphone 7. Fig. 2 shows an example in which “Ookini” spoken by the first language speaker is converted to the second language “Thank youj” and displayed on the display 6. Fig. 3 shows the first language speaker. The voice spoken by the Japanese (Japanese “Ookini”) is converted to an intermediate language code, then converted back to Japanese “Thank you” and confirmed, and converted to a second language (English “Thank you”) FIG. 6 is a diagram conceptually showing a process that is displayed on the display 6. Figure 4 shows that the speech spoken by a second language speaker (English “You are welcome”) is converted to an intermediate language code and then converted back to English “You are welcome”. It is a figure which shows notionally the process in which it is converted into 1 national language (Japanese "youtashimashi") and is output by voice.
[0029] まず、第 1国語 (以下、日本語)話者の音声は、マイク 1によって電気信号に変換さ れ、第 1音声分析モジュール 2に供給される。第 1音声分析モジュール 2では、音声 信号を認識し、第 1中間言語 C1を生成する。例えば、図 2に示すように、 日本語話者 がマイク 1に向力つて「おおきに」と音声を発した場合、図 3に示すように、この電気信 号列が(日本語)辞書 20に基づ 、て音声分析モジュール 2によって分析された結果 、感謝を意味する第 1中間言語としての符号(図 3では" CF5A88Dh")が生成される  First, the voice of a first language (hereinafter, Japanese) speaker is converted into an electrical signal by the microphone 1 and supplied to the first voice analysis module 2. The first speech analysis module 2 recognizes the speech signal and generates the first intermediate language C1. For example, as shown in FIG. 2, when a Japanese speaker utters “Ookini” as he or she urges microphone 1, this electric signal string is stored in (Japanese) dictionary 20 as shown in FIG. Based on the result of the analysis by the speech analysis module 2, a code ("CF5A88Dh" in Fig. 3) is generated as the first intermediate language meaning thanks.
[0030] その後、この第 1中間言語 C1は、第 1翻訳モジュール 4によって第 2国語 (英語)に 変換されると同時に、第 1音声合成モジュール 3に供給され、再度、 日本語の音声に 変換される。 日本語話者は、図 2に示すように、この再変換された音声を耳に近接し て装着したスピーカー 11から聞くことができる。第 1音声合成モジュール 3は、感謝の 意味を持つ第 1中間言語 (符号)を、辞書 20を用いて日本語に変換する。ここで話言 葉と中間言語とは必ずしも 1 : 1に対応している必要はない。たとえば、「ありがとう」も「 おおきに」も「だんだん」も(辞書 20に登録されていれば)感謝の意を表す同一の中 間言語" CF5A88Dh"に変換されてもよい。逆に、共通の第 1中間言語 C1から日本 語に変換する場合には、どの日本語を選択するのかが問題となるが、標準語の「あり 力 Sとう」を選択すれば、少なくとも日本語話者にはその意味が通じる。 [0030] After that, the first intermediate language C1 is converted into the second language (English) by the first translation module 4, and simultaneously supplied to the first speech synthesis module 3. Converted. As shown in Fig. 2, a Japanese speaker can hear the reconverted speech from the speaker 11 attached close to the ear. The first speech synthesis module 3 converts the first intermediate language (code) having the meaning of gratitude into Japanese using the dictionary 20. Here, the spoken word and the intermediate language do not necessarily correspond to 1: 1. For example, “Thank you”, “Ookini”, and “Dandan” (if registered in dictionary 20) may be converted to the same intermediate language “CF5A88Dh” that expresses gratitude. On the other hand, when converting from the common first intermediate language C1 to Japanese, the question is which Japanese to select. The meaning is understood by the speaker.
[0031] 一方、第 1翻訳モジュール 4によって変換された第 2国語 (英語)は、文字合成モジ ユール 5によって文書ィ匕されディスプレイ 6で表示される。  On the other hand, the second national language (English) converted by the first translation module 4 is documented by the character synthesis module 5 and displayed on the display 6.
[0032] 図 6のフローチャートを用いて、第 1国語話者の話す音声を第 2国語に変換し、ディ スプレイ表示する上記の自動翻訳方法を説明する。  [0032] The automatic translation method described above, which converts the speech spoken by the first language speaker into the second language and displays it on the display, will be described using the flowchart of FIG.
(a)まず、処理を開始する(S01)。処理の開始の指示については特に図示されてい ないが、例えば、装置本体若しくはマイクロホンに設けられたボタンを押すといった方 法がある。  (a) First, processing is started (S01). The instruction to start the process is not particularly shown, but there is a method of pressing a button provided on the apparatus main body or the microphone, for example.
(b)次に、本装置を操作する操作者 (第 1国語話者)は、マイク 1に向かって第一国語 を音声入力する(S02)。  (b) Next, the operator (first language speaker) who operates the device inputs the first language into the microphone 1 (S02).
(c)次いで、第 1音声分析モジュール 2によって、入力された音声が第 1中間言語 C1 に変換される (S03)。なお、音声分析 (認識)については、特定話者認識技術を使用 することができる。例えば、上記第 1国語話者が予め本装置に対して所定の基本セン テンスを口述し、癖や独特の節回し等を本装置に覚え込ませておくことができる。音 声認識が完了すると第 1中間言語 C1が生成される。ここで、第 1中間言語 C1とは、 第 1国語についての名詞、動詞等の属性を含め、意味が確定した情報ストリームをい  (c) Next, the input speech is converted to the first intermediate language C1 by the first speech analysis module 2 (S03). For speech analysis (recognition), specific speaker recognition technology can be used. For example, the first language speaker can dictate a predetermined basic sentence to the apparatus in advance, and can learn habits and unique passages in the apparatus. When the speech recognition is completed, the first intermediate language C1 is generated. Here, the first intermediate language C1 is an information stream whose meaning is fixed including attributes such as nouns and verbs for the first language.
(d)この第 1中間言語 C1を第 1音声合成モジュール 3に供給して、第 1国語に変換し 、スピーカー 11を介して音声出力できる(S04)。ここで、スピーカー 11は、広義に、 電気信号を音声に変換するもの、例えば、ヘッドホン、イヤホンも含むものとする。 (d) The first intermediate language C1 is supplied to the first speech synthesis module 3, converted into the first language, and output through the speaker 11 (S04). Here, the speaker 11 broadly includes a device that converts an electrical signal into sound, for example, a headphone and an earphone.
[0033] (e)上記第 1国語話者は、上記スピーカー音声を聴いて、自分の話した内容が正確 に認識されたかどうか確認する(S05)。誤って認識されたと判断した場合には、第 1 国語話者は、訂正ボタン(図示せず)を押す。一方、正しく認識されたと判断した場合 には、話者は、確定ボタン(図示せず)を押す。 [0033] (e) The first language speaker listens to the speaker voice and the content of his speech is accurate. (S05). If it is determined that the recognition has been made in error, the first language speaker presses a correction button (not shown). On the other hand, if it is determined that the speaker is correctly recognized, the speaker presses a confirmation button (not shown).
このように、一旦音声認識'中間言語生成された第 1中間言語 C1を再度日本語に 変換した「ありがとう」を日本語話者が"聞ぐ'ことによって、自分が発した言葉が正しく 認識されたカゝどうかを確認することができる。  In this way, once a Japanese speaker “listens” to “Thank you” that once converted the first intermediate language C1 generated in speech recognition into Japanese again, the speech that was spoken is recognized correctly. You can check if it ’s bad.
[0034] なお、上記の訂正ボタンは、特に図示して!/ヽな 、が、例えば、ディスプレイ 6に設け られたタツチセンサ、装置筐体に設けられた押しボタン等が考えられる。また、ヘッド ホン型のスピーカーを用いる場合には、そこに加速度センサを設けておき、首を横に 振った場合、すなわち水平方向の加速度を検出した場合に訂正ボタンを押したのと 同じ処理を行うものとすることができる。一方、口述内容が正しく認識されたと判断さ れた場合に押される確定ボタンについても上記訂正ボタンと同様の態様が考えられ る。特に、上記加速度センサを用いた場合には、首を縦に振ることによって、すなわ ち垂直方向の加速度を検出した場合、確定ボタンを押したのと同様の処理を行うもの とすることができる。 [0034] Note that the correction button is not particularly shown in the figure, but for example, a touch sensor provided on the display 6, a push button provided on the apparatus housing, and the like are conceivable. In addition, when using a headphone-type speaker, an acceleration sensor is installed there, and the same processing as pressing the correction button is performed when the head is shaken sideways, that is, when horizontal acceleration is detected. Can be done. On the other hand, the confirmation button that is pressed when it is determined that the dictation content has been correctly recognized can have the same mode as the correction button. In particular, when the acceleration sensor is used, when the vertical acceleration is detected by swinging the head vertically, the same processing as when the enter button is pressed can be performed. .
[0035] (f)そこで、装置では、訂正ボタン Z確定ボタンが押されたか否力判断する(S06)。  (F) Therefore, the apparatus determines whether or not the correction button Z confirmation button has been pressed (S06).
訂正ボタンが押された場合には、口述内容が誤って認識されたと判断された場合で あるので、ステップ S02に戻る。  If the correction button is pressed, it is determined that the dictation content has been erroneously recognized, and the process returns to step S02.
(g)—方、確定ボタンが押された場合には、上記第 1中間言語 C1は、確定され、第 1 翻訳モジュール 4によって第 2国語に翻訳される (S07)。  (g) —If the confirm button is pressed, the first intermediate language C1 is confirmed and translated into the second language by the first translation module 4 (S07).
(h)さらに、翻訳が完了すれば、翻訳結果の第 2国語が文字合成モジュール 5によつ てディスプレイ 6に表示される(S08)。  (h) Further, when the translation is completed, the second national language of the translation result is displayed on the display 6 by the character synthesis module 5 (S08).
以上によって、第 1国語話者の話した音声は、第 2国語に変換され、ディスプレイ 6 に表示されるので、第二国語話者はビジュアルで認識できるので、周囲の雑音に影 響されることなく翻訳結果を理解できる。  As described above, the voice spoken by the first language speaker is converted into the second language and displayed on the display 6, so that the second language speaker can be visually recognized and is therefore affected by ambient noise. Without understanding the translation results.
[0036] ここで、第 1中間言語 C1は、第 1翻訳モジュール 4によって第二国語 (英語)の第 2 中間言語 C2に翻訳される。なお、日本語と英語とでは文法体系が異なるので、中間 言語の構造体 (修飾、接続関係など)を所定の規則に基づいて変換する必要がある 。翻訳された第 2中間言語 C2は、辞書 80に基づいて文字合成モジュール 5によって 文書化され、ディスプレイ 6に表示される。本実施の形態の場合、感謝の意を表す「T hank you」に翻訳される。ディスプレイ 6は図 2に示すように、第二国語 (英語)話者 が自然な目線で判読できるように、第一国語(日本語)話者の胸力 方のあたりに装 着されるようにするのがよい。図 2の配置では、日本語話者は、ディスプレイ 6を(自然 な格好で)見ることはできないが、既に、自分の話言葉が正確に意味認識されたこと をスピーカー 11からの音声で把握できているので、後は、その中間言語が機械的に 英語に翻訳されるものと思っていればよい。なお、仮にディスプレイ 6の表示が読めた ところで、第二国語が自分が全く理解できな!/、言語であった場合は全く用をなさな!ヽ 。なお、ディスプレイ 6の具体的構成として低電力タイプの反射型液晶素子、又は電 気泳動素子などを用いてもよい。また、携帯電話や PDA等のモパイル情報端末や、 ビデオムービーカメラ又はデジタルスチルカメラといったモパイル AV機器の液晶表 示部分を利用するものであってもよい。 [0036] Here, the first intermediate language C1 is translated by the first translation module 4 into the second intermediate language C2 of the second language (English). Since Japanese and English have different grammatical systems, intermediate language structures (modifications, connection relationships, etc.) must be converted based on the prescribed rules. . The translated second intermediate language C2 is documented by the character synthesis module 5 based on the dictionary 80 and displayed on the display 6. In the case of this embodiment, it is translated into “T hank you” representing gratitude. As shown in Fig. 2, display 6 should be worn around the chest of the first language (Japanese) speaker so that the second language (English) speaker can read with a natural perspective. It is good to do. In the arrangement shown in Fig. 2, the Japanese speaker cannot see the display 6 (natural appearance), but can already grasp the meaning of the spoken word accurately from the speaker 11. After that, you can think that the intermediate language is mechanically translated into English. If the display 6 can be read, the second language cannot be understood at all! / If the language is a language, do not use it! As a specific configuration of the display 6, a low power type reflective liquid crystal element, an electrophoretic element, or the like may be used. Also, it may use a liquid crystal display part of a mopile AV device such as a mobile phone or PDA, a video movie camera or a digital still camera.
以上により、上記一連の動作は完了する。  Thus, the series of operations is completed.
次に、第 2国語話者が応答する場合にっ 、て、図 7に示すフローチャートを用いて 説明する。  Next, the case where a second language speaker responds will be described with reference to the flowchart shown in FIG.
(a)まず、処理を開始する(Sl l)。処理の開始の指示については特に図示されてい ないが、例えば、装置本体若しくはマイクロホンに設けられたボタンを押すといった方 法がある。また、上述のような加速度センサを用いた場合も考えられる。つまり、首を 縦に振れば、すなわち垂直方向の加速度を検出すれば処理を開始するようにしても よい。なお、この開始処理は、図 6のステップ 8と同様のものである力 処理を先に進 めるといった意味では、共通の目的を持っており、第 1国語話者が混乱することはな いと考えられる。  (a) First, processing is started (Sl l). The instruction to start the process is not particularly shown, but there is a method of pressing a button provided on the apparatus main body or the microphone, for example. Further, the case where the acceleration sensor as described above is used is also conceivable. That is, the processing may be started when the head is swung vertically, that is, when acceleration in the vertical direction is detected. Note that this start process has a common purpose in the sense that the power process, which is the same as step 8 in FIG. 6, is advanced, and the first language speaker should not be confused. Conceivable.
(b)次に、第 2国語話者 (英語話者)は、マイク 7に向カゝつて第 2国語で音声を発し、マ イク 7で音声信号に変換される(S12)。図 4の例では、「You are welcome」である  (b) Next, the second language speaker (English speaker) speaks to the microphone 7 in the second language, and is converted into a voice signal by the microphone 7 (S12). In the example in Figure 4, it is “You are welcome”
(c)上記第 2国語の音声信号は、図 4に示すように、第 2音声分析モジュール 8によつ て辞書 80に基づいて音声分析 (8)され、第 2中間言語 C2 ("FAC653BOh")が生 成される(S13)。なお、音声分析 (音声認識)については、第 1国語の場合と異なり、 第 2国語話者は、随時変わる場合が多いため認識効率を向上させることができる特 定話者認識技術を利用することができない。しかし、第 2国語話者が話す内容は、第 1国語話者に対する「返事」であることが多ぐ認識すべきボキャブラリーが限定される ので、不特定話者認識でも十分に足りると考えられる。ここで、上述の第 1中間言語 C 1と同様に、第 2中間言語 C2とは、第 2国語についての名詞、動詞等の属性を含め、 意味が確定した情報ストリームを 、う。 (c) The second language speech signal is subjected to speech analysis (8) based on the dictionary 80 by the second speech analysis module 8, as shown in FIG. 4, and the second intermediate language C2 ("FAC653BOh" ) Is raw (S13). For speech analysis (speech recognition), unlike in the case of the first language, speakers of the second language often change from time to time, so use specific speaker recognition technology that can improve recognition efficiency. I can't. However, since the vocabulary that the second language speaker speaks is often “reply” to the first language speaker, the recognition of unspecified speakers is considered to be sufficient. Here, like the first intermediate language C 1 described above, the second intermediate language C 2 is an information stream whose meaning is fixed including attributes such as nouns and verbs for the second language.
(d)第 2中間言語 C2は、第 2国語に変換され、文字合成モジュール 5によって文書ィ匕 されて、ディスプレイ 6に表示される(S 14)。  (d) The second intermediate language C2 is converted into a second national language, documented by the character synthesis module 5, and displayed on the display 6 (S14).
[0038] (e)第 2国語話者は、上記ディスプレイ表示によって自分の話した言葉が正しく認識 された力どうかを確認することができる(S 15)。ここで、誤って認識されたと判断した 場合には、第 2国語話者は、訂正ボタン(図示せず)を押す。または、第 2国語話者は 、ジェスチャー等によって訂正ボタンを押すように第 1国語話者に指示する。一方、正 しく認識されたと判断した場合には、話者は確定ボタン (図示せず)を押す。  [0038] (e) The second language speaker can confirm whether or not the word spoken by the second language speaker is correctly recognized (S15). Here, if it is determined that the recognition has been mistaken, the second language speaker presses a correction button (not shown). Alternatively, the second language speaker instructs the first language speaker to press the correction button by a gesture or the like. On the other hand, if it is determined that the speaker has been correctly recognized, the speaker presses a confirmation button (not shown).
このように、一旦音声認識'中間言語生成された第 2中間言語 C2を再度英語に変 換した「You are welcome」のディスプレイ表示を英語話者が"見る"ことによって、 自分が発した言葉が正しく認識された力どうかを確認することができる。  In this way, when the English speaker "sees" the display of "You are welcome", which is the speech recognition 'second intermediate language C2 generated once again and converted into English again, the words spoken by me It is possible to check whether the force is recognized correctly.
[0039] なお、この訂正ボタンは、特に図示していないが、ディスプレイ 6に設けられたタツチ センサ、装置筐体に設けられた押しボタン等によって実現できる。また、ヘッドホン型 のスピーカーを用いる場合には、そこに加速度センサを設けておき、第 1国語話者が 首を横に振った場合に、すなわち水平方向の加速度が検出された場合に、上記訂 正ボタンが押されたと同様の処理を行うようにしてもよい。また、否定を表す特定語( 英語 。!」、ドイツ語「Nein!」、中国語「Bu!」、等)を認識すれば、訂正スィッチ を押したのと同様の処理を行うようにしてもよい。なお、上記の確定ボタンについても 上記訂正ボタンと同様の形態で実現できる。なお、上記の加速度センサを使用する 場合には、首を縦に振った場合、すなわち垂直方向の加速度が検出された場合に、 確認ボタンが押されたのと同様の処理を行うようにしてもよい。  Note that this correction button is not particularly shown, but can be realized by a touch sensor provided on the display 6, a push button provided on the apparatus housing, or the like. In addition, when using a headphone-type speaker, an acceleration sensor is provided there, and the above correction is made when a first language speaker shakes his / her head, that is, when a horizontal acceleration is detected. The same processing as when the normal button is pressed may be performed. Also, if it recognizes a specific word that represents negation (English.!), German “Nein!”, Chinese “Bu!”, Etc., the same processing as pressing the correction switch may be performed. Good. The above confirmation button can be realized in the same form as the above correction button. When the above acceleration sensor is used, the same processing as when the confirmation button is pressed may be performed when the head is swung vertically, that is, when acceleration in the vertical direction is detected. Good.
[0040] (f)そこで、装置では、訂正ボタン Z確定ボタンが押されたか否力判断する(S 16)。 上述のように、第 2国語話者の口述内容が間違って認識されたと判断された場合に は訂正ボタンが押され、又はそれに相当する手段が行われ、上記ステップ S12に戻 つて再度上記処理が実行される。 (F) Therefore, the apparatus determines whether or not the correction button Z confirmation button has been pressed (S 16). As described above, when it is determined that the dictation content of the second language speaker has been mistakenly recognized, the correction button is pressed or a corresponding means is taken, and the above processing is performed again after returning to step S12. Executed.
(g)—方、確定ボタンが押された場合には、第 2中間言語 C2は、確定され、第 2翻訳 モジュール 10によって第 1中間言語 C1に翻訳される(S17)。  (g) —On the other hand, when the confirm button is pressed, the second intermediate language C2 is confirmed and translated into the first intermediate language C1 by the second translation module 10 (S17).
(h)さらに、翻訳が完了すれば、翻訳結果の第 1国語は、第 1音声合成モジュール 3 によって第 1国語で音声変換され、スピーカー 11力も第 1国語で「どういたしまして」と いう音声が発せられる(S18)。  (h) Furthermore, if the translation is completed, the first national language of the translation result will be converted into the first national language by the first speech synthesis module 3, and the speaker 11 will also be able to produce a voice saying “you are welcome” in the first national language. (S18).
以上によって、第 2国語話者の話した音声は、第 1国語に変換され、第 1国語話者 の耳の近傍に設けたスピーカー 11から音声出力されるので、第 1国語話者は、周囲 の雑音に影響されることなく翻訳結果をオーラルに認識できる。  As described above, the speech spoken by the second language speaker is converted into the first language and output from the speaker 11 provided in the vicinity of the ear of the first language speaker. Translation results can be recognized orally without being affected by noise.
[0041] 以上、本実施の形態によれば、モパイルに適した構成で、より自然な形で二国語間 の自動翻訳を実行することができる。つまり、 日本語話者は認識された自己の言葉と 、 日本語に翻訳された英語話者の言葉を、スピーカー 11からリアルタイムで聞くこと ができる。ここで「スピーカー」とあるのは、いわゆるイヤホンのようなものであってもよ い。また、英語話者は、認識された自己の言葉と、英語に翻訳された日本語話者の 言葉をディスプレイ 6の表示として見ることができる。これは野外の騒音下にお ヽても 可能である。 [0041] As described above, according to the present embodiment, automatic translation between two languages can be executed in a more natural manner with a configuration suitable for mopile. In other words, a Japanese speaker can hear the recognized self-language and the English speaker's language translated into Japanese from the speaker 11 in real time. Here, the “speaker” may be a so-called earphone. In addition, the English speaker can see the recognized self-word and the Japanese speaker's word translated into English on the display 6. This is possible even in the open air.
[0042] なお、図 1において、セレクタ 9、 12は、自己の音声情報を通すか、相手話者の音 声情報を通すかを切り替えるものである。具体的には、(図示していないが)マイクロ プロセッサーによって制御され、例えば、マイク 1からの音声入力と、マイク 2からの音 声入力とをそれぞれの音声信号における振幅の大小に応じて、入力元を切り替える ものであってもよい。  In FIG. 1, selectors 9 and 12 are used to switch between passing the own voice information and passing the other party's voice information. Specifically, it is controlled by a microprocessor (not shown), and for example, the audio input from microphone 1 and the audio input from microphone 2 are input according to the amplitude of each audio signal. The source may be switched.
[0043] (実施の形態 2) [Embodiment 2]
図 8は、本発明の実施の形態 2に係る自動翻訳装置の構成を示すブロック図である 。図 9は、その一部外観図である。この自動翻訳装置は、実施の形態 1に係る自動翻 訳装置と比較すると、第 2音声合成モジュール 51及び第 2スピーカー 61を有する点 で相違する。図 8において、図 1と同じ構成要素については同じ符号を用い、説明を 省略する。図 8において、第 2音声合成モジュール 51は、第 1翻訳モジュール 4によ つて変換された英語のみを音声信号に変換するものである。この音声信号は、図 9に 示されるように、ディスプレイ 6の近くに設けられた第 2スピーカー 61によって音声化 される。以上の動作により、英語話者は、日本語話者の話し言葉の翻訳をディスプレ ィ 6を介して目視によって確認することもでき、さらに、比較的静かな環境下であれば 、第 2スピーカー 61からの音声として聞くこともでき、合わせて、より確度の高い会話 を実現することが可能となる。 FIG. 8 is a block diagram showing the configuration of the automatic translation apparatus according to Embodiment 2 of the present invention. Fig. 9 is a partial external view. This automatic translation apparatus is different from the automatic translation apparatus according to Embodiment 1 in that it includes a second speech synthesis module 51 and a second speaker 61. In FIG. 8, the same components as those in FIG. Omitted. In FIG. 8, the second speech synthesis module 51 converts only English converted by the first translation module 4 into a speech signal. As shown in FIG. 9, this audio signal is converted into audio by a second speaker 61 provided near the display 6. With the above operation, the English speaker can also visually confirm the translation of the spoken language of the Japanese speaker through the display 6, and in a relatively quiet environment, the second speaker 61 It is also possible to listen to the voice as a voice, and at the same time, it is possible to realize a conversation with higher accuracy.
[0044] ここで、本実施の形態では第 2スピーカー 61は、図 9に示すように、ディスプレイ 6の 近傍に別途設ける構成にしているが、これらの場合に限られず、例えば、透明圧電ス ピーカーをディスプレイ 6の表示面に貼り付けたものであってもよい。  Here, in the present embodiment, the second speaker 61 is separately provided in the vicinity of the display 6 as shown in FIG. 9. However, the present invention is not limited to these, and for example, a transparent piezoelectric speaker is used. May be pasted on the display surface of the display 6.
[0045] 上述の通り、本発明は好ましい実施形態により詳細に説明されている力 本発明は 、これらに限定されるものではなぐ特許請求の範囲に記載された本発明の技術的範 囲内において多くの好ましい変形例及び修正例が可能であることは当業者にとって 自明なことであろう。  [0045] As described above, the present invention is described in detail with reference to the preferred embodiments. The present invention is not limited to these, but is within the technical scope of the present invention described in the claims. It will be apparent to those skilled in the art that preferred variations and modifications can be made.
産業上の利用可能性  Industrial applicability
[0046] 本発明に係る自動翻訳装置は、日本語話者に対してはオーラルに、対する英語話 者に対してはビジュアルに会話情報を伝達する機能を有し、モパイル用電子通訳機 等として、特に海外旅行中の使用する際に有用である。また国際会議等の用途にも 応用できる。 [0046] The automatic translation apparatus according to the present invention has a function of transmitting conversation information to an oral speaker for a Japanese speaker and visually to an English speaker, and as an electronic interpreter for mopile, etc. Especially useful when used during overseas travel. It can also be used for international conferences.

Claims

請求の範囲 The scope of the claims
[I] 第 1国語で話された音声を分析し、第 1中間言語を生成する第 1音声分析モジユー ルと、  [I] a first speech analysis module that analyzes speech spoken in a first language and generates a first intermediate language;
前記第 1中間言語を第 2国語に変換する第 1翻訳モジュールと、  A first translation module for converting the first intermediate language into a second language;
変換された前記第 2国語をディスプレイに表示する文字合成モジュールと を備えたことを特徴とする自動翻訳装置。  An automatic translation apparatus, comprising: a character synthesis module that displays the converted second language on a display.
[2] 前記第 1中間言語を第 1国語に再変換し、再変換された前記第 1国語の音声信号 を生成する第 1音声合成モジュールをさらに備えることを特徴とする請求項 1に記載 の自動翻訳装置。 [2] The first speech synthesis module according to claim 1, further comprising a first speech synthesis module that reconverts the first intermediate language into a first national language and generates the reconverted speech signal of the first national language. Automatic translation device.
[3] 変換された前記第 2国語の音声信号を生成する第 2音声合成モジュールをさらに 備えることを特徴とする請求項 1又は 2に記載の自動翻訳装置。  [3] The automatic translation device according to [1] or [2], further comprising a second speech synthesis module that generates the converted speech signal of the second language.
[4] 前記第 1中間言語を確定する確定ボタンをさらに備えることを特徴とする請求項 1か ら 3の 、ずれか一項に記載の自動翻訳装置。 [4] The automatic translation device according to any one of claims 1 to 3, further comprising a confirm button for confirming the first intermediate language.
[5] 前記第 1中間言語を不確定として、第 1国語の音声の再入力を促す訂正ボタンをさ らに備えることを特徴とする請求項 1から 3のいずれか一項に記載の自動翻訳装置。 [5] The automatic translation according to any one of claims 1 to 3, further comprising a correction button that makes the first intermediate language uncertain and prompts re-input of the speech of the first language. apparatus.
[6] 第 2国語で話された音声を分析し、第 2中間言語を生成する第 2音声分析モジユー ルをさらに備えることを特徴とする請求項 1から 5のいずれか一項に記載の自動翻訳 装置。 [6] The automatic speech processing device according to any one of claims 1 to 5, further comprising a second speech analysis module that analyzes speech spoken in a second language and generates a second intermediate language. Translation device.
[7] 前記第 2中間言語を第 1国語に変換する第 2翻訳モジュールをさらに備えることを 特徴とする請求項 6に記載の自動翻訳装置。  7. The automatic translation apparatus according to claim 6, further comprising a second translation module that converts the second intermediate language into a first national language.
[8] 前記文字合成モジュールは、前記第 2中間言語を第 2国語に再変換し、ディスプレ ィに表示することを特徴とする請求項 6又は 7に記載の自動翻訳装置。 8. The automatic translation device according to claim 6 or 7, wherein the character synthesis module reconverts the second intermediate language into a second national language and displays it on a display.
[9] 前記第 2中間言語を確定する確定ボタンをさらに備えることを特徴とする請求項 8に 記載の自動翻訳装置。 9. The automatic translation apparatus according to claim 8, further comprising a confirm button for confirming the second intermediate language.
[10] 前記第 2中間言語を不確定として、第 2国語の音声の再入力を促す訂正ボタンをさ らに備えることを特徴とする請求項 8に記載の自動翻訳装置。  10. The automatic translation apparatus according to claim 8, further comprising a correction button that makes the second intermediate language uncertain and prompts the user to re-input the second language speech.
[II] 前記文字合成モジュールの入力元を第 1翻訳モジュールと前記第 2音声分析モジ ユールとの間で切り替える第 1選択部をさらに備えることを特徴とする請求項 6に記載 の自動翻訳装置。 7. The apparatus of claim 6, further comprising a first selection unit that switches an input source of the character synthesis module between a first translation module and the second speech analysis module. Automatic translation device.
[12] 前記第 1音声合成モジュールの入力元を第 1音声分析モジュールと前記第 2翻訳 モジュールとの間で切り替える第 2選択部をさらに備えることを特徴とする請求項 7に 記載の自動翻訳装置。  12. The automatic translation device according to claim 7, further comprising a second selection unit that switches an input source of the first speech synthesis module between the first speech analysis module and the second translation module. .
[13] 第 1国語で話された音声を分析し、第 1中間言語を生成するステップと、 [13] analyzing speech spoken in a first language and generating a first intermediate language;
前記第 1中間言語を第 2国語に変換するステップと、  Converting the first intermediate language into a second language;
変換された前記第 2国語をディスプレイに表示するステップと  Displaying the converted second language on a display;
を含むことを特徴とする自動翻訳方法。  The automatic translation method characterized by including.
[14] 前記第 1中間言語を第 1国語に再変換し、再変換された前記第 1国語の音声信号 を生成するステップをさらに含むことを特徴とする請求項 13に記載の自動翻訳方法 14. The automatic translation method according to claim 13, further comprising the step of reconverting the first intermediate language into a first national language and generating the reconverted speech signal of the first national language.
[15] 変換された前記第 2国語の音声信号を生成するステップをさらに含むことを特徴と する請求項 13又は 14に記載の自動翻訳方法。 15. The automatic translation method according to claim 13, further comprising a step of generating a converted speech signal in the second language.
[16] 前記第 1国語の音声信号を聞いて前記第 1中間言語を確定するステップをさらに含 むことを特徴とする請求項 14に記載の自動翻訳方法。 16. The automatic translation method according to claim 14, further comprising the step of confirming the first intermediate language by listening to the audio signal of the first language.
[17] 前記第 1国語の音声信号を聞いて前記第 1中間言語を不確定として、第 1国語の 音声の再入力を促すステップをさらに含むことを特徴とする請求項 14に記載の自動 翻訳方法。 17. The automatic translation according to claim 14, further comprising the step of listening to the first national language speech signal and making the first intermediate language uncertain and prompting re-input of the first national language speech. Method.
[18] 第 2国語で話された音声を分析し、第 2中間言語を生成するステップをさらに含むこ とを特徴とする請求項 13から 17のいずれか一項に記載の自動翻訳方法。  [18] The automatic translation method according to any one of [13] to [17], further comprising the step of analyzing the speech spoken in the second language and generating a second intermediate language.
[19] 前記第 2中間言語を第 1国語に変換するステップをさらに含むことを特徴とする請 求項 18に記載の自動翻訳方法。 [19] The automatic translation method according to claim 18, further comprising a step of converting the second intermediate language into a first national language.
[20] 前記第 2中間言語を第 2国語に再変換し、ディスプレイに表示するステップをさらに 含むことを特徴とする請求項 18に記載の自動翻訳方法。 20. The automatic translation method according to claim 18, further comprising the step of reconverting the second intermediate language into a second national language and displaying the second intermediate language on a display.
[21] 前記ディスプレイに表示された前記第 2国語を見て、前記第 2中間言語を確定させ るステップをさらに含むことを特徴とする請求項 20に記載の自動翻訳方法。 21. The automatic translation method according to claim 20, further comprising the step of confirming the second intermediate language by looking at the second national language displayed on the display.
[22] 前記ディスプレイに表示された前記第 2国語を見て、前記第 2中間言語を不確定と して、第 2国語の音声の再入力を促すステップをさらに含むことを特徴とする請求項 2 に記載の自動翻訳方法。 [22] The method may further include a step of urging the re-input of the second language speech by observing the second language displayed on the display and making the second intermediate language uncertain. 2 Automatic translation method described in 1.
PCT/JP2005/010946 2004-06-23 2005-06-15 Automatic translation device and automatic translation method WO2006001204A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2004-184713 2004-06-23
JP2004184713A JP2007272260A (en) 2004-06-23 2004-06-23 Automatic translation device

Publications (1)

Publication Number Publication Date
WO2006001204A1 true WO2006001204A1 (en) 2006-01-05

Family

ID=35781706

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2005/010946 WO2006001204A1 (en) 2004-06-23 2005-06-15 Automatic translation device and automatic translation method

Country Status (2)

Country Link
JP (1) JP2007272260A (en)
WO (1) WO2006001204A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109344411A (en) * 2018-09-19 2019-02-15 深圳市合言信息科技有限公司 A kind of interpretation method for listening to formula simultaneous interpretation automatically
WO2022214991A1 (en) * 2020-04-08 2022-10-13 Rajiv Trehan Multilingual concierge systems and method thereof

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10152476B2 (en) 2015-03-19 2018-12-11 Panasonic Intellectual Property Management Co., Ltd. Wearable device and translation system
WO2018008227A1 (en) 2016-07-08 2018-01-11 パナソニックIpマネジメント株式会社 Translation device and translation method

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04319769A (en) * 1991-04-18 1992-11-10 Toshiba Corp Interpretation system
JPH06348751A (en) * 1992-11-09 1994-12-22 Ricoh Co Ltd Language translating device
JPH1021241A (en) * 1996-07-03 1998-01-23 Sony Corp Method and system for intention communication
JP2003288339A (en) * 2001-01-24 2003-10-10 Matsushita Electric Ind Co Ltd Device and method for voice conversion, program, and medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04319769A (en) * 1991-04-18 1992-11-10 Toshiba Corp Interpretation system
JPH06348751A (en) * 1992-11-09 1994-12-22 Ricoh Co Ltd Language translating device
JPH1021241A (en) * 1996-07-03 1998-01-23 Sony Corp Method and system for intention communication
JP2003288339A (en) * 2001-01-24 2003-10-10 Matsushita Electric Ind Co Ltd Device and method for voice conversion, program, and medium

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109344411A (en) * 2018-09-19 2019-02-15 深圳市合言信息科技有限公司 A kind of interpretation method for listening to formula simultaneous interpretation automatically
WO2022214991A1 (en) * 2020-04-08 2022-10-13 Rajiv Trehan Multilingual concierge systems and method thereof

Also Published As

Publication number Publication date
JP2007272260A (en) 2007-10-18

Similar Documents

Publication Publication Date Title
US9111545B2 (en) Hand-held communication aid for individuals with auditory, speech and visual impairments
KR102108500B1 (en) Supporting Method And System For communication Service, and Electronic Device supporting the same
US20140171036A1 (en) Method of communication
US8082152B2 (en) Device for communication for persons with speech and/or hearing handicap
US20050192811A1 (en) Portable translation device
US20200012724A1 (en) Bidirectional speech translation system, bidirectional speech translation method and program
JP2005513619A (en) Real-time translator and method for real-time translation of multiple spoken languages
WO2003079328A1 (en) Audio video conversion apparatus and method, and audio video conversion program
JP2019175426A (en) Translation system, translation method, translation device, and voice input/output device
JP2010026220A (en) Voice translation device and voice translation method
WO2006001204A1 (en) Automatic translation device and automatic translation method
JP6457706B1 (en) Translation system, translation method, and translation apparatus
JP2015041101A (en) Foreign language learning system using smart spectacles and its method
JP6832503B2 (en) Information presentation method, information presentation program and information presentation system
US20110116608A1 (en) Method of providing two-way communication between a deaf person and a hearing person
CN105913841A (en) Voice recognition method, voice recognition device and terminal
KR20210158369A (en) Voice recognition device
JP2004129174A (en) Information communication instrument, information communication program, and recording medium
JP2011150657A (en) Translation voice reproduction apparatus and reproduction method thereof
JPH10224520A (en) Multi-media public telephone system
US20030171912A1 (en) Translation apparatus
JP2020119043A (en) Voice translation system and voice translation method
KR20200003529A (en) Digital device for recognizing voice and method for controlling the same
JPH0863185A (en) Speech recognition device
KR102496398B1 (en) A voice-to-text conversion device paired with a user device and method therefor

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DPEN Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed from 20040101)
NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP