WO2024042963A1 - Error correcting translation device, error correcting translation method, program and storage medium for same - Google Patents

Error correcting translation device, error correcting translation method, program and storage medium for same Download PDF

Info

Publication number
WO2024042963A1
WO2024042963A1 PCT/JP2023/027192 JP2023027192W WO2024042963A1 WO 2024042963 A1 WO2024042963 A1 WO 2024042963A1 JP 2023027192 W JP2023027192 W JP 2023027192W WO 2024042963 A1 WO2024042963 A1 WO 2024042963A1
Authority
WO
WIPO (PCT)
Prior art keywords
text
terminal
language
error correction
chunk
Prior art date
Application number
PCT/JP2023/027192
Other languages
French (fr)
Japanese (ja)
Inventor
史昭 菅谷
将夫 内山
英一郎 隅田
照二 小林
Original Assignee
マインドワード株式会社
国立研究開発法人情報通信研究機構
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by マインドワード株式会社, 国立研究開発法人情報通信研究機構 filed Critical マインドワード株式会社
Publication of WO2024042963A1 publication Critical patent/WO2024042963A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/40Processing or translation of natural language
    • G06F40/58Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation

Definitions

  • the present invention relates to an error correction translation device, an error correction translation method, a program, and a storage medium thereof.
  • Patent Document 1 discloses a data segmentation device that can perform sequence data segmentation processing (for example, text segmentation processing) in real time.
  • sequence data segmentation processing for example, text segmentation processing
  • the present invention has been made in view of the above points, and is an error correction method that can reliably feed back errors in simultaneous interpretation to the source language side, makes error correction efficient, and improves the quality of conversation.
  • the purpose of the present invention is to provide a translation device, an error-correcting translation method, a program, and a storage medium thereof.
  • An error correction translation device includes: a first terminal into which text in a first language is input; a text divider that divides the first language text into chunks to generate divided text; a translator that individually translates each chunk included in the divided text to generate translated chunks in a second language; a second terminal; and an error correction control unit that causes each of the translation chunks from the translator to be individually displayed on the second terminal.
  • the second terminal receives an input of an identifier specifying a mistranslated chunk among the translated chunks
  • the error correction control unit causes the mistranslation chunk to be displayed on the first terminal in a manner different from other translation chunks based on the identifier input to the second terminal, and the error correction control unit causes the mistranslation chunk to be displayed in a manner different from other translation chunks.
  • the first terminal is configured to accept input of paraphrased text of the text in the first language.
  • An error correction translation method includes: inputting text in a first language into a first terminal; a text dividing step of dividing the first language text into chunks to generate divided text; a translation step of individually translating each chunk included in the divided text to generate translated chunks in a second language; an error correction control step of individually displaying each of the translation chunks from the translator on the second terminal; receiving, at the second terminal, an input of an identifier specifying a mistranslated chunk among the translated chunks; a mistranslation chunk displaying step of displaying the mistranslation chunk on the first terminal in a manner different from other translated chunks based on the identifier input to the second terminal; after the mistranslation chunk displaying step, receiving input of a paraphrase text of the text in the first language to the first terminal; has.
  • An error correction translation program includes: computer to translate text data, a first terminal means into which text in a first language is input; text dividing means for dividing the first language text into chunks to generate divided text; Translation means for individually translating each chunk included in the divided text to generate translated chunks in a second language; second terminal means; functioning as error correction control means for individually displaying each of the translation chunks from the translation means on the second terminal means;
  • the second terminal means receives an input of an identifier specifying a mistranslated chunk among the translated chunks,
  • the error correction control means causes the first terminal means to display the mistranslated chunk in a manner different from other translation chunks based on the identifier input to the second terminal means,
  • the method is characterized in that, in response to the display, the first terminal means accepts input of paraphrased text of the text in the first language.
  • FIG. 1 is a block diagram showing the configuration of an error correction translation device 10 according to a first embodiment of the present invention.
  • FIG. 3 is a flowchart showing a flow of error correction control executed by an error correction control unit 25.
  • FIG. FIG. 3 is a diagram schematically showing a display screen of the first terminal 11 on which spoken text is displayed.
  • FIG. 3 is a diagram schematically showing a display screen of the second terminal 21 on which translation chunks are separated and displayed in individual frames.
  • FIG. 6 is a diagram schematically showing the display screen of the second terminal 21 when the designation of mistranslated chunks is received.
  • FIG. 3 is a diagram schematically showing a display screen of the first terminal 11 on which mistranslated chunks are highlighted.
  • FIG. 7 is a diagram schematically showing another example of the display screen of the first terminal 11 on which mistranslated chunks are highlighted. It is a figure which shows typically the display screen of the 2nd terminal 21 on which the paraphrase text was translated and displayed.
  • 3D is a diagram schematically showing a modification example of the display of the first terminal 11 shown in FIG. 3D.
  • FIG. 2 is a block diagram showing an example of the configuration of a text divider 13.
  • FIG. It is a figure showing the result of morphological analysis of Japanese text using MeCab. It is a figure which shows the morphological analysis result of the English text using NLTK.
  • 3 is a flowchart showing a chunk boundary determination processing procedure executed by the threshold value determination manager 34.
  • FIG. FIG. 2 is a block diagram showing the configuration of an error correction translation device 30 according to a second embodiment of the present invention.
  • FIG. 3 is a block diagram showing the configuration of an error correction translation device 50 according to a third embodiment of the present invention.
  • FIG. 1 is a block diagram showing the configuration of an error correction translation device 10 according to a first embodiment of the present invention.
  • the error correction translation device 10 is configured as a translation device that translates a conversation of a Japanese (first language) speaker to an English (second language) speaker.
  • two-way simultaneous translation of Japanese (first language) speaker's conversation to English (second language) speaker's and English speaker's conversation to Japanese speaker's It may be configured as a translation device for an interpreter.
  • the first language (original language) and the second language (translation language) are not limited to Japanese and English, and can be applied to other languages as well.
  • the error correction translation device 10 includes a first terminal (terminal #1 in the figure) 11 which is a tablet terminal (input/output device), a text divider 14, and a Japanese-English/English-Japanese translator (hereinafter simply referred to as a translator). ) 15, a second terminal (terminal #2 in the figure) 21 which is a tablet terminal (input/output device), and an error correction control section 25.
  • the error correction control unit 25 also includes a first terminal controller 12 , a divided text correction manager (hereinafter simply referred to as a correction manager) 13 , and a second terminal controller 22 .
  • the first terminal 11 is equipped with a screen keyboard (KB) 11A.
  • Kana characters input from the screen keyboard (hereinafter simply referred to as the keyboard) 11A by the Japanese speaker 1 are converted into Japanese text data (sequence data) by a processor (not shown) in the first terminal 11. is output from the first terminal 11 as .
  • kana-kanji conversion which is a known technique, may be used for the kana characters input from the keyboard 11A, if necessary.
  • the first terminal 11 also has a display that receives and displays data from the first terminal controller 12. Furthermore, the display has a function as a touch panel that receives operations from the Japanese speaker 1.
  • the second terminal 21 is equipped with a keyboard (KB) 21A. Characters input from the keyboard 21A by the English speaker 2 are output from the second terminal 21 as English text data (sequence data) by a processor (not shown) in the second terminal 21.
  • the second terminal 21 also has a display that receives and displays data from the second terminal controller 22. Furthermore, the display has a function as a touch panel that receives operations from the English speaker 2.
  • FIG. 2 is a flowchart showing the flow of the error correction translation method executed by the error correction control unit 25. Further, FIGS. 3A to 3G are diagrams schematically showing display screens of the first terminal 11 and the second terminal 21.
  • the Japanese speaker 1 inputs the text "I'm going to Tokyo tomorrow" (hereinafter also referred to as uttered text) using the keyboard (KB) 11A.
  • the first terminal 11 sends the input spoken text to the first terminal controller 12 .
  • the first terminal controller 12 generates the following data signal DS1 from the input text received from the first terminal 11 (step S1), and sends it to the correction manager 13.
  • the first terminal controller 12 also sends the data signal DS1 to the first terminal 11, and as shown in FIG. 3A, the first terminal 11 receives the spoken text data [ 'I'm going to Tokyo tomorrow'] is displayed in the format of the text widget 11W (step S2).
  • JSON JavaScript Object Notation
  • Data signal DS1 ⁇ 'id':10, 'jap': ['I'm going to Tokyo tomorrow'] ⁇
  • 'id' is the ID of the utterance
  • the correction manager 13 of the error correction control unit 25 sends the data signal DS1 to the text divider 14 (step S2).
  • the text divider 14 divides the spoken text data included in the data signal DS1 into chunks (partial sentences). The operation of the text divider 13 will be described in detail below with reference to FIG.
  • FIG. 4 is a block diagram showing an example of the configuration of the text divider 13.
  • a text ST (indicated as ⁇ ST> in the figure) input to the input end 13I of the text divider 13 is divided into words by the word divider 31.
  • the word divider 31 performs, for example, morphological analysis to divide the text ST into words w 1 to w m .
  • FIG. 5A shows the results of morphological analysis of Japanese text using MeCab.
  • m 5
  • the word sequences divided by the word divider 31 are 'Asu', 'Tokyo', 'ni', 'Go', and 'Masu'. .
  • FIG. 5B shows the results of morphological analysis of an English text using NLTK.
  • threshold values ⁇ i are stored in advance in the threshold value register 35 in correspondence with each of the registers r 1 to r m of the word register 32.
  • the initial value i is set to 1 (step S21), and subsequent steps S22 to S25 are executed.
  • step S22 the comparator 36 determines whether P(wr i )> ⁇ i. If it is determined that P(wr i )> ⁇ i, the end of the word w i is determined to be the division position, that is, the boundary (chunk boundary) where the spoken text is divided into chunks (step S25). . If it is determined that P(wr i )> ⁇ i is not satisfied, the process advances to the next step S23.
  • step S23 it is determined whether the current word is the last word of the text (i ⁇ m). If it is determined that the current word is the last word of the text, the process proceeds to step S23, where the end of the word is determined to be the division position.
  • step S23 If it is determined in step S23 that the current word is not the last word of the text, the process proceeds to step S24, where i is incremented by 1, and the above steps are repeated for the next word.
  • the divided text output device 37 adds a division mark (separator) to the series of words w 1 to w m based on the chunk boundary determined by the threshold value judgment manager 34 and outputs the result.
  • the text divider 14 divides the input text ST ("I'm going to Tokyo tomorrow") using a division engine, and creates the divided text STT as ' ⁇ STT>Tomorrow ⁇ SEP>I'm going to Tokyo ⁇ SEP>'. is generated and output (FIG. 2, step S3).
  • the divided text STT is sent to the translation machine 15 via a divided text correction manager (hereinafter simply referred to as a correction manager) 13.
  • a correction manager hereinafter simply referred to as a correction manager 13
  • the ' ⁇ SEP>' mark is used as a delimiter in translation processing.
  • the text divider 14 divides the uttered text data ['I'm going to Tokyo tomorrow]' in the data signal DS1 into two chunks 'Tomorrow' and 'I'm going to Tokyo', and generates the following data.
  • a signal DS2 is generated and sent to the segmented text correction manager 13.
  • chunk may be any group of words that can be translated. Therefore, not limited to the above case, for example, the text 'I'm going to Tokyo tomorrow' may be divided into three chunks: 'Tomorrow', 'To Tokyo', and 'I'm going', or 'Tomorrow'. , 'Tokyo', 'ni', and 'go'.
  • the correction manager 13 also has a function of transmitting the data signal DS2 to the first terminal controller 12 in response to a request signal from the first terminal controller 12.
  • the divided text correction manager 13 sends the data signal DS2, which is the output of the text divider 14, to the translator 15, and translates the two chunks ⁇ Asu'' and ⁇ I'm going to Tokyo'' included in the data signal DS2, respectively.
  • the translator 15 translates each of the two chunks into English. More specifically, the translator 15 outputs the translation result for the input according to the function translate (input, 'dir').
  • 'dir' defines the direction of translation.
  • 'dir' 'je' to specify translation from Japanese to English.
  • Japanese-English translation uses translate(input, 'je') as a function
  • English-Japanese translation uses translate(input, 'ej') as a function.
  • the translator 15 translates the 'jap' element of the data signal DS2 according to the following procedure, and generates the data signal DS2A.
  • the correction manager 13 incorporates the data signal DS2A into the data signal DS2, generates the following data signal DS3 (step S4), and sends it to the second terminal controller 22 (FIG. 1).
  • Data signal DS3 ⁇ 'id':10, 'eng': ['ASU', 'I go to Tokyo'] ⁇ That is, the data signal DS3 includes translated chunks (referred to as translated chunks) 'ASU' and 'I go to Tokyo' of the two chunks included in the data signal DS2.
  • translated chunks referred to as translated chunks
  • the second terminal controller 22 sends the data signal DS3 to the second terminal 21, and the second terminal 21 receives the two translation chunks 'ASU' and 'ASU' included in the data signal DS3, as shown in FIG. 3B. 'I go to Tokyo' is separated and displayed in a separate frame.
  • each translation chunk is displayed as an image display component called a text widget 21W surrounded by a frame line.
  • the text widget 21W displays the text of the translation result, and the second terminal 21 receives a designation (mistranslation designation) of mistranslation or meaningless chunks from the user (English speaker 2) (step S5).
  • mistranslation chunk When the English speaker 2 finds any mistranslation or meaningless chunk in any of the translated chunks, the English speaker 2 can remove the mistranslation or meaningless chunk (hereinafter referred to as mistranslation chunk) by touching the text widget (frame) 21W. specify.
  • the specified text widget 21W is changed from the standard display to the specified display.
  • mistranslated chunk includes a translated chunk that is mistranslated or has no meaning due to an error in the context judgment of the translator.
  • step S5 if the English speaker 2 determines that there is no mistranslated chunk among the translated chunks, for example, the English speaker 2 can simply touch the SEND button, etc., and the designation (identifier) of the mistranslated chunk can be sent to the data. Not included in the signal.
  • the error correction control unit 25 determines that there is no mistranslation chunk specification, and control shifts to RETURN, and shifts to processing of the next spoken text (step S1) (step S6, No).
  • step S1 the second terminal 21 is controlled to receive the spoken text (English text) in response to the previous spoken text (Japanese text), but the first terminal 11 Control may be performed to receive the next new spoken text (Japanese text) from.
  • the designation (identifier) of the mistranslated chunk is included in the data signal by the following operation by the English speaker 2 (step S6, Yes).
  • FIG. 3C shows the display on the second terminal 21 when the designation of mistranslated chunks by the English speaker 2 is accepted.
  • this shows the display when the translation chunk 'ASU' is mistranslated and a mistranslation designation is accepted.
  • the text widget 21W for 'ASU' is highlighted (designated display), and the text widget 21W for 'I go to Tokyo' is not highlighted (standard display). Note that when the text widget 21W is repeatedly touched, the display is alternately changed to no highlighting, highlighting, and no highlighting.
  • change the text color or frame line color, etc. such as displaying the text or frame line of the text widget 21W for 'ASU' in red (specified display) and 'I go to Tokyo' in black (standard display), Alternatively, it is sufficient if the mistranslation designation can be visually recognized by blinking characters or the like. Alternatively, the system may be configured to notify the user by audio or the like so that the text widget designated as an incorrect translation can be confirmed.
  • step S5 when the English speaker 2 presses the SEND button in FIG. (FIG. 1) (step S7).
  • chunk ID 0 is assigned to the specified displayed (touched) chunk 'ASU', and the data signal DS4 has an identifier 'error':[ 0] has been added.
  • an identifier (also referred to as a mistranslation identifier) is generated according to a signal (identification signal) specifying a mistranslation chunk inputted to the second terminal 21, and the error correction control unit 25 corrects the mistranslation based on the identifier.
  • the chunk is displayed on the first terminal 11 in a manner different from other translated chunks (ie correctly translated chunks).
  • the original language chunk chk[j] is translated one by one in chunk order. Therefore, translation chunks can be output without making the listener wait. Note that at this time, one chunk at a time, or two or more consecutive chunks may be translated one by one in chunk order. That is, at least one chunk may be translated sequentially.
  • the correction manager 13 receives the data signal DS4, extracts the element (translation chunk) of 'eng': ['ASU','I go to Tokyo'], and generates the following data signal DS4A.
  • the data signal DS4A is the content of the English text displayed on the display screen (touch screen) of the second terminal 21.
  • the translator 15 has a back translation function that back translates the translation result (English) of the speaker's (Japanese speaker) text and generates a back translation chunk (Japanese).
  • the first terminal 11 displays the mistranslation chunk widget 11W ('ASU') with a highlight, as shown in FIG. 3D, based on the detection result of the first terminal controller 12 (step S8 ). Note that, as described above, the font color or frame color of the mistranslated chunk 'ASU' may be changed to a different color from the standard display.
  • the English chunk translated by the translator 15 is back-translated by the translator 15 and displayed on the first terminal 11. That is, the first terminal 11 displays 'I'm going to Tokyo', which is the result (back translation chunk) of the translation (English) of the speaker's text (Japanese) back translated into Japanese. Therefore, the speaker can easily determine whether or not the translation has been correct.
  • the English chunk (translation chunk) translated by the translator 15 may be displayed as is on the first terminal 11.
  • the Japanese speaker 1 can confirm that there is a mistranslation or meaningless translation (collectively referred to as "mistranslation") by highlighting or other specified display. It can be seen that the English speaker 2 has designated the chunk 'ASU' corresponding to 'tomorrow' as a chunk that is a mistranslation or a meaningless translation (mistranslation chunk).
  • the Japanese speaker 1 paraphrases the Japanese chunk corresponding to the chunk specified by the English speaker 2, and inputs the paraphrased text into the first terminal 11 (step S9).
  • the paraphrased text is translated through the flow from step S1 described above, and translated chunks of the paraphrased text are displayed on the second terminal 21 (step S5).
  • all chunks included in the divided text may be translated, but only the paraphrased chunk may be translated, and the previous translated chunk may be used for the other translated chunks. It has advantages in terms of reducing translation capacity and improving translation speed.
  • Japanese speaker 1 can translate the Japanese part corresponding to the chunk specified by English speaker 2 into English by saying it in a different way, using a simpler expression, etc. can be conveyed to speaker 2.
  • the Japanese speaker 1 can convey the accurate translated text to the English speaker 2 by rephrasing 'Asu' as 'Ashita'.
  • FIG. 3F is a diagram showing the second terminal 21 on which the paraphrased text is translated and displayed. This time, the translation is accurate, so by touching the SEND button, the translation of this spoken text, ⁇ I'm going to Tokyo tomorrow'', is completed.
  • FIG. 3G shows a display that is a modified example of the display of the first terminal 11 shown in FIG. 3D. That is, the data signal DS2 is sent from the correction manager 13 to the first terminal 11 along with the data signal DS5 via the first terminal controller 12, and the first terminal 11 is sent the data signal DS2 together with the error 'ASU'.
  • the utterance content 'tomorrow' of speaker 1 may be added and displayed.
  • FIG. 7 is a block diagram showing the configuration of an error correction translation device 30 according to the second embodiment of the present invention.
  • a speech recognition section 11B and a speech recognition section 21B are provided in the first terminal 11 and the second terminal 21, respectively, in place of the keyboard 11A and the keyboard 21A. , is different from the error correction translation device 10 according to the first embodiment.
  • the first terminal 11 includes a speech recognition unit 11B that performs speech recognition on the utterance of the Japanese speaker 1 and outputs it as text data.
  • the second terminal 21 also includes a speech recognition unit 21B that performs speech recognition on the utterances of the English speaker 2 and outputs the speech as text data.
  • the voice recognition unit 11B of the first terminal 11 performs voice recognition of the utterance and sends text data that is the voice recognition result to the first terminal controller 12.
  • the second terminal 21 receives voice input from the English speaker 2, and sends the voice recognition result to the second terminal controller 22.
  • the speech recognition unit 11B and the speech recognition unit 21B have a speech recognition engine that can encode speech and convert it into data.
  • a method of encoding audio such as WAV or MP3, which is a known technology as a speech recognition method, can be used.
  • mistranslation chunks are specified in the same way as in the first embodiment. That is, an identifier is generated that specifies that there is a mistranslation or a meaningless translation, and based on the identifier, the mistranslation chunk is displayed on the speaker's terminal in a manner different from other translation chunks.
  • the translation chunk is also processed as a mistranslation chunk. Ru.
  • either one of the first terminal 11 and the second terminal 21 may be configured so that text data is input through voice recognition, and the other terminal is configured such that text data is input through a keyboard. Further, the first terminal 11 and the second terminal 21 may have a function of inputting text data by keyboard input and voice recognition.
  • FIG. 8 is a block diagram showing the configuration of an error correction translation device 50 according to the third embodiment of the present invention.
  • the error correction translation device 50 has the translation function from Japanese (first language) to English (second language) of the error correction translation device 10 according to the first embodiment.
  • a translation function is provided from Japanese (second language) to Japanese (first language). That is, the error correction translation device 50 is configured as a translation device capable of bidirectional simultaneous translation.
  • first language and the second language are not limited to Japanese and English, and can be applied to other languages as well.
  • the text divider 54 divides the spoken text data included in the data signal DS1 in the Japanese-English translation as well as the data signal ES1 in the English-Japanese translation into chunks.
  • the translator 55 is configured as a Japanese-English/English-Japanese translator.
  • the correction manager 53 is configured to manage data exchange of data signals DS1 to DS5 during Japanese-English translation and data signals ES1 to ES5 during English-Japanese translation.
  • the correction manager 53 is provided with an error correction completion determination unit 53A that determines whether error correction of one utterance text is completed.
  • the error correction completion determining unit 53A determines whether error correction of the one utterance text is completed based on the presence or absence of an identifier that specifies that there is a mistranslation or a meaningless translation.
  • the identifier input from the second terminal 21 first It is determined whether the error correction of the utterance text is completed based on the utterance identifier), and in the error correction of the utterance text from the speaker (English speaker 2) to the receiver (Japanese speaker 1), Based on the identifier (second identifier) input from the second terminal 11, it is determined whether or not the error correction of the uttered text has been completed.
  • the error correction completion determination unit 53A has a timer that measures the elapsed time.
  • the error correction completion determination unit 53A determines that error correction of the utterance text has been completed. You may.
  • the mistranslation identifier is present in the transmission/reception signal of the correction manager 53, it means that the error correction for the inputted utterance text has not been completed yet.
  • the transmitted/received signal is determined to be a transfer data signal for error correction.
  • the correction manager 53 functions to prohibit input of the next new utterance text and prioritize completion of error correction.
  • data signals ES1 to ES5 during English-Japanese translation correspond to data signals DS1 to DS5 during Japanese-English translation, respectively.
  • This data signal is configured as an equivalent signal to the data signals DS1 to DS5 except that the language of the speaker and the direction of translation are different.
  • the following data signals ES1 to ES5 can be exemplified.
  • ⁇ Data signal ES1 ⁇ 'id':11, 'eng': ['I'm heading to Boston tomorrow'] ⁇
  • ⁇ Data signal ES2 ⁇ 'id':11,'eng': ['I'm heading to','Boston','tomorrow'] ⁇
  • ⁇ Data signal ES3 ⁇ 'id':11,'jap': ['I'm going','To Boston','Tomorrow'] ⁇
  • ⁇ Data signal ES4 ⁇ 'id':11,'jap': ['I'm going','To Boston','Tomorrow'],'error':[1] ⁇
  • ⁇ Data signal ES5 ⁇ 'id':11,'eng': ['I'm heading to','BOSTON','tomorrow'],'error': [1] ⁇
  • the recipient (Japanese speaker 1) specifies 'To Boston' as meaningless and adds the identifier 'error':[1] (second identifier) containing chunk ID[1].
  • the resulting data signal ES4 is sent to the speaker (English speaker 2).
  • the second terminal 21 of the speaker receives an identifier in the data signal ES5 as a designated indication that 'BOSTON' is a meaningless translation (mistranslation), unlike other translation chunks. It is displayed in a different manner (for example, in uppercase letters), and the English speaker 2 can confirm that 'Boston' in the speech chunk is meaningless to the listener (Japanese speaker 1).
  • the translator 55 has a context determination function. In this example, the translator 55 determines that 'Boston' means 'Boston Library' from the context of the conversation between Japanese speaker 1 and English speaker 2, and translates it. This is a mistranslation of "Boston" when it should have been translated.
  • the English speaker 2 can convey the accurate translated text to the Japanese speaker 1 by rephrasing 'BOSTON' to 'the Boston Library'.
  • bidirectional simultaneous translation is possible between speakers of the first language and speakers of the second language. Therefore, both parties can confirm whether the meaning is unclear or there is a mistranslation by the designated display, and conversation can be smoothly carried out through accurate translation.
  • the present invention by managing units of utterances with meaningful chunk IDs as translations, and by linking utterance information in the source language and translation information in the target language to the chunk IDs, errors can be avoided on the source language side. can provide reliable feedback.
  • chunk-based error correction of the present invention errors can be localized and corrected locally, making error correction conversations more efficient and improving the quality of conversations in simultaneous interpretation. can be improved.
  • the first terminal controller 12 and the second terminal controller 22 receive signals indicating input confirmation such as 'SEND' and 'Enter' transmitted from the first terminal 11 and the second terminal 21.
  • the end of the text can be determined based on .
  • the error correction translation device described above may be configured as a CPU (Central Processing Unit), a GPU (Graphics Processing Unit), or a server. Furthermore, it may be implemented in computers, information terminals, mobile phones, and the like.
  • CPU Central Processing Unit
  • GPU Graphics Processing Unit
  • server a server
  • it may be implemented in computers, information terminals, mobile phones, and the like.
  • error correction translation device and error correction translation method are configured as hardware, they may be configured as computer software (program). Alternatively, the program may be configured as a computer-readable storage medium in which the program is stored.
  • any one of the first and second terminal controllers, text divider, translator, speech recognition unit, and error correction control unit may be configured as software. Alternatively, these may be configured as one software.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Artificial Intelligence (AREA)
  • Human Computer Interaction (AREA)
  • Machine Translation (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

This error correcting translation device comprises: a first terminal into which text in a first language is input; a text divider which divides the text into chunks to generate divided text; a translation machine which individually translates each chunk included in the divided text to generate translated chunks in a second language; and an error correction control unit which causes each translated chunk from the translation machine to be displayed individually on a second terminal. The second terminal accepts input of an identifier specifying a mistranslated chunk, and the error correction control unit causes the mistranslated chunk to be displayed on the first terminal in a different manner to other translated chunks, on the basis of the identifier input into the second terminal, and causes the first terminal to accept input of paraphrased text of the text in the first language in response to the display of the mistranslated chunk.

Description

誤り訂正翻訳装置、誤り訂正翻訳方法及びプログラム並びにその格納媒体Error correction translation device, error correction translation method, program, and storage medium thereof
 本発明は、誤り訂正翻訳装置、誤り訂正翻訳方法及びプログラム並びにその格納媒体に関する。 The present invention relates to an error correction translation device, an error correction translation method, a program, and a storage medium thereof.
 従来、機械翻訳を用いた同時通訳のために、入力されたシーケンスデータを分割する文章分割方法が知られている。 Conventionally, there is a known sentence division method for dividing input sequence data for simultaneous interpretation using machine translation.
 例えば、特許文献1には、リアルタイムでシーケンスデータ分割処理(例えば、文章分割処理)を実行することができるデータ・セグメンテーション装置が開示されている。 For example, Patent Document 1 discloses a data segmentation device that can perform sequence data segmentation processing (for example, text segmentation processing) in real time.
特開2020-24277号公報JP2020-24277A
 しかしながら、従来技術においては、通訳で生じる誤りの範囲を、発話者と聞き手で共通に理解して進めることが困難であった。そのため、誤り訂正会話が複雑で冗長なものになっていた。 However, in the prior art, it has been difficult for the speaker and the listener to have a common understanding of the scope of errors that occur during interpretation. As a result, error correction conversations became complicated and redundant.
 本発明は上記した点に鑑みてなされたものであり、同時通訳における誤りを原言語側に確実にフィードバックすることができ、誤り訂正が効率的で、会話の質を向上することができる誤り訂正翻訳装置、誤り訂正翻訳方法及びプログラム並びにその格納媒体を提供することを目的とする。 The present invention has been made in view of the above points, and is an error correction method that can reliably feed back errors in simultaneous interpretation to the source language side, makes error correction efficient, and improves the quality of conversation. The purpose of the present invention is to provide a translation device, an error-correcting translation method, a program, and a storage medium thereof.
 本発明の1実施形態による誤り訂正翻訳装置は、
 第1言語のテキストが入力される第1の端末と、
 前記第1言語のテキストをチャンクに分割して分割テキストを生成するテキスト分割器と、
 前記分割テキストに含まれるチャンクの各々を個別に翻訳して第2言語の翻訳チャンクを生成する翻訳機と、
 第2の端末と、
 前記翻訳機からの前記翻訳チャンクの各々を前記第2の端末に個別に表示させる誤り訂正制御部と、を有している。
An error correction translation device according to an embodiment of the present invention includes:
a first terminal into which text in a first language is input;
a text divider that divides the first language text into chunks to generate divided text;
a translator that individually translates each chunk included in the divided text to generate translated chunks in a second language;
a second terminal;
and an error correction control unit that causes each of the translation chunks from the translator to be individually displayed on the second terminal.
 前記第2の端末は、前記翻訳チャンクのうちの誤訳チャンクを指定する識別子の入力を受け付け、
 前記誤り訂正制御部は、前記第2の端末に入力された前記識別子に基づいて、前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第1の端末に表示させ、前記誤訳チャンクの表示に応答して、前記第1の端末に前記第1言語のテキストの言い換えテキストの入力を受け付けさせる、ことを特徴とする。
The second terminal receives an input of an identifier specifying a mistranslated chunk among the translated chunks,
The error correction control unit causes the mistranslation chunk to be displayed on the first terminal in a manner different from other translation chunks based on the identifier input to the second terminal, and the error correction control unit causes the mistranslation chunk to be displayed in a manner different from other translation chunks. In response, the first terminal is configured to accept input of paraphrased text of the text in the first language.
 本発明の他の実施形態による誤り訂正翻訳方法は、
 第1言語のテキストを第1の端末に入力するステップと、
 前記第1言語のテキストをチャンクに分割して分割テキストを生成するテキスト分割ステップと、
 前記分割テキストに含まれるチャンクの各々を個別に翻訳して第2言語の翻訳チャンクを生成する翻訳ステップと、
 前記翻訳機からの前記翻訳チャンクの各々を前記第2の端末に個別に表示させる誤り訂正制御ステップと、
 前記第2の端末において、前記翻訳チャンクのうちの誤訳チャンクを指定する識別子の入力を受け付けるステップと、
 前記第2の端末に入力された前記識別子に基づいて、前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第1の端末に表示させる誤訳チャンク表示ステップと、
 前記誤訳チャンク表示ステップの後に、前記第1の端末への、前記第1言語のテキストの言い換えテキストの入力を受け付けるステップと、
を有する。
An error correction translation method according to another embodiment of the present invention includes:
inputting text in a first language into a first terminal;
a text dividing step of dividing the first language text into chunks to generate divided text;
a translation step of individually translating each chunk included in the divided text to generate translated chunks in a second language;
an error correction control step of individually displaying each of the translation chunks from the translator on the second terminal;
receiving, at the second terminal, an input of an identifier specifying a mistranslated chunk among the translated chunks;
a mistranslation chunk displaying step of displaying the mistranslation chunk on the first terminal in a manner different from other translated chunks based on the identifier input to the second terminal;
after the mistranslation chunk displaying step, receiving input of a paraphrase text of the text in the first language to the first terminal;
has.
 本発明のさらに他の実施形態による誤り訂正翻訳プログラムは、
 テキストデータを翻訳するためにコンピュータを、
 第1言語のテキストが入力される第1の端末手段、
 前記第1言語のテキストをチャンクに分割して分割テキストを生成するテキスト分割手段、
 前記分割テキストに含まれるチャンクの各々を個別に翻訳して第2言語の翻訳チャンクを生成する翻訳手段、
 第2の端末手段、
 前記翻訳手段からの前記翻訳チャンクの各々を前記第2の端末手段に個別に表示させる誤り訂正制御手段、として機能させ、
 前記第2の端末手段は、前記翻訳チャンクのうちの誤訳チャンクを指定する識別子の入力を受け付け、
 前記誤り訂正制御手段は、前記第2の端末手段に入力された前記識別子に基づいて、前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第1の端末手段に表示させ、前記誤訳チャンクの表示に応答して、前記第1の端末手段に前記第1言語のテキストの言い換えテキストの入力を受け付けさせる、ことを特徴とする。
An error correction translation program according to yet another embodiment of the present invention includes:
computer to translate text data,
a first terminal means into which text in a first language is input;
text dividing means for dividing the first language text into chunks to generate divided text;
Translation means for individually translating each chunk included in the divided text to generate translated chunks in a second language;
second terminal means;
functioning as error correction control means for individually displaying each of the translation chunks from the translation means on the second terminal means;
The second terminal means receives an input of an identifier specifying a mistranslated chunk among the translated chunks,
The error correction control means causes the first terminal means to display the mistranslated chunk in a manner different from other translation chunks based on the identifier input to the second terminal means, The method is characterized in that, in response to the display, the first terminal means accepts input of paraphrased text of the text in the first language.
本発明の第1の実施形態による誤り訂正翻訳装置10の構成を示すブロック図である。1 is a block diagram showing the configuration of an error correction translation device 10 according to a first embodiment of the present invention. FIG. 誤り訂正制御部25が実行する誤り訂正制御のフローを示すフローチャートである。3 is a flowchart showing a flow of error correction control executed by an error correction control unit 25. FIG. 発話テキストが表示される第1の端末11の表示画面を模式的に示す図である。FIG. 3 is a diagram schematically showing a display screen of the first terminal 11 on which spoken text is displayed. 翻訳チャンクが分離して個別に枠表示される第2の端末21の表示画面を模式的に示す図である。FIG. 3 is a diagram schematically showing a display screen of the second terminal 21 on which translation chunks are separated and displayed in individual frames. 誤訳チャンクの指定を受け付けたときの第2の端末21の表示画面を模式的に示す図である。FIG. 6 is a diagram schematically showing the display screen of the second terminal 21 when the designation of mistranslated chunks is received. 誤訳チャンクがハイライト表示された第1の端末11の表示画面を模式的に示す図である。FIG. 3 is a diagram schematically showing a display screen of the first terminal 11 on which mistranslated chunks are highlighted. 誤訳チャンクがハイライト表示された第1の端末11の表示画面の他の例を模式的に示す図である。FIG. 7 is a diagram schematically showing another example of the display screen of the first terminal 11 on which mistranslated chunks are highlighted. 言い換えテキストが翻訳されて表示された第2の端末21の表示画面を模式的に示す図である。It is a figure which shows typically the display screen of the 2nd terminal 21 on which the paraphrase text was translated and displayed. 図3Dに示す第1の端末11の表示の改変例を模式的に示す図である。3D is a diagram schematically showing a modification example of the display of the first terminal 11 shown in FIG. 3D. FIG. テキスト分割器13の構成の一例を示すブロック図である。2 is a block diagram showing an example of the configuration of a text divider 13. FIG. MeCabを用いた日本語テキストの形態素解析結果を示す図である。It is a figure showing the result of morphological analysis of Japanese text using MeCab. NLTKを用いた英語テキストの形態素解析結果を示す図である。It is a figure which shows the morphological analysis result of the English text using NLTK. 閾値判定管理器34が実行するチャンク境界の決定処理手順を示すフローチャートである。3 is a flowchart showing a chunk boundary determination processing procedure executed by the threshold value determination manager 34. FIG. 本発明の第2の実施形態による誤り訂正翻訳装置30の構成を示すブロック図である。FIG. 2 is a block diagram showing the configuration of an error correction translation device 30 according to a second embodiment of the present invention. 本発明の第3の実施形態による誤り訂正翻訳装置50の構成を示すブロック図である。FIG. 3 is a block diagram showing the configuration of an error correction translation device 50 according to a third embodiment of the present invention.
 以下においては、本発明の好適な実施例について説明するが、これらを適宜改変し、組合せてもよい。また、以下の説明及び添付図面において、実質的に同一又は等価な部分には同一の参照符を付して説明する。 In the following, preferred embodiments of the present invention will be described, but these may be modified and combined as appropriate. Further, in the following description and the accompanying drawings, substantially the same or equivalent parts are designated by the same reference numerals.
[第1の実施形態]
 図1は、本発明の第1の実施形態による誤り訂正翻訳装置10の構成を示すブロック図である。
[First embodiment]
FIG. 1 is a block diagram showing the configuration of an error correction translation device 10 according to a first embodiment of the present invention.
 図1に示すように、誤り訂正翻訳装置10は、日本語(第1の言語)の話者の会話を英語(第2の言語)の話者に翻訳する翻訳装置として構成されている。しかし、日本語(第1の言語)の話者の会話を英語(第2の言語)の話者に翻訳する、及び、英語話者の会話を日本語話者に翻訳する双方向的な同時通訳の翻訳装置として構成されていてもよい。 As shown in FIG. 1, the error correction translation device 10 is configured as a translation device that translates a conversation of a Japanese (first language) speaker to an English (second language) speaker. However, two-way simultaneous translation of Japanese (first language) speaker's conversation to English (second language) speaker's and English speaker's conversation to Japanese speaker's It may be configured as a translation device for an interpreter.
 当該第1の言語(原言語)及び当該第2の言語(翻訳言語)は日本語及び英語に限定されず、他の言語に対しても適用することができる。 The first language (original language) and the second language (translation language) are not limited to Japanese and English, and can be applied to other languages as well.
 誤り訂正翻訳装置10は、タブレット端末(入出力装置)である第1の端末(図中、端末#1)11、テキスト分割器14、日英/英日翻訳機(以下、単に翻訳機という。)15、タブレット端末(入出力装置)である第2の端末(図中、端末#2)21及び誤り訂正制御部25を有している。 The error correction translation device 10 includes a first terminal (terminal #1 in the figure) 11 which is a tablet terminal (input/output device), a text divider 14, and a Japanese-English/English-Japanese translator (hereinafter simply referred to as a translator). ) 15, a second terminal (terminal #2 in the figure) 21 which is a tablet terminal (input/output device), and an error correction control section 25.
 また、誤り訂正制御部25は、第1の端末制御器12、分割テキスト訂正管理器(以下、単に訂正管理器という。)13、第2の端末制御器22を有している。 The error correction control unit 25 also includes a first terminal controller 12 , a divided text correction manager (hereinafter simply referred to as a correction manager) 13 , and a second terminal controller 22 .
 第1の端末11は、スクリーンキーボード(KB)11Aを具備している。日本語話者1の操作によってスクリーンキーボード(以下、単にキーボードという。)11Aから入力されたかな文字は、第1の端末11内のプロセッサ(図示しない)により、日本語のテキストデータ(シーケンスデータ)として第1の端末11から出力される。なお、この際、キーボード11Aから入力されたかな文字について、必要に応じて公知の技術であるかな漢字変換が利用されてもよい。 The first terminal 11 is equipped with a screen keyboard (KB) 11A. Kana characters input from the screen keyboard (hereinafter simply referred to as the keyboard) 11A by the Japanese speaker 1 are converted into Japanese text data (sequence data) by a processor (not shown) in the first terminal 11. is output from the first terminal 11 as . At this time, kana-kanji conversion, which is a known technique, may be used for the kana characters input from the keyboard 11A, if necessary.
 また、第1の端末11は、第1の端末制御器12からのデータを受信して表示するディスプレイを有する。さらに、当該ディスプレイは日本語話者1の操作を受け付けるタッチパネルとしての機能を有している。 The first terminal 11 also has a display that receives and displays data from the first terminal controller 12. Furthermore, the display has a function as a touch panel that receives operations from the Japanese speaker 1.
 同様に、第2の端末21は、キーボード(KB)21Aを具備している。英語話者2の操作によってキーボード21Aから入力された文字は、第2の端末21内のプロセッサ(図示しない)により、英語のテキストデータ(シーケンスデータ)として第2の端末21から出力される。 Similarly, the second terminal 21 is equipped with a keyboard (KB) 21A. Characters input from the keyboard 21A by the English speaker 2 are output from the second terminal 21 as English text data (sequence data) by a processor (not shown) in the second terminal 21.
 また、第2の端末21は、第2の端末制御器22からのデータを受信して表示するディスプレイを有する。さらに、当該ディスプレイは英語話者2の操作を受け付けるタッチパネルとしての機能を有している。 The second terminal 21 also has a display that receives and displays data from the second terminal controller 22. Furthermore, the display has a function as a touch panel that receives operations from the English speaker 2.
 以下に、日本語話者の会話を英語話者に翻訳する同時通訳の場合について、図を参照して具体的に説明する。図2は、誤り訂正制御部25が実行する誤り訂正翻訳方法のフローを示すフローチャートである。また、図3A~図3Gは、第1の端末11及び第2の端末21のディスプレイ画面を模式的に示す図である。 Below, a case of simultaneous interpretation in which a conversation between a Japanese speaker and an English speaker is translated will be specifically explained with reference to the diagram. FIG. 2 is a flowchart showing the flow of the error correction translation method executed by the error correction control unit 25. Further, FIGS. 3A to 3G are diagrams schematically showing display screens of the first terminal 11 and the second terminal 21.
 なお、以下においては、誤り訂正翻訳装置10に1つの文からなるテキストが入力され、翻訳される場合について説明する。 In the following, a case will be described in which a text consisting of one sentence is input to the error correction translation device 10 and translated.
 日本語話者1は、キーボード(KB)11Aを用いて「あす東京に行きます」というテキスト(以下、発話テキストともいう。)を入力する。第1の端末11は、入力された発話テキストを第1の端末制御器12に送る。 The Japanese speaker 1 inputs the text "I'm going to Tokyo tomorrow" (hereinafter also referred to as uttered text) using the keyboard (KB) 11A. The first terminal 11 sends the input spoken text to the first terminal controller 12 .
 第1の端末制御器12は、第1の端末11から受信した入力テキストから以下のデータ信号DS1を生成し(ステップS1)、訂正管理器13に送る。また、第1の端末制御器12は、データ信号DS1は第1の端末11にも送られ、図3Aに示すように、第1の端末11には、データ信号DS1に含まれる発話テキストデータ[‘あす東京に行きます’]がテキストウィジェット11Wの形式で表示される(ステップS2)。 The first terminal controller 12 generates the following data signal DS1 from the input text received from the first terminal 11 (step S1), and sends it to the correction manager 13. The first terminal controller 12 also sends the data signal DS1 to the first terminal 11, and as shown in FIG. 3A, the first terminal 11 receives the spoken text data [ 'I'm going to Tokyo tomorrow'] is displayed in the format of the text widget 11W (step S2).
 なお、データ信号のデータフォーマットとして、JSON(JavaScript Object Notation)フォーマットを用いた場合を例に説明する。 Note that an example will be described in which the JSON (JavaScript Object Notation) format is used as the data format of the data signal.
 データ信号DS1:{‘id’:10, ‘jap’:[‘あす東京に行きます’]}
ここで、‘id’は発話のIDであり、‘id’はid ={0, 1, ・・・}なるゼロ以上の整数である。したがって、「‘id’:10」は11番目の発話であることを意味している。
Data signal DS1: {'id':10, 'jap': ['I'm going to Tokyo tomorrow']}
Here, 'id' is the ID of the utterance, and 'id' is an integer greater than or equal to zero such that id = {0, 1, . . .}. Therefore, "'id':10" means the 11th utterance.
 誤り訂正制御部25の訂正管理器13は、データ信号DS1をテキスト分割器14に送る(ステップS2)。テキスト分割器14は、データ信号DS1に含まれる発話テキストデータをチャンク(部分文)に分割する。以下に図4を参照してテキスト分割器13の動作について詳細に説明する。 The correction manager 13 of the error correction control unit 25 sends the data signal DS1 to the text divider 14 (step S2). The text divider 14 divides the spoken text data included in the data signal DS1 into chunks (partial sentences). The operation of the text divider 13 will be described in detail below with reference to FIG.
 図4は、テキスト分割器13の構成の一例を示すブロック図である。テキスト分割器13の入力端13Iに入力されたテキストST(図中、<ST>と示す。)は単語分割器31によって単語に分割される。 FIG. 4 is a block diagram showing an example of the configuration of the text divider 13. A text ST (indicated as <ST> in the figure) input to the input end 13I of the text divider 13 is divided into words by the word divider 31.
 単語分割器31は、例えば、形態素解析を行ってテキストSTを単語w~wに分割する。分割により得られた各単語はFIFO(First In First Out)レジスタである単語レジスタ32のレジスタr~rに順次レジスタ値wr~wrとして保持される。すなわち、レジスタrj には、単語w(i=1, 2, ・, m)が保持される(mは1以上の整数)。 The word divider 31 performs, for example, morphological analysis to divide the text ST into words w 1 to w m . Each word obtained by the division is sequentially held as register values wr 1 to wr m in registers r 1 to r m of a word register 32, which is a FIFO (First In First Out) register. That is, the word w i (i=1, 2, ·, m) is held in the register r j (m is an integer of 1 or more).
 具体的には、公知技術のオープンソースの形態素解析エンジンであるMeCabを用いた。図5Aは、MeCabを用いた日本語テキストの形態素解析結果を示している。 Specifically, MeCab, which is a publicly known open source morphological analysis engine, was used. FIG. 5A shows the results of morphological analysis of Japanese text using MeCab.
 単語分割器31に‘あす東京に行きます’という日本語テキストが入力されると、例えば、図5Aに示すように、形態素解析により5つの単語に分割される。なお、図中の「名詞」等は日本語の文法における品詞を示している。また、EOS(End Of Sentence)は、分割されたテキストの最後であることを示す記号である。 When the Japanese text ``I'm going to Tokyo tomorrow'' is input to the word divider 31, it is divided into five words by morphological analysis, for example, as shown in FIG. 5A. Note that "nouns" and the like in the diagram indicate parts of speech in Japanese grammar. Furthermore, EOS (End Of Sentence) is a symbol indicating the end of the divided text.
 すなわち、本実施例の場合では、m=5であり、単語分割器31により分割された単語の系列として、‘あす’、‘東京’ 、‘に’、‘行き’、‘ます’が得られる。 That is, in the case of this embodiment, m=5, and the word sequences divided by the word divider 31 are 'Asu', 'Tokyo', 'ni', 'Go', and 'Masu'. .
 なお、英語の単語分割器31としては、公知の技術である形態素解析技術であるNLTK(Natural Language Toolkit)などを用いることができる。図5Bは、NLTKを用いた英語テキストの形態素解析結果を示している。 Note that as the English word divider 31, a well-known morphological analysis technique such as NLTK (Natural Language Toolkit) can be used. FIG. 5B shows the results of morphological analysis of an English text using NLTK.
 この場合、単語分割器31に‘I’m going to Tokyo tomorrow’という英語テキストが入力されると、例えば、図5Bに示すように、’I’、‘m’、‘going’、’to’、’Tokyo’、’tomorrow’の6つの単語に分割される。なお、図中のPRPなどは、英文法における品詞を示している。また、EOSは分割されたテキストの最後であることを示す記号である。 In this case, when the English text 'I'm going to Tokyo tomorrow' is input to the word divider 31, for example, as shown in FIG. 5B, 'I', 'm', 'going', 'to' , 'Tokyo', and 'tomorrow'. Note that PRP in the diagram indicates the part of speech in English grammar. Also, EOS is a symbol indicating the end of the divided text.
 次に、図6のフローチャートを参照しつつ、閾値判定管理器34が実行するチャンク境界の決定処理手順について説明する。 Next, with reference to the flowchart of FIG. 6, the chunk boundary determination processing procedure executed by the threshold value determination manager 34 will be described.
 図4に示すように、閾値レジスタ35には単語レジスタ32のレジスタr~rの各々に対応して閾値θi(i=1, 2, ・・・, m)が予め保持されている。 As shown in FIG. 4, threshold values θi (i=1, 2, . . . , m) are stored in advance in the threshold value register 35 in correspondence with each of the registers r 1 to r m of the word register 32.
 単語レジスタ32にレジスタ値wr~wrが保持されると、深層学習器33による深層学習によって単語w(i=1, 2, ・・・, m)の終わりが分割位置である確率が計算される。この確率をP(wr)とする。 When the register values wr 1 to wr m are held in the word register 32, the probability that the end of the word w m (i=1, 2, ..., m) is the division position is determined by deep learning by the deep learning device 33. calculated. Let this probability be P(wr i ).
 なお、各単語の末尾でテキストが分割される確率P(wr)(i=1, 2, ・, m)は、単語系列が分割される確率の総和が1であるように付与される。 Note that the probability P(wr i ) (i=1, 2, ·, m) that the text is divided at the end of each word is given so that the sum of the probabilities that the word series is divided is 1.
 まず、初期値i=1とし(ステップS21)、以降のステップS22~S25が実行される。 First, the initial value i is set to 1 (step S21), and subsequent steps S22 to S25 are executed.
 ステップS22において、比較器36によりP(wr)>θiであるか否かが判定される。P(wr)>θiであると判定された場合には、単語wの終わりが分割位置、すなわち、発話テキストをチャンクに分割する境界(チャンク境界)であると決定される(ステップS25)。P(wr)>θiではないと判定された場合には、次のステップS23に進む。 In step S22, the comparator 36 determines whether P(wr i )>θi. If it is determined that P(wr i )>θi, the end of the word w i is determined to be the division position, that is, the boundary (chunk boundary) where the spoken text is divided into chunks (step S25). . If it is determined that P(wr i )>θi is not satisfied, the process advances to the next step S23.
 ステップS23において、現在の単語がテキストの最後の単語か否か(i≧m)が判定される。現在の単語がテキストの最後の単語であると判定された場合には、ステップS23に進み、当該単語の終わりが分割位置であると決定される。 In step S23, it is determined whether the current word is the last word of the text (i≧m). If it is determined that the current word is the last word of the text, the process proceeds to step S23, where the end of the word is determined to be the division position.
 ステップS23において、現在の単語がテキストの最後の単語ではないと判定された場合には、ステップS24に進み、iが1だけインクリメントされ、次の単語について上記したステップが繰り返される。 If it is determined in step S23 that the current word is not the last word of the text, the process proceeds to step S24, where i is incremented by 1, and the above steps are repeated for the next word.
 最後の単語までチャンク境界の決定処理が終了した場合には本処理を終了する(END)。 If the chunk boundary determination process has been completed up to the last word, this process ends (END).
 分割テキスト出力器37は、閾値判定管理器34により決定されたチャンク境界に基づいて、単語w~wの系列に分割マーク(セパレータ)を付与して出力する。 The divided text output device 37 adds a division mark (separator) to the series of words w 1 to w m based on the chunk boundary determined by the threshold value judgment manager 34 and outputs the result.
 例えば、単語wと単語wの間に'<SEP>'で表される分割マークを付与する場合、w,w,w,w,'<SEP>',w,・・・,wが分割マーク付きのテキストSTT(図4、<STT>と示す。)としてテキスト分割器14の入力端13Oから出力される。 For example, when adding a division mark represented by '<SEP>' between word w 4 and word w 5 , w 1 , w 2 , w 3 , w 4 , '<SEP>', w 5 , . .
 すなわち、テキスト分割器14は、入力されたテキストST(「あす東京に行きます」)を分割エンジンで分割し、分割テキストSTTとして、‘<STT>あす<SEP>東京に行きます<SEP>’を生成し、出力する(図2、ステップS3)。 That is, the text divider 14 divides the input text ST ("I'm going to Tokyo tomorrow") using a division engine, and creates the divided text STT as '<STT>Tomorrow<SEP>I'm going to Tokyo<SEP>'. is generated and output (FIG. 2, step S3).
 また、分割テキストSTTは、分割テキスト訂正管理器(以下、単に訂正管理器という)13を介して翻訳機15へ送られる。翻訳機15では、'<SEP>'マークが翻訳処理の区切りとして利用される。 Further, the divided text STT is sent to the translation machine 15 via a divided text correction manager (hereinafter simply referred to as a correction manager) 13. In the translator 15, the '<SEP>' mark is used as a delimiter in translation processing.
 具体的には、テキスト分割器14は、データ信号DS1中の発話テキストデータ[‘あす東京に行きます’]を2つのチャンク‘あす’、及び‘東京に行きます’に分割し、以下のデータ信号DS2を生成し、分割テキスト訂正管理器13に送る。 Specifically, the text divider 14 divides the uttered text data ['I'm going to Tokyo tomorrow]' in the data signal DS1 into two chunks 'Tomorrow' and 'I'm going to Tokyo', and generates the following data. A signal DS2 is generated and sent to the segmented text correction manager 13.
 なお、本明細書において用語「チャンク」は、翻訳可能なひとまとまりの語であればよい。従って、上記した場合に限らず、例えば、テキスト‘あす東京に行きます’を、‘あす’、‘東京に’、‘行きます’ の3つのチャンクに分割してもよく、あるいは、‘あす’、‘東京’、‘に’、‘行きます’ の4つのチャンクに分割してもよい。 Note that in this specification, the term "chunk" may be any group of words that can be translated. Therefore, not limited to the above case, for example, the text 'I'm going to Tokyo tomorrow' may be divided into three chunks: 'Tomorrow', 'To Tokyo', and 'I'm going', or 'Tomorrow'. , 'Tokyo', 'ni', and 'go'.
 データ信号DS2:{‘id’:10, ‘jap’:[‘あす’,‘東京に行きます’]}
 また、訂正管理器13は、第1の端末制御器12からの要求信号に応じて、データ信号DS2を第1の端末制御器12に送信する機能を有する。
Data signal DS2: {'id':10, 'jap': ['tomorrow', 'I'm going to Tokyo']}
The correction manager 13 also has a function of transmitting the data signal DS2 to the first terminal controller 12 in response to a request signal from the first terminal controller 12.
 分割テキスト訂正管理器13は、テキスト分割器14の出力であるデータ信号DS2を翻訳機15に送り、データ信号DS2に含まれる2つのチャンク‘あす’及び‘東京に行きます’ をそれぞれ翻訳する。 The divided text correction manager 13 sends the data signal DS2, which is the output of the text divider 14, to the translator 15, and translates the two chunks ``Asu'' and ``I'm going to Tokyo'' included in the data signal DS2, respectively.
 翻訳機15は、2つのチャンクをそれぞれ英語に翻訳する。より詳細には、翻訳機15は、translate (input, ‘dir’)の関数にしたがって、inputに対して、翻訳結果を出力する。ここで、’dir‘は、翻訳の方向を定義する。ここでは、’dir’=’je’として、日本語から英語に翻訳することを指定する。つまり、日英翻訳は、translate(input, ‘je’)を関数とし、一方、英日翻訳は、translate(input, ‘ej’) を関数とする。 The translator 15 translates each of the two chunks into English. More specifically, the translator 15 outputs the translation result for the input according to the function translate (input, 'dir'). Here, 'dir' defines the direction of translation. Here, 'dir' = 'je' to specify translation from Japanese to English. In other words, Japanese-English translation uses translate(input, 'je') as a function, while English-Japanese translation uses translate(input, 'ej') as a function.
 翻訳機15は、データ信号DS2の‘jap’要素を以下の手順で翻訳し、データ信号DS2Aを生成する。
output=translate(‘jap’: [‘あす’, ‘東京に行きます’], ‘je’)
=translate(‘jap’:[‘あす’], ‘je’) + translate(‘jap’: [‘東京に行きます’], ‘je’)
=’eng’: [‘ASU’] + ‘eng: [‘I go to Tokyo’]
=’eng’: [‘ASU’, ‘I go to Tokyo’]   
 データ信号DS2A:’eng’: [‘ASU’, ‘I go to Tokyo’]
The translator 15 translates the 'jap' element of the data signal DS2 according to the following procedure, and generates the data signal DS2A.
output=translate('jap': ['tomorrow', 'I'm going to Tokyo'], 'je')
= translate('jap':['tomorrow'], 'je') + translate('jap': ['I'm going to Tokyo'], 'je')
='eng': ['ASU'] + 'eng: ['I go to Tokyo']
='eng': ['ASU', 'I go to Tokyo']
Data signal DS2A: 'eng': ['ASU', 'I go to Tokyo']
 訂正管理器13は、データ信号DS2Aをデータ信号DS2に組み込んで、以下のデータ信号DS3を生成し(ステップS4)、第2の端末制御器22に送る(図1)。 The correction manager 13 incorporates the data signal DS2A into the data signal DS2, generates the following data signal DS3 (step S4), and sends it to the second terminal controller 22 (FIG. 1).
 データ信号DS3:{‘id’:10, ’eng’: [‘ASU’, ‘I go to Tokyo’]}
 すなわち、データ信号DS3には、データ信号DS2に含まれる2つのチャンクについてそれぞれ翻訳されたチャンク(翻訳チャンクという。)‘ASU’及び‘I go to Tokyo’が含まれている。
Data signal DS3: {'id':10, 'eng': ['ASU', 'I go to Tokyo']}
That is, the data signal DS3 includes translated chunks (referred to as translated chunks) 'ASU' and 'I go to Tokyo' of the two chunks included in the data signal DS2.
 第2の端末制御器22は、データ信号DS3を第2の端末21に送り、第2の端末21には、図3Bに示すように、データ信号DS3に含まれる2つの翻訳チャンク‘ASU’及び‘I go to Tokyo’が分離して個別に枠表示される。 The second terminal controller 22 sends the data signal DS3 to the second terminal 21, and the second terminal 21 receives the two translation chunks 'ASU' and 'ASU' included in the data signal DS3, as shown in FIG. 3B. 'I go to Tokyo' is separated and displayed in a separate frame.
 より具体的には、翻訳チャンクはそれぞれ枠線で囲まれたテキストウィジェット21Wと呼ばれる画像表示の部品として表示される。テキストウィジェット21Wは、翻訳結果のテキストを表示し、第2の端末21は、利用者(英語話者2)から誤訳又は意味不明なチャンクの指定(誤訳指定)を受け付ける(ステップS5)。 More specifically, each translation chunk is displayed as an image display component called a text widget 21W surrounded by a frame line. The text widget 21W displays the text of the translation result, and the second terminal 21 receives a designation (mistranslation designation) of mistranslation or meaningless chunks from the user (English speaker 2) (step S5).
 英語話者2は、翻訳チャンクのいずれかに誤訳又は意味不明なチャンクがあったときにはそのテキストウィジェット(枠)21Wにタッチすることによって、誤訳又は意味不明なチャンク(以下、誤訳チャンクという。)を指定する。指定されたテキストウィジェット21Wは基準表示から指定表示に変更される。 When the English speaker 2 finds any mistranslation or meaningless chunk in any of the translated chunks, the English speaker 2 can remove the mistranslation or meaningless chunk (hereinafter referred to as mistranslation chunk) by touching the text widget (frame) 21W. specify. The specified text widget 21W is changed from the standard display to the specified display.
 なお、本明細書において、用語「誤訳チャンク」は、翻訳機の文脈判断の誤りによる誤訳又は意味不明となった翻訳チャンクを含む。 Note that in this specification, the term "mistranslated chunk" includes a translated chunk that is mistranslated or has no meaning due to an error in the context judgment of the translator.
 ステップS5において、英語話者2が翻訳チャンク中に誤訳チャンクが無いと判断した場合には、例えば、英語話者2が単にSENDボタン等にタッチすることにより、誤訳チャンクの指定(識別子)はデータ信号中に含まれない。 In step S5, if the English speaker 2 determines that there is no mistranslated chunk among the translated chunks, for example, the English speaker 2 can simply touch the SEND button, etc., and the designation (identifier) of the mistranslated chunk can be sent to the data. Not included in the signal.
 この場合、誤り訂正制御部25は誤訳チャンクの指定が無いと判別し、制御はRETURNに移行し、次の発話テキストの処理(ステップS1)に移行する(ステップS6、No)。 In this case, the error correction control unit 25 determines that there is no mistranslation chunk specification, and control shifts to RETURN, and shifts to processing of the next spoken text (step S1) (step S6, No).
 なお、この場合、ステップS1において、第2の端末21から先の発話テキスト(日本語テキスト)に応答する発話テキスト(英語テキスト)を受信するように制御がなされるが、引き続き第1の端末11から次の新たな発話テキスト(日本語テキスト)を受信するように制御がなされてもよい。 In this case, in step S1, the second terminal 21 is controlled to receive the spoken text (English text) in response to the previous spoken text (Japanese text), but the first terminal 11 Control may be performed to receive the next new spoken text (Japanese text) from.
 一方、英語話者2が翻訳チャンク中に誤訳チャンクが有ると判断した場合、英語話者2の以下の操作によって誤訳チャンクの指定(識別子)がデータ信号中に含まれる(ステップS6、Yes)。 On the other hand, if the English speaker 2 determines that there is a mistranslated chunk among the translated chunks, the designation (identifier) of the mistranslated chunk is included in the data signal by the following operation by the English speaker 2 (step S6, Yes).
 具体的には、図3Cは、英語話者2による誤訳チャンクの指定を受け付けたときの第2の端末21の表示を示している。すなわち、翻訳チャンク‘ASU’が誤訳であるとして誤訳指定を受け付けたときの表示を示している。 Specifically, FIG. 3C shows the display on the second terminal 21 when the designation of mistranslated chunks by the English speaker 2 is accepted. In other words, this shows the display when the translation chunk 'ASU' is mistranslated and a mistranslation designation is accepted.
 すなわち、’ASU’ のテキストウィジェット21Wについてはハイライト表示され(指定表示)、’I go to Tokyo’ のテキストウィジェット21Wにはハイライトが付されていない(基準表示)場合を示している。なお、テキストウィジェット21Wのタッチを連打すると、ハイライト無し、ハイライト有り、ハイライト無し、と交互に変更される。 In other words, the text widget 21W for 'ASU' is highlighted (designated display), and the text widget 21W for 'I go to Tokyo' is not highlighted (standard display). Note that when the text widget 21W is repeatedly touched, the display is alternately changed to no highlighting, highlighting, and no highlighting.
 あるいは、’ASU’ のテキストウィジェット21Wの文字又は枠線を赤色(指定表示)、’I go to Tokyo’を黒色(基準表示)で表示するなど、文字色又は枠線色等を変更すること、又は文字の点滅などにより誤訳指定が視認できるようになっていればよい。あるいは、誤訳指定されたテキストウィジェットが確認できるよう音声などによって報知されるように構成されていてもよい。 Alternatively, change the text color or frame line color, etc., such as displaying the text or frame line of the text widget 21W for 'ASU' in red (specified display) and 'I go to Tokyo' in black (standard display), Alternatively, it is sufficient if the mistranslation designation can be visually recognized by blinking characters or the like. Alternatively, the system may be configured to notify the user by audio or the like so that the text widget designated as an incorrect translation can be confirmed.
 ステップS5において、英語話者2が図3CのSENDボタンを押すと、テキストウィジェット21Wの状態を示す下記のデータ信号DS4が生成され、第2の端末制御器22を介して訂正管理器13に送られる(図1)(ステップS7)。 In step S5, when the English speaker 2 presses the SEND button in FIG. (FIG. 1) (step S7).
 データ信号DS4:{‘id’:10, ‘eng’:[‘ASU’,’I go to Tokyo’], ‘error’:[0]} Data signal DS4: {‘id’:10, ‘eng’:[‘ASU’,’I go to Tokyo’], ‘error’:[0]}
 ここで、指定表示された(タッチされた)チャンク’ASU’にはチャンクID=0が割り当てられ、データ信号DS4には、誤訳である翻訳チャンク(誤訳チャンク)を指定する識別子’error’:[0]が追記されている。 Here, chunk ID = 0 is assigned to the specified displayed (touched) chunk 'ASU', and the data signal DS4 has an identifier 'error':[ 0] has been added.
 すなわち、第2の端末21に入力された誤訳チャンクを指定する信号(識別信号)に応じて識別子(誤訳識別子ともいう。)が生成され、誤り訂正制御部25は、当該識別子に基づいて、誤訳チャンクを他の翻訳チャンク(すなわち、正しく翻訳されたチャンク)とは異なる態様で第1の端末11に表示される。 That is, an identifier (also referred to as a mistranslation identifier) is generated according to a signal (identification signal) specifying a mistranslation chunk inputted to the second terminal 21, and the error correction control unit 25 corrects the mistranslation based on the identifier. The chunk is displayed on the first terminal 11 in a manner different from other translated chunks (ie correctly translated chunks).
 なお、チャンクIDである[j]は、例えば、発話テキスト(原言語、日本語テキスト)のチャンク順にj=0,1,2・・・が割り当てられる。また、識別子‘error’:[0]におけるチャンクIDである[k]は、原言語(日本語)のチャンクID[j]に対応する翻訳チャンクのIDが割り当てられる。 Note that the chunk ID [j] is assigned, for example, j=0, 1, 2, etc. in the chunk order of the uttered text (original language, Japanese text). Furthermore, the chunk ID [k] in the identifier 'error':[0] is assigned the ID of the translated chunk corresponding to the source language (Japanese) chunk ID [j].
  また、発話テキストの各原語チャンクをchk[j](j=1, 2, ・jをchk1, chk2,・として表したとき、翻訳関数は、translate(chk1, chk2, ・ = translate(chk1) + translate(chk2) + ・・・で表される。 Also, when each source language chunk of the spoken text is expressed as chk[j] (j=1, 2, ·j as chk1, chk2,·, the translation function is translate(chk1, chk2, · = translate(chk1) + It is expressed as translate(chk2) + ....
 すなわち、原語チャンクchk[j]は、チャンク順に逐次翻訳される。したがって、受話者を待たせることなく翻訳チャンクを出力することができる。なお、この際、1のチャンク毎に、又は連続する2以上のチャンクがチャンク順に逐次翻訳されてもよい。すなわち、少なくとも1のチャンク毎に逐次的に翻訳されてもよい。 That is, the original language chunk chk[j] is translated one by one in chunk order. Therefore, translation chunks can be output without making the listener wait. Note that at this time, one chunk at a time, or two or more consecutive chunks may be translated one by one in chunk order. That is, at least one chunk may be translated sequentially.
 訂正管理器13は、データ信号DS4を受信して、‘eng’: [‘ASU’,’I go to Tokyo’]の要素(翻訳チャンク)を取り出し、下記のデータ信号DS4Aを生成する。データ信号DS4Aは、第2の端末21の表示画面(タッチ画面)に表示されていた英文の内容である。 The correction manager 13 receives the data signal DS4, extracts the element (translation chunk) of 'eng': ['ASU','I go to Tokyo'], and generates the following data signal DS4A. The data signal DS4A is the content of the English text displayed on the display screen (touch screen) of the second terminal 21.
 データ信号DS4A:‘eng’: [‘ASU’,’I go to Tokyo’] Data signal DS4A: ‘eng’: [‘ASU’,’I go to Tokyo’]
 次に、翻訳機15により日本語に戻すため、以下の手順で翻訳を行い、下記のデータ信号DS5を生成する。翻訳機15は、発話者(日本語話者)テキストの翻訳結果(英語)を逆翻訳して逆翻訳チャンク(日本語)を生成する逆翻訳機能を有している。 Next, in order to convert the text back into Japanese using the translator 15, translation is performed using the following procedure to generate the following data signal DS5. The translator 15 has a back translation function that back translates the translation result (English) of the speaker's (Japanese speaker) text and generates a back translation chunk (Japanese).
 すなわち、発話者テキスト(日本語)の翻訳結果(英語)を逆翻訳した逆翻訳チャンク(日本語)が発話者の第1の端末11に戻される。
translate(‘eng’: [‘ASU’,’I go to Tokyo’], ‘ej’)
=translate(‘eng’: [‘ASU’], ‘ej) + translate(‘eng’: [‘I go to Tokyo’]‘ej’)
=’jap’:[‘ASU’] + ‘jap’:[‘東京に行きます’]
=’jap’: [‘ASU’, ‘東京に行きます’] 
 データ信号DS5:{‘id’:10, ‘jap’: [’'ASU’,’東京に行きます’], ‘error’: [0]} 
That is, a back-translation chunk (Japanese) obtained by back-translating the translation result (English) of the speaker's text (Japanese) is returned to the speaker's first terminal 11.
translate('eng': ['ASU','I go to Tokyo'], 'ej')
= translate('eng': ['ASU'], 'ej) + translate('eng': ['I go to Tokyo']'ej')
='jap':['ASU'] + 'jap':['I'm going to Tokyo']
='jap': ['ASU', 'I'm going to Tokyo']
Data signal DS5: {'id':10, 'jap': [''ASU','I'm going to Tokyo'], 'error': [0]}
 第1の端末制御器12は、データ信号DS5中のエラー(誤訳)を示す識別子‘error’:[0]及びそのチャンクID=[0]を参照し、逆翻訳結果‘jap’: [’ASU’,’東京に行きます’]の‘ASU’がエラーであることを検出する。 The first terminal controller 12 refers to the identifier 'error':[0] indicating an error (mistranslation) and its chunk ID=[0] in the data signal DS5, and refers to the back translation result 'jap': ['ASU Detects that 'ASU' in ','I'm going to Tokyo'] is an error.
 第1の端末11は、第1の端末制御器12の検出結果に基づいて、図3Dに示すように、誤訳チャンクのウィジェット11W(‘ASU’)にハイライトを付して表示する(ステップS8)。なお、上記したように、誤訳チャンク‘ASU’ の文字色又は枠線色等を基準表示とは異なる色に変更して表示してもよい。 The first terminal 11 displays the mistranslation chunk widget 11W ('ASU') with a highlight, as shown in FIG. 3D, based on the detection result of the first terminal controller 12 (step S8 ). Note that, as described above, the font color or frame color of the mistranslated chunk 'ASU' may be changed to a different color from the standard display.
 また、この際、図3Dに示すように、第1の端末11には、翻訳機15により翻訳された英語チャンクが、翻訳機15によって逆翻訳されて表示される。すなわち、第1の端末11には、発話者テキスト(日本語)の翻訳(英語)が日本語に逆翻訳された結果(逆翻訳チャンク)である‘東京に行きます’が表示される。したがって、発話者は、正しく翻訳されたか否かを容易に判断することができる。 Also, at this time, as shown in FIG. 3D, the English chunk translated by the translator 15 is back-translated by the translator 15 and displayed on the first terminal 11. That is, the first terminal 11 displays 'I'm going to Tokyo', which is the result (back translation chunk) of the translation (English) of the speaker's text (Japanese) back translated into Japanese. Therefore, the speaker can easily determine whether or not the translation has been correct.
 なお、図3Eに示すように、翻訳機15により翻訳された英語チャンク(翻訳チャンク)が第1の端末11にそのまま表示されてもよい。 Note that, as shown in FIG. 3E, the English chunk (translation chunk) translated by the translator 15 may be displayed as is on the first terminal 11.
 したがって、日本語話者1は、発話者は誤訳又は意味不明な翻訳(「誤訳)と総称する。)があることをハイライト表示等の指定表示によって確認することができる。さらに、発話内容のうち ‘あす’に対応するチャンク‘ASU’が誤訳又は意味不明な翻訳であるチャンク(誤訳チャンク)として英語話者2が指定したことを知ることができる。 Therefore, the Japanese speaker 1 can confirm that there is a mistranslation or meaningless translation (collectively referred to as "mistranslation") by highlighting or other specified display. It can be seen that the English speaker 2 has designated the chunk 'ASU' corresponding to 'tomorrow' as a chunk that is a mistranslation or a meaningless translation (mistranslation chunk).
 日本語話者1は、英語話者2によって指定されたチャンクに対応する日本語チャンクを言い換え、第1の端末11に言い換えテキストを入力する(ステップS9)。言い換えテキストは上記したステップS1以降のフローを経て翻訳され、当該言い換えテキストの翻訳チャンクが第2の端末21に表示される(ステップS5)。 The Japanese speaker 1 paraphrases the Japanese chunk corresponding to the chunk specified by the English speaker 2, and inputs the paraphrased text into the first terminal 11 (step S9). The paraphrased text is translated through the flow from step S1 described above, and translated chunks of the paraphrased text are displayed on the second terminal 21 (step S5).
 なお、当該フローにおいて、分割テキストに含まれる全てのチャンクが翻訳されてもよいが、当該言い換えられたチャンクのみが翻訳され、他の翻訳チャンクについては前回の翻訳チャンクが用いられてもよい。翻訳容量の削減、翻訳速度の向上の点で利点を有する。 Note that in this flow, all chunks included in the divided text may be translated, but only the paraphrased chunk may be translated, and the previous translated chunk may be used for the other translated chunks. It has advantages in terms of reducing translation capacity and improving translation speed.
 すなわち、日本語話者1が、英語話者2によって指定されたチャンクに対応する日本語部分を別の表現で発話したり、より簡単な表現を使うなど、言い換えることによって正確な翻訳テキストを英語話者2に伝えることができる。 In other words, Japanese speaker 1 can translate the Japanese part corresponding to the chunk specified by English speaker 2 into English by saying it in a different way, using a simpler expression, etc. can be conveyed to speaker 2.
 例えば、日本語話者1は、‘あす’を‘あした’と言い換えることによって正確な翻訳テキストを英語話者2に伝えることができる。 For example, the Japanese speaker 1 can convey the accurate translated text to the English speaker 2 by rephrasing 'Asu' as 'Ashita'.
 図3Fは、言い換えテキストが翻訳されて表示された第2の端末21を示す図である。今度は正確に翻訳されたのでSENDボタンをタッチすることで、この発話テキスト‘あす東京に行きます’の翻訳が完了する。 FIG. 3F is a diagram showing the second terminal 21 on which the paraphrased text is translated and displayed. This time, the translation is accurate, so by touching the SEND button, the translation of this spoken text, ``I'm going to Tokyo tomorrow'', is completed.
 以上説明したように、翻訳誤りが原言語側に確実にフィードバックされるので、誤り訂正が効率的で正確な翻訳を行うことができ、確実に会話を行うことができる。 As explained above, since translation errors are reliably fed back to the source language side, error correction is efficient, accurate translation can be performed, and conversation can be carried out reliably.
 なお、図3Dに示す第1の端末11の表示の改変例である表示を図3Gに示す。すなわち、第1の端末11には、訂正管理器13から第1の端末制御器12を介してデータ信号DS5とともにデータ信号DS2が送られ、第1の端末11には、エラー‘ASU’ とともに日本語話者1の発話内容‘あす’が付加されて表示されるようにしてもよい。 Note that FIG. 3G shows a display that is a modified example of the display of the first terminal 11 shown in FIG. 3D. That is, the data signal DS2 is sent from the correction manager 13 to the first terminal 11 along with the data signal DS5 via the first terminal controller 12, and the first terminal 11 is sent the data signal DS2 together with the error 'ASU'. The utterance content 'tomorrow' of speaker 1 may be added and displayed.
 この場合、意味不明な翻訳チャンク(‘ASU’) とともに、これに対応する発話チャンク(‘あす’)が第1の端末11に明示されるので、日本語話者1の発話内容である‘あす’が‘ASU’に変換されて、英語話者2が意味不明であると判断したということが簡潔かつ明確に表示される。したがって、正確な翻訳により円滑な会話に役立つ。 In this case, along with the meaningless translation chunk ('ASU'), the corresponding utterance chunk ('Asu') is displayed on the first terminal 11, so 'ASU', which is the utterance content of the Japanese speaker 1, is displayed on the first terminal 11. ' is converted to 'ASU', and it is clearly and concisely displayed that the English speaker 2 has judged the sentence to be meaningless. Therefore, accurate translation helps in smooth conversation.
 すなわち、チャンクを単位とした誤り訂正により、誤りの箇所を局所的に限定して、局所的に訂正することができるので、誤り訂正会話が効率的になり、同時通訳における会話の質を向上することができる。 In other words, by correcting errors in chunks, errors can be localized and corrected locally, making error correction conversations more efficient and improving the quality of conversations during simultaneous interpretation. be able to.
[第2の実施形態]
 図7は、本発明の第2の実施形態による誤り訂正翻訳装置30の構成を示すブロック図である。
[Second embodiment]
FIG. 7 is a block diagram showing the configuration of an error correction translation device 30 according to the second embodiment of the present invention.
 本実施形態における誤り訂正翻訳装置30においては、キーボード11A及びキーボード21Aに代わって、音声認識部11B及び音声認識部21Bがそれぞれ第1の端末11及び第2の端末21に設けられている点において、第1の実施形態による誤り訂正翻訳装置10とは異なっている。 In the error correction translation device 30 according to the present embodiment, a speech recognition section 11B and a speech recognition section 21B are provided in the first terminal 11 and the second terminal 21, respectively, in place of the keyboard 11A and the keyboard 21A. , is different from the error correction translation device 10 according to the first embodiment.
 より詳細には、誤り訂正翻訳装置30において、第1の端末11は、日本語話者1の発話を音声認識し、テキストデータとして出力する音声認識部11Bを有する。また、第2の端末21は、英語話者2の発話を音声認識し、テキストデータとして出力する音声認識部21Bを有する。 More specifically, in the error correction translation device 30, the first terminal 11 includes a speech recognition unit 11B that performs speech recognition on the utterance of the Japanese speaker 1 and outputs it as text data. The second terminal 21 also includes a speech recognition unit 21B that performs speech recognition on the utterances of the English speaker 2 and outputs the speech as text data.
 日本語話者1が発話すると、第1の端末11の音声認識部11Bは、当該発話の音声認識を行い、音声認識結果であるテキストデータを第1の端末制御器12に送る。 When the Japanese speaker 1 speaks, the voice recognition unit 11B of the first terminal 11 performs voice recognition of the utterance and sends text data that is the voice recognition result to the first terminal controller 12.
 同様に、第2の端末21は、英語話者2による入力を音声によって受け付け、音声認識した結果を第2の端末制御器22に送る。 Similarly, the second terminal 21 receives voice input from the English speaker 2, and sends the voice recognition result to the second terminal controller 22.
 音声認識部11B及び音声認識部21Bは、音声を符号化してデータに変換できる音声認識エンジンを有している。例えば、音声認識法として公知の技術であるWAVやMP3などの音声を符号化する方法を用いることができる。 The speech recognition unit 11B and the speech recognition unit 21B have a speech recognition engine that can encode speech and convert it into data. For example, a method of encoding audio such as WAV or MP3, which is a known technology as a speech recognition method, can be used.
 なお、本実施形態の誤り訂正翻訳装置30においても第1の実施形態の場合と同様に誤訳チャンクの指定が行われる。すなわち、誤訳又は意味不明な翻訳があることを指定する識別子が生成され、当該識別子に基づいて、誤訳チャンクを他の翻訳チャンクとは異なる態様で発話者側の端末に表示させる。 Note that in the error correction translation device 30 of this embodiment as well, mistranslation chunks are specified in the same way as in the first embodiment. That is, an identifier is generated that specifies that there is a mistranslation or a meaningless translation, and based on the identifier, the mistranslation chunk is displayed on the speaker's terminal in a manner different from other translation chunks.
 本実施形態においては、音声認識部11B又は音声認識部21Bによる音声認識の誤りに起因して意味不明な又は不明瞭な翻訳チャンクが生成された場合には、当該翻訳チャンクも誤訳チャンクとして処理される。 In this embodiment, if a meaningless or unclear translation chunk is generated due to an error in speech recognition by the speech recognition unit 11B or the speech recognition unit 21B, the translation chunk is also processed as a mistranslation chunk. Ru.
 なお、第1の端末11及び第2の端末21のいずれか一方が音声認識によりテキストデータが入力され、他方の端末がキーボードによりテキストデータが入力されるように構成されていてもよい。また、第1の端末11及び第2の端末21がキーボード入力及び音声認識によるテキストデータ入力の機能を有していてもよい。 Note that either one of the first terminal 11 and the second terminal 21 may be configured so that text data is input through voice recognition, and the other terminal is configured such that text data is input through a keyboard. Further, the first terminal 11 and the second terminal 21 may have a function of inputting text data by keyboard input and voice recognition.
[第3の実施形態]
 図8は、本発明の第3の実施形態による誤り訂正翻訳装置50の構成を示すブロック図である。
[Third embodiment]
FIG. 8 is a block diagram showing the configuration of an error correction translation device 50 according to the third embodiment of the present invention.
 本実施形態における誤り訂正翻訳装置50においては、第1の実施形態による誤り訂正翻訳装置10の日本語(第1の言語)から英語(第2の言語)への翻訳機能に加え、英語(第2の言語)から日本語(第1の言語)への翻訳機能が設けられている。すなわち、誤り訂正翻訳装置50は、双方向に同時翻訳が可能な翻訳装置として構成されている。 The error correction translation device 50 according to the present embodiment has the translation function from Japanese (first language) to English (second language) of the error correction translation device 10 according to the first embodiment. A translation function is provided from Japanese (second language) to Japanese (first language). That is, the error correction translation device 50 is configured as a translation device capable of bidirectional simultaneous translation.
 なお、当該第1の言語及び当該第2の言語は日本語及び英語に限定されず、他の言語に対しても適用することができる。 Note that the first language and the second language are not limited to Japanese and English, and can be applied to other languages as well.
 具体的には、テキスト分割器54は、日英翻訳におけるデータ信号DS1に加えて英日翻訳におけるデータ信号ES1に含まれる発話テキストデータをチャンクに分割する。 Specifically, the text divider 54 divides the spoken text data included in the data signal DS1 in the Japanese-English translation as well as the data signal ES1 in the English-Japanese translation into chunks.
 また、翻訳機55は、日英/英日翻訳機として構成されている。また、訂正管理器53は、日英翻訳時のデータ信号DS1~DS5及び英日翻訳時のデータ信号ES1~ES5のデータ交換を管理するように構成されている。 Furthermore, the translator 55 is configured as a Japanese-English/English-Japanese translator. Further, the correction manager 53 is configured to manage data exchange of data signals DS1 to DS5 during Japanese-English translation and data signals ES1 to ES5 during English-Japanese translation.
 さらに、訂正管理器53には、1の発話テキストの誤り訂正が完了したか否かを判別する誤り訂正完了判別部53Aが設けられている。誤り訂正完了判別部53Aは、当該1の発話テキストの誤り訂正が完了したか否かを、誤訳又は意味不明な翻訳があることを指定する識別子の有無によって判別する。 Further, the correction manager 53 is provided with an error correction completion determination unit 53A that determines whether error correction of one utterance text is completed. The error correction completion determining unit 53A determines whether error correction of the one utterance text is completed based on the presence or absence of an identifier that specifies that there is a mistranslation or a meaningless translation.
 例えば、双方向翻訳の場合、発話者(日本語話者1)から受話者(英語話者2)への発話テキストの誤り訂正においては、第2の端末21から入力された識別子(第1の識別子)に基づいて当該発話テキストの誤り訂正が完了したか否かが判別され、発話者(英語話者2)から受話者(日本語話者1)への発話テキストの誤り訂正においては、第1の端末11から入力された識別子(第2の識別子)に基づいて当該発話テキストの誤り訂正が完了したか否かが判別され
 あるいは、誤り訂正完了判別部53Aは、第2の端末21又は第1の端末11に翻訳チャンクの各々が表示されてから、所定の経過時間が経過したとき、当該1の発話テキストの誤り訂正が完了したと判別してもよい。翻訳チャンク中に誤訳チャンクが含まれている場合であっても、正しい意味を理解できる場合があるからである。この場合、誤り訂正完了判別部53Aは当該経過時間を計時するタイマを有している。
For example, in the case of bidirectional translation, when correcting errors in the uttered text from the speaker (Japanese speaker 1) to the receiver (English speaker 2), the identifier input from the second terminal 21 (first It is determined whether the error correction of the utterance text is completed based on the utterance identifier), and in the error correction of the utterance text from the speaker (English speaker 2) to the receiver (Japanese speaker 1), Based on the identifier (second identifier) input from the second terminal 11, it is determined whether or not the error correction of the uttered text has been completed. When a predetermined elapsed time has elapsed since each of the translation chunks was displayed on the terminal 11 of the terminal 11, it may be determined that the error correction of the uttered text of the one has been completed. This is because even if a mistranslated chunk is included in the translated chunk, the correct meaning may be understood. In this case, the error correction completion determination unit 53A has a timer that measures the elapsed time.
 あるいは、誤訳チャンクの指定入力が行われず、単にSENDボタン(又は、RETURNキー、ENTERキー)による入力があった場合に、誤り訂正完了判別部53Aが当該発話テキストの誤り訂正が完了したことを判別してもよい。 Alternatively, if the mistranslation chunk is not specified and input is simply made using the SEND button (or RETURN key or ENTER key), the error correction completion determination unit 53A determines that error correction of the utterance text has been completed. You may.
 すなわち、図2のステップS6等に関して説明したとおり、訂正管理器53の送受信信号中に当該誤訳識別子が有る場合には、入力された当該1の発話テキストについての誤り訂正がまだ完了しておらず、当該送受信信号は誤り訂正のための転送データ信号であると判別する。この場合、訂正管理器53は次の新たな発話テキストの入力を禁止し、誤り訂正の完了を優先するように機能する。 In other words, as explained in relation to step S6 etc. in FIG. 2, if the mistranslation identifier is present in the transmission/reception signal of the correction manager 53, it means that the error correction for the inputted utterance text has not been completed yet. , the transmitted/received signal is determined to be a transfer data signal for error correction. In this case, the correction manager 53 functions to prohibit input of the next new utterance text and prioritize completion of error correction.
 ここで、英語(第2の話者)から日本語(第1の話者)へ翻訳する場合、英日翻訳時のデータ信号ES1~ES5は日英翻訳時のデータ信号DS1~DS5にそれぞれ対応するデータ信号であり、データ信号DS1~DS5とは発話者の言語及び翻訳方向が異なる点を除いて等価な信号として構成されている。具体的には、上記した場合と同様に、下記のようなデータ信号ES1~ES5を例示することができる。
・データ信号ES1:{‘id’:11, ‘eng’:[‘I’m heading to Boston tomorrow’]}
・データ信号ES2:{‘id’:11,‘eng’:[‘I’m heading to’,‘Boston’,‘tomorrow’]}
・データ信号ES3:{‘id’:11,‘jap’: [‘私は行きます’,‘ボストンに’,‘明日’]}
・データ信号ES4:{‘id’:11,‘jap’: [‘私は行きます’,‘ボストンに’,‘明日’],‘error’:[1]}
・データ信号ES5:{‘id’:11,‘eng’: [‘I’m heading to’,‘BOSTON’,‘tomorrow’],‘error’: [1]}
Here, when translating from English (second speaker) to Japanese (first speaker), data signals ES1 to ES5 during English-Japanese translation correspond to data signals DS1 to DS5 during Japanese-English translation, respectively. This data signal is configured as an equivalent signal to the data signals DS1 to DS5 except that the language of the speaker and the direction of translation are different. Specifically, similar to the case described above, the following data signals ES1 to ES5 can be exemplified.
・Data signal ES1: {'id':11, 'eng': ['I'm heading to Boston tomorrow']}
・Data signal ES2: {'id':11,'eng': ['I'm heading to','Boston','tomorrow']}
・Data signal ES3: {'id':11,'jap': ['I'm going','To Boston','Tomorrow']}
・Data signal ES4: {'id':11,'jap': ['I'm going','To Boston','Tomorrow'],'error':[1]}
・Data signal ES5: {'id':11,'eng': ['I'm heading to','BOSTON','tomorrow'],'error': [1]}
 この場合、受話者(日本語話者1)は、‘ボストンに’ が意味不明であるとして指定し、チャンクID[1] を含む識別子‘error’:[1] (第2の識別子)が付されたデータ信号ES4が発話者(英語話者2)に送られる。 In this case, the recipient (Japanese speaker 1) specifies 'To Boston' as meaningless and adds the identifier 'error':[1] (second identifier) containing chunk ID[1]. The resulting data signal ES4 is sent to the speaker (English speaker 2).
 発話者(英語話者2)の第2の端末21には、 データ信号ES5中の識別子によって、‘BOSTON’が意味不明な翻訳(誤訳)であることを示す指定表示として他の翻訳チャンクとは異なる態様(例えば、大文字)で表示され、英語話者2は、受話者(日本語話者1)にとって発話チャンクのうちの‘Boston’が意味不明であることを確認することができる。 The second terminal 21 of the speaker (English speaker 2) receives an identifier in the data signal ES5 as a designated indication that 'BOSTON' is a meaningless translation (mistranslation), unlike other translation chunks. It is displayed in a different manner (for example, in uppercase letters), and the English speaker 2 can confirm that 'Boston' in the speech chunk is meaningless to the listener (Japanese speaker 1).
 なお、翻訳機55は、文脈判断機能を有している。この例の場合では、翻訳機55は、それまでの日本語話者1及び英語話者2の間の会話内容の文脈から‘Boston’が「ボストン図書館」を意味することを判断して翻訳すべきところを「ボストン」と翻訳した誤訳である。 Note that the translator 55 has a context determination function. In this example, the translator 55 determines that 'Boston' means 'Boston Library' from the context of the conversation between Japanese speaker 1 and English speaker 2, and translates it. This is a mistranslation of "Boston" when it should have been translated.
 従って、この場合では、英語話者2が‘BOSTON’ を‘the Boston Library’に言い換えることによって正確な翻訳テキストを日本語話者1に伝えることができる。 Therefore, in this case, the English speaker 2 can convey the accurate translated text to the Japanese speaker 1 by rephrasing 'BOSTON' to 'the Boston Library'.
 本実施形態の場合では、第1言語の話者及び第2言語の話者間で双方向に同時翻訳が可能である。したがって、互いに、意味不明又は誤訳があることを指定表示によって確認することができ、正確な翻訳により円滑に会話を行うことができる。 In the case of this embodiment, bidirectional simultaneous translation is possible between speakers of the first language and speakers of the second language. Therefore, both parties can confirm whether the meaning is unclear or there is a mistranslation by the designated display, and conversation can be smoothly carried out through accurate translation.
 本発明によれば、発話の単位を翻訳として意味のあるチャンクIDで管理することにより、チャンクIDに対して原言語の発話情報と目的言語の翻訳情報を紐づけることにより、誤りを原言語側に確実にフィードバックすることができる。 According to the present invention, by managing units of utterances with meaningful chunk IDs as translations, and by linking utterance information in the source language and translation information in the target language to the chunk IDs, errors can be avoided on the source language side. can provide reliable feedback.
 また、本発明のチャンクを単位とした誤り訂正により、誤りの箇所を局所的に限定して、局所的に訂正することができるので、誤り訂正会話が効率的になり、同時通訳における会話の質を向上することができる。 In addition, by using the chunk-based error correction of the present invention, errors can be localized and corrected locally, making error correction conversations more efficient and improving the quality of conversations in simultaneous interpretation. can be improved.
 なお、上記した例では、誤り訂正翻訳装置にテキストとして1つの文が入力された場合について説明したが、複数の文からなるテキストが入力され、当該入力テキストに対して翻訳が行われるように構成されていてもよい。 Note that in the above example, a case was explained in which one sentence was input as text to the error correction translation device, but it is also configured such that text consisting of multiple sentences is input and translation is performed on the input text. may have been done.
 この場合、例えば、第1の端末制御器12及び第2の端末制御器22が第1の端末11及び第2の端末21から送信された’SEND‘、’Enter‘等の入力確定を示す信号に基づいて当該テキストの終わりを判別することができる。 In this case, for example, the first terminal controller 12 and the second terminal controller 22 receive signals indicating input confirmation such as 'SEND' and 'Enter' transmitted from the first terminal 11 and the second terminal 21. The end of the text can be determined based on .
 また、データ信号DS1~DS5等のデータの送受信によりデータ交換を行う場合についてJSONフォーマットを用いた場合を例に説明したが、XML(Extensible Markup Language)等を用いることもできる。 Furthermore, although the case where data is exchanged by sending and receiving data such as data signals DS1 to DS5 is explained using the JSON format as an example, XML (Extensible Markup Language) or the like may also be used.
 また、上記した誤り訂正翻訳装置はCPU(Central Processing Unit),GPU(Graphics Processing Unit)等、あるいはサーバとして構成されていてもよい。また、コンピュータ、情報端末及び携帯電話等に実装されていてもよい。 Furthermore, the error correction translation device described above may be configured as a CPU (Central Processing Unit), a GPU (Graphics Processing Unit), or a server. Furthermore, it may be implemented in computers, information terminals, mobile phones, and the like.
 上記した誤り訂正翻訳装置及び誤り訂正翻訳方法はハードウエアとして構成されている場合を例に説明したが、コンピュータのソフトウエア(プログラム)として構成されていてもよい。あるいは、当該プログラムが格納されたコンピュータ読み取り可能な格納媒体として構成されていてもよい。 Although the above-described error correction translation device and error correction translation method are configured as hardware, they may be configured as computer software (program). Alternatively, the program may be configured as a computer-readable storage medium in which the program is stored.
 例えば、第1及び第2の端末制御器、テキスト分割器、翻訳機、音声認識部及び誤り訂正制御部のいずれかはソフトウエアとして構成されていてもよい。あるいは、これらが1つのソフトウエアとして構成されていてもよい。 For example, any one of the first and second terminal controllers, text divider, translator, speech recognition unit, and error correction control unit may be configured as software. Alternatively, these may be configured as one software.
 以上、本発明の実施形態について詳細に説明したが、本発明の趣旨を逸脱しない範囲で種々の改変を行うことが可能である。 Although the embodiments of the present invention have been described in detail above, various modifications can be made without departing from the spirit of the present invention.
 10,30,50:誤り訂正翻訳装置、11:第1の端末、11A,21A:キーボード、11B,21B:音声認識部、11W,21W:テキストウィジェット、12:第1の端末制御器、13,53:分割テキスト訂正管理器、53A:誤り訂正完了判別部、14,54:テキスト分割器、15,55:翻訳機、21:第2の端末、22:第2の端末制御器、25:誤り訂正制御部、53A 10, 30, 50: error correction translation device, 11: first terminal, 11A, 21A: keyboard, 11B, 21B: speech recognition unit, 11W, 21W: text widget, 12: first terminal controller, 13, 53: Divided text correction manager, 53A: Error correction completion determination unit, 14, 54: Text divider, 15, 55: Translator, 21: Second terminal, 22: Second terminal controller, 25: Error Correction control unit, 53A

Claims (14)

  1.  第1言語のテキストが入力される第1の端末と、
     前記第1言語のテキストをチャンクに分割して分割テキストを生成するテキスト分割器と、
     前記分割テキストに含まれるチャンクの各々を個別に翻訳して第2言語の翻訳チャンクを生成する翻訳機と、
     第2の端末と、
     前記翻訳機からの前記翻訳チャンクの各々を前記第2の端末に個別に表示させる誤り訂正制御部と、を有し、
     前記第2の端末は、前記翻訳チャンクのうちの誤訳チャンクを指定する識別子の入力を受け付け、
     前記誤り訂正制御部は、前記第2の端末に入力された前記識別子に基づいて、前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第1の端末に表示させ、前記誤訳チャンクの表示に応答して、前記第1の端末に前記第1言語のテキストの言い換えテキストの入力を受け付けさせる、ことを特徴とする誤り訂正翻訳装置。
    a first terminal into which text in a first language is input;
    a text divider that divides the first language text into chunks to generate divided text;
    a translator that individually translates each chunk included in the divided text to generate translated chunks in a second language;
    a second terminal;
    an error correction control unit that causes each of the translation chunks from the translator to be individually displayed on the second terminal;
    The second terminal receives an input of an identifier specifying a mistranslated chunk among the translated chunks,
    The error correction control unit causes the mistranslation chunk to be displayed on the first terminal in a manner different from other translation chunks based on the identifier input to the second terminal, and the error correction control unit causes the mistranslation chunk to be displayed in a manner different from other translation chunks. An error correction translation device characterized in that, in response, the first terminal receives input of paraphrased text of the text in the first language.
  2.  前記誤り訂正制御部は、前記言い換えテキストを前記第1言語のテキストとし、前記言い換えテキストの前記第2言語の翻訳チャンクの各々を前記第2の端末に表示させる請求項1に記載の誤り訂正翻訳装置。 The error correction translation according to claim 1, wherein the error correction control unit makes the paraphrase text the text in the first language, and displays each of the second language translation chunks of the paraphrase text on the second terminal. Device.
  3.  前記翻訳機は、前記第2言語の前記翻訳チャンクを逆翻訳して逆翻訳チャンクを生成し、
     前記誤り訂正制御部は、前記逆翻訳チャンクを前記第1の端末に表示させる、請求項1に記載の誤り訂正翻訳装置。
    the translator back-translates the translated chunks in the second language to generate back-translated chunks;
    The error correction translation device according to claim 1, wherein the error correction control unit causes the back translation chunk to be displayed on the first terminal.
  4.  前記第2の端末はタッチパネル機能を有するディスプレイを有し、
     前記翻訳チャンク及び前記誤訳チャンクは、テキストウィジェットとして前記ディスプレイに表示され、
     前記第2の端末は前記識別子の入力を前記ディスプレイへのタッチによって受け付ける、請求項1に記載の誤り訂正翻訳装置。
    The second terminal has a display with a touch panel function,
    the translated chunk and the mistranslated chunk are displayed on the display as a text widget;
    The error correction translation device according to claim 1, wherein the second terminal receives the input of the identifier by touching the display.
  5.  前記テキスト分割器は形態素解析を行って前記第1言語のテキストを前記チャンクに分割する請求項1に記載の誤り訂正翻訳装置。 The error correction translation device according to claim 1, wherein the text divider performs morphological analysis to divide the text in the first language into the chunks.
  6.  前記テキスト分割器は深層学習により前記チャンクの境界を決定する請求項5に記載の誤り訂正翻訳装置。 The error correction translation device according to claim 5, wherein the text divider determines boundaries of the chunks by deep learning.
  7.  前記第1の端末は、発話された音声の認識を行い、前記第1言語のテキストを生成する音声認識部を有する請求項1に記載の誤り訂正翻訳装置。 The error correction translation device according to claim 1, wherein the first terminal includes a speech recognition unit that recognizes spoken speech and generates text in the first language.
  8.  前記誤訳チャンクは、前記音声認識部による音声認識誤りによる翻訳誤りを含む請求項7に記載の誤り訂正翻訳装置。 The error correction translation device according to claim 7, wherein the mistranslation chunk includes a translation error due to a speech recognition error by the speech recognition unit.
  9.  第1の端末及び第2の端末と、
     前記第1の端末から入力された第1言語のテキスト及び前記第2の端末から入力された第2言語のテキストをチャンクに分割してそれぞれ第1の分割テキスト及び第2の分割テキストを生成するテキスト分割器と、
     前記第1の分割テキスト及び前記第2の分割テキストに含まれるチャンクの各々を個別に翻訳して、それぞれ前記第2言語の翻訳チャンク及び前記第1言語の翻訳チャンクを生成する翻訳機と、
     前記翻訳機からの前記第1言語の前記翻訳チャンクの各々を前記第2の端末に個別に表示させるとともに、前記翻訳機からの前記第2言語の前記翻訳チャンクの各々を前記第1の端末に個別に表示させる誤り訂正制御部と、を有し、
     前記第2の端末は、前記第1言語の前記翻訳チャンクのうちの誤訳チャンクを指定する第1の識別子の入力を受け付け、
     前記第1の端末は、前記第2言語の前記翻訳チャンクのうちの誤訳チャンクを指定する第2の識別子の入力を受け付け、
     前記誤り訂正制御部は、
     前記第2の端末に前記第1の識別子が入力されたとき、前記第1の識別子に基づいて、前記第1言語の前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第1の端末に表示させ、
     前記第1の端末に前記第2の識別子が入力されたとき、前記第2の識別子に基づいて、前記第2言語の前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第2の端末に表示させ、
     前記第1言語の前記誤訳チャンクの表示に応答して、前記第1の端末に前記第1言語のテキストの言い換えテキストの入力を受け付けさせ、
     前記第2言語の前記誤訳チャンクの表示に応答して、前記第2の端末に前記第2言語のテキストの言い換えテキストの入力を受け付けさせる、誤り訂正翻訳装置。
    a first terminal and a second terminal;
    Divide the text in a first language input from the first terminal and the text in a second language input from the second terminal into chunks to generate a first divided text and a second divided text, respectively. a text divider,
    a translation machine that individually translates each of the chunks included in the first divided text and the second divided text to generate translated chunks in the second language and translated chunks in the first language, respectively;
    individually displaying each of the translated chunks in the first language from the translator on the second terminal; and displaying each of the translated chunks in the second language from the translator on the first terminal. an error correction control unit that displays the information individually;
    The second terminal receives an input of a first identifier specifying a mistranslated chunk among the translated chunks in the first language,
    The first terminal receives an input of a second identifier specifying a mistranslated chunk among the translated chunks in the second language,
    The error correction control unit includes:
    When the first identifier is input to the second terminal, the mistranslated chunk in the first language is sent to the first terminal in a manner different from other translated chunks based on the first identifier. display,
    When the second identifier is input to the first terminal, the mistranslated chunk in the second language is transmitted to the second terminal in a manner different from other translated chunks based on the second identifier. display,
    in response to displaying the mistranslated chunk in the first language, causing the first terminal to accept input of paraphrased text of the text in the first language;
    An error correction translation device that causes the second terminal to accept input of paraphrased text of the text in the second language in response to displaying the mistranslated chunk in the second language.
  10.  前記誤り訂正制御部は、前記第1の識別子及び前記第2の識別子に基づいて、又は、前記第2の端末又は前記第1の端末に前記誤訳チャンクが表示されてからの経過時間に基づいて、前記第1の端末又は前記第2の端末から入力された1の発話テキストの誤り訂正が完了したか否かを判別する誤り訂正完了判別部を有する請求項9に記載の誤り訂正翻訳装置。 The error correction control unit is configured to perform an error correction based on the first identifier and the second identifier, or based on the elapsed time since the mistranslation chunk was displayed on the second terminal or the first terminal. 10. The error correction translation device according to claim 9, further comprising an error correction completion determination unit that determines whether error correction of one utterance text inputted from the first terminal or the second terminal is completed.
  11.  前記誤り訂正完了判別部は、前記1の発話テキストの誤り訂正が完了していないことを判別した場合、次の新たな発話テキストの入力を禁止する請求項10に記載の誤り訂正翻訳装置。 11. The error correction translation device according to claim 10, wherein the error correction completion determining unit prohibits input of a next new utterance text when determining that error correction of the first utterance text is not completed.
  12.  第1言語のテキストを第1の端末に入力するステップと、
     前記第1言語のテキストをチャンクに分割して分割テキストを生成するテキスト分割ステップと、
     前記分割テキストに含まれるチャンクの各々を個別に翻訳して第2言語の翻訳チャンクを生成する翻訳ステップと、
     前記翻訳機からの前記翻訳チャンクの各々を前記第2の端末に個別に表示させる誤り訂正制御ステップと、
     前記第2の端末において、前記翻訳チャンクのうちの誤訳チャンクを指定する識別子の入力を受け付けるステップと、
     前記第2の端末に入力された前記識別子に基づいて、前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第1の端末に表示させる誤訳チャンク表示ステップと、
     前記誤訳チャンク表示ステップの後に、前記第1の端末への、前記第1言語のテキストの言い換えテキストの入力を受け付けるステップと、
    を有する誤り訂正翻訳方法。
    inputting text in a first language into a first terminal;
    a text dividing step of dividing the first language text into chunks to generate divided text;
    a translation step of individually translating each chunk included in the divided text to generate translated chunks in a second language;
    an error correction control step of individually displaying each of the translation chunks from the translator on the second terminal;
    receiving, at the second terminal, an input of an identifier specifying a mistranslated chunk among the translated chunks;
    a mistranslation chunk displaying step of displaying the mistranslation chunk on the first terminal in a manner different from other translated chunks based on the identifier input to the second terminal;
    after the mistranslation chunk displaying step, receiving input of a paraphrase text of the text in the first language to the first terminal;
    An error correction translation method having the following.
  13.  テキストデータを翻訳するためにコンピュータを、
     第1言語のテキストが入力される第1の端末手段、
     前記第1言語のテキストをチャンクに分割して分割テキストを生成するテキスト分割手段、
     前記分割テキストに含まれるチャンクの各々を個別に翻訳して第2言語の翻訳チャンクを生成する翻訳手段と、
     第2の端末手段、
     前記翻訳手段からの前記翻訳チャンクの各々を前記第2の端末手段に個別に表示させる誤り訂正制御手段、として機能させ、
     前記第2の端末手段は、前記翻訳チャンクのうちの誤訳チャンクを指定する識別子の入力を受け付け、
     前記誤り訂正制御手段は、前記第2の端末手段に入力された前記識別子に基づいて、前記誤訳チャンクを他の翻訳チャンクとは異なる態様で前記第1の端末手段に表示させ、前記誤訳チャンクの表示に応答して、前記第1の端末手段に前記第1言語のテキストの言い換えテキストの入力を受け付けさせる、ことを特徴とする誤り訂正翻訳プログラム。
    computer to translate text data,
    a first terminal means into which text in a first language is input;
    text dividing means for dividing the first language text into chunks to generate divided text;
    Translation means for individually translating each chunk included in the divided text to generate translated chunks in a second language;
    second terminal means;
    functioning as error correction control means for individually displaying each of the translation chunks from the translation means on the second terminal means;
    The second terminal means receives an input of an identifier specifying a mistranslated chunk among the translated chunks,
    The error correction control means causes the first terminal means to display the mistranslation chunk in a manner different from other translation chunks based on the identifier input to the second terminal means, and displays the mistranslation chunk on the first terminal means in a manner different from other translation chunks. An error correction translation program characterized in that, in response to a display, the first terminal means accepts input of paraphrased text of the text in the first language.
  14.  請求項13に記載の誤り訂正翻訳プログラムを記録したコンピュータ読み取り可能な格納媒体。 A computer-readable storage medium storing the error correction translation program according to claim 13.
PCT/JP2023/027192 2022-08-26 2023-07-25 Error correcting translation device, error correcting translation method, program and storage medium for same WO2024042963A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-135277 2022-08-26
JP2022135277A JP2024031617A (en) 2022-08-26 2022-08-26 Error correction translation device, error correction translation method, program, and storage medium thereof

Publications (1)

Publication Number Publication Date
WO2024042963A1 true WO2024042963A1 (en) 2024-02-29

Family

ID=90013262

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/027192 WO2024042963A1 (en) 2022-08-26 2023-07-25 Error correcting translation device, error correcting translation method, program and storage medium for same

Country Status (2)

Country Link
JP (1) JP2024031617A (en)
WO (1) WO2024042963A1 (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008243076A (en) * 2007-03-28 2008-10-09 Toshiba Corp Interpretation device, method and program

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008243076A (en) * 2007-03-28 2008-10-09 Toshiba Corp Interpretation device, method and program

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
釜谷 聡史, 音声処理技術 音声インタフェースのトータルソリューションを提供, 東芝レビュー, 01 September 2013, vol. 68, no. 9, pp. 18-21, (KAMATANI, Satoshi. Speech Processing Technologies, Toshiba Review.), non-official translation (Providing overall solution of voice interface.) *

Also Published As

Publication number Publication date
JP2024031617A (en) 2024-03-07

Similar Documents

Publication Publication Date Title
JP6334354B2 (en) Machine translation apparatus, method and program
US8954333B2 (en) Apparatus, method, and computer program product for processing input speech
US20180144747A1 (en) Real-time caption correction by moderator
US9484034B2 (en) Voice conversation support apparatus, voice conversation support method, and computer readable medium
CN102193913B (en) Translation apparatus and translation method
CN108228574B (en) Text translation processing method and device
JP6150268B2 (en) Word registration apparatus and computer program therefor
CN108986820B (en) Method, device, electronic equipment and storage medium for speech translation
JP7199061B2 (en) translation device
CN108304389B (en) Interactive voice translation method and device
WO2018198807A1 (en) Translation device
KR20160080711A (en) Apparatus, Method and System for Translation based on Communication
KR20170055630A (en) Multilingual translation method
WO2024042963A1 (en) Error correcting translation device, error correcting translation method, program and storage medium for same
JPH04319769A (en) Interpretation system
CN116933806A (en) Concurrent translation system and concurrent translation terminal
JP6353860B2 (en) Speech translation device, speech translation method, and speech translation program
CN114330311A (en) Translation method, translation device, electronic equipment and computer readable storage medium
KR20190046124A (en) Method and apparatus for real-time automatic interpretation based on context information
JP2002279353A (en) Character recognition device, method therefor, and recording medium
JP2018195248A (en) Translation display device, computer terminal, and translation display method
US20220092277A1 (en) Translation device
JP5628485B2 (en) Translation support system and method and program thereof
JP2006309469A (en) Retrieval device, retrieval method, program, and computer-readable recording medium
JP2018005368A (en) Output mode determination system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23857079

Country of ref document: EP

Kind code of ref document: A1