WO2018186416A1 - Translation processing method, translation processing program, and recording medium - Google Patents

Translation processing method, translation processing program, and recording medium Download PDF

Info

Publication number
WO2018186416A1
WO2018186416A1 PCT/JP2018/014317 JP2018014317W WO2018186416A1 WO 2018186416 A1 WO2018186416 A1 WO 2018186416A1 JP 2018014317 W JP2018014317 W JP 2018014317W WO 2018186416 A1 WO2018186416 A1 WO 2018186416A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
language
translation
basic
voice
Prior art date
Application number
PCT/JP2018/014317
Other languages
French (fr)
Japanese (ja)
Inventor
旋造 田代
Original Assignee
旋造 田代
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 旋造 田代 filed Critical 旋造 田代
Publication of WO2018186416A1 publication Critical patent/WO2018186416A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M11/00Telephonic communication systems specially adapted for combination with other electrical systems

Definitions

  • the present invention relates to a translation processing method, a translation processing program, and a recording medium for mutual translation between a basic language and another language when communicating with each other via a telecommunication line.
  • Patent Document 1 is merely a category of simultaneous interpretation that realizes listening and reading at the time of a voice call, and has a problem of low versatility.
  • An object of the present invention is to provide a translation processing method, a translation processing program, and a recording medium that can make a variation of translation multifunctional and improve versatility.
  • a communication content receiving step for receiving at least one of speech data and text data in the basic language as transmission reference data via the input means, and starting the translation application and sending reference data received in the communication content receiving step.
  • Other language translation step for translating into speech data and text data in other languages to generate transmission data transmission step for transmitting the transmission data translated in the other language translation step, and voice data in other languages
  • Receive text data as reception reference data by receiving means A basic verbalization translation step of translating the reception reference data received in the reception step by activating the translation application into speech data and text data in a basic language, and a basic languageization translation step
  • a communication content output step for outputting the translated received data via the output means.
  • the transmission data is transmitted by a transmission means through a telecommunication line
  • the reception reference data is received by a reception means through a telecommunication line.
  • the telecommunications line is an internet line.
  • the telecommunications line is a telephone line.
  • the transmission step is executed without going through a telecommunication line
  • the receiving step is executed without going through a telecommunication line
  • the other language to be translated before executing the other language translation step, the other language to be translated based on address data including a registered name of a communication partner registered in advance in the storage means And a different language specifying step for specifying.
  • the listening application installed in the storage means in advance is started and the other language is started.
  • the transmission reference data received by the input means is text data in the basic language
  • the speech data in the other language translated in the other language translation step A transmission step of transmitting text data in the basic language received by the input means in addition to the transmission data by the transmission means.
  • the period includes an interrupt canceling step for canceling reception by the receiving means.
  • the sign language analysis application installed in the storage means in advance is activated to perform the sign language operation.
  • a basic analysis data generating step for generating basic analysis data in the basic language for generating the transmission data composed of at least one of voice data, text data, and image data in the other language, , Are provided.
  • the translation processing method described in any one of claims 1 to 6 is executed by a calculation means of a communication terminal.
  • the invention according to claim 13 is a recording medium on which a translation processing program for causing a computing means of a communication terminal to execute the translation processing method according to any one of claims 1 to 6 is recorded.
  • the input data in the basic language received via the input means can be generated with the output data after being converted into another language, and the output data can be output simultaneously.
  • output data after being converted into a basic language can be generated and output through the output means at the same time.
  • both voice and text data can be used effectively, and translation variations can be made multifunctional to improve versatility. be able to.
  • the transmission data is transmitted by the transmission means through the telecommunication line
  • the reception reference data is received by the reception means through the telecommunication line.
  • Transmission data in other languages
  • reception reference data in other languages
  • the telecommunication line is an internet line, it can be used for mutual translation using SNS or the like, and if the telecommunication line is a telephone line, a TV including a normal voice call or a video signal such as a video signal. It can be used for mutual translation using telephone.
  • the sending step is executed without going through the telecommunications line and the receiving step is run without going through the telecommunications line, it can be used for mutual translation such as face-to-face interpretation without using the telecommunications line.
  • both voice data and text data after being converted to another language are generated, and at the same time, both the data are transmitted to the other party.
  • both voice data and text data after being converted into a basic language can be generated, and at the same time, both data can be output as voice and text.
  • the other language translation step by providing another language identification step for identifying another language to be translated based on the address data including the registered name of the communication partner registered in advance in the storage means, Communication relations without having to select / designate other languages using phone numbers (country codes, etc.) or e-mail addresses (domain names, etc.) included in the address data for specifying the destination or receiver in advance.
  • the translation function can be validated by specifying the country or language before establishing.
  • the transmission reference data received by the input means is voice data in a basic language
  • a basic language for generating transmission data composed of text data in another language by starting a listening application installed in advance in the storage means By providing the basic text data generation step for generating the basic text data according to the above, it is possible to effectively use the text data of the basic language as the translation source.
  • the transmission reference data received by the input means is text data in a basic language
  • a basic language for generating a transmission data composed of voice data in another language by starting a reading application installed in the storage means in advance By providing the basic speech data generation step for generating the basic speech data according to the above, it is possible to effectively use the text data of the basic language as the translation source.
  • the transmission reference data received by the input means is text data in the basic language
  • the text data in the basic language received by the input means in addition to the transmission data by the speech data of other languages translated in the other language translation step
  • the text data of the basic language as the translation source can be effectively used as a substitute for mistranslation or memo.
  • an interrupt canceling step for canceling reception by the receiving means is provided during a period from when the input means accepts the voice data in the basic language to when the sending means sends the outgoing data in another language translated in the other language translation step.
  • the voice input of the other party can be rejected until the intentional utterance is completed, and the turbidity (overlap) of the conversation can be suppressed.
  • the sign language action is analyzed by starting a sign language analysis application installed in the storage means in advance, and then voice data / text data in another language
  • a language is used as a sign language operation by providing a translation processing method including a basic analysis data generation step for generating basic analysis data in a basic language for generating transmission data composed of at least one of image data In some cases, it can be used for mutual translation.
  • the above-described translation processing method can be executed by the computing means of the communication terminal.
  • a translation processing program for causing the calculation means of the communication terminal to execute the translation processing method described above can be easily used.
  • the translation processing method concerning one embodiment of the present invention, a translation processing program, and a system application example concerning a recording medium are shown,
  • (A) is an explanatory view of an example which installs a translation application in a communication terminal
  • (B) is It is a block block diagram of the principal part of a communication terminal. It is explanatory drawing which shows the example of simultaneous interpretation of the text data by a chat system. It is explanatory drawing of the usage example of the speech data and text data after other languages based on the speech input by a basic language.
  • (A) is an explanatory diagram of an example of using voice data and text data after other languages based on text input in a basic language
  • (B) is voice data and text data after being converted into a basic language based on voice input in another language. It is explanatory drawing of the usage example. It is a flowchart of an example of the translation routine which the control part which concerns on one embodiment of this invention performs. It is explanatory drawing of an example of the translation mode selection screen of the communication terminal which concerns on one embodiment of this invention. It is explanatory drawing which shows the example of the sign language interpretation of the text data by a chat system.
  • the communication terminal 1 causes a single communication terminal 1 having storage means in which a translation application for performing mutual translation between a basic language and another language is installed in advance.
  • the communication terminal 1 realizes a voice input unit 2 and a voice output unit 3 for realizing a telephone call by a telephone function through various communication methods of the telecommunication line NT, and a display function.
  • Display unit 4 a storage unit 5 that stores a program related to overall functions as communication terminal 1 including a translation processing program for realizing a translation processing method to be described later, and a communication terminal according to a program stored in storage unit 5
  • a control unit (CPU) 6 that processes one of various functions.
  • the communication terminal 1 shown in FIG. 1A is a smartphone, but has a computer function such as a tablet terminal or a personal computer.
  • the text data chat function E -Any terminal that can use the Mail function, SNS (Social Networking Service) function, etc. is not particularly limited.
  • the various communication methods can include mainly a wireless telephone communication line method, an Internet connection line method, a short-distance wireless communication method such as infrared ray or Bluetooth (registered trademark), and the like.
  • the voice input unit 2 allows voice input from a built-in microphone (microphone) provided below the communication terminal 1.
  • the voice input unit 2 is connected to a tag, a connector (not shown) or the like provided on the peripheral surface of the communication terminal 1 or a short-distance wireless communication connection such as Bluetooth (registered trademark). Voice input from a microphone can be included.
  • the audio output unit 3 allows audio output from a built-in speaker provided above the communication terminal 1.
  • the audio output unit 3 is connected to a tag, a connector (not shown) or the like provided on the peripheral surface of the communication terminal 1 or a short-distance wireless communication connection such as Bluetooth (registered trademark). Audio output from the speaker can be included.
  • the display unit 4 can permit various operations using a touch panel system in addition to image display using a color liquid crystal panel or the like.
  • Various operations by the user can include operations by a switch 7 provided on the peripheral surface or the surface of the communication terminal 1 in addition to the so-called touch operation using the display unit 4.
  • the display unit 4 activates, for example, a telephone icon AP1 for executing a telephone function, an E-Mail icon AP2 for executing an E-Mail function, a chat icon AP3 for executing a chat function, and a translation application.
  • Various function icons such as a translation icon AP4 are displayed.
  • the storage unit 5 is a so-called ROM (Read only memory) or RAM (Random access memory), and realizes a computer function with the control unit 6.
  • the storage unit 5 stores internal data, removable external storage using various media, online storage including cloud C by Internet connection, etc. as a storage location for data according to the usage form such as primary storage, etc. Can be included.
  • the storage unit 5 may be referred to as an execution program (hereinafter referred to as “translation driver”) of a translation application (hereinafter simply referred to as “application”) downloaded through a telecommunication line (Internet connection line) NT, for example. Is stored in advance.
  • translation driver an execution program
  • application a translation application
  • NT Internet connection line
  • the storage unit 5 stores address information (address book) for displaying the name of the other party on the display unit 4 in a table system when a call is made or received by the telephone function.
  • address book address book
  • An SNS function that is a community-based membership system service can include member registration information (appropriate member public data under personal information protection) with the service provider.
  • Pieces of personal information are appropriate and indispensable information when using each function, and can identify the country and language.
  • the control unit 6 of the communication terminal 1 receives, from the reception interface (I / F) 9, reception data when a radio wave is received by a built-in antenna (not shown), for example. . Similarly, the control unit 6 transmits transmission data when transmitting a radio wave by the built-in antenna to the transmission interface (I / F) 10.
  • the control unit 6 controls the speech recognition unit 11, the speech generation unit 12, the character recognition unit 13, and the character generation unit 14 for translation processing using a translation driver, and transmits / receives voice data or text data.
  • the control unit 6 displays an image (for example, JPG method or the like) or video (moving image or moving image with sound: MP4 or the like) photographed by the photographing camera 15 provided on the front side (and / or the back side) of the communication terminal 1.
  • An image control function is provided for storing in the storage unit 5 or for use in a TV phone or sign language translation described later.
  • the voice recognition unit 11 receives voice data based on the voice input by the user to the voice input unit 2 or voice data received via the reception interface 9.
  • the voice recognition unit 11 has a reading function for analyzing the voice data.
  • the voice generation unit 12 uses voice data output from the voice output unit 3 or voice data transmitted from the transmission interface 10 based on voice data output from the voice recognition unit 11 or text data output from the character recognition unit 13. To generate.
  • the voice generation unit 12 has a reading function for outputting the generated voice data from the voice output unit 3.
  • the text recognition unit 13 receives text data based on a character input touched by the user on the display unit 4 or text data received via the reception interface 9.
  • the character recognition unit 13 recognizes the character data code of the text data.
  • the character generation unit 14 generates text data to be displayed on the display unit 4 or text data to be transmitted from the transmission interface 10 based on voice data output from the speech recognition unit 11 or text data output from the character recognition unit 13. Generate a character data code.
  • the photographing camera 15 can photograph the front side of the communication terminal 1 in the self-shooting mode, for example.
  • the camera function itself is well-known, the detailed description is abbreviate
  • the control unit 6 can cause the sign language analysis unit 16 to analyze the video imaged by the imaging camera 15.
  • the sign language analysis unit 16 has a function of analyzing operations related to sign language photographed by the photographing camera 15 and generating, for example, character data and voice data.
  • the sign language analysis unit 16 can analyze the motion related to the sign language photographed by the photographing camera 15 and generate substitute image data such as a simple motion image including a hand image and a face image. A detailed example of the generation of the character / sound / image data will be described later.
  • control unit 6 sets the language (for example, Japanese) set by the operating system stored in the storage unit 5 as a basic language that is normally used by the user, and stores the address stored in the storage unit 5 by the translation driver.
  • the language for example, English
  • the language used by the communication partner included in the personal information of the information is set as another language to be translated.
  • the other language is not limited to one language and can be a language according to the type of translation driver. Therefore, the translation driver can use one or more stored in the storage unit 5.
  • the control unit 6 activates the chat function and causes the display unit 4 to display a chat screen.
  • the control unit 7 executes a function of translating the corresponding used language into another language.
  • the communication terminal 21 used by the other party is also a smartphone including an audio input unit 22 and an audio output unit 23, and a display unit 24 for realizing a display function. At this time, it is assumed that no translation application is installed in the communication terminal 21.
  • the text data is a character code, it is also referred to as a simple “code” or “character code” in the following description.
  • the voice data or text data is also simply referred to as “voice” or “text” (or “character” converted for display).
  • the user refers to the character input pad displayed at the bottom of the screen of the display unit 4 and performs a touch operation. For example, “Tone Doo? ”Is entered in the upper right of the screen of the display unit 4 with a frame [a tone? ] Is displayed.
  • the character code data processing at this time uses known processing.
  • control unit 6 recognizes the character code input by the character recognition unit 13, and generates a character code (for example, ⁇ How have you > been>) of another language from the recognized character code using a translation driver.
  • the control unit 6 transmits the generated character code ⁇ How have you been> to the communication terminal 21 of the other party via the transmission interface 10.
  • the partner communication terminal 21 (the control unit thereof) displays the character [How [have you been] on the upper left side of the screen of the display unit 24 in a chat mode based on the received character code ⁇ How have you been>.
  • the control unit 6 of the communication terminal 1 that has received the character code ⁇ so and so> causes the character recognition unit 13 to recognize the character code ⁇ so and so>, and then translates the character code ⁇ maaama>. > Is generated. Further, the control unit 7 displays a character [maama] on the upper left of the screen of the display unit 4.
  • the language used by the partner is another language, and the translation application is installed in the partner communication terminal 21. Even if not, communication can be easily established.
  • control unit 6 that enables such a basic translation function includes a speech translated into another language by simply inputting either speech or characters in the basic language using its own communication terminal 1. Output and character notation can be realized in the communication terminal 21 of the other party.
  • control unit 6 causes the communication terminal 1 to realize the voice output and the character notation translated into the basic language only by inputting either one of speech or characters in another language using the communication terminal 21 by the other party. be able to.
  • the communication terminal 1 uses a chat application with voice or a telephone application with character display.
  • a translation with a character output function may be additionally programmed in a mode to a normal telephone application (phone icon AP1).
  • chat icon AP3 may be programmed to add translation with voice function in a mode. Further, it is assumed that when the user starts up these applications and designates a partner, the control unit 6 determines to translate based on information stored in the address data.
  • the control unit 6 causes the voice recognition unit 11 to read the voice and generate a voice ⁇ tone>. To do.
  • control unit 6 causes the character generation unit 14 to generate a character ⁇ tone> on the basis of the read voice ⁇ tone> and causes the display unit 4 to display the character [tone].
  • the translated character ⁇ How have you been> is generated and the character [How have you been] is displayed on the display unit 4.
  • control unit 6 causes the voice generation unit 12 to generate a voice ⁇ How have you been> based on the generated character ⁇ How have you been>.
  • control unit 6 superimposes the generated three data of the character ⁇ tone>, the character ⁇ How ⁇ have you been>, and the voice ⁇ How have you been> from the transmission interface 10 to the communication terminal 21. Send to.
  • the communication terminal 21 displays the pre-translation characters [Tone Doo] and the post-translation characters [How have you been] on the display unit 24 and the voice “How have you been”. Output from the output unit 23.
  • the communication terminal 21 can acquire three data of the character ⁇ tone>, the character ⁇ How have you been>, and the voice ⁇ How have you been> without installing the translation application. Can do.
  • the control unit 6 of the communication terminal 1 causes the voice recognition unit 11 to recognize the voice ⁇ so ⁇ ⁇ and so> when the reception interface 9 receives the voice ⁇ so and so>.
  • control unit 6 causes the character generation unit 14 to generate the character ⁇ so ⁇ and so> and causes the display unit 4 to display the character [so] and so], and also generates and displays the translated character ⁇ maamaa>.
  • the character “maamaa” is displayed in part 4.
  • control unit 6 causes the voice generation unit 12 to generate a voice ⁇ maamaa> based on the generated character ⁇ maaama>, and causes the voice output unit 3 to output the transliterated voice “maaamaa”.
  • the communication terminal 1 does not have the translation application installed on the communication terminal 21, so that the pre-translation character ⁇ so and so>, the post-translation character ⁇ ma-ama>, and the post-translation voice ⁇ ma-ama> Can be generated on the display unit 4 and the audio output unit 3 to be recognized by the user. Note that these three pieces of data may be generated by the communication terminal 21 as long as the translation application is installed in the communication terminal 21.
  • the control unit 6 causes the character recognition unit 13 to recognize the character code.
  • the control unit 6 causes the character generator 14 to generate the character ⁇ tone> and display the character [tone] on the display unit 4 and translate it.
  • the character ⁇ How have you been> is generated and the character [How have you been] is displayed on the display unit 4.
  • control unit 6 causes the voice generation unit 12 to generate a voice ⁇ How have you been> based on the generated character ⁇ How have you been>.
  • control unit 6 superimposes the generated three data of the character ⁇ tone>, the character ⁇ How ⁇ have you been>, and the voice ⁇ How have you been> from the transmission interface 10 to the communication terminal 21. Send to.
  • the communication terminal 21 displays the pre-translation characters [Tone Doo] and the post-translation characters [How have you been] on the display unit 24 and the voice “How have you been”. Output from the output unit 23.
  • the communication terminal 21 can acquire three data of the character ⁇ tone>, the character ⁇ How have you been>, and the voice ⁇ How have you been> without installing the translation application. Can do.
  • the voice recognition unit 11 causes the voice ⁇ so and so> (or the character recognition unit 13). To recognize the characters ⁇ so and so>).
  • control unit 6 causes the character generation unit 14 to generate the character ⁇ so ⁇ and so> and causes the display unit 4 to display the character [so] and so], and also generates and displays the translated character ⁇ maamaa>.
  • the character “maamaa” is displayed in part 4.
  • control unit 6 causes the voice generation unit 12 to generate a voice ⁇ maamaa> based on the generated character ⁇ maamaa>, and causes the voice output unit 3 to output the translated voice “maaama”.
  • the communication terminal 1 does not have the translation application installed on the communication terminal 21, so that the pre-translation character ⁇ so and so>, the post-translation character ⁇ ma-ama>, and the post-translation voice ⁇ ma-ama> Can be generated on the display unit 4 and the audio output unit 3 to be recognized by the user. Note that these three pieces of data may be generated by the communication terminal 21 as long as the translation application is installed in the communication terminal 21.
  • the communication terminal 1 is preinstalled with a translation application that executes mutual translation between the basic language and another language in the storage unit 5 and stores an address book.
  • Step S1 In step S ⁇ b> 1, the control unit 6 receives the designation of the communication application by the operation of the operation unit 8 by the user, refers to the address book corresponding to the communication application from the storage unit 5, and causes the display unit 4 to display the address book. The process proceeds to step S2.
  • Step S2 the control unit 6 determines whether or not a call destination (communication partner) designation operation has been performed by a user operation. When it is determined that the call destination has been designated (Yes), the control unit 6 proceeds to step S3.
  • control unit 6 determines that the call destination has been specified (No)
  • the control unit 6 continues this routine until the call destination is specified (or until another application is operated). Monitor. Then, the control unit 6 establishes communication (here, a telephone function) with the designated call destination.
  • step S3 the control unit 6 specifies the call destination based on the call destination designated by the user, analyzes the address information corresponding to the call destination, and proceeds to step S4.
  • Step S4 Step for identifying other languages
  • the control unit 6 specifies the country or language including whether the language used by the communication partner is a basic language or another language based on the analyzed address information, and proceeds to step S5.
  • step S4 the language used by the communication partner in step S4 is specified as another language.
  • Step S5 Communication content reception step
  • the control unit 6 accepts at least one input of voice data or text data in the basic language as transmission reference data via the voice input unit 2 or the display unit 4 as input means, and the process proceeds to step S6. Transition.
  • Step S6 Translation into other languages
  • the control unit 6 activates the translation application, translates the transmission reference data received in step S5 into voice data and text data in other languages, generates transmission data, and proceeds to step S7.
  • step S5 If the voice input from the voice input unit 2 is input in step S5, the control unit 6 recognizes the voice data in the basic language, which is the transmission reference data received by the voice input unit 2, and stores it in the storage unit 5. Start the listening application installed in advance.
  • control part 6 will perform the basic text data production
  • step S5 the control unit 6 recognizes the text data in the basic language, which is the transmission reference data received from the display unit 4, and is installed in the storage unit 5 in advance. Start the reading application.
  • control part 6 will perform the basic audio
  • Step S7 Transmission step
  • the control unit 6 transmits the transmission data translated in step S6 from the transmission interface 10 as a transmission means through the telecommunication line, and proceeds to step S8.
  • step S7 when the transmission reference data received by the display unit 4 is text data in the basic language, the control unit 6 adds the text data in the basic language in addition to the transmission data in the speech data in the other languages that have been translated. May be transmitted in a superimposed manner.
  • step S2 since the communication state (call state) with the other party is established in step S2, the user completes the transmission in steps S3 to S7 (until the word is exhausted or the other party receives data). ) A cancel step may be executed (accepted) to depress the switch 7 and prevent interruption from the other party. As a result, voice turbidity (translation turbidity) can be suppressed.
  • Step S8 the control unit 6 determines whether or not there is reception from the other party.
  • the control part 6 can also receive the communication content from a user as interruption by the routine of step S5 continuously. If it is determined that there is a reception from the other party (Yes), the control unit 6 proceeds to step S9.
  • control unit 6 determines that there is a reception from the other party (No)
  • the control unit 6 continues to monitor this routine or the interruption in step S5.
  • Step S9 Reception step
  • the control unit 6 receives voice data or text data in another language from the communication partner as reception reference data by the reception interface 9 as a reception unit through the telecommunication line, and proceeds to step S10.
  • step S10 the control unit 6 specifies whether the received data is text data or voice data, and proceeds to step S11.
  • step S11 the control unit 6 determines an output method, that is, whether to perform character analysis or voice analysis, based on the received data, and proceeds to step S12.
  • Step S12 Basic verbalization translation step
  • the control unit 6 activates the translation application, translates the reception reference data received in step S9 into speech data and text data in a basic language, generates reception data, and proceeds to step S13.
  • Step S13 Communication content output step
  • the control unit 6 outputs the received data translated in step S12 through the audio output unit 3 and the display unit 4 as output means and outputs the characters and ends this routine.
  • the control unit 6 thereafter repeats the above routine until the communication (call) is completed.
  • the translation processing method, the translation processing program, and the recording medium according to the present embodiment are translated into the communication terminal 21 of the communication partner with respect to either the voice or the character input to the communication terminal 1. Voice and text can be output.
  • the translated voice and characters can be output to the communication terminal 1.
  • the characters before translation which is the basic language of the communication terminal 1 and the communication terminal 21, can be transmitted and received with each other.
  • sign language is a visual language that performs finger movement and non-finger movement at the same time, and is recognized as a language along with speech language. Therefore, mutual translation with other languages can be realized by converting the visual language based on this sign language into the basic language and converting it into character data, voice data, or alternative image data.
  • other languages may include translations in the case of converting the native language into characters, sounds, and images, as well as re-converting into other languages based on the characters, sounds, and images of the native language. it can.
  • the storage unit 5 stores data that can be translated with respect to the sign language action as well as the language.
  • Sign language although depending on the level of proficiency, is relatively slower than conversation but is often faster than character input. Therefore, convenience can be improved for some users by analyzing the sign language action and generating character data and voice data.
  • substitute image data such as a simple motion image including hand and face images
  • the responsiveness can be secured, but the amount of communication data increases. End up. Accordingly, the sign language motion video may be used as it is.
  • the substitute image expressed in the screen of the display unit 4 as easily as possible is replaced with the substitute image data displayed on the display unit 4. Is desirable.
  • the burden can be reduced when the other party is a normal hearing person.
  • a translation mode selection menu is displayed on the display unit 4 as shown in FIG.
  • the translation mode selection menu displayed on the display unit 4 includes “1) offline face-to-face translation”, “2) SNS translation”, “3) general call translation”, “ There are 6 types: “4) Simultaneous translation of conference”, “5) Video phone translation” and “6) Sign language translation”.
  • One communication terminal 1 shown in the present embodiment includes a voice input unit 2 for voice input, an operation unit 8 for character input, and a photographing camera 15 for video input as input means. Further, one communication terminal 1 has an audio output unit 3 for outputting audio and a display unit 4 for outputting characters / images as output means. Further, one communication terminal 1 is capable of translating a basic language and another language.
  • one communication terminal 11 generates output data after being converted into another language for input data in a basic language received via the input means, and simultaneously outputs the output data (to the output means).
  • Offline face-to-face translation is a translation mode that can be used when two or more foreign language speakers have a conversation using one communication terminal 1.
  • one communication terminal 1 can use SNS on the cloud through an Internet line as a telecommunication line. Therefore, since the language used can be specified using the address book on the cloud as an SNS, for example, on the timeline, etc., the display language can be changed using the mutual translation function between the basic language and other languages. Mutual translation is possible.
  • the basic language and the other language may translate the translated text by tapping a screen display portion of a text (including a sign language image or the like), for example, Depending on the system, the entire text display may be translated at the labeling timing or the like.
  • one communication terminal 11 generates output data after being converted into another language for input data in the basic language received via the input means, and at the same time outputs the output data (on the cloud).
  • SNS translation is a translation that can be used, for example, when a translation that specifies a mutual language is performed by text or voice using a cloud-type SNS member list that can be managed on the Internet. Mode. .
  • one communication terminal 1 can make a call through a telephone line or an internet line as a telecommunication line.
  • one communication terminal 11 generates output data after being converted into another language for input data in the basic language received via the input means, and at the same time outputs the output data (to a telephone line or the like).
  • output data it is possible to generate output data after being converted into a basic language with respect to input data in another language, and at the same time, output it via the output means (the audio output unit 3 or the display unit 4).
  • General call translation is a translation mode that can be used for conversations between foreigners and other parties on a normal telephone line.
  • one communication terminal 1 can communicate through a short-range wireless communication line, a telephone line, and an Internet line as an electric communication line.
  • one communication terminal 11 generates output data after being converted into another language for input data in a basic language received via the input means, and at the same time outputs the output data (short-range wireless communication line / telephone). (Line, internet line, etc.) can be output, and output data after being converted into a basic language is generated for input data in other languages, and at the same time, output via output means (voice output unit 3 or display unit 4) can do.
  • output data short-range wireless communication line / telephone
  • the basic language and the other language can be translated into each other not only by the same conference room but also by two or more persons (two or more communication terminals) in a remote place.
  • the number of other languages is not limited to one language, and a plurality of languages can be translated.
  • Simultaneous Conference Translation is a translation mode that can be used when two or more users have conversations between foreign languages. .
  • the simultaneous translation mode can be activated in the TV phone using the photographing camera 15.
  • the other party's communication terminal 21 is also equipped with a photographing camera 25.
  • sign language translation for analyzing the sign language motion photographed by the photographing camera 15 and converting it into character data, voice data, and substitute image data can be activated.
  • the control unit 6 captures an image photographed by the photographing camera 15. Execute sign language mode including the analyzed translation.
  • the control unit 6 uses, for example, a sign language action unit (corresponding to a word vocabulary unit) from an image captured by the photographing camera 15.
  • the sign language analysis unit 16 is made to analyze the content of the operation.
  • the sign language analysis unit 16 analyzes each sign language operation.
  • each sign language operation unit is stored in the storage unit 5.
  • the video of each sign language operation unit may be temporarily stored in a frame memory or the like different from the storage unit 5, for example.
  • the control unit 6 Based on the sign language action ⁇ I love you> analyzed by the sign language analysis unit 16, the control unit 6 causes the character generation unit 14 to generate Japanese characters ⁇ I love you> that are masculine expressions. A character [I love you] is displayed on the display unit 4 and a translated English character ⁇ I love you> is generated and the character [I love you] is displayed on the display unit 4.
  • control unit 6 activates a sign language analysis application installed in the storage unit 5 in advance when the sign language operation as the transmission reference data received by the photographing camera 15 as the input means is used as a basic language. Analyzing the sign language movement with the basic analysis data in the basic language for generating outgoing data consisting of at least one of speech data, text data, and image data in other languages A basic analysis data generation step (corresponding to step S6) for generating “I love you>) is executed.
  • the translation by this sign language action includes the case where the sign language action is analyzed to generate ⁇ I love you>, and the case where it is secondarily translated into another language.
  • the sign language analysis application of the present embodiment includes an analysis conversion function for analyzing sign language movement and converting it into Japanese, a translation function for translating the converted Japanese into English, and an English-based version based on the translated English. It can also be said that three language conversion functions (translation functions), that is, a conversion function for converting to a sign language action of the above, are executed.
  • control unit 6 causes the voice generation unit 12 to generate a voice ⁇ I love you> based on the generated character ⁇ I love you>.
  • control unit 6 causes the display unit 4 to display an image of the analyzed sign language action.
  • the sign language analysis unit 16 uses the image data stored in the storage unit 5 in advance to sign language operations such as an English-language sign language operation or a similar icon / stamp (JPG, video GIF, etc.). ) To enable transmission.
  • control unit 6 superimposes the generated four data of the character ⁇ I love you>, the character ⁇ I love you>, the voice ⁇ I love you>, and the sign language image ⁇ (I love you)>. Then, the data is transmitted from the transmission interface 10 to the communication terminal 21.
  • the communication terminal 21 displays the pre-translation character [I love you], the post-translation character [I love you], and the post-translation sign language image [(I love you)] on the display unit 24.
  • the voice “I love you” is output from the voice output unit 23 while being displayed.
  • the communication terminal 21 has the character ⁇ I love you>, the character ⁇ I love you>, the voice ⁇ I love you>, and the sign language image [(I love you) ] Can be acquired. Note that these four pieces of data may be generated by the communication terminal 21 if a translation application is installed in the communication terminal 21.
  • the control unit 6 of the communication terminal 1 causes the voice recognition unit 11 to recognize the voice ⁇ I am happy> when the reception interface 9 receives the voice ⁇ I > am happy>.
  • control unit 6 causes the character generation unit 14 to generate the character ⁇ I am happy>, displays the character [I am happy] on the display unit 4, and translates the character ⁇ happy! > Is generated on the display unit 4 and the character [happy! ], And a sign language image [(happy)] of the corresponding sign language action is displayed.
  • the control unit 6 can also generate the speech ⁇ joyful> by the speech generation unit 12 based on the generated character ⁇ joyful> and output the transliterated speech “happy” from the speech output unit 3.
  • the communication terminal 1 does not have a translation application installed on the communication terminal 21, the character before translation ⁇ I am happy> and the character after translation ⁇ happy! >, Translated sign language image ⁇ (happy)>, translated speech ⁇ joyful>, and can be expanded on the display unit 4 and the voice output unit 3 to be recognized by the user. Note that these four pieces of data may be generated by the communication terminal 21 if a translation application is installed in the communication terminal 21.
  • the translation data stored in the storage unit (including on the cloud) 5 is updated by learning speech recognition, video recognition, natural language processing, etc. by a learning function (AI function) such as deep-learning. It is also possible to store data sequentially.
  • AI function learning function
  • the translation processing method, the translation processing program, and the recording medium according to the present invention are functions that are executed by a single communication terminal, but the translation variations can be made multifunctional to improve versatility. And is useful for a translation processing method, a translation processing program, and a general recording medium for mutual translation between a basic language and another language when communicating with each other via a telecommunication line. .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Signal Processing (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • General Health & Medical Sciences (AREA)
  • Telephonic Communication Services (AREA)
  • Information Transfer Between Computers (AREA)
  • Telephone Function (AREA)
  • Machine Translation (AREA)

Abstract

[Problem] To provide a translation processing method that enables translation variations to be made multifunctional and that can improve versatility. [Solution] As steps to be executed by one communication terminal that includes a storage means onto which is pre-installed a translation application that executes mutual translation between a base language and a second language, the following steps are provided: a communication content reception step for receiving at least one input, via an input means, as transmission reference data, such input being audio data, text data, or image data in the base language; a to-second-language translation step in which the translation application is initiated, the transmission reference data received in the communication content reception step is translated into at least one of audio data, text data, and image data in the second language to thereby generate transmission data; a transmission step for transmitting, by a transmission means, the transmission data which was translated in the to-second-language translation step; a reception step for receiving, using a reception means, at least one the audio data, text data, and image data in the second language as reception reference data; a to-base-language translation step in which the translation application is initiated, the reception reference data received in the reception step is translated into at least one of audio data, text data, and image data in the base language to thereby generate reception data; and a communication content output step for outputting, via an output means, the reception data which was translated in the to-base-language translation step.

Description

翻訳処理方法、翻訳処理プログラム、及び、記録媒体Translation processing method, translation processing program, and recording medium
 本発明は、電気通信回線を通じて相互通信している際に、基本言語と他言語とで相互翻訳するための翻訳処理方法、翻訳処理プログラム、及び、記録媒体に関する。 The present invention relates to a translation processing method, a translation processing program, and a recording medium for mutual translation between a basic language and another language when communicating with each other via a telecommunication line.
 従来、電話回線を通じた通話(会話)や、インターネット回線を通じたチャット或はE-Mailによる往復書簡など、電気通信回線を通じて相互通信している際に双方の言語が異なっている場合に、内容の理解のために翻訳する各種翻訳機能が周知である。 Conventionally, if the two languages are different when communicating with each other via a telecommunication line, such as a call (conversation) via a telephone line, a chat via an internet line or a round-trip letter by E-Mail, Various translation functions for translating for understanding are well known.
 この際、対面した相互会話の場合、一方が翻訳機を持参していれば、互いの言語が異なりかつ相手方の言語に対する語学力を有していなくても会話を成立させ、コミュニケーションをとることが可能である。 In this case, in the case of face-to-face conversations, if one of them has a translator, it is possible to establish a conversation and communicate even if they do not have language skills for the other party's language. Is possible.
 一方、受信から返信までに時間的な余裕があり、その間に翻訳を行うことができるE-Mail等とは異なり、通話やチャット(各種SNSのタイムライン等を含む)などのように、即時性を有する相互通信では、発信側における基本言語から他言語への翻訳と、受信時における他言語から基本言語への翻訳、の2通りを実現する必要がある。 On the other hand, unlike E-Mail etc., where there is a time lag between receiving and replying, and translation is possible during that time, such as calls and chats (including timelines of various SNSs), etc. It is necessary to realize two kinds of communication, that is, translation from a basic language to another language on the transmission side and translation from another language to the basic language at the time of reception.
 そこで、相互に翻訳機(翻訳機能)を有していなくても、一方側で発信時の他言語化と受信時の基本言語化とを実現するために、音声を聞き取ってテキストデータ化したうえで翻訳後のテキストデータを生成し、そのテキストデータをさらに音声に読み上げることにより、リアルタイムに近い通話を可能とした技術が知られている(例えば、特許文献1参照)。 Therefore, even if they do not have a translator (translation function), they can listen to the voice and convert it into text data in order to realize a different language when sending and a basic language when receiving. A technique is known in which text data after translation is generated, and the text data is further read aloud to enable near real-time communication (see, for example, Patent Document 1).
特表2005-513619号公報JP 2005-513619 A
 しかしながら、特許文献1に開示の翻訳機は、音声通話時における聞き取りと読み上げとを実現する同時通訳の範疇に過ぎず、汎用性が低いという問題が生じていた。 However, the translator disclosed in Patent Document 1 is merely a category of simultaneous interpretation that realizes listening and reading at the time of a voice call, and has a problem of low versatility.
 本発明は、翻訳のバリエーションを多機能化し得て汎用性を向上することができる翻訳処理方法、翻訳処理プログラム、及び、記録媒体を提供することを目的とする。 An object of the present invention is to provide a translation processing method, a translation processing program, and a recording medium that can make a variation of translation multifunctional and improve versatility.
 上記目的を達成するため、請求項1記載の発明にあっては、基本言語と他言語との相互翻訳を実行する翻訳アプリケーションを予めインストールした記憶手段を有する1つの通信端末に実行させるステップとして、基本言語による音声データ又はテキストデータの少なくとも何れか一方の入力を入力手段を介して発信基準データとして受け付ける通信内容受付ステップと、翻訳アプリケーションを起動させて、通信内容受付ステップで受け付けた発信基準データを他言語の音声データ及びテキストデータに翻訳して発信データを生成する他言語化翻訳ステップと、他言語化翻訳ステップで翻訳した発信データを発信手段により発信する発信ステップと、他言語による音声データ又はテキストデータを受信手段により受信基準データとして受信する受信ステップと、翻訳アプリケーションを起動させて受信ステップで受信した受信基準データを基本言語による音声データ及びテキストデータに翻訳して受信データを生成する基本言語化翻訳ステップと、基本言語化翻訳ステップで翻訳した受信データを出力手段を介して出力する通信内容出力ステップと、を備えるものである。 In order to achieve the above object, in the invention described in claim 1, as a step of causing a single communication terminal having a storage unit, in which a translation application for performing mutual translation between a basic language and another language is installed, to be executed, A communication content receiving step for receiving at least one of speech data and text data in the basic language as transmission reference data via the input means, and starting the translation application and sending reference data received in the communication content receiving step. Other language translation step for translating into speech data and text data in other languages to generate transmission data, transmission step for transmitting the transmission data translated in the other language translation step, and voice data in other languages Receive text data as reception reference data by receiving means A basic verbalization translation step of translating the reception reference data received in the reception step by activating the translation application into speech data and text data in a basic language, and a basic languageization translation step And a communication content output step for outputting the translated received data via the output means.
 請求項2記載の発明にあっては、前記発信ステップにおいて、前記発信データは発信手段により電気通信回線を通じて発信され、前記受信ステップにおいて、前記受信基準データは電気通信回線を通じて受信手段により受信される、ものである。 In the invention according to claim 2, in the transmission step, the transmission data is transmitted by a transmission means through a telecommunication line, and in the reception step, the reception reference data is received by a reception means through a telecommunication line. , That is.
 請求項3記載の発明にあっては、前記電気通信回線がインターネット回線である。 In the invention according to claim 3, the telecommunications line is an internet line.
 請求項4記載の発明にあっては、前記電気通信回線が電話回線である。 In the invention according to claim 4, the telecommunications line is a telephone line.
 請求項5記載の発明にあっては、前記発信ステップは電気通信回線を介することなく実行され、前記受信ステップは電気通信回線を介することなく実行される、ものである。 In the invention according to claim 5, the transmission step is executed without going through a telecommunication line, and the receiving step is executed without going through a telecommunication line.
 請求項6記載の発明にあっては、前記他言語化翻訳ステップを実行させる前に、前記記憶手段に予め登録した通信相手先の登録名を含むアドレスデータに基づいて翻訳対象となる前記他言語を特定する他言語特定ステップと、を備えるものである。 In the invention according to claim 6, before executing the other language translation step, the other language to be translated based on address data including a registered name of a communication partner registered in advance in the storage means And a different language specifying step for specifying.
 請求項7記載の発明にあっては、前記入力手段により受け付けた前記発信基準データが前記基本言語による音声データである場合に、前記記憶手段に予めインストールした聞き取りアプリケーションを起動させて、前記他言語によるテキストデータからなる前記発信データを生成するための前記基本言語による基本テキストデータを生成する基本テキストデータ生成ステップと、を備えるものである。 In the invention according to claim 7, when the transmission reference data received by the input means is voice data in the basic language, the listening application installed in the storage means in advance is started and the other language is started. A basic text data generating step for generating basic text data in the basic language for generating the outgoing data composed of text data according to the above.
 請求項8記載の発明にあっては、前記入力手段により受け付けた前記発信基準データが前記基本言語によるテキストデータである場合に、前記記憶手段に予めインストールした読み上げアプリケーションを起動させて、前記他言語による音声データからなる前記発信データを生成するための前記基本言語による基本音声データを生成する基本音声データ生成ステップと、を備えるものである。 In the invention according to claim 8, when the transmission reference data received by the input means is text data in the basic language, a reading application installed in advance in the storage means is started to activate the other language. A basic voice data generation step of generating basic voice data in the basic language for generating the outgoing data composed of voice data according to.
 請求項9記載の発明にあっては、前記入力手段により受け付けた前記発信基準データが前記基本言語によるテキストデータである場合に、前記他言語化翻訳ステップで翻訳した前記他言語の音声データによる前記発信データに加えて前記入力手段により受け付けた前記基本言語によるテキストデータを前記発信手段により発信する発信ステップと、を備えるものである。 In the invention according to claim 9, when the transmission reference data received by the input means is text data in the basic language, the speech data in the other language translated in the other language translation step A transmission step of transmitting text data in the basic language received by the input means in addition to the transmission data by the transmission means.
 請求項10記載の発明にあっては、前記入力手段が前記基本言語による音声データを受け付けてから前記他言語化翻訳ステップで翻訳した前記他言語による前記発信データを前記発信手段により発信するまでの期間は前記受信手段による受信をキャンセルする割り込みキャンセルステップと、を備えるものである。 In the invention according to claim 10, from the time when the input means accepts the voice data in the basic language to the time when the sending means in the other language translated in the other language is sent by the sending means. The period includes an interrupt canceling step for canceling reception by the receiving means.
 請求項11記載の発明にあっては、前記入力手段により受け付けた前記発信基準データが前記基本言語による手話動作である場合に、前記記憶手段に予めインストールした手話解析アプリケーションを起動させて前記手話動作を解析したうえで、前記他言語による音声データ・テキストデータ・画像データの少なくとも何れか一つからなる前記発信データを生成するための前記基本言語による基本解析データを生成する基本解析データ生成ステップと、を備えるものである。 In the invention according to claim 11, when the transmission reference data received by the input means is a sign language operation in the basic language, the sign language analysis application installed in the storage means in advance is activated to perform the sign language operation. A basic analysis data generating step for generating basic analysis data in the basic language for generating the transmission data composed of at least one of voice data, text data, and image data in the other language, , Are provided.
 請求項12記載の発明にあっては、請求項1乃至請求項6の何れか1項に記載の翻訳処理方法を通信端末の演算手段に実行させる、ものである。 In the invention described in claim 12, the translation processing method described in any one of claims 1 to 6 is executed by a calculation means of a communication terminal.
 請求項13記載の発明にあっては、請求項1乃至請求項6の何れか1項に記載の翻訳処理方法を通信端末の演算手段に実行させる翻訳処理プログラムを記録した記録媒体である。 The invention according to claim 13 is a recording medium on which a translation processing program for causing a computing means of a communication terminal to execute the translation processing method according to any one of claims 1 to 6 is recorded.
 本発明によれば、入力手段を介して受け付けた基本言語による入力データに対して、他言語化した後の出力データを生成すると同時に、その出力データを出力することができると共に、他言語による入力データに対して、基本言語化した後の出力データ生成すると同時に、出力手段を介して出力することができる。 According to the present invention, the input data in the basic language received via the input means can be generated with the output data after being converted into another language, and the output data can be output simultaneously. With respect to the data, output data after being converted into a basic language can be generated and output through the output means at the same time.
 これにより、目の不自由な人や耳の不自由な人であっても、基本言語と他言語との相互翻訳を1つの通信端末で両立させることが可能となる。また、目や耳が不自由でなくても、テキストデータを相互で有することができ、通話メモの代わりや、契約等の大事な通話時における通話内容の誤認等を抑制することが可能となる。 This makes it possible for a single communication terminal to achieve mutual translation between the basic language and another language, even for visually impaired people or hearing impaired people. Moreover, even if the eyes and ears are not inconvenient, it is possible to have text data mutually, and it is possible to suppress misrecognition of the contents of a call instead of a call memo or during an important call such as a contract. .
 したがって、一方側にのみ翻訳機能を有している状態であるにもかかわらず、音声及びテキストの両方のデータを有効利用することができ、翻訳のバリエーションを多機能化し得て汎用性を向上することができる。 Therefore, even if it has a translation function only on one side, both voice and text data can be used effectively, and translation variations can be made multifunctional to improve versatility. be able to.
 また、発信ステップにおいて、発信データは発信手段により電気通信回線を通じて発信され、受信ステップにおいて、受信基準データは電気通信回線を通じて受信手段により受信される、ことにより、電気通信回線を通じて、翻訳後の(他言語の)発信データを発信することができるともに、翻訳前の(他言語の)受信基準データを受信することができる。 In the transmission step, the transmission data is transmitted by the transmission means through the telecommunication line, and in the reception step, the reception reference data is received by the reception means through the telecommunication line. Transmission data (in other languages) can be transmitted, and reception reference data (in other languages) before translation can be received.
 ここで、電気通信回線をインターネット回線とすれば、SNS等を利用した相互翻訳に供することができ、電気通信回線を電話回線とすれば、通常の音声通話やビデオ信号等の動画等を含むTV電話を利用した相互翻訳に供することができる。 Here, if the telecommunication line is an internet line, it can be used for mutual translation using SNS or the like, and if the telecommunication line is a telephone line, a TV including a normal voice call or a video signal such as a video signal. It can be used for mutual translation using telephone.
 また、発信ステップは電気通信回線を介することなく実行され、受信ステップは電気通信回線を介することなく実行される、ことにより、電気通信回線を用いない対面通訳等の相互翻訳に供することができる。 Also, since the sending step is executed without going through the telecommunications line and the receiving step is run without going through the telecommunications line, it can be used for mutual translation such as face-to-face interpretation without using the telecommunications line.
 このように、基本言語による音声入力又はテキスト(画像・映像を含む)入力に対して、他言語化した後の音声データ及びテキストデータの両方を生成すると同時に、その両データを相手方に送信することができると共に、他言語による音声データ又はテキストデータの受信に対して、基本言語化した後の音声データ及びテキストデータの両方を生成すると同時に、その両データを音声及びテキストとして出力することができる。 In this way, for voice input or text (including images / videos) input in a basic language, both voice data and text data after being converted to another language are generated, and at the same time, both the data are transmitted to the other party. In addition, when receiving voice data or text data in another language, both voice data and text data after being converted into a basic language can be generated, and at the same time, both data can be output as voice and text.
 また、他言語化翻訳ステップを実行させる前に、記憶手段に予め登録した通信相手先の登録名を含むアドレスデータに基づいて翻訳対象となる他言語を特定する他言語特定ステップを備えることにより、予め発信先或は受信元を特定するためのアドレスデータに含まれる電話番号(国番号等)やメールアドレス(ドメイン名等)等を利用して他言語の選択・指定をすることなく、通信関係を成立させる前に国又は言語を特定して翻訳機能を有効化させることができる。 In addition, before executing the other language translation step, by providing another language identification step for identifying another language to be translated based on the address data including the registered name of the communication partner registered in advance in the storage means, Communication relations without having to select / designate other languages using phone numbers (country codes, etc.) or e-mail addresses (domain names, etc.) included in the address data for specifying the destination or receiver in advance. The translation function can be validated by specifying the country or language before establishing.
 また、入力手段により受け付けた発信基準データが基本言語による音声データである場合に、記憶手段に予めインストールした聞き取りアプリケーションを起動させて、他言語によるテキストデータからなる発信データを生成するための基本言語による基本テキストデータを生成する基本テキストデータ生成ステップを備えることにより、翻訳元となる基本言語のテキストデータを有効利用することができる。 In addition, when the transmission reference data received by the input means is voice data in a basic language, a basic language for generating transmission data composed of text data in another language by starting a listening application installed in advance in the storage means By providing the basic text data generation step for generating the basic text data according to the above, it is possible to effectively use the text data of the basic language as the translation source.
 また、入力手段により受け付けた発信基準データが基本言語によるテキストデータである場合に、記憶手段に予めインストールした読み上げアプリケーションを起動させて、他言語による音声データからなる発信データを生成するための基本言語による基本音声データを生成する基本音声データ生成ステップを備えることにより、翻訳元となる基本言語のテキストデータを有効利用することができる。 In addition, when the transmission reference data received by the input means is text data in a basic language, a basic language for generating a transmission data composed of voice data in another language by starting a reading application installed in the storage means in advance. By providing the basic speech data generation step for generating the basic speech data according to the above, it is possible to effectively use the text data of the basic language as the translation source.
 また、入力手段により受け付けた発信基準データが基本言語によるテキストデータである場合に、他言語化翻訳ステップで翻訳した他言語の音声データによる発信データに加えて入力手段により受け付けた基本言語によるテキストデータを発信手段により発信する発信ステップを備えることにより、翻訳元となる基本言語のテキストデータを誤訳やメモ等の代わりとして有効利用することができる。 In addition, when the transmission reference data received by the input means is text data in the basic language, the text data in the basic language received by the input means in addition to the transmission data by the speech data of other languages translated in the other language translation step By providing a sending step for sending the message by the sending means, the text data of the basic language as the translation source can be effectively used as a substitute for mistranslation or memo.
 また、入力手段が基本言語による音声データを受け付けてから他言語化翻訳ステップで翻訳した他言語による発信データを発信手段により発信するまでの期間は受信手段による受信をキャンセルする割り込みキャンセルステップを備えることにより、意図的な発話完了まで相手方の音声入力を拒否することができ、会話の混濁(重なり)を抑制することができる。 In addition, an interrupt canceling step for canceling reception by the receiving means is provided during a period from when the input means accepts the voice data in the basic language to when the sending means sends the outgoing data in another language translated in the other language translation step. Thus, the voice input of the other party can be rejected until the intentional utterance is completed, and the turbidity (overlap) of the conversation can be suppressed.
 また、入力手段により受け付けた発信基準データが基本言語による手話動作である場合に、記憶手段に予めインストールした手話解析アプリケーションを起動させて手話動作を解析したうえで、他言語による音声データ・テキストデータ・画像データの少なくとも何れか一つからなる発信データを生成するための基本言語による基本解析データを生成する基本解析データ生成ステップと、を備える翻訳処理方法とすることにより、言語を手話動作とした場合にも相互翻訳の利用に供することができる。 In addition, when the call reference data received by the input means is a sign language action in a basic language, the sign language action is analyzed by starting a sign language analysis application installed in the storage means in advance, and then voice data / text data in another language A language is used as a sign language operation by providing a translation processing method including a basic analysis data generation step for generating basic analysis data in a basic language for generating transmission data composed of at least one of image data In some cases, it can be used for mutual translation.
 また、本発明の翻訳処理プログラムによれば、上述した翻訳処理方法を通信端末の演算手段に実行させることができる。 Moreover, according to the translation processing program of the present invention, the above-described translation processing method can be executed by the computing means of the communication terminal.
 本発明の記録媒体によれば、上述した翻訳処理方法を通信端末の演算手段に実行させるための翻訳処理プログラムを容易に利用することができる。 According to the recording medium of the present invention, a translation processing program for causing the calculation means of the communication terminal to execute the translation processing method described above can be easily used.
本発明の一実施の形態に係る翻訳処理方法、翻訳処理プログラム、及び、記録媒体に係るシステム適用例を示し、(A)は通信端末に翻訳アプリケーションをインストールする一例の説明図、(B)は通信端末の要部のブロック構成図である。The translation processing method concerning one embodiment of the present invention, a translation processing program, and a system application example concerning a recording medium are shown, (A) is an explanatory view of an example which installs a translation application in a communication terminal, (B) is It is a block block diagram of the principal part of a communication terminal. チャット方式によるテキストデータの同時通訳例を示す説明図である。It is explanatory drawing which shows the example of simultaneous interpretation of the text data by a chat system. 基本言語による音声入力に基づく他言語化後の音声データとテキストデータとの利用例の説明図である。It is explanatory drawing of the usage example of the speech data and text data after other languages based on the speech input by a basic language. (A)は基本言語によるテキスト入力に基づく他言語化後の音声データとテキストデータとの利用例の説明図、(B)は他言語による音声入力に基づく基本言語化後の音声データとテキストデータとの利用例の説明図である。(A) is an explanatory diagram of an example of using voice data and text data after other languages based on text input in a basic language, and (B) is voice data and text data after being converted into a basic language based on voice input in another language. It is explanatory drawing of the usage example. 本発明の一実施の形態に係る制御部が実行する翻訳ルーチンの一例のフロー図である。It is a flowchart of an example of the translation routine which the control part which concerns on one embodiment of this invention performs. 本発明の一実施の形態に係る通信端末の翻訳モード選択画面の一例の説明図である。It is explanatory drawing of an example of the translation mode selection screen of the communication terminal which concerns on one embodiment of this invention. チャット方式によるテキストデータの手話通訳例を示す説明図である。It is explanatory drawing which shows the example of the sign language interpretation of the text data by a chat system.
 次に、本発明に係る一実施の形態について図面を参照して説明する。なお、以下の説明において、例えば、図1(A)に示す通信端末1における上下左右表裏等の方向は、一般的な使用状態である図示の状態を基準として、紙面の上下左右表裏等と同じものとして説明する。本実施形態にかかる通信端末1は、基本言語と他言語との相互翻訳を実行する翻訳アプリケーションを予めインストールした記憶手段を有する1つの通信端末1に実行させるものである。 Next, an embodiment according to the present invention will be described with reference to the drawings. In the following description, for example, the directions of the top, bottom, left, and right sides of the communication terminal 1 shown in FIG. 1A are the same as the top, bottom, left, right, and back sides of the page with reference to the state shown in FIG. It will be explained as a thing. The communication terminal 1 according to the present embodiment causes a single communication terminal 1 having storage means in which a translation application for performing mutual translation between a basic language and another language is installed in advance.
 図1(A)に示すように、通信端末1は、各種通信方式の電気通信回線NTを通じた電話機能による通話を実現するための音声入力部2及び音声出力部3と、表示機能を実現するための表示部4と、後述する翻訳処理方法を実現するための翻訳処理プログラムを含む通信端末1としての機能全般に関するプログラムを記憶した記憶部5と、記憶部5に記憶したプログラムにしたがって通信端末1の各種機能を処理する制御部(CPU)6と、を備える。 As shown in FIG. 1A, the communication terminal 1 realizes a voice input unit 2 and a voice output unit 3 for realizing a telephone call by a telephone function through various communication methods of the telecommunication line NT, and a display function. Display unit 4, a storage unit 5 that stores a program related to overall functions as communication terminal 1 including a translation processing program for realizing a translation processing method to be described later, and a communication terminal according to a program stored in storage unit 5 And a control unit (CPU) 6 that processes one of various functions.
 なお、図1(A)に示した通信端末1は、スマートフォンであるが、タブレット端末やパーソナルコンピュータ等、コンピュータ機能を有していて、音声データによる通話機能に加え、テキストデータによるチャット機能・E-Mail機能・SNS(Social Networking Service)機能等を利用することができる端末であれば特に限定されるものではない。 Note that the communication terminal 1 shown in FIG. 1A is a smartphone, but has a computer function such as a tablet terminal or a personal computer. In addition to the voice data call function, the text data chat function E -Any terminal that can use the Mail function, SNS (Social Networking Service) function, etc. is not particularly limited.
 なお、各種通信方式には、主として無線による、電話通信回線方式、インターネット接続回線方式、赤外線やBluetooth(登録商標)等の近距離無線通信方式、等を含ませることができる。 It should be noted that the various communication methods can include mainly a wireless telephone communication line method, an Internet connection line method, a short-distance wireless communication method such as infrared ray or Bluetooth (registered trademark), and the like.
 音声入力部2は、通信端末1の下方に設けた内蔵マイク(マイクロフォン)からの音声入力を許容する。また、音声入力部2は、例えば、通信端末1の周面に設けたプタグやコネクタ(図示せず)等への接続、或は、Bluetooth(登録商標)等の近距離無線通信接続、による外部マイクからの音声入力を含ませることができる。 The voice input unit 2 allows voice input from a built-in microphone (microphone) provided below the communication terminal 1. The voice input unit 2 is connected to a tag, a connector (not shown) or the like provided on the peripheral surface of the communication terminal 1 or a short-distance wireless communication connection such as Bluetooth (registered trademark). Voice input from a microphone can be included.
 音声出力部3は、通信端末1の上方に設けた内蔵スピーカからの音声出力を許容する。また、音声出力部3は、例えば、通信端末1の周面に設けたプタグやコネクタ(図示せず)等への接続、或は、Bluetooth(登録商標)等の近距離無線通信接続、による外部スピーカからの音声出力を含ませることができる。 The audio output unit 3 allows audio output from a built-in speaker provided above the communication terminal 1. The audio output unit 3 is connected to a tag, a connector (not shown) or the like provided on the peripheral surface of the communication terminal 1 or a short-distance wireless communication connection such as Bluetooth (registered trademark). Audio output from the speaker can be included.
 表示部4は、カラー液晶パネル等を用いた画像表示に加え、タッチパネル方式を用いた各種操作を許容することができる。なお、ユーザによる各種操作には、表示部4を用いた所謂タッチ操作に加え、通信端末1の周面或は表面等に設けたスイッチ7による操作を含ませることができる。 The display unit 4 can permit various operations using a touch panel system in addition to image display using a color liquid crystal panel or the like. Various operations by the user can include operations by a switch 7 provided on the peripheral surface or the surface of the communication terminal 1 in addition to the so-called touch operation using the display unit 4.
 したがって、以下の説明では、表示部4の表示機能に特化した説明の場合は表示部4として説明するとともに、特に限定しないタッチパネル操作及びスイッチ操作以外の操作全般は、図1(B)に示すように、制御部6に操作信号(命令信号)を出力する操作部8による操作として総称して説明する。 Therefore, in the following description, in the case of the description specialized for the display function of the display unit 4, it will be described as the display unit 4, and all operations other than the touch panel operation and the switch operation that are not particularly limited are shown in FIG. As described above, the operation by the operation unit 8 that outputs the operation signal (command signal) to the control unit 6 will be generically described.
 表示部4には、例えば、電話機能を実行するための電話アイコンAP1、E-Mail機能を実行するためのE-MailアイコンAP2、チャット機能を実行するためのチャットアイコンAP3、翻訳アプリケーションを起動させる翻訳アイコンAP4、といった各種機能アイコンが表示される。 The display unit 4 activates, for example, a telephone icon AP1 for executing a telephone function, an E-Mail icon AP2 for executing an E-Mail function, a chat icon AP3 for executing a chat function, and a translation application. Various function icons such as a translation icon AP4 are displayed.
 記憶部5は、所謂ROM(Read only memory)やRAM(Random access memory)であり、制御部6とでコンピュータ機能を実現する。 The storage unit 5 is a so-called ROM (Read only memory) or RAM (Random access memory), and realizes a computer function with the control unit 6.
 また、記憶部5には、一次記憶等の利用形態に応じたデータの保管場所として、内部ストレージの他、各種メディアを用いた着脱可能な外部ストレージ、インターネット接続によるクラウドCを含むオンラインストレージ、等を含ませることができる。 In addition, the storage unit 5 stores internal data, removable external storage using various media, online storage including cloud C by Internet connection, etc. as a storage location for data according to the usage form such as primary storage, etc. Can be included.
 また、記憶部5には、例えば、電気通信回線(インターネット接続回線)NTを通じてダウンロードした翻訳アプリケーション(以下、アプリケーションを単に「アプリ」とも称する)の実行プログラム(以下、「翻訳ドライバ」と称する場合もある)を予め記憶している。 In addition, the storage unit 5 may be referred to as an execution program (hereinafter referred to as “translation driver”) of a translation application (hereinafter simply referred to as “application”) downloaded through a telecommunication line (Internet connection line) NT, for example. Is stored in advance.
 また、記憶部5には、電話機能による発呼や着呼の際に表示部4に相手先の氏名等を表示させるためのアドレス情報(アドレス帳)をテーブル方式で記憶している。 Further, the storage unit 5 stores address information (address book) for displaying the name of the other party on the display unit 4 in a table system when a call is made or received by the telephone function.
 このアドレス情報には、相手先を特定するための氏名の他、電話機能であれば相手先の電話番号、インターネット接続によるE-Mail機能であれば相手先のE-Mailアドレス(ドメイン名を含む)、コミュニティ型会員制のサービスであるSNS機能であればサービス提供元への会員登録情報(個人情報保護のもとでの適正な会員公開データ)を含ませることができる。 In this address information, in addition to the name for identifying the other party, the telephone number of the other party in the case of a telephone function, and the E-Mail address (including the domain name) of the other party in the case of an E-Mail function by Internet connection ), An SNS function that is a community-based membership system service can include member registration information (appropriate member public data under personal information protection) with the service provider.
 これらの個人情報は、適正かつ各機能を利用する際の必須の情報であり、かつ、国や言語を特定することができる情報となっている。 These pieces of personal information are appropriate and indispensable information when using each function, and can identify the country and language.
 図1(B)に示すように、通信端末1の制御部6は、例えば、内蔵アンテナ(図示せず)によって無線電波を受信した際の受信データを受信インターフェース(I/F)9から受信する。同様に、制御部6は、内蔵アンテナによって無線電波を送信する際の送信データを送信インターフェース(I/F)10に送信する。 As shown in FIG. 1B, the control unit 6 of the communication terminal 1 receives, from the reception interface (I / F) 9, reception data when a radio wave is received by a built-in antenna (not shown), for example. . Similarly, the control unit 6 transmits transmission data when transmitting a radio wave by the built-in antenna to the transmission interface (I / F) 10.
 制御部6は、翻訳ドライバを用いた翻訳処理のための、音声認識部11、音声生成部12、文字認識部13、文字生成部14、を制御するとともに音声データ又はテキストデータを送受信する。また、制御部6は、通信端末1の表面側(及び/又は裏面側)に設けた撮影カメラ15によって撮影した画像(例えば、JPG方式など)や映像(動画又は音声付動画:MP4など)を、記憶部5に記憶する際、あるいは、後述するTV電話或いは手話翻訳に利用する際、の画像制御機能を具備している。 The control unit 6 controls the speech recognition unit 11, the speech generation unit 12, the character recognition unit 13, and the character generation unit 14 for translation processing using a translation driver, and transmits / receives voice data or text data. In addition, the control unit 6 displays an image (for example, JPG method or the like) or video (moving image or moving image with sound: MP4 or the like) photographed by the photographing camera 15 provided on the front side (and / or the back side) of the communication terminal 1. , An image control function is provided for storing in the storage unit 5 or for use in a TV phone or sign language translation described later.
 音声認識部11には、音声入力部2にユーザが入力した音声に基づく音声データ、或は、受信インターフェース9を介して受信した音声データが入力される。音声認識部11は、その音声データを解析する読み取り機能を有する。 The voice recognition unit 11 receives voice data based on the voice input by the user to the voice input unit 2 or voice data received via the reception interface 9. The voice recognition unit 11 has a reading function for analyzing the voice data.
 音声生成部12は、音声出力部3から出力する音声データ、或は、送信インターフェース10から送信する音声データを、音声認識部11から出力した音声データ又は文字認識部13から出力したテキストデータに基づいて生成する。音声生成部12は、その生成した音声データを音声出力部3から出力する読み上げ機能を有する。 The voice generation unit 12 uses voice data output from the voice output unit 3 or voice data transmitted from the transmission interface 10 based on voice data output from the voice recognition unit 11 or text data output from the character recognition unit 13. To generate. The voice generation unit 12 has a reading function for outputting the generated voice data from the voice output unit 3.
 文字認識部13には、表示部4にユーザがタッチ入力した文字入力に基づくテキストデータ、或は、受信インターフェース9を介して受信したテキストデータが入力される。文字認識部13は、そのテキストデータの文字データコードを認識する。 The text recognition unit 13 receives text data based on a character input touched by the user on the display unit 4 or text data received via the reception interface 9. The character recognition unit 13 recognizes the character data code of the text data.
 文字生成部14は、表示部4に表示するテキストデータ、或は、送信インターフェース10から送信するテキストデータを、音声認識部11から出力した音声データ又は文字認識部13から出力したテキストデータに基づいて文字データコードを生成する。 The character generation unit 14 generates text data to be displayed on the display unit 4 or text data to be transmitted from the transmission interface 10 based on voice data output from the speech recognition unit 11 or text data output from the character recognition unit 13. Generate a character data code.
 撮影カメラ15は、例えば、自撮りモードにより通信端末1の表面側を撮影することができるようになっている。なお、カメラ機能そのものは公知であるため、ここでは、その詳細な説明は省略する。 The photographing camera 15 can photograph the front side of the communication terminal 1 in the self-shooting mode, for example. In addition, since the camera function itself is well-known, the detailed description is abbreviate | omitted here.
 制御部6は、撮影カメラ15で撮影した映像を手話解析部16によって解析させることができる。 The control unit 6 can cause the sign language analysis unit 16 to analyze the video imaged by the imaging camera 15.
 手話解析部16は、撮影カメラ15で撮影した手話に関する動作を解析し、例えば、文字データや音声データを生成する機能を有している。 The sign language analysis unit 16 has a function of analyzing operations related to sign language photographed by the photographing camera 15 and generating, for example, character data and voice data.
 また、手話解析部16は、撮影カメラ15で撮影した手話に関する動作を解析し、手や顔の画像を含む簡易動作画像等の代替え画像データを生成することも可能である。なお、この文字・音声・画像の各データの生成の詳細な例は後述する。 Also, the sign language analysis unit 16 can analyze the motion related to the sign language photographed by the photographing camera 15 and generate substitute image data such as a simple motion image including a hand image and a face image. A detailed example of the generation of the character / sound / image data will be described later.
 このような基本構成において、制御部6は、記憶部5に記憶したオペレーティングシステムで設定した言語(例えば、日本語)をユーザが通常利用する基本言語とし、翻訳ドライバにより記憶部5に記憶したアドレス情報の個人情報に含まれる通信相手が使用する言語(例えば、英語)を翻訳対象とする他言語とする。 In such a basic configuration, the control unit 6 sets the language (for example, Japanese) set by the operating system stored in the storage unit 5 as a basic language that is normally used by the user, and stores the address stored in the storage unit 5 by the translation driver. The language (for example, English) used by the communication partner included in the personal information of the information is set as another language to be translated.
 なお、他言語は1つの言語に限定されず、翻訳ドライバの種類に応じた言語とすることができる。したがって、翻訳ドライバは記憶部5に記憶した一つ以上を利用することができる。 Note that the other language is not limited to one language and can be a language according to the type of translation driver. Therefore, the translation driver can use one or more stored in the storage unit 5.
 図2に示すように、制御部6は、ユーザが通信端末1の表示部4に表示しているチャットアイコンAP3を選択した場合、チャット機能を立ち上げて表示部4にチャット画面を表示させる。 As shown in FIG. 2, when the user selects the chat icon AP3 displayed on the display unit 4 of the communication terminal 1, the control unit 6 activates the chat function and causes the display unit 4 to display a chat screen.
 この際、相手先はチャット機能に付随している相手先選択画面から相手先を選択すれば、基本言語を使用する相手なのか他言語を使用する相手なのかは自動で識別することができる。これにより、制御部7は、相手先が他言語を使用するユーザである場合、その対応した使用言語の他言語への翻訳機能を実行する。 At this time, if the other party selects the other party from the other party selection screen attached to the chat function, it is possible to automatically identify whether the other party uses the basic language or another party. Thereby, when the other party is a user who uses another language, the control unit 7 executes a function of translating the corresponding used language into another language.
 以下の説明において、相手先が利用する通信端末21も、音声入力部22及び音声出力部23と、表示機能を実現するための表示部24と、を含むスマートフォンである。この際、通信端末21に翻訳アプリケーションはインストールされていないものとする。 In the following description, the communication terminal 21 used by the other party is also a smartphone including an audio input unit 22 and an audio output unit 23, and a display unit 24 for realizing a display function. At this time, it is assumed that no translation application is installed in the communication terminal 21.
 また、以下の説明においては、説明の便宜上、音声や文字等の言葉上の入出力に関しては『○○』で表記し、画面表示上の入出力に関しては[○○]で表記し、データ処理上の入出力に関しては<○○>で表記する。 Also, in the following explanation, for convenience of explanation, input / output on words such as voice and characters is indicated by “XX”, input / output on screen display is indicated by [XX], and data processing The above input / output is indicated by <XX>.
 また、テキストデータは文字コード化したものであるため、以下の説明では、単なる「コード」或は「文字コード」とも称する。また、音声データ又はテキストデータを単に「音声」又は「テキスト」(又は表示用に変換した「文字」)とも称する。 In addition, since the text data is a character code, it is also referred to as a simple “code” or “character code” in the following description. The voice data or text data is also simply referred to as “voice” or “text” (or “character” converted for display).
 ここで、ユーザは、表示部4の画面下に表示した文字入力パッドを参照してタッチ操作を行い、例えば、『調子どお?』と入力すると、表示部4の画面右上に枠付きで文字[調子どお?]を表示させる。なお、この際の文字コードのデータ処理は公知の処理を利用している。 Here, the user refers to the character input pad displayed at the bottom of the screen of the display unit 4 and performs a touch operation. For example, “Tone Doo? ”Is entered in the upper right of the screen of the display unit 4 with a frame [a tone? ] Is displayed. The character code data processing at this time uses known processing.
 一方、制御部6は、文字認識部13により入力した文字コードを認識させ、認識した文字コードから翻訳ドライバを用いて、他言語の文字コード(例えば、<How have you been>)を生成する。制御部6は、生成した文字コード<How have you been>を、送信インターフェース10を介して相手先の通信端末21に送信する。 On the other hand, the control unit 6 recognizes the character code input by the character recognition unit 13, and generates a character code (for example, <How have you > been>) of another language from the recognized character code using a translation driver. The control unit 6 transmits the generated character code <How have you been> to the communication terminal 21 of the other party via the transmission interface 10.
 相手先の通信端末21(の制御部)は、受信した文字コード<How have you been>に基づいて、チャット方式で表示部24の画面左上側に文字[How have you been]を表示する。 The partner communication terminal 21 (the control unit thereof) displays the character [How [have you been] on the upper left side of the screen of the display unit 24 in a chat mode based on the received character code <How have you been>.
 これを受けて、相手先は、返事となる文字『so and so』を入力すると、相手先が所有する通信端末21の表示部24の画面右上に文字[so and so]が表示されるともに、文字コード<so and so>が送信される。 In response to this, when the partner inputs the reply character “so and so”, the character [so and so] is displayed at the upper right of the screen of the display unit 24 of the communication terminal 21 owned by the partner, The character code <so and so> is sent.
 文字コード<so and so>を受信した通信端末1の制御部6は、文字コード<so and so>を文字認識部13で認識させた後、文字生成部14で翻訳化した文字コード<まぁまぁ>を生成する。さらに、制御部7は、表示部4の画面左上に文字[まぁまぁ]を表示させる。 The control unit 6 of the communication terminal 1 that has received the character code <so and so> causes the character recognition unit 13 to recognize the character code <so and so>, and then translates the character code <maaama>. > Is generated. Further, the control unit 7 displays a character [maama] on the upper left of the screen of the display unit 4.
 このように、通信端末1と通信端末21との間で、テキストデータ-テキストデータの相互通信に関し、相手先の使用言語が他言語であって、相手先の通信端末21に翻訳アプリケーションがインストールされていない場合であっても、コミュニケーションを容易に確立することができる。 As described above, regarding the mutual communication between the text data and the text data between the communication terminal 1 and the communication terminal 21, the language used by the partner is another language, and the translation application is installed in the partner communication terminal 21. Even if not, communication can be easily established.
 このような基本的な翻訳機能を可能とする制御部6は、上記に加え、自身の通信端末1を用いた基本言語による音声又は文字の何れか一方の入力だけで、他言語に翻訳した音声出力並びに文字表記を相手先の通信端末21に実現させることができる。 In addition to the above, the control unit 6 that enables such a basic translation function includes a speech translated into another language by simply inputting either speech or characters in the basic language using its own communication terminal 1. Output and character notation can be realized in the communication terminal 21 of the other party.
 また、制御部6は、相手先が通信端末21を用いた他言語による音声又は文字の何れか一方の入力だけで、基本言語に翻訳した音声出力並びに文字表記を自身の通信端末1に実現させることができる。 In addition, the control unit 6 causes the communication terminal 1 to realize the voice output and the character notation translated into the basic language only by inputting either one of speech or characters in another language using the communication terminal 21 by the other party. be able to.
 以下、通信端末1における音声入力の場合(図3参照)及び通信端末1における文字入力の場合(図4参照)の一例を説明する。ここで、通信端末1は、音声付きチャットアプリ又は文字表示付き電話アプリを用いるものとする。なお、通常の電話アプリ(電話アイコンAP1)に文字出力機能付き翻訳をモード的に追加プログラムしてもよい。 Hereinafter, an example of voice input in the communication terminal 1 (see FIG. 3) and character input in the communication terminal 1 (see FIG. 4) will be described. Here, the communication terminal 1 uses a chat application with voice or a telephone application with character display. Note that a translation with a character output function may be additionally programmed in a mode to a normal telephone application (phone icon AP1).
 同様に、通常のチャットアプリ(チャットアイコンAP3)に音声機能付き翻訳をモード的に追加プログラムしてもよい。また、制御部6は、ユーザがこれらのアプリケーションを立ち上げて相手先を指定すると、そのアドレスデータに記憶した情報に基づいて翻訳をする旨を決定しているものとする。 Similarly, a normal chat application (chat icon AP3) may be programmed to add translation with voice function in a mode. Further, it is assumed that when the user starts up these applications and designates a partner, the control unit 6 determines to translate based on information stored in the address data.
 図3に示すように、制御部6は、ユーザが音声入力部2に向けて『調子どお』と音声入力すると、音声認識部11にその音声を読み取らせて音声<調子どお>を生成する。 As shown in FIG. 3, when the user inputs “tone” to the voice input unit 2, the control unit 6 causes the voice recognition unit 11 to read the voice and generate a voice <tone>. To do.
 次に、制御部6は、読み取った音声<調子どお>に基づいて、文字生成部14により文字<調子どお>を生成させて表示部4に文字[調子どお]を表示させるとともに、翻訳した文字<How have you been>を生成させて表示部4に文字[How have you been]を表示させる。また、制御部6は、生成した文字<How have you been>に基づいて、音声生成部12により音声<How have you been>を生成させる。 Next, the control unit 6 causes the character generation unit 14 to generate a character <tone> on the basis of the read voice <tone> and causes the display unit 4 to display the character [tone]. The translated character <How have you been> is generated and the character [How have you been] is displayed on the display unit 4. Further, the control unit 6 causes the voice generation unit 12 to generate a voice <How have you been> based on the generated character <How have you been>.
 これにより、制御部6は、生成した、文字<調子どお>、文字<How have you been>、音声<How have you been>、の3つのデータを重畳して送信インターフェース10から通信端末21に向けて送信する。 Thereby, the control unit 6 superimposes the generated three data of the character <tone>, the character <How < have you been>, and the voice <How have you been> from the transmission interface 10 to the communication terminal 21. Send to.
 通信端末21は、受信したデータに基づいて、翻訳前の文字[調子どお]と翻訳後の文字[How have you been]を表示部24に表示させるとともに、音声『How have you been』を音声出力部23から出力させる。 Based on the received data, the communication terminal 21 displays the pre-translation characters [Tone Doo] and the post-translation characters [How have you been] on the display unit 24 and the voice “How have you been”. Output from the output unit 23.
 このように、通信端末21では、翻訳アプリをインストールしていなくても、文字<調子どお>、文字<How have you been>、音声<How have you been>、の3つのデータを取得することができる。 In this way, the communication terminal 21 can acquire three data of the character <tone>, the character <How have you been>, and the voice <How have you been> without installing the translation application. Can do.
 一方、通信端末21のユーザは、音声『How have you been』に対して、音声入力部22に『so and so』と音声入力すると、その音声<so and so>を通信端末1に向けて送信する。 On the other hand, when the user of the communication terminal 21 inputs “so and so” to the voice input unit 22 in response to the voice “How have you been”, the voice <so and so> is transmitted to the communication terminal 1. To do.
 通信端末1の制御部6は、受信インターフェース9で音声<so and so>を受信させると、音声認識部11により音声<so and so>を認識させる。 The control unit 6 of the communication terminal 1 causes the voice recognition unit 11 to recognize the voice <so 受 信 and so> when the reception interface 9 receives the voice <so and so>.
 これにより、制御部6は、文字生成部14により文字<so and so>を生成させて表示部4に文字[so and so]を表示させるとともに、翻訳した文字<まぁまぁ>を生成させて表示部4に文字[まぁまぁ]を表示させる。また、制御部6は、生成した文字<まぁまぁ>に基づいて、音声生成部12により音声<まぁまぁ>を生成させ、音声出力部3から音訳した音声『まぁまぁ』を出力させる。 Thereby, the control unit 6 causes the character generation unit 14 to generate the character <so < and so> and causes the display unit 4 to display the character [so] and so], and also generates and displays the translated character <maamaa>. The character “maamaa” is displayed in part 4. Further, the control unit 6 causes the voice generation unit 12 to generate a voice <maamaa> based on the generated character <maaama>, and causes the voice output unit 3 to output the transliterated voice “maaamaa”.
 このように、通信端末1は、通信端末21に翻訳アプリをインストールしていなくても、翻訳前の文字<so and so>、翻訳後の文字<まぁまぁ>、翻訳後の音声<まぁまぁ>、の3つのデータを生成し、それぞれ表示部4及び音声出力部3に展開してユーザに認識させることができる。なお、これら3つの各データは、通信端末21に翻訳アプリがインストールされていれば、通信端末21が生成してもよい。 As described above, the communication terminal 1 does not have the translation application installed on the communication terminal 21, so that the pre-translation character <so and so>, the post-translation character <ma-ama>, and the post-translation voice <ma-ama> Can be generated on the display unit 4 and the audio output unit 3 to be recognized by the user. Note that these three pieces of data may be generated by the communication terminal 21 as long as the translation application is installed in the communication terminal 21.
 図4に示すように、制御部6は、ユーザが表示部4に表示した文字入力パッドを用いて『調子どお』と文字入力すると、文字認識部13にその文字コードを認識させる。 As shown in FIG. 4, when the user inputs a character “tone” using the character input pad displayed on the display unit 4, the control unit 6 causes the character recognition unit 13 to recognize the character code.
 次に、制御部6は、文字<調子どお>に基づいて、文字生成部14により文字<調子どお>を生成させて表示部4に文字[調子どお]を表示させるとともに、翻訳した文字<How have you been>を生成させて表示部4に文字[How have you been]を表示させる。 Next, based on the character <tone>, the control unit 6 causes the character generator 14 to generate the character <tone> and display the character [tone] on the display unit 4 and translate it. The character <How have you been> is generated and the character [How have you been] is displayed on the display unit 4.
 また、制御部6は、生成した文字<How have you been>に基づいて、音声生成部12により音声<How have you been>を生成させる。 Also, the control unit 6 causes the voice generation unit 12 to generate a voice <How have you been> based on the generated character <How have you been>.
 これにより、制御部6は、生成した、文字<調子どお>、文字<How have you been>、音声<How have you been>、の3つのデータを重畳して送信インターフェース10から通信端末21に向けて送信する。 Thereby, the control unit 6 superimposes the generated three data of the character <tone>, the character <How < have you been>, and the voice <How have you been> from the transmission interface 10 to the communication terminal 21. Send to.
 通信端末21は、受信したデータに基づいて、翻訳前の文字[調子どお]と翻訳後の文字[How have you been]を表示部24に表示させるとともに、音声『How have you been』を音声出力部23から出力させる。 Based on the received data, the communication terminal 21 displays the pre-translation characters [Tone Doo] and the post-translation characters [How have you been] on the display unit 24 and the voice “How have you been”. Output from the output unit 23.
 このように、通信端末21では、翻訳アプリをインストールしていなくても、文字<調子どお>、文字<How have you been>、音声<How have you been>、の3つのデータを取得することができる。 In this way, the communication terminal 21 can acquire three data of the character <tone>, the character <How have you been>, and the voice <How have you been> without installing the translation application. Can do.
 一方、通信端末21のユーザは、音声『How have you been』に対して、音声入力部22に『so and so』と音声入力(又は文字『so and so』と入力)すると、その音声<so and so>(又は文字<so and so>)を通信端末1に向けて送信する。 On the other hand, when the user of the communication terminal 21 inputs “so and so” to the voice input unit 22 (or inputs the characters “so に and so”) to the voice “How have you been”, the voice <so and so> (or characters <so and so>) are transmitted to the communication terminal 1.
 通信端末1の制御部6は、受信インターフェース9で音声<so and so>(又は文字<so and so>)を受信させると、音声認識部11により音声<so and so>(又は文字認識部13により文字<so and so>)を認識させる。 When the control unit 6 of the communication terminal 1 receives the voice <so and so> (or the character <so and so>) by the reception interface 9, the voice recognition unit 11 causes the voice <so and so> (or the character recognition unit 13). To recognize the characters <so and so>).
 これにより、制御部6は、文字生成部14により文字<so and so>を生成させて表示部4に文字[so and so]を表示させるとともに、翻訳した文字<まぁまぁ>を生成させて表示部4に文字[まぁまぁ]を表示させる。 Thereby, the control unit 6 causes the character generation unit 14 to generate the character <so < and so> and causes the display unit 4 to display the character [so] and so], and also generates and displays the translated character <maamaa>. The character “maamaa” is displayed in part 4.
 また、制御部6は、生成した文字<まぁまぁ>に基づいて、音声生成部12により音声<まぁまぁ>を生成させ、音声出力部3から翻訳した音声『まぁまぁ』を出力させる。 In addition, the control unit 6 causes the voice generation unit 12 to generate a voice <maamaa> based on the generated character <maamaa>, and causes the voice output unit 3 to output the translated voice “maaama”.
 このように、通信端末1は、通信端末21に翻訳アプリをインストールしていなくても、翻訳前の文字<so and so>、翻訳後の文字<まぁまぁ>、翻訳後の音声<まぁまぁ>、の3つのデータを生成し、それぞれ表示部4及び音声出力部3に展開してユーザに認識させることができる。なお、これら3つの各データは、通信端末21に翻訳アプリがインストールされていれば、通信端末21が生成してもよい。 As described above, the communication terminal 1 does not have the translation application installed on the communication terminal 21, so that the pre-translation character <so and so>, the post-translation character <ma-ama>, and the post-translation voice <ma-ama> Can be generated on the display unit 4 and the audio output unit 3 to be recognized by the user. Note that these three pieces of data may be generated by the communication terminal 21 as long as the translation application is installed in the communication terminal 21.
 次に、制御部7で実行する具体的なルーチンを図5のフロー図に基づいて説明する。 Next, a specific routine executed by the control unit 7 will be described based on the flowchart of FIG.
 上述したように、通信端末1は、記憶部5に基本言語と他言語との相互翻訳を実行する翻訳アプリケーションを予めインストールしているとともに、アドレス帳を記憶している。 As described above, the communication terminal 1 is preinstalled with a translation application that executes mutual translation between the basic language and another language in the storage unit 5 and stores an address book.
 (ステップS1)
 ステップS1において、制御部6は、ユーザによる操作部8の操作による通信アプリケーションの指定を受けて、その通信アプリケーションに対応したアドレス帳を記憶部5から参照し、表示部4にアドレス帳を表示させ、ステップS2へと移行する。
(Step S1)
In step S <b> 1, the control unit 6 receives the designation of the communication application by the operation of the operation unit 8 by the user, refers to the address book corresponding to the communication application from the storage unit 5, and causes the display unit 4 to display the address book. The process proceeds to step S2.
 (ステップS2)
 ステップS2において、制御部6は、ユーザの操作により発呼先(通信相手)の指定操作があったか否かの判定をする。制御部6は、発呼先の指定があったと判定した場合には(Yes)ステップS3へと移行する。
(Step S2)
In step S2, the control unit 6 determines whether or not a call destination (communication partner) designation operation has been performed by a user operation. When it is determined that the call destination has been designated (Yes), the control unit 6 proceeds to step S3.
 一方、制御部6は、発呼先の指定があったと判定しなかった場合(No)には、発呼先の指定があるまで(又は他のアプリケーション等の操作があるまで)引き続きこのルーチンを監視する。そして、制御部6は、指定した発呼先との通信(ここでは電話機能)を確立させる。 On the other hand, if the control unit 6 does not determine that the call destination has been specified (No), the control unit 6 continues this routine until the call destination is specified (or until another application is operated). Monitor. Then, the control unit 6 establishes communication (here, a telephone function) with the designated call destination.
 (ステップS3)
 ステップS3において、制御部6は、ユーザが指定した発呼先により、発呼先を特定するとともに発呼先に対応したアドレス情報を解析し、ステップS4へと移行する。
(Step S3)
In step S3, the control unit 6 specifies the call destination based on the call destination designated by the user, analyzes the address information corresponding to the call destination, and proceeds to step S4.
(ステップS4:他言語特定ステップ)
 ステップS4において、制御部6は、解析したアドレス情報に基づいて、通信相手が使用する言語が基本言語であるか他言語であるかを含め国又は言語を特定し、ステップS5へと移行する。
(Step S4: Step for identifying other languages)
In step S4, the control unit 6 specifies the country or language including whether the language used by the communication partner is a basic language or another language based on the analyzed address information, and proceeds to step S5.
 なお、以下の説明では、このステップS4にて通信相手が使用する言語が他言語である旨を特定したものとして説明する。 In the following description, it is assumed that the language used by the communication partner in step S4 is specified as another language.
 (ステップS5:通信内容受付ステップ)
 ステップS5において、制御部6は、基本言語による音声データ又はテキストデータの少なくとも何れか一方の入力を入力手段としての音声入力部2又は表示部4を介して発信基準データとして受け付け、ステップS6へと移行する。
(Step S5: Communication content reception step)
In step S5, the control unit 6 accepts at least one input of voice data or text data in the basic language as transmission reference data via the voice input unit 2 or the display unit 4 as input means, and the process proceeds to step S6. Transition.
 (ステップS6:他言語化翻訳ステップ)
 ステップS6において、制御部6は、翻訳アプリケーションを起動させて、ステップS5で受け付けた発信基準データを他言語の音声データ及びテキストデータに翻訳して発信データを生成し、ステップS7へと移行する。
(Step S6: Translation into other languages)
In step S6, the control unit 6 activates the translation application, translates the transmission reference data received in step S5 into voice data and text data in other languages, generates transmission data, and proceeds to step S7.
 なお、上記ステップS5において、音声入力部2からの音声入力であった場合、制御部6は、音声入力部2で受け付けた発信基準データである基本言語による音声データとして認識し、記憶部5に予めインストールした聞き取りアプリケーションを起動させる。 If the voice input from the voice input unit 2 is input in step S5, the control unit 6 recognizes the voice data in the basic language, which is the transmission reference data received by the voice input unit 2, and stores it in the storage unit 5. Start the listening application installed in advance.
 そして、このステップS6において、制御部6は、他言語によるテキストデータからなる発信データを生成するための基本言語による基本テキストデータを生成する基本テキストデータ生成ステップを実行することとなる。 And in this step S6, the control part 6 will perform the basic text data production | generation step which produces | generates the basic text data by the basic language for producing | generating the transmission data which consists of text data by another language.
 一方、上記ステップS5において、表示部4からの文字入力であった場合、制御部6は、表示部4から受け付けた発信基準データである基本言語によるテキストデータとして認識し、記憶部5に予めインストールした読み上げアプリケーションを起動させる。 On the other hand, if the character input is from the display unit 4 in step S5, the control unit 6 recognizes the text data in the basic language, which is the transmission reference data received from the display unit 4, and is installed in the storage unit 5 in advance. Start the reading application.
 そして、このステップS6において、制御部6は、他言語による音声データからなる発信データを生成するための基本言語による基本音声データを生成する基本音声データ生成ステップを実行することとなる。 And in this step S6, the control part 6 will perform the basic audio | voice data production | generation step which produces | generates the basic audio | voice data by the basic language for producing | generating the transmission data which consist of audio | voice data by another language.
 (ステップS7:発信ステップ)
 ステップS7において、制御部6は、ステップS6で翻訳した発信データを電気通信回線を通じて発信手段としての送信インターフェース10から発信し、ステップS8へと移行する。
(Step S7: Transmission step)
In step S7, the control unit 6 transmits the transmission data translated in step S6 from the transmission interface 10 as a transmission means through the telecommunication line, and proceeds to step S8.
 なお、制御部6は、このステップS7において、表示部4により受け付けた発信基準データが基本言語によるテキストデータである場合に、翻訳した他言語の音声データによる発信データに加えて基本言語によるテキストデータを重畳して発信してもよい。 In addition, in step S7, when the transmission reference data received by the display unit 4 is text data in the basic language, the control unit 6 adds the text data in the basic language in addition to the transmission data in the speech data in the other languages that have been translated. May be transmitted in a superimposed manner.
 なお、ユーザは、ステップS2において相手先との通信状態(通話状態)が確立しているため、ステップS3~ステップS7の発信が完了するまで(言葉を言い尽くすか相手先がデータを受信するまで)スイッチ7を押下して相手からの割り込みができないようにするキャンセルステップを実行(受け付け)してもよい。これにより、音声の混濁(翻訳の混濁)を抑制することができる。 Note that since the communication state (call state) with the other party is established in step S2, the user completes the transmission in steps S3 to S7 (until the word is exhausted or the other party receives data). ) A cancel step may be executed (accepted) to depress the switch 7 and prevent interruption from the other party. As a result, voice turbidity (translation turbidity) can be suppressed.
 (ステップS8)
 次に、制御部6は、相手先からの受信があったか否かを判定する。なお、制御部6は、引き続きステップS5のルーチンによりユーザからの通信内容を割り込みとして受け付けることも可能である。制御部6は、相手先から受信があったと判定した場合(Yes)には、ステップS9へと移行する。
(Step S8)
Next, the control unit 6 determines whether or not there is reception from the other party. In addition, the control part 6 can also receive the communication content from a user as interruption by the routine of step S5 continuously. If it is determined that there is a reception from the other party (Yes), the control unit 6 proceeds to step S9.
 一方、制御部6は、相手先から受信があったと判定しなかった場合(No)には、引き続きこのルーチン若しくはステップS5の割り込みを監視する。 On the other hand, if the control unit 6 does not determine that there is a reception from the other party (No), the control unit 6 continues to monitor this routine or the interruption in step S5.
 (ステップS9:受信ステップ)
 ステップS9において、制御部6は、通信相手からの他言語による音声データ又はテキストデータを電気通信回線を通じて受信手段としての受信インターフェース9により受信基準データとして受信し、ステップS10へと移行する。
(Step S9: Reception step)
In step S9, the control unit 6 receives voice data or text data in another language from the communication partner as reception reference data by the reception interface 9 as a reception unit through the telecommunication line, and proceeds to step S10.
 (ステップS10)
 ステップS10において、制御部6は、受信データがテキストデータであるのか音声データであるのかを特定し、ステップS11へと移行する。
(Step S10)
In step S10, the control unit 6 specifies whether the received data is text data or voice data, and proceeds to step S11.
 (ステップS11)
 ステップS11において、制御部6は、受信データに基づいて、出力方式、すなわち、文字解析を実行するのか音声解析を実行するのかを決定し、ステップS12へと移行する。
(Step S11)
In step S11, the control unit 6 determines an output method, that is, whether to perform character analysis or voice analysis, based on the received data, and proceeds to step S12.
 (ステップS12:基本言語化翻訳ステップ)
 ステップS12において、制御部6は、翻訳アプリケーションを起動させてステップS9で受信した受信基準データを基本言語による音声データ及びテキストデータに翻訳して受信データを生成し、ステップS13へと移行する。
(Step S12: Basic verbalization translation step)
In step S12, the control unit 6 activates the translation application, translates the reception reference data received in step S9 into speech data and text data in a basic language, generates reception data, and proceeds to step S13.
 (ステップS13:通信内容出力ステップ)
 ステップS13において、制御部6は、ステップS12で翻訳した受信データを出力手段としての音声出力部3及び表示部4を介して音声出力並びに文字出力してこのルーチンを終了する。なお、制御部6は、以降、通信(通話)が終了するまで上記ルーチンを繰り返す。
(Step S13: Communication content output step)
In step S13, the control unit 6 outputs the received data translated in step S12 through the audio output unit 3 and the display unit 4 as output means and outputs the characters and ends this routine. The control unit 6 thereafter repeats the above routine until the communication (call) is completed.
 このように、本実施の形態に係る翻訳処理方法、翻訳処理プログラム、及び、記録媒体は、通信端末1に入力した音声又は文字の何れか一方に対して、通信相手の通信端末21に翻訳した音声と文字とを出力させることができる。 As described above, the translation processing method, the translation processing program, and the recording medium according to the present embodiment are translated into the communication terminal 21 of the communication partner with respect to either the voice or the character input to the communication terminal 1. Voice and text can be output.
 また、通信相手の通信端末から音声又は文字の何れか一方の送信に対して、通信端末1に翻訳した音声と文字とを出力させることができる。 Also, in response to transmission of either voice or characters from the communication terminal of the communication partner, the translated voice and characters can be output to the communication terminal 1.
 ここで、通信端末1及び通信端末21の基本言語である翻訳前の文字を互いに送受信することもできる。 Here, the characters before translation, which is the basic language of the communication terminal 1 and the communication terminal 21, can be transmitted and received with each other.
 また、以上既に述べた以外にも、上記実施の形態による手法を適宜組み合わせて利用しても良い。 In addition to those already described above, the methods according to the above embodiments may be used in appropriate combination.
 その他、本発明は、その趣旨を逸脱しない範囲内において、種々の変更が加えられて実施されるものである。 In addition, the present invention is implemented with various modifications within a range not departing from the gist thereof.
 例えば、手話は手指動作と非手指動作を同時に行う視覚言語であり、音声言語と並ぶ言語として認識されている。したがって、この手話による視覚言語を基本言語として、文字データや音声データ、或いは、代替え画像データに変換することで、他言語との相互翻訳を実現することができる。 For example, sign language is a visual language that performs finger movement and non-finger movement at the same time, and is recognized as a language along with speech language. Therefore, mutual translation with other languages can be realized by converting the visual language based on this sign language into the basic language and converting it into character data, voice data, or alternative image data.
 この際、他言語には、母国語を文字・音声・画像に変換する場合のほか、その母国語の文字・音声・画像を基準としてさらに他国語に再変換する場合の翻訳を含ませることができる。 In this case, other languages may include translations in the case of converting the native language into characters, sounds, and images, as well as re-converting into other languages based on the characters, sounds, and images of the native language. it can.
 また、手話の動作形態は、各国によって異なるため、言語と同様の翻訳が必要となる場合がある。したがって、記憶部5には、言語と同様に手話動作に関しても翻訳可能なデータが記憶されている。 In addition, since the sign language behavior varies from country to country, translation similar to the language may be required. Accordingly, the storage unit 5 stores data that can be translated with respect to the sign language action as well as the language.
 手話は、熟練度にもよるが、会話よりも相対的に遅いものの、文字入力よりも早い場合が多い。したがって、手話動作を解析して文字データや音声データを生成することにより、利用者によっては利便性を向上することができる。 Sign language, although depending on the level of proficiency, is relatively slower than conversation but is often faster than character input. Therefore, convenience can be improved for some users by analyzing the sign language action and generating character data and voice data.
 上述した「手や顔の画像を含む簡易動作画像等の代替え画像データ」とは、手話動作の動画をそのまま利用したのでは、即応性は確保することができるものの、通信データ量が増大してしまう。したがって、手話動作の動画をそのまま利用してもよいが、通信データ量を少なくするために、表示部4の画面中にできるだけわかりやすく表現した代替え画像を表示部4に表示する代替え画像データに置き換えるのが望ましい。 The above-mentioned “substitute image data such as a simple motion image including hand and face images” can be used as it is, but the responsiveness can be secured, but the amount of communication data increases. End up. Accordingly, the sign language motion video may be used as it is. However, in order to reduce the amount of communication data, the substitute image expressed in the screen of the display unit 4 as easily as possible is replaced with the substitute image data displayed on the display unit 4. Is desirable.
 なお、現実的には、手話を必要とする利用者は、視覚は有しているため、少なくとも、言語入力時において手話動作を文字データに変換することができれば、出力時において上述した文字によるテキスト表示を行えば、問題なく相互理解の翻訳機能として成立する。 In reality, a user who requires sign language has vision, so at least if the sign language action can be converted into character data at the time of language input, the above-described text by characters at the time of output If displayed, it will be established as a translation function for mutual understanding without problems.
 そのうえで、音声データに変換して相手方に送信することができれば、相手方が健聴者である場合に、負担を少なくすることができる。 Furthermore, if it can be converted into voice data and transmitted to the other party, the burden can be reduced when the other party is a normal hearing person.
 さらに、手話を覚えてコミュニケーションを高めたい、あるいは、絵文字(スタンプ)的な利用をしたい(例えば、周知の「バイバイ」を模した手のひらを振っているような静止画像やJIF動画)、という場合に、上述した代替え画像データは便利である。 In addition, when you want to learn sign language and improve communication, or you want to use it like a pictograph (stamp) (for example, a still image or JIF movie that shakes the palm of a known "bye-bye") The above-described substitute image data is convenient.
 そこで、例えば、図1に示す、翻訳アイコンAP4を選択操作すると、図6に示すように、表示部4に翻訳モード選択メニューが表示される。 Therefore, for example, when the translation icon AP4 shown in FIG. 1 is selected, a translation mode selection menu is displayed on the display unit 4 as shown in FIG.
 表示部4に表示された翻訳モード選択メニューとしては、1つの通信端末1で実行する翻訳機能として、「1)オフライン対面翻訳」、「2)SNS翻訳」、「3)一般通話翻訳」、「4)会議同時翻訳」、「5)TV電話翻訳」、「6)手話翻訳」の6通りがある。 The translation mode selection menu displayed on the display unit 4 includes “1) offline face-to-face translation”, “2) SNS translation”, “3) general call translation”, “ There are 6 types: “4) Simultaneous translation of conference”, “5) Video phone translation” and “6) Sign language translation”.
 本実施の形態に示す1つの通信端末1は、音声入力用の音声入力部2、文字入力用の操作部8、映像入力用の撮影カメラ15、を入力手段として有する。また、1つの通信端末1は、音声出力用の音声出力部3、文字・画像出力用の表示部4、を出力手段として有する。さらに、1つの通信端末1は、基本言語と他言語とを相互に翻訳可能とするものである。 One communication terminal 1 shown in the present embodiment includes a voice input unit 2 for voice input, an operation unit 8 for character input, and a photographing camera 15 for video input as input means. Further, one communication terminal 1 has an audio output unit 3 for outputting audio and a display unit 4 for outputting characters / images as output means. Further, one communication terminal 1 is capable of translating a basic language and another language.
 すなわち、一つの通信端末11は、入力手段を介して受け付けた基本言語による入力データに対して、他言語化した後の出力データを生成すると同時に、その出力データを(出力手段に)出力することができると共に、他言語による入力データに対して、基本言語化した後の出力データ生成すると同時に、出力手段を介して出力することができる。 That is, one communication terminal 11 generates output data after being converted into another language for input data in a basic language received via the input means, and simultaneously outputs the output data (to the output means). In addition, it is possible to generate output data after making it into a basic language with respect to input data in another language, and at the same time, output it through the output means.
 したがって、当然に、2つの通信端末1,21を用いた通信方式ではなく、例えば、基本言語を使うユーザと、他言語を使うユーザと、が同じ場所にいて対話をする場合の対面翻訳の場合に、1つの通信端末1での利用が可能である。 Therefore, as a matter of course, it is not a communication method using two communication terminals 1 and 21, but for example, in the case of face-to-face translation when a user who uses a basic language and a user who uses another language are in the same place and have a conversation Furthermore, it can be used with one communication terminal 1.
 このように、「1)オフライン対面翻訳」は、1つの通信端末1を用いて2名以上の異国言語者同士の会話の際に用いることができる翻訳モードである。 As described above, “1) Offline face-to-face translation” is a translation mode that can be used when two or more foreign language speakers have a conversation using one communication terminal 1.
 また、1つの通信端末1は、電気通信回線としてのインターネット回線を通じてクラウド上のSNSを利用することができる。したがって、SNSとしての、例えば、タイムライン等において、クラウド上のアドレス帳等を利用して使用言語を特定することができることから、基本言語と他言語との相互翻訳機能を用いて、表示言語を相互翻訳することができる。 Further, one communication terminal 1 can use SNS on the cloud through an Internet line as a telecommunication line. Therefore, since the language used can be specified using the address book on the cloud as an SNS, for example, on the timeline, etc., the display language can be changed using the mutual translation function between the basic language and other languages. Mutual translation is possible.
 この際、基本言語と他言語とは、例えば、テキスト(手話画像等を含む)表示部分を画面タップすることにより翻訳後のテキストを翻訳してもよいし、例えば、通信端末1,21のオペレーティングシステムに応じて、テキスト表示全体を標示タイミング等で翻訳してもよい。 At this time, the basic language and the other language may translate the translated text by tapping a screen display portion of a text (including a sign language image or the like), for example, Depending on the system, the entire text display may be translated at the labeling timing or the like.
 すなわち、一つの通信端末11は、入力手段を介して受け付けた基本言語による入力データに対して、他言語化した後の出力データを生成すると同時に、その出力データを(クラウド上に)出力することができると共に、他言語による入力データに対して、基本言語化した後の出力データ生成すると同時に、出力手段(表示部4)を介して出力することができる。 That is, one communication terminal 11 generates output data after being converted into another language for input data in the basic language received via the input means, and at the same time outputs the output data (on the cloud). In addition, it is possible to generate output data after being converted into a basic language with respect to input data in another language, and at the same time, output it via the output means (display unit 4).
 このように、「2)SNS翻訳」は、例えば、インターネットで管理できるクラウド型のSNS会員名簿を用いて相互の言語を特定した翻訳を文字化や音声化にて行う際に用いることができる翻訳モードである。。 In this way, “2) SNS translation” is a translation that can be used, for example, when a translation that specifies a mutual language is performed by text or voice using a cloud-type SNS member list that can be managed on the Internet. Mode. .
 また、1つの通信端末1は、電気通信回線としての電話回線又はインターネット回線を通じた通話も可能である。 Further, one communication terminal 1 can make a call through a telephone line or an internet line as a telecommunication line.
 すなわち、一つの通信端末11は、入力手段を介して受け付けた基本言語による入力データに対して、他言語化した後の出力データを生成すると同時に、その出力データを(電話回線等に)出力することができると共に、他言語による入力データに対して、基本言語化した後の出力データ生成すると同時に、出力手段(音声出力部3又は表示部4)を介して出力することができる。 That is, one communication terminal 11 generates output data after being converted into another language for input data in the basic language received via the input means, and at the same time outputs the output data (to a telephone line or the like). In addition, it is possible to generate output data after being converted into a basic language with respect to input data in another language, and at the same time, output it via the output means (the audio output unit 3 or the display unit 4).
 このように、「3)一般通話翻訳」は、通常の電話回線による相手方との異国言語者同士の会話の際に用いることができる翻訳モードである。 As described above, “3) General call translation” is a translation mode that can be used for conversations between foreigners and other parties on a normal telephone line.
 また、1つの通信端末1は、電気通信回線としての近距離無線通信回線・電話回線・インターネット回線を通じた通信も可能である。 Also, one communication terminal 1 can communicate through a short-range wireless communication line, a telephone line, and an Internet line as an electric communication line.
 すなわち、一つの通信端末11は、入力手段を介して受け付けた基本言語による入力データに対して、他言語化した後の出力データを生成すると同時に、その出力データを(近距離無線通信回線・電話回線・インターネット回線等)出力することができると共に、他言語による入力データに対して、基本言語化した後の出力データ生成すると同時に、出力手段(音声出力部3又は表示部4)を介して出力することができる。 That is, one communication terminal 11 generates output data after being converted into another language for input data in a basic language received via the input means, and at the same time outputs the output data (short-range wireless communication line / telephone). (Line, internet line, etc.) can be output, and output data after being converted into a basic language is generated for input data in other languages, and at the same time, output via output means (voice output unit 3 or display unit 4) can do.
 これにより、例えば、同一会議室内のみならず遠隔地にいる2名以上(2台以上の通信端末)で基本言語と他言語とを相互に翻訳することができる。この際、他言語の数は1つの言語に限定されず、複数言語の翻訳が可能である。 Thus, for example, the basic language and the other language can be translated into each other not only by the same conference room but also by two or more persons (two or more communication terminals) in a remote place. At this time, the number of other languages is not limited to one language, and a plurality of languages can be translated.
 このように、「4)会議同時翻訳」は、2名以上の利用者による異国言語者同士の会話の際に用いることができる翻訳モードである。。 In this way, “4) Simultaneous Conference Translation” is a translation mode that can be used when two or more users have conversations between foreign languages. .
 同様にして、「5)TV電話翻訳」を選択すれば、撮影カメラ15を用いたTV電話において、同時翻訳モードを起動させることができる。この際、例えば、図7に示すように、相手方の通信端末21においても、撮影カメラ25が搭載されているのが望ましい。 Similarly, if “5) TV phone translation” is selected, the simultaneous translation mode can be activated in the TV phone using the photographing camera 15. At this time, for example, as shown in FIG. 7, it is desirable that the other party's communication terminal 21 is also equipped with a photographing camera 25.
 そして、「6)手話翻訳」を選択すれば、撮影カメラ15で撮影した手話動作を解析し、文字データ、音声データ、代替え画像データ、に変換する手話翻訳モードを起動させることができる。 If “6) sign language translation” is selected, the sign language translation mode for analyzing the sign language motion photographed by the photographing camera 15 and converting it into character data, voice data, and substitute image data can be activated.
 具体的には、通信端末1のユーザが翻訳アイコンAP4の「6)手話翻訳」を選択したチャット(又は電話)機能付き翻訳モードを選択すると、制御部6は、撮影カメラ15で撮影した画像を解析した翻訳を含む手話モードを実行する。制御部6は、電話回線等の電気通信回線を通じて通信端末21のユーザとの通話環境が確立すると、撮影カメラ15で撮影した画像から、例えば、手話動作単位(言葉の語彙単位に相当)で手話動作の内容を手話解析部16に解析させる。 Specifically, when the user of the communication terminal 1 selects the translation mode with chat (or telephone) function in which “6) sign language translation” of the translation icon AP4 is selected, the control unit 6 captures an image photographed by the photographing camera 15. Execute sign language mode including the analyzed translation. When the communication environment with the user of the communication terminal 21 is established through an electric communication line such as a telephone line, the control unit 6 uses, for example, a sign language action unit (corresponding to a word vocabulary unit) from an image captured by the photographing camera 15. The sign language analysis unit 16 is made to analyze the content of the operation.
 例えば、通信端末1のユーザが、自分を指さす動作(私は)、相手を指さす動作(あなたを)、両掌を上下に離して上手を回転させる動作(愛しています)、を行うと、これらの各手話動作を手話解析部16が解析する。 For example, when the user of the communication terminal 1 performs an action pointing at himself (I), an action pointing at the other party (you), an action of rotating both hands up and down (loved), The sign language analysis unit 16 analyzes each sign language operation.
 この際、各手話動作単位の映像は、記憶部5に記憶される。なお、各手話動作単位の映像は、例えば、記憶部5とは別のフレームメモリ等に一時記憶してもよい。 At this time, an image of each sign language operation unit is stored in the storage unit 5. The video of each sign language operation unit may be temporarily stored in a frame memory or the like different from the storage unit 5, for example.
 制御部6は、手話解析部16で解析した手話動作<私はあなたを愛しています>に基づいて、文字生成部14により男性的表現である日本語文字<愛しているよ>を生成させて表示部4に文字[愛しているよ]を表示させるとともに、翻訳した英語文字<I love you>を生成させて表示部4に文字[I love you]を表示させる。 Based on the sign language action <I love you> analyzed by the sign language analysis unit 16, the control unit 6 causes the character generation unit 14 to generate Japanese characters <I love you> that are masculine expressions. A character [I love you] is displayed on the display unit 4 and a translated English character <I love you> is generated and the character [I love you] is displayed on the display unit 4.
 すなわち、制御部6は、入力手段としての撮影カメラ15により受け付けた発信基準データとしての手話動作を基本言語とする場合に、記憶部5に予めインストールした手話解析アプリケーションを起動させて手話解析部16で手話動作を解析させたうえで、他言語による音声データ・テキストデータ・画像データの少なくとも何れか一つからなる発信データを生成するための基本言語による基本解析データ(手話動作<私はあなたを愛しています>)を生成する基本解析データ生成ステップ(ステップS6に相当)を実行する。 That is, the control unit 6 activates a sign language analysis application installed in the storage unit 5 in advance when the sign language operation as the transmission reference data received by the photographing camera 15 as the input means is used as a basic language. Analyzing the sign language movement with the basic analysis data in the basic language for generating outgoing data consisting of at least one of speech data, text data, and image data in other languages A basic analysis data generation step (corresponding to step S6) for generating “I love you>) is executed.
 なお、この手話動作による翻訳には、上記の場合、手話動作を解析して<私はあなたを愛しています>を生成する場合を含み、さらに、他言語へと二次翻訳する場合を含む。 In addition, the translation by this sign language action includes the case where the sign language action is analyzed to generate <I love you>, and the case where it is secondarily translated into another language.
 より現実的に、本実施形態の手話解析アプリケーションは、手話動作を解析して日本語に変換する解析変換機能と、変換した日本語を英語に翻訳する翻訳機能と、翻訳した英語に基づく英語用の手話動作に変換する変換機能と、の3つの言語変換機能(翻訳機能)を実行するともいえる。 More realistically, the sign language analysis application of the present embodiment includes an analysis conversion function for analyzing sign language movement and converting it into Japanese, a translation function for translating the converted Japanese into English, and an English-based version based on the translated English. It can also be said that three language conversion functions (translation functions), that is, a conversion function for converting to a sign language action of the above, are executed.
 また、制御部6は、生成した文字<I love you>に基づいて、音声生成部12により音声<I love you>を生成させる。 Also, the control unit 6 causes the voice generation unit 12 to generate a voice <I love you> based on the generated character <I love you>.
 さらに、制御部6は、表示部4に、解析した手話動作の画像を表示させる。この際、手話解析部16は、翻訳した語が英語であれば、予め記憶部5に格納した画像データから、英語表現の手話動作又は類似するアイコン・スタンプ等の手話画像(JPGや動画GIF等)を呼び出して送信可能とする。 Furthermore, the control unit 6 causes the display unit 4 to display an image of the analyzed sign language action. At this time, if the translated word is English, the sign language analysis unit 16 uses the image data stored in the storage unit 5 in advance to sign language operations such as an English-language sign language operation or a similar icon / stamp (JPG, video GIF, etc.). ) To enable transmission.
 これにより、制御部6は、生成した、文字<愛しているよ>、文字<I love you>、音声<I love you>、手話画像<(I love you)>、の4つのデータを重畳して送信インターフェース10から通信端末21に向けて送信する。 Thereby, the control unit 6 superimposes the generated four data of the character <I love you>, the character <I love you>, the voice <I love you>, and the sign language image <(I love you)>. Then, the data is transmitted from the transmission interface 10 to the communication terminal 21.
 通信端末21は、受信したデータに基づいて、翻訳前の文字[愛しているよ]と翻訳後の文字[I love you]並びに翻訳後の手話画像[(I love you)]を表示部24に表示させるとともに、音声『I love you』を音声出力部23から出力させる。 Based on the received data, the communication terminal 21 displays the pre-translation character [I love you], the post-translation character [I love you], and the post-translation sign language image [(I love you)] on the display unit 24. The voice “I love you” is output from the voice output unit 23 while being displayed.
 このように、通信端末21では、翻訳アプリをインストールしていなくても、文字<愛しているよ>、文字<I love you>、音声<I love you>、並びに手話画像[(I love you)]の4つのデータを取得することができる。なお、これら4つの各データは、通信端末21に翻訳アプリがインストールされていれば、通信端末21が生成してもよい。 In this manner, even if the translation application is not installed, the communication terminal 21 has the character <I love you>, the character <I love you>, the voice <I love you>, and the sign language image [(I love you) ] Can be acquired. Note that these four pieces of data may be generated by the communication terminal 21 if a translation application is installed in the communication terminal 21.
 一方、通信端末21のユーザは、音声『I love you』に対して、音声入力部22に『I am happy』と音声入力すると、その音声<I am happy>を通信端末1に向けて送信する。 On the other hand, when the user of the communication terminal 21 inputs “I am happy” to the voice input unit 22 in response to the voice “I love you”, the user transmits the voice <I am happy> to the communication terminal 1. .
 通信端末1の制御部6は、受信インターフェース9で音声<I am happy>を受信させると、音声認識部11により音声<I am happy>を認識させる。 The control unit 6 of the communication terminal 1 causes the voice recognition unit 11 to recognize the voice <I am happy> when the reception interface 9 receives the voice <I > am happy>.
 これにより、制御部6は、文字生成部14により文字<I am happy>を生成させて表示部4に文字[I am happy]を表示させるとともに、翻訳した文字<嬉しい!>を生成させて表示部4に文字[嬉しい!]を表示させ、さらに、嬉しいに対応する手話動作の手話画像[(嬉しい)]を表示させる。なお、制御部6は、生成した文字<嬉しい>に基づいて、音声生成部12により音声<嬉しい>を生成させ、音声出力部3から音訳した音声『嬉しい』を出力させることも可能である。 Thus, the control unit 6 causes the character generation unit 14 to generate the character <I am happy>, displays the character [I am happy] on the display unit 4, and translates the character <happy! > Is generated on the display unit 4 and the character [happy! ], And a sign language image [(happy)] of the corresponding sign language action is displayed. The control unit 6 can also generate the speech <joyful> by the speech generation unit 12 based on the generated character <joyful> and output the transliterated speech “happy” from the speech output unit 3.
 このように、通信端末1は、通信端末21に翻訳アプリをインストールしていなくても、翻訳前の文字<I am happy>、翻訳後の文字<嬉しい!>、翻訳後の手話画像<(嬉しい)>、翻訳後の音声<嬉しい>、の4つのデータを生成し、それぞれ表示部4及び音声出力部3に展開してユーザに認識させることができる。なお、これら4つの各データは、通信端末21に翻訳アプリがインストールされていれば、通信端末21が生成してもよい。 Thus, even if the communication terminal 1 does not have a translation application installed on the communication terminal 21, the character before translation <I am happy> and the character after translation <happy! >, Translated sign language image <(happy)>, translated speech <joyful>, and can be expanded on the display unit 4 and the voice output unit 3 to be recognized by the user. Note that these four pieces of data may be generated by the communication terminal 21 if a translation application is installed in the communication terminal 21.
 尚、本発明は、上記実施の形態に示した構成・用途等に限られるものではなく、種々の変更が可能である。例えば、記憶部(クラウド上を含む)5に記憶した翻訳データは、ディープラーニング(deep-learning)等の学習機能(AI機能)によって音声認識、動画認識、自然言語処理、等を学習して更新データを逐次記憶することも可能である。 It should be noted that the present invention is not limited to the configuration and application shown in the above embodiment, and various modifications are possible. For example, the translation data stored in the storage unit (including on the cloud) 5 is updated by learning speech recognition, video recognition, natural language processing, etc. by a learning function (AI function) such as deep-learning. It is also possible to store data sequentially.
 以上説明したように、本発明に係る翻訳処理方法、翻訳処理プログラム、及び、記録媒体は、1つの通信端末に実行させる機能でありながら、翻訳のバリエーションを多機能化し得て汎用性を向上することができるという効果を有し、電気通信回線を通じて相互通信している際に、基本言語と他言語とで相互翻訳するための翻訳処理方法、翻訳処理プログラム、及び、記録媒体全般に有用である。 As described above, the translation processing method, the translation processing program, and the recording medium according to the present invention are functions that are executed by a single communication terminal, but the translation variations can be made multifunctional to improve versatility. And is useful for a translation processing method, a translation processing program, and a general recording medium for mutual translation between a basic language and another language when communicating with each other via a telecommunication line. .
 1   通信端末
 2   音声入力部
 3   音声出力部
 4   表示部
 5   記憶部
 6   制御部
 7   スイッチ
 8   操作部
 9   受信インターフェース
 10  送信インターフェース
 11  音声認識部
 12  音声生成部
 13  文字認識部
 14  文字生成部
 15  撮影カメラ
 16  手話解析部(画像生成部)
 21  通信端末
 22  音声入力部
 23  音声出力部
 24  表示部
 25  撮影カメラ
 AP1 電話アイコン
 AP2 E-Mailアイコン
 AP3 チャットアイコン
 C   クラウド
 NT  電気通信回線
DESCRIPTION OF SYMBOLS 1 Communication terminal 2 Voice input part 3 Voice output part 4 Display part 5 Memory | storage part 6 Control part 7 Switch 8 Operation part 9 Reception interface 10 Transmission interface 11 Voice recognition part 12 Voice generation part 13 Character recognition part 14 Character generation part 15 Shooting camera 16 Sign language analyzer (image generator)
21 Communication Terminal 22 Audio Input Unit 23 Audio Output Unit 24 Display Unit 25 Shooting Camera AP1 Telephone Icon AP2 E-Mail Icon AP3 Chat Icon C Cloud NT Electric Communication Line

Claims (13)

  1.  基本言語と他言語との相互翻訳を実行する翻訳アプリケーションを予めインストールした記憶手段を有する1つの通信端末に実行させるステップとして、
     前記基本言語による音声データ・テキストデータ・画像データの少なくとも何れか一つの入力を入力手段を介して発信基準データとして受け付ける通信内容受付ステップと、
     前記翻訳アプリケーションを起動させて、前記通信内容受付ステップで受け付けた前記発信基準データを前記他言語の音声データ・テキストデータ・画像データの少なくとも何れか一つに翻訳して発信データを生成する他言語化翻訳ステップと、
     前記他言語化翻訳ステップで翻訳した前記発信データを発信手段により発信する発信ステップと、
     前記他言語による音声データ・テキストデータ・画像データの少なくとも何れか一つを受信手段により受信基準データとして受信する受信ステップと、
     前記翻訳アプリケーションを起動させて前記受信ステップで受信した前記受信基準データを前記基本言語による音声データ・テキストデータ・画像データの少なくとも何れか一つに翻訳して受信データを生成する基本言語化翻訳ステップと、
     前記基本言語化翻訳ステップで翻訳した前記受信データを出力手段を介して出力する通信内容出力ステップと、
     を備えることを特徴とする翻訳処理方法。
    As a step of causing a single communication terminal having a storage means preinstalled with a translation application that performs mutual translation between a basic language and another language,
    A communication content receiving step for receiving at least one of voice data, text data, and image data in the basic language as transmission reference data via an input means;
    Another language that activates the translation application and translates the transmission reference data received in the communication content reception step into at least one of the speech data, text data, and image data of the other language to generate transmission data A translation step,
    A sending step of sending the sending data translated in the other language translation step by sending means;
    A reception step of receiving at least one of the voice data, text data, and image data in the other language as reception reference data by a reception unit;
    A basic languageized translation step of generating reception data by translating the reception reference data received in the reception step by starting the translation application into at least one of voice data, text data, and image data in the basic language When,
    A communication content output step for outputting the received data translated in the basic language translation step through an output means;
    A translation processing method comprising:
  2.  前記発信ステップにおいて、前記発信データは発信手段により電気通信回線を通じて発信され、
     前記受信ステップにおいて、前記受信基準データは電気通信回線を通じて受信手段により受信される、
     ことを特徴とする請求項1に記載の翻訳処理方法。
    In the transmission step, the transmission data is transmitted through a telecommunication line by a transmission means,
    In the receiving step, the reception reference data is received by a receiving means through a telecommunication line.
    The translation processing method according to claim 1, wherein:
  3.  前記電気通信回線がインターネット回線である、
     ことを特徴とする請求項2に記載の翻訳処理方法。
    The telecommunication line is an internet line;
    The translation processing method according to claim 2, wherein:
  4.  前記電気通信回線が電話回線である、
     ことを特徴とする請求項2に記載の翻訳処理方法。
    The telecommunication line is a telephone line;
    The translation processing method according to claim 2, wherein:
  5.  前記発信ステップは電気通信回線を介することなく実行され、
     前記受信ステップは電気通信回線を介することなく実行される、
     ことを特徴とする請求項1に記載の翻訳処理方法。
    The calling step is performed without going through a telecommunication line,
    The receiving step is performed without going through a telecommunication line;
    The translation processing method according to claim 1, wherein:
  6.  前記他言語化翻訳ステップを実行させる前に、前記記憶手段に予め登録した通信相手先の登録名を含むアドレスデータに基づいて翻訳対象となる前記他言語を特定する他言語特定ステップと、
     を備えることを特徴とする請求項1乃至請求項5の何れか1項に記載の翻訳処理方法。
    Before executing the other language translation step, another language specifying step for specifying the other language to be translated based on address data including a registered name of a communication partner registered in advance in the storage unit;
    The translation processing method according to any one of claims 1 to 5, further comprising:
  7.  前記入力手段により受け付けた前記発信基準データが前記基本言語による音声データである場合に、前記記憶手段に予めインストールした聞き取りアプリケーションを起動させて、前記他言語によるテキストデータからなる前記発信データを生成するための前記基本言語による基本テキストデータを生成する基本テキストデータ生成ステップと、
     を備えることを特徴とする請求項1乃至請求項6の何れか1項に記載の翻訳処理方法。
    When the transmission reference data received by the input unit is voice data in the basic language, the listening application installed in advance in the storage unit is activated to generate the transmission data including text data in the other language Generating basic text data in the basic language for generating basic text data;
    The translation processing method according to claim 1, further comprising:
  8.  前記入力手段により受け付けた前記発信基準データが前記基本言語によるテキストデータである場合に、前記記憶手段に予めインストールした読み上げアプリケーションを起動させて、前記他言語による音声データからなる前記発信データを生成するための前記基本言語による基本音声データを生成する基本音声データ生成ステップと、
     を備えることを特徴とする請求項1乃至請求項6の何れか1項に記載の翻訳処理方法。
    When the transmission reference data received by the input unit is text data in the basic language, a reading application installed in advance in the storage unit is activated to generate the transmission data including voice data in the other language Generating basic voice data in the basic language for generating basic voice data;
    The translation processing method according to claim 1, further comprising:
  9.  前記入力手段により受け付けた前記発信基準データが前記基本言語によるテキストデータである場合に、前記他言語化翻訳ステップで翻訳した前記他言語の音声データによる前記発信データに加えて前記入力手段により受け付けた前記基本言語によるテキストデータを前記発信手段により発信する発信ステップと、
     を備えることを特徴とする請求項8に記載の翻訳処理方法。
    When the transmission reference data received by the input means is text data in the basic language, it is received by the input means in addition to the transmission data by the voice data of the other language translated in the other language translation step. A sending step of sending the text data in the basic language by the sending means;
    The translation processing method according to claim 8, further comprising:
  10.  前記入力手段が前記基本言語による音声データを受け付けてから前記他言語化翻訳ステップで翻訳した前記他言語による前記発信データを前記発信手段により発信するまでの期間は前記受信手段による受信をキャンセルする割り込みキャンセルステップと、
     を備えることを特徴とする請求項1乃至請求項9の何れか1項に記載の翻訳処理方法。
    An interrupt for canceling reception by the receiving means during a period from when the input means accepts the speech data in the basic language to when the sending means in the other language is transmitted by the sending means. Cancellation step,
    The translation processing method according to claim 1, further comprising:
  11.  前記入力手段により受け付けた前記発信基準データが前記基本言語による手話動作である場合に、前記記憶手段に予めインストールした手話解析アプリケーションを起動させて前記手話動作を解析したうえで、前記他言語による音声データ・テキストデータ・画像データの少なくとも何れか一つからなる前記発信データを生成するための前記基本言語による基本解析データを生成する基本解析データ生成ステップと、
     を備えることを特徴とする請求項1乃至請求項6の何れか1項に記載の翻訳処理方法。
    When the transmission reference data received by the input means is a sign language action in the basic language, a sign language analysis application installed in the storage means is activated to analyze the sign language action, and then the voice in the other language A basic analysis data generation step of generating basic analysis data in the basic language for generating the transmission data consisting of at least one of data, text data, and image data;
    The translation processing method according to claim 1, further comprising:
  12.  請求項1乃至請求項11何れか1項に記載の翻訳処理方法を通信端末の演算手段に実行させる翻訳処理プログラム。 A translation processing program for causing a computing means of a communication terminal to execute the translation processing method according to any one of claims 1 to 11.
  13.  請求項1乃至請求項11の何れか1項に記載の翻訳処理方法を通信端末の演算手段に実行させる翻訳処理プログラムを記録した記録媒体。 A recording medium recording a translation processing program for causing a computing means of a communication terminal to execute the translation processing method according to any one of claims 1 to 11.
PCT/JP2018/014317 2017-04-03 2018-04-03 Translation processing method, translation processing program, and recording medium WO2018186416A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2017073698A JP6243071B1 (en) 2017-04-03 2017-04-03 Communication content translation processing method, communication content translation processing program, and recording medium
JP2017-073698 2017-04-03

Publications (1)

Publication Number Publication Date
WO2018186416A1 true WO2018186416A1 (en) 2018-10-11

Family

ID=60570344

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2018/014317 WO2018186416A1 (en) 2017-04-03 2018-04-03 Translation processing method, translation processing program, and recording medium

Country Status (2)

Country Link
JP (1) JP6243071B1 (en)
WO (1) WO2018186416A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113726750A (en) * 2021-08-18 2021-11-30 中国联合网络通信集团有限公司 Voice real-time translation method, device and storage medium
KR20220099083A (en) * 2021-01-05 2022-07-12 한국전자통신연구원 System, user device and method for providing automatic interpretation service based on speaker separation
JP2022105982A (en) * 2021-01-05 2022-07-15 エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート Automatic interpretation method based on speaker separation, user terminal providing automatic interpretation service based on speaker separation, and automatic interpretation service providing system based on speaker separation

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6927177B2 (en) 2018-09-26 2021-08-25 信越化学工業株式会社 Phase shift photomask blank and phase shift photomask
US11776286B2 (en) 2020-02-11 2023-10-03 NextVPU (Shanghai) Co., Ltd. Image text broadcasting
CN110991455B (en) * 2020-02-11 2023-05-05 上海肇观电子科技有限公司 Image text broadcasting method and equipment, electronic circuit and storage medium thereof

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001067287A (en) * 1999-08-30 2001-03-16 Fujitsu Ltd Electronic mail device, electronic mail transmitting method, and recording medium
JP2009527811A (en) * 2006-02-17 2009-07-30 マイクロソフト コーポレーション Machine translation in instant messaging applications
JP2015076774A (en) * 2013-10-10 2015-04-20 みずほ情報総研株式会社 Communication support system, communication support method, and communication support program
WO2016006354A1 (en) * 2014-07-08 2016-01-14 Necソリューションイノベータ株式会社 Information processing device, and translation-data provision method

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001265700A (en) * 2000-03-16 2001-09-28 Matsushita Electric Ind Co Ltd Portable communication terminal equipment
JP2003188948A (en) * 2001-12-19 2003-07-04 Nec Corp Portable terminal device
JP2005202884A (en) * 2004-01-19 2005-07-28 Toshiba Corp Transmission device, reception device, relay device, and transmission/reception system
JP2011118690A (en) * 2009-12-03 2011-06-16 Fujitsu Toshiba Mobile Communications Ltd Translating device
KR20130071958A (en) * 2011-12-21 2013-07-01 엔에이치엔(주) System and method for providing interpretation or translation of user message by instant messaging application
KR101271285B1 (en) * 2011-12-28 2013-06-04 (주)카카오 A method of providing a multi language translation for messages included in chatting window
KR102108500B1 (en) * 2013-02-22 2020-05-08 삼성전자 주식회사 Supporting Method And System For communication Service, and Electronic Device supporting the same
JP6218568B2 (en) * 2013-11-20 2017-10-25 エヌ・ティ・ティ・コミュニケーションズ株式会社 COMMUNICATION DEVICE, COMMUNICATION SYSTEM, COMMUNICATION METHOD, AND COMMUNICATION PROGRAM
WO2016018004A1 (en) * 2014-07-31 2016-02-04 Samsung Electronics Co., Ltd. Method, apparatus, and system for providing translated content
JP5856708B1 (en) * 2015-08-31 2016-02-10 株式会社Wing of Freedom Translation system and server
JP5998298B1 (en) * 2016-01-13 2016-09-28 株式会社リクルートライフスタイル Speech translation device, speech translation method, and speech translation program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001067287A (en) * 1999-08-30 2001-03-16 Fujitsu Ltd Electronic mail device, electronic mail transmitting method, and recording medium
JP2009527811A (en) * 2006-02-17 2009-07-30 マイクロソフト コーポレーション Machine translation in instant messaging applications
JP2015076774A (en) * 2013-10-10 2015-04-20 みずほ情報総研株式会社 Communication support system, communication support method, and communication support program
WO2016006354A1 (en) * 2014-07-08 2016-01-14 Necソリューションイノベータ株式会社 Information processing device, and translation-data provision method

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20220099083A (en) * 2021-01-05 2022-07-12 한국전자통신연구원 System, user device and method for providing automatic interpretation service based on speaker separation
JP2022105982A (en) * 2021-01-05 2022-07-15 エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート Automatic interpretation method based on speaker separation, user terminal providing automatic interpretation service based on speaker separation, and automatic interpretation service providing system based on speaker separation
JP7333371B2 (en) 2021-01-05 2023-08-24 エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート Automatic Interpretation Method Based on Speaker Separation, User Terminal Providing Automatic Interpretation Service Based on Speaker Separation, and Automatic Interpretation Service Providing System Based on Speaker Separation
KR102584436B1 (en) * 2021-01-05 2023-10-05 한국전자통신연구원 System, user device and method for providing automatic interpretation service based on speaker separation
CN113726750A (en) * 2021-08-18 2021-11-30 中国联合网络通信集团有限公司 Voice real-time translation method, device and storage medium

Also Published As

Publication number Publication date
JP2018180590A (en) 2018-11-15
JP6243071B1 (en) 2017-12-06

Similar Documents

Publication Publication Date Title
WO2018186416A1 (en) Translation processing method, translation processing program, and recording medium
US10943158B2 (en) Translation and display of text in picture
KR101942308B1 (en) Method for providing message function and an electronic device thereof
JP2009527811A (en) Machine translation in instant messaging applications
JP7467635B2 (en) User terminal, video calling device, video calling system, and control method thereof
KR20200115625A (en) How to learn personalized intent
WO2021006538A1 (en) Avatar visual transformation device expressing text message as v-moji and message transformation method
US20120215521A1 (en) Software Application Method to Translate an Incoming Message, an Outgoing Message, or an User Input Text
KR20150113652A (en) smart device easy to convert of Multilingual.
JP2010026686A (en) Interactive communication terminal with integrative interface, and communication system using the same
JP5856708B1 (en) Translation system and server
WO2019196645A1 (en) Conversational information processing method, device, mobile terminal, and storage medium
KR20110012491A (en) System, management server, terminal and method for transmitting of message using image data and avatar
CN103003874A (en) Provisioning text services based on assignment of language attributes to contact entry
KR20090054609A (en) Voip telephone communication system and method for providing users with telephone communication service comprising emotional contents effect
TW201346597A (en) Multiple language real-time translation system
JP2020119043A (en) Voice translation system and voice translation method
JP2016091195A (en) Information transmission/reception program and system
JP2005222316A (en) Conversation support device, conference support system, reception work support system, and program
JP7462995B1 (en) Information processing system, information processing method, and program
JP7411369B2 (en) Communication systems, reception terminal devices and their programs
WO2022091675A1 (en) Program, method, information processing device, and system
JP6957067B1 (en) Systems for communicating with people and programs for them
CN210402846U (en) Sign language translation terminal and sign language translation server
TW202334858A (en) Various sign language translation system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18781849

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18781849

Country of ref document: EP

Kind code of ref document: A1