WO2020009261A1 - Dispositif numérique assurant une reconnaissance vocale et son procédé de commande - Google Patents

Dispositif numérique assurant une reconnaissance vocale et son procédé de commande Download PDF

Info

Publication number
WO2020009261A1
WO2020009261A1 PCT/KR2018/007797 KR2018007797W WO2020009261A1 WO 2020009261 A1 WO2020009261 A1 WO 2020009261A1 KR 2018007797 W KR2018007797 W KR 2018007797W WO 2020009261 A1 WO2020009261 A1 WO 2020009261A1
Authority
WO
WIPO (PCT)
Prior art keywords
language
audio signal
text
digital device
memory
Prior art date
Application number
PCT/KR2018/007797
Other languages
English (en)
Korean (ko)
Inventor
황정환
조택일
민동옥
Original Assignee
엘지전자 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 엘지전자 주식회사 filed Critical 엘지전자 주식회사
Publication of WO2020009261A1 publication Critical patent/WO2020009261A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination
    • G10L13/10Prosody rules derived from text; Stress or intonation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Definitions

  • the present invention relates to a digital device capable of speech recognition, and is applicable to, for example, a portable audio device such as a headset, a neckband, or a mobile device such as a mobile phone.
  • an embodiment of the present invention is to propose a technique for fundamentally eliminating errors in speech recognition and translation when translating audio of a first language into a second language and outputting the same through a speaker.
  • another embodiment of the present invention is to propose a technique for detecting not only the text of a language but also the state information of a user who speaks through a speech recognition engine.
  • the control method of the digital device capable of speech recognition includes.
  • the comparing may include displaying different portions of the same and different portions when the text of the retranslated first language and the audio signal of the received first language are different, and user selection or data stored in a memory of the digital device.
  • the method further includes providing an option for automatically modifying the other portion accordingly.
  • the receiving may further include extracting feature points of the audio signal of the first language, wherein the feature points include at least one of stress, intonation, voice, and volume.
  • the converting may further include outputting the translated second language text in different accent, intonation, voice, or volume according to a feature point extracted from the audio signal of the first language.
  • the first language and the second language are different from each other, and the digital device includes at least one of a wearable device and a mobile device.
  • a digital device capable of speech recognition includes a memory, a microphone for receiving an audio signal of a first language, and a controller for controlling the memory and the microphone.
  • the controller may be configured to translate the received audio signal of the first language into text of a second predetermined language by referring to the memory, and to translate the text of the translated second language into the first language by referring to the memory. Retranslate into a second language, compare the text of the retranslated first language with the audio signal of the received first language, and modify some or all of the text of the translated second language according to the comparison result. And converting it into a format that can be output as a signal.
  • the digital device displays the same part and a different part differently, and displays the user part or data stored in the memory of the digital device.
  • the display module may further include a display module configured to automatically display an option for automatically modifying the other part.
  • the controller extracts a feature point of the audio signal of the first language, and the feature point includes at least one of stress, intonation, voice, and volume.
  • the controller may be configured to control the speaker to output the translated text of the second language in another accent, intonation, voice, or volume according to a feature point extracted from the audio signal of the first language.
  • FIG. 1 is a block diagram illustrating a portable audio device as an example of a digital device according to an embodiment of the present invention.
  • FIG. 2 is a perspective view of a portable audio device as an example of a digital device according to an embodiment of the present invention.
  • FIG. 3 is an exploded perspective view of a portable audio device as an example of a digital device according to an embodiment of the present invention.
  • FIG. 4 illustrates a process in which a digital device corrects an error through a first translation / second translation or the like according to an embodiment of the present invention.
  • Fig. 5 shows a detailed flowchart of the digital device performing post-interpretation correction according to an embodiment of the present invention.
  • FIG. 6 illustrates a process in which a digital device executes N to N interpretation according to an embodiment of the present invention.
  • FIG. 7 illustrates a scenario in which a digital device according to another embodiment of the present invention masks a voice / emotion of a talker.
  • FIG. 8 shows a flow chart embodying the scenario shown in FIG. 7 in more detail.
  • FIG. 9 is a block diagram showing a main configuration of the digital device shown in FIGS. 1 to 8.
  • the portable audio device 100 includes a wireless communication unit 110, an input unit 120, a sensing unit 130, an output unit 140, an interface unit 150, a controller 160, a power supply unit 170, and the like. can do.
  • the portable acoustic apparatus refers to a device that receives a sound signal from a terminal (for example, a mobile phone) and transmits sound information collected through a microphone to the terminal.
  • a portable audio device is connected to a terminal jack of a terminal using a wired method for receiving an acoustic signal, but recently, a portable communication device of a wireless communication method has also been commercialized.
  • various portable audio devices have been developed in consideration of portability, such as a headphone type that fits in a band shape on the head, a type that hangs on the ear, and a type that fits in the ear so as to be carried on a user's body.
  • a headphone type that fits in a band shape on the head
  • a type that hangs on the ear and a type that fits in the ear so as to be carried on a user's body.
  • neck-back portable sound devices that can be worn on the user's neck. 1 illustrates this.
  • FIG. 1 The components shown in FIG. 1 are not essential to implementing a portable audio device, so the portable audio device described herein may have more or fewer components than those listed above.
  • the wireless communication unit 110 of the components between the portable sound device 100 and the wireless communication system, between the portable sound device 100 and another mobile terminal, or between the portable sound device 100 and an external server It may include one or more modules to enable wireless communication of.
  • the wireless communication unit 110 may include one or more modules for connecting the portable audio device 100 to one or more networks.
  • the wireless communication unit 110 may include at least one of the short range communication module 111 and the location information module 112. Or it may include a mobile communication module or a wireless Internet module, if necessary.
  • the short range communication module 111 is for short range communication, and includes Bluetooth TM, Radio Frequency Identification (RFID), Infrared Data Association (IrDA), Ultra Wideband (UWB), ZigBee, NFC (Near Field Communication), at least one of Wi-Fi (Wireless-Fidelity), Wi-Fi Direct, Wireless USB (Wireless Universal Serial Bus) technology can be used to support short-range communication.
  • RFID Radio Frequency Identification
  • IrDA Infrared Data Association
  • UWB Ultra Wideband
  • ZigBee ZigBee
  • NFC Near Field Communication
  • Wi-Fi Wireless-Fidelity
  • Wi-Fi Direct Wireless USB (Wireless Universal Serial Bus) technology
  • the short-range communication module 111 may be provided between the portable sound device 100 and the wireless communication system, between the portable sound device 100 and another mobile terminal, or the portable sound device 100 through local area network (Wireless Area Networks). ) And a wireless communication between the network where the other mobile terminal (or external server) is located.
  • the short range wireless communication network may be short range wireless personal area networks.
  • the short range communication module 111 may detect (or recognize) a terminal or the like that can communicate around the portable sound device 100. Further, when the detected terminal is a device authorized to communicate with the portable acoustic apparatus 100 according to the present invention, the controller 160 transmits at least a part of the data processed by the mobile terminal through the short range communication module 111. Can be sent. Accordingly, the user of the portable acoustic apparatus 100 may use the data processed by the terminal through the wearable device.
  • the user when a telephone is received in a terminal or the like, the user can perform a telephone call through the portable acoustic apparatus 100.
  • the location information module 112 is a module for acquiring the location (or current location) of the portable audio device 100.
  • Examples of the location information module 112 include a GPS (Global Positioning System) module or a WiFi (Wireless Fidelity) module.
  • the mobile terminal may acquire the location of the portable acoustic apparatus 100 using a signal transmitted from a GPS satellite.
  • the portable audio device 100 utilizes the Wi-Fi module
  • the portable audio device 100 is based on information of the wireless access point (AP) that transmits or receives a Wi-Fi module and a wireless signal. To obtain the position of.
  • AP wireless access point
  • the location information module 112 may perform any function of other modules of the wireless communication unit 110 to substitute or additionally obtain data regarding the location of the portable sound device 100.
  • the location information module 115 is a module used to acquire the location (or current location) of the portable audio device 100, and is not limited to a module that directly calculates or acquires the location of the portable audio device 100. .
  • the input unit 120 may be a microphone (MIC) 121 for inputting an audio signal, or an audio input unit, a user input unit 122 (eg, a touch key or a push key) for receiving information from a user. key), etc.).
  • the voice data or the image data collected by the input unit 120 may be analyzed and processed as a control command of the user.
  • the user input unit 122 is a configuration for the user to control the portable audio device 100, for example, a call button, a button for adjusting the volume, a power button and a storage button for accommodating the sound cable into the main body as an example. Can be.
  • the user input unit 122 may include only a call button, a pair of volume control buttons, and may further include a play / stop button and a song order change button.
  • the size of the portable audio device 100 is limited, and the user input unit 122 is often inputted by the user without directly viewing the number of buttons, so that it is difficult to distinguish the functions of each button. Time and number of times of pressing and combinations of a plurality of buttons can be expanded to expand the control commands that can be input.
  • the microphone 121 processes an external sound signal as electrical voice data.
  • the processed voice data may be utilized according to a function (or an application program being executed) performed by the portable sound device 100 or transmitted to an external terminal or an external server through the wireless communication unit 110.
  • various microphones for removing noise may be implemented in the microphone 122 to remove noise generated in the process of receiving an external sound signal.
  • the sensing unit 130 may include one or more sensors for sensing at least one of information in the portable sound device, surrounding environment information surrounding the portable sound device, and user information.
  • the sensing unit 130 may include a proximity sensor 131, an illumination sensor 132, an illumination sensor, a touch sensor, an acceleration sensor, a magnetic sensor, and gravity.
  • the portable audio device disclosed herein may utilize a combination of information sensed by at least two or more of these sensors.
  • it may include a sensor for detecting whether the earphone to be described later is located in the holder, such a sensor may be representatively applied.
  • the output unit 140 is used to generate an output related to visual, auditory, or tactile, and may include at least one of an audio output unit 141, a haptic module 142, and an optical output unit 143.
  • the sound output unit 141 is a device for outputting sound according to an audio signal, and representative examples thereof may include earphones that are inserted into a user's ear and deliver sound, and a speaker that outputs sound without the earphones being inserted.
  • the interface unit 150 serves as a path to various types of external devices connected to the portable sound device 100.
  • the interface unit 150 may include at least one of an external charger port and a wired / wireless data port.
  • appropriate control associated with the connected external device may be performed.
  • the controller 160 In addition to the operation related to the application program, the controller 160 generally controls the overall operation of the portable acoustic apparatus 100.
  • the controller 160 may process signals, data, information, or the like, which are input or output through the above-described components.
  • the power supply unit 170 receives power from an external power source and an internal power source under the control of the controller 160 to supply power to each component included in the portable sound device 100.
  • the power supply unit 170 includes a battery, which may be a built-in battery or a replaceable battery.
  • At least some of the components may operate in cooperation with each other in order to implement an operation, control, or control method of the portable acoustic apparatus according to various embodiments described below.
  • FIG. 2 is a perspective view of a portable audio device as an example of a digital device according to an embodiment of the present invention.
  • 3 is an exploded perspective view of the portable audio device 100 as an example of a digital device according to an embodiment of the present invention.
  • the portable sound device 100 may largely include a configuration of the neckband wire 310 and the main body 200.
  • the neckband wire 310 may surround the neck area of the user and serve to mount the portable audio device 100 on the neck of the user.
  • the main body 200 and the neckband wire 310 may be combined to form a 'U' shape, or both ends may be detached and optionally formed to form an 'O' shape as needed.
  • the main body 200 may be coupled to both ends of the necked wire, respectively.
  • the main body 200 may be divided into a first body coupled to one end of the neckband wire 310 and a second body coupled to the other end of the neckband wire 310.
  • the main body 200 may be provided integrally, but may be provided in the form of a combination of the upper case 210 and the lower case 220 for the purpose of component mounting.
  • One side of the upper case 210 or the lower case 220 may include a side region of the main body 200.
  • the side case may be provided as a separate member to be combined with the upper case 210 and the lower case 220. In the present invention, it will be described on the premise that the upper case 210 forms a side region of the main body 200.
  • the portable acoustic device 100 Since the portable acoustic device 100 is easily exposed to moisture such as sweat, it may add a waterproof function. A rib covering a gap between the upper case 210 and the lower case 220 may be formed or water may be prevented from penetrating through the waterproof member.
  • the main body 200 may be manufactured by injection molding a polymer material.
  • a plastic product having a strength such as polystyrene (PS) may be used.
  • PS polystyrene
  • it may be partially composed of a release material including metal, glass, leather, and the like.
  • the main body 200 may protect the internal parts and at the same time may be a polyurethane coating on the surface of the main body 200 to be in close contact with the user's body. Applying a polyurethane coating on the surface of the main body 200 can form a unitary appearance of the portable audio device 100. The main body 200 is in close contact with the user's skin and does not shake in accordance with the user's movement. There is an advantage.
  • the main body 200 may mount most components capable of performing the functions of the portable acoustic apparatus 100 of the present invention.
  • various components such as the main board, the wireless communication unit, the battery 260, and the recovery module may be inserted into the main body 200.
  • the main board is mounted with a wireless communication unit, a microphone 201, etc., and is connected to a battery 260, a user input unit, a sound output unit, and the like.
  • Components mounted on the main body 200 may be provided symmetrically on both sides of the body, that is, the first body and the second body, or may be provided on only one side.
  • the neckband wire 310 may be elastically deformed within a predetermined range when a force is applied to the neckband wire 310 and then restored to its original shape when the force is removed.
  • the shape memory alloy may be a representative form having elastic properties. Further, the first substrate 240, the fixing bracket 250, the battery 260, and the like are additionally provided.
  • the present invention is not limited thereto and may be applied to any other type of digital device capable of speech recognition.
  • FIG. 4 illustrates a process in which a digital device corrects an error through a first translation / second translation or the like according to an embodiment of the present invention. It is possible to initially receive a user's voice through the microphone described above with reference to FIGS. 1 to 3, and to analyze and translate the voice received through a controller, a memory, a network interface, and the like.
  • the main technical idea of the present invention is a technology of translating a speech recognized language into a preset language (for example, Korean-English) and interpreting (English-Korean) again to correct the difference. Learning the difference (Diff) of the corrected content by the feature (feature), and then designed to enable automatic correction through the AI (Artificial Intelligence) also belongs to another scope of the present invention.
  • a device according to an embodiment of the present invention is a device A shown in FIG. 4.
  • a voice of “see at 6 o'clock tomorrow” is input (S401).
  • the A device detects the difference between the Korean text (S401) and the retranslated Korean sentence (S404) received in step S401. Then, it provides a list for correcting the "look" corresponding to the difference (S405).
  • a GUI (S409) for re-checking the modified “See you at the clock tower at 6 o'clock tomorrow” (S408) is displayed, and when a user confirmation (transmission) is received, the final English “See you at” is displayed.
  • the clock tower at 6 o'clock tomorrow ” is transmitted to the B device (S411), and the corresponding Korean language (S410) is stored together in memory, so that the same error is prevented in the future.
  • Fig. 5 shows a detailed flowchart of the digital device performing post-interpretation correction according to an embodiment of the present invention.
  • the digital device performs translation in a predetermined language (S502).
  • the preset language may be English, Korean, Japanese, Chinese, etc., but is not limited thereto.
  • the digital device retranslates the original language back to the original language based on the text translated in step S502 (S503).
  • step S501 it is determined whether the original text input in step S501 and the text retranslated in step S503 are the same (S504).
  • the translated content is output through the speaker or transmitted to another device (S514).
  • the digital device determines whether there is a user input (S509).
  • the translated content is output as it is without error correction or transmitted to another device (S514).
  • the difference may be that the other party has no difficulty in understanding even if the translation error is not corrected.
  • step S512 If the determination result (S512) AI is not applied, the process returns to step S514. If the determination result (S512) AI is applied, the correction word is learned and stored in the memory (S513), and the process moves to step S514.
  • one feature of the present invention is to use only N translation engines by using official language translation instead of using N ⁇ N translation engines. That is, the official language translation can be "Korean” or "English” here.
  • FIG. 6 proposes a one-to-n or many-to-many (N to N) interpretation service. .
  • the translation network is configured by connecting all connected headsets, including communication between terminals (mobile phones).
  • two headsets are connected to the first mobile phone and two headsets are also connected to the second mobile phone.
  • the language of the first headset connected to the first mobile phone is set to “English”, the language of the first headset connected to the first mobile phone is set to “Chinese”, and the language of the third headset connected to the second mobile phone Is also set to “Japanese” and the language of the fourth headset connected to the second mobile phone is also set to “Korean”.
  • FIG. 7 illustrates a scenario in which a digital device according to another embodiment of the present invention masks a voice / emotion of a talker.
  • FIG. 7 proposes a technique for more precisely detecting the intent of the talker.
  • the listener can identify who the talker is.
  • the language before translation is given so that the guide can be used to identify which country is speaking.
  • the stressed voices are mapped to words in the speaker's voice, and the translated sentences are conveyed for feelings through stressed stresses.
  • expressing and displaying the stress or emotion in text will be said to belong to another scope of the present invention.
  • FIG. 8 shows a flow chart embodying the scenario shown in FIG. 7 in more detail.
  • the digital device capable of speech recognition receives a user voice input (S801).
  • a user voice input S801
  • feature points of the input voice are extracted (S802).
  • the feature point corresponds to, for example, stress, intonation, voice, etc., and other additional information may be extracted.
  • the recognized voice is converted into text (S803) and translated into another preset language (S804).
  • the extracted feature points as well as the translated content are mapped to each other and transmitted to another device or an internal controller (S805).
  • step S805 it is determined whether a feature point exists in the data received in step S805 (S806).
  • step S809 After the feature point masked text is applied to the text to speech (TTS), the process moves to step S809. Therefore, there is a technical effect that the emphasis portion of the first spoken first language can be reflected almost the same or similarly in the interpreted second language.
  • FIG. 9 is a block diagram showing a main configuration of the digital device shown in FIGS. 1 to 8. Accordingly, FIG. 9 may be interpreted independently, and supplementary interpretation with reference to the previous drawings is also within the scope of the present invention. However, in particular, the multilateral interpretation service illustrated in FIG. 6 is specifically illustrated in FIG.
  • the A device 900, the server 910, and the B device group 920 to which the present invention is applied are shown separately, but some or all of the functions of the server 910 are A device 900 or B device. Designing to perform in group 920 is also within the scope of the present invention.
  • the voice message of the first language received through the A device 900 is transmitted to the message receiving unit 911 of the server 910.
  • the message receiving unit 911 receives a voice message from the A device 900 (S902), and translates the message into an official language (for example, English) (S903).
  • the language that you want to receive from the B device group 920 through an interpreter is transmitted to the server 910 (S900), the translation engine 912 is translated into English as a public language with reference to the translation manager 913 (S904). Retranslation is performed in a specific language received in step S900 (S905).
  • the message transmitter 914 transmits the translated message to respective receiving devices in the B device group 920 (S906).
  • the present invention described above can be embodied as computer readable codes on a medium on which a program is recorded.
  • the computer-readable medium includes all kinds of recording devices in which data that can be read by a computer system is stored.
  • Examples of computer-readable media include hard disk drives (HDDs), solid state disks (SSDs), silicon disk drives (SDDs), ROMs, RAMs, CD-ROMs, magnetic tapes, floppy disks, optical data storage devices, and the like. This also includes implementations in the form of carrier waves (eg, transmission over the Internet).
  • the computer may include a control unit of the terminal.
  • the present invention is applicable to any type of digital device capable of speech recognition, and there is industrial applicability.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Machine Translation (AREA)

Abstract

La présente invention concerne un dispositif numérique assurant une reconnaissance vocale. Un dispositif numérique selon un mode de réalisation comprend une mémoire, un microphone pour recevoir un signal audio dans une première langue, et un dispositif de commande pour commander la mémoire et le microphone. Plus particulièrement, le dispositif de commande est caractérisé en ce qu'il traduit le signal audio reçu dans la première langue en un texte dans une seconde langue préconfigurée en référence à la mémoire, en ce qu'il retraduit le texte traduit dans la seconde langue dans la première langue en référence à la mémoire, en ce qu'il compare le texte retraduit dans la première langue et le signal audio reçu dans la première langue, en ce qu'il modifie une partie ou la totalité du texte traduit dans la seconde langue en fonction du résultat de la comparaison, et en ce qu'il convertit celui-ci en un format permettant la sortie en un signal audio.
PCT/KR2018/007797 2018-07-02 2018-07-10 Dispositif numérique assurant une reconnaissance vocale et son procédé de commande WO2020009261A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020180076422A KR20200003529A (ko) 2018-07-02 2018-07-02 음성 인식이 가능한 디지털 디바이스 및 그 제어 방법
KR10-2018-0076422 2018-07-02

Publications (1)

Publication Number Publication Date
WO2020009261A1 true WO2020009261A1 (fr) 2020-01-09

Family

ID=69059223

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2018/007797 WO2020009261A1 (fr) 2018-07-02 2018-07-10 Dispositif numérique assurant une reconnaissance vocale et son procédé de commande

Country Status (2)

Country Link
KR (1) KR20200003529A (fr)
WO (1) WO2020009261A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112614482A (zh) * 2020-12-16 2021-04-06 平安国际智慧城市科技股份有限公司 移动端外语翻译方法、系统及存储介质

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006127356A (ja) * 2004-11-01 2006-05-18 Brother Ind Ltd 機械翻訳装置、プログラム及び記録媒体
KR20140079598A (ko) * 2012-12-17 2014-06-27 한국전자통신연구원 문맥 검증 장치 및 방법
KR20140121580A (ko) * 2013-04-08 2014-10-16 한국전자통신연구원 자동 번역 및 통역 장치 및 그 방법
KR20170103209A (ko) * 2016-03-03 2017-09-13 한국전자통신연구원 원시 발화자의 목소리와 유사한 특성을 갖는 합성음을 생성하는 자동 통역 시스템 및 그 동작 방법
JP2018005368A (ja) * 2016-06-29 2018-01-11 株式会社Nttドコモ 出力態様決定システム

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2006127356A (ja) * 2004-11-01 2006-05-18 Brother Ind Ltd 機械翻訳装置、プログラム及び記録媒体
KR20140079598A (ko) * 2012-12-17 2014-06-27 한국전자통신연구원 문맥 검증 장치 및 방법
KR20140121580A (ko) * 2013-04-08 2014-10-16 한국전자통신연구원 자동 번역 및 통역 장치 및 그 방법
KR20170103209A (ko) * 2016-03-03 2017-09-13 한국전자통신연구원 원시 발화자의 목소리와 유사한 특성을 갖는 합성음을 생성하는 자동 통역 시스템 및 그 동작 방법
JP2018005368A (ja) * 2016-06-29 2018-01-11 株式会社Nttドコモ 出力態様決定システム

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112614482A (zh) * 2020-12-16 2021-04-06 平安国际智慧城市科技股份有限公司 移动端外语翻译方法、系统及存储介质

Also Published As

Publication number Publication date
KR20200003529A (ko) 2020-01-10

Similar Documents

Publication Publication Date Title
WO2015111818A1 (fr) Dispositif de synthèse de parole émotionnelle, son procédé de fonctionnement et terminal mobile le comprenant
WO2018034552A1 (fr) Dispositif et procédé de traduction de langue
WO2013047968A1 (fr) Procédé d'interface utilisateur et dispositif correspondant
WO2020159288A1 (fr) Dispositif électronique et son procédé de commande
WO2015199288A1 (fr) Terminal du type lunettes, et procédé de commande de ce terminal
WO2020080635A1 (fr) Dispositif électronique permettant d'effectuer une reconnaissance vocale à l'aide de microphones sélectionnés d'après un état de fonctionnement, et procédé de fonctionnement associé
WO2021060728A1 (fr) Dispositif électronique permettant de traiter un énoncé d'utilisateur et procédé permettant de faire fonctionner celui-ci
WO2020009261A1 (fr) Dispositif numérique assurant une reconnaissance vocale et son procédé de commande
WO2015030340A1 (fr) Dispositif terminal et dispositif mains libres pour service d'interprétation automatique mains libres, et procédé de service d'interprétation automatique mains libres
WO2021071271A1 (fr) Appareil électronique et procédé de commande associé
WO2020101174A1 (fr) Procédé et appareil pour produire un modèle de lecture sur les lèvres personnalisé
WO2019004762A1 (fr) Procédé et dispositif permettant de fournir une fonction d'interprétation à l'aide d'un écouteur
WO2022216059A1 (fr) Dispositif électronique et procédé de fourniture d'informations audio personnalisées
WO2022191395A1 (fr) Appareil de traitement d'une instruction utilisateur et son procédé de fonctionnement
WO2021080362A1 (fr) Système de traitement de langue utilisant un écouteur
WO2011028051A2 (fr) Dispositif électronique et procédé de reconnaissance vocale faisant intervenir ce dispositif
WO2020138943A1 (fr) Appareil et procédé de reconnaissance vocale
WO2020076087A1 (fr) Dispositif électronique et son procédé de fonctionnement
WO2024014869A1 (fr) Procédé de traitement de traduction et dispositif électronique
WO2024071946A1 (fr) Procédé de traduction basé sur une caractéristique vocale et dispositif électronique associé
WO2022177063A1 (fr) Dispositif électronique et son procédé de commande
WO2022186471A1 (fr) Procédé pour fournir un service d'appel de groupe et dispositif électronique le prenant en charge
WO2021235890A1 (fr) Procédé de sortie de texte dans un service d'assistant virtuel à intelligence artificielle et dispositif électronique prenant en charge ledit procédé
WO2021246689A1 (fr) Dispositif électronique et procédé de traduction de voix ou de texte
WO2022177186A1 (fr) Dispositif électronique comprenant un haut-parleur et un microphone, et son procédé de fonctionnement

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 18925387

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 18925387

Country of ref document: EP

Kind code of ref document: A1