US20180150458A1 - User terminal device for providing translation service, and method for controlling same - Google Patents
User terminal device for providing translation service, and method for controlling same Download PDFInfo
- Publication number
- US20180150458A1 US20180150458A1 US15/572,400 US201615572400A US2018150458A1 US 20180150458 A1 US20180150458 A1 US 20180150458A1 US 201615572400 A US201615572400 A US 201615572400A US 2018150458 A1 US2018150458 A1 US 2018150458A1
- Authority
- US
- United States
- Prior art keywords
- message
- user terminal
- terminal device
- touch gesture
- displayed
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000013519 translation Methods 0.000 title claims abstract description 91
- 238000000034 method Methods 0.000 title claims description 25
- 238000004891 communication Methods 0.000 claims abstract description 50
- 230000004044 response Effects 0.000 claims description 26
- 230000033001 locomotion Effects 0.000 claims description 17
- 238000010586 diagram Methods 0.000 description 16
- 230000005540 biological transmission Effects 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 229920001621 AMOLED Polymers 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000004888 barrier function Effects 0.000 description 1
- 244000240602 cacao Species 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 239000007788 liquid Substances 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 238000000060 site-specific infrared dichroism spectroscopy Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
Images
Classifications
-
- G06F17/289—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/40—Processing or translation of natural language
- G06F40/58—Use of machine translation, e.g. for multi-lingual retrieval, for server-side translation for client devices or for real-time translation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
- G06F9/454—Multi-language systems; Localisation; Internationalisation
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/04—Details of speech synthesis systems, e.g. synthesiser structure or memory management
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/06—Message adaptation to terminal or network requirements
- H04L51/063—Content adaptation, e.g. replacement of unsuitable content
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W4/00—Services specially adapted for wireless communication networks; Facilities therefor
- H04W4/12—Messaging; Mailboxes; Announcements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04808—Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/52—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail for supporting social networking services
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W88/00—Devices specially adapted for wireless communication networks, e.g. terminals, base stations or access point devices
- H04W88/02—Terminal devices
Definitions
- the present disclosure relates to a user terminal device and a method for controlling the same, and more particularly, to a user terminal device for providing translation service and a method for controlling the same.
- instant message transmitting applications such as mobile messenger, social network service (SNS) or the like are expanded and users thereof are exponentially increased.
- services such as instant message transmitting applications and so on are actively implemented, even the general users need free communication with foreigners using different languages.
- an object of the present disclosure is to provide a user terminal device for translating messages transmitted and received between user terminal devices and providing a result more conveniently and a method for controlling the same.
- the present invention provides a user terminal device for providing translation service, including a communication unit configured to perform communication with an external device, a display configured to display a message transmitted and received in a communication with the external device, a sensing unit configured to sense a gesture for the user terminal device, and a processor configured to provide the translation service for at least one of the displayed messages when a preset gesture is sensed.
- the processor may distinctively display the messages being transmitted and received on a message unit basis, and provide the translation service for the displayed messages on the message unit basis.
- the processor may control such that a translated message of the message for which touch gesture is inputted is displayed in a preset language.
- the processor may control such that the translated message for which the touch gesture is inputted is displayed in a source language before the translation.
- the touch gesture for the message may be a touch and drag in a first direction
- the touch gesture for the translated message may be a touch and drag in a second direction opposite the first direction
- the processor may control such that, the message for which the touch gesture is inputted is replaced by the translated message and displayed, or the translated message is displayed together with the message for which the touch gesture is inputted.
- the processor may control such that pronunciation information with respect to the translated message for which touch gesture is inputted may be displayed.
- the user terminal device may further include a speaker and, in response to a touch gesture with respect to the pronunciation information being input, the processor may converts the pronunciation information for which the touch gesture is inputted into voice and outputs the converted voice through the speaker.
- the processor may control such that all of the displayed messages may be translated into a preset language and displayed.
- the communication unit may perform communication with an external server for providing translation service
- the processor may controls such that at least one of the displayed message is transmitted to the external server, and a translated message of the at least one message received from the external server may be displayed.
- a control method of a user terminal device for providing translation service may include displaying a message transmitted and received in a communication with an external device, sensing a gesture for the user terminal device, and providing the translation service for at least one of the displayed messages.
- the displaying may include distinctively displaying the transmitted and received message on a message unit basis
- the providing the translation service may include providing the translation service for the displayed message on a message unit basis.
- the providing the translation service may include displaying a translated message of the message for which touch gesture is inputted, in a preset language.
- the providing the translation service may include displaying the translated message for which the touch gesture is inputted in a source language before translation.
- the touch gesture inputted for the message may be a touch and drag in a first direction
- the touch gesture inputted for the translated message may be a touch and drag in a second direction opposite the first direction
- the providing the translation service may include replacing the message for which the touch gesture is inputted by the translated message and displaying a result, or displaying the translated message together with the message for which the touch gesture is inputted.
- pronunciation information with respect to the translated message for which touch gesture is inputted may be displayed.
- the pronunciation information for which the touch gesture is inputted may be converted into voice, and the converted voice may be outputted through the speaker.
- all of the displayed messages may be translated into a preset target language and displayed.
- the providing translation service may include transmitting at least one of the displayed messages to the external server, receiving a translated message with respect to at least one message from the external server, and displaying the received translated message.
- a message transmitted and received between user terminal devices can be instantly translated, and communication between users using different languages from each other can be performed further facilitated.
- FIG. 1 is a block diagram briefly illustrating a constitution of a user terminal device according to an embodiment
- FIG. 2 is a diagram illustrating a system in which a user terminal device in a communication with an external server performs translation according to an embodiment
- FIGS. 3 to 5 are diagrams provided to explain a touch gesture for translating messages according to an embodiment
- FIG. 6 is a diagram provided to explain a touch gesture for translating messages according to another embodiment
- FIG. 7 is a diagram provided to explain a method for displaying pronunciation information of a message according to an embodiment
- FIGS. 8 and 9 are diagrams provided to explain a motion gesture for translating all of displayed messages according to an embodiment
- FIG. 10 is a diagram provided to explain a touch gesture for translating messages posted on the social network service according to an embodiment
- FIG. 11 is a block diagram illustrating a constitution of a user terminal device in detail according to another embodiment.
- FIG. 12 is a flowchart provided to explain a method of a user terminal device according to an embodiment.
- FIG. 1 is a block diagram briefly illustrating a constitution of a user terminal device according to an embodiment.
- the user terminal device 100 includes a communication unit 110 , a display 120 , a sensing unit 130 and a processor 140 .
- the communication unit 110 is configured to perform communication with various types of external devices according to various types of communication methods.
- the external devices may include at least one among a messaging service providing server 200 , a translation service providing server 300 and a counterpart user terminal device.
- the communication unit 110 may transmit a message written on the user terminal device 100 or receive a message from the counterpart user terminal device or the messaging service providing server 200 providing the message transmitting and receiving services, in a communication with the counterpart user terminal device or the messaging service providing server 200 .
- the messaging service providing server 200 refers to a server for providing service to relay message transmission and reception with respect to the counterpart user terminal device.
- the communication unit 110 may perform communication with the translation service providing server 300 .
- the communication unit 110 may generate a translation request message to request translation of a message selected by a user into a language according to preset translation options, and transmit the generated translation request message to the translation service providing server 300 .
- the communication unit 110 may receive the translated message of the selected message from the translation service providing server 300 .
- the communication unit 110 may include Wi-Fi chip, Bluetooth chip, wireless communication chip, NFC chip or the like.
- the controller 140 may perform communication with the external devices described above by using the communication unit 110 .
- Wi-Fi chip and Bluetooth chip perform communication according to Wi-Fi method and Bluetooth method, respectively.
- various connection information such as SSID and session key
- the wireless communication chip refers to a chip for performing communication according to various communication standards, such as IEEE, ZigBee, 3rd Generation (3G), 3rd Generation Partnership Project (3GPP), Long Term Evolution (LTE), or the like.
- the NFC chip refers to a chip that operates according to Near Field Communication (NFC) utilizing 13.56 MHz bandwidth among various RF-ID frequency bandwidths such as 135 kHz, 13.56 MHz, 433 MHz, 860-960 MHz, 2.45 GHz, or the like.
- NFC Near Field Communication
- the display 120 may provide various content screens.
- content screen may include various contents such as image, video, and text, application running screen including various contents, graphic user interface (GUI) screen, or the like.
- GUI graphic user interface
- the display 120 may display a message transmitted to, or received from an external device, and a translated message of the transmitted and received message for the notice of a user.
- a method for implementing the display 120 is not strictly limited.
- the display may be implemented as various forms of displays, such as liquid crystal display (LCD), organic light emitting diodes (OLED) display, active-matrix organic light-emitting diode (AM-OLED), plasma display panel (PDP), or the like.
- the display 120 may additionally include additional configurations depending on the method for implementing the same.
- the display 120 may include a LCD display panel (not illustrated), a backlight unit (not illustrated) for providing light, and a panel drive substrate (not illustrated) for driving a panel (not illustrated).
- the display 120 may be preferably combined with a touch sensing unit of the sensing unit 130 to be thus provided as a touch screen.
- the sensing unit 130 may sense various user interactions.
- the sensing unit 130 may be configured to include a motion gesture sensing unit (not illustrated) and a touch gesture sensing unit (not illustrated).
- the motion gesture sensing unit may include at least one of an acceleration sensor and a gyro sensor, which can sense a motion of the user terminal device 100 .
- the touch gesture sensing unit may include a touch sensor.
- the touch gesture sensing unit may sense a touch input of a user, using the touch sensor attached on a back side of the display panel.
- the processor 140 may obtain information such as touch coordinates, touch time or the like from the touch gesture sensor to determine the type of the sensed touch input (e.g., tap gesture, double tap gesture, panning gesture, flick gesture, touch and drag gesture, and so on). Further, the touch gesture sensing unit may directly determine the type of the touch input, using the obtained touch coordinates and touch time.
- the processor 140 is configured to control overall operation of the user terminal device 100 .
- the processor 140 may provide the translation service for at least one message among the messages displayed on the display 120 , when a preset gesture is sensed by the sensing unit 130 . Specifically, the processor 140 may transmit at least one message of the displayed messages to the external server for providing translation service, and when a translated message is received from the external device, control the display 120 to display the received translated message.
- the processor 140 may distinctively display the messages being transmitted and received on a message unit basis, and provide the translation service for the displayed messages on the message unit basis.
- the processor 140 may provide the translation service for a message selected by a user by a unit of a message box that includes word balloon, comment window or the like, based on which separately transmitted and received messages are divided and displayed.
- the processor 140 may control such that the message for which touch gesture is inputted is translated and the corresponding translated message in a preset language is displayed.
- the touch gesture for the message may be a touch and drag in one direction.
- the processor 140 may control the display 120 to replace the message for which the touch gesture is inputted with a translated message, or display the translated message together with the message for which the touch gesture is inputted.
- the processor 140 may control the display 120 to translate all of displayed messages into a preset language and display the result.
- the motion gesture may be various motions of the user terminal device 100 including rotating movement, linear movement, reciprocal movement such as shaking, and so on.
- FIG. 2 is a diagram illustrating a system in which a user terminal device in a communication with an external server performs translation.
- a network 20 may include a messaging service providing server 200 and a translation service providing server 300 .
- the network 20 may be a single network or a combination of networks, which may wirelessly connect the translation service providing server 300 , the user terminal device 100 , and the messaging service providing server 200 for mutual communication of message-related data.
- the user terminal device 100 may be generally implemented as a small-sized device such as a smartphone, and so on, and accordingly, has a limit in storing data in the user terminal device 100 in view of a memory capacity. Therefore, the user terminal device 100 according to an embodiment may be provided with the translation data from the translation service providing server 300 via the communication unit 110 .
- the translation service providing server 300 may receive a selected message from the user terminal device 100 and perform translation of the received message. Specifically, the translation service providing server 300 may perform translation of a source message to be translated based on the translation data included in a translation DB loaded therein and transmit a translated message to the user terminal device 100 .
- the translation DB may store data for performing translation according to various national languages.
- the user terminal device 100 may transmit language setting information in which a target national language of the translation is set, when transmitting a source message to be translated to the translation service providing server 300 .
- a user may set at the user terminal device 100 a Korean language as a target language into which a message is to be translated, in which case, the language setting information may be transmitted together when a source message to be translated is transmitted.
- the translation service providing server 300 may perform translation according to the target language of the translation based on the received language setting information.
- the messaging service providing server 200 is a mobile carrier server, which provides the messaging service.
- the messaging service providing server 200 may include at least one among a server for relaying transmission of messages, such as short messaging service, multimedia messaging service or the like, a server for relaying transmission of messages by mobile messenger service, and a server for providing social network service.
- the user terminal device 100 may transmit and receive a message with the counterpart user terminal device through the messaging service providing server 200 or transmit a message to a server for providing social network service.
- FIGS. 3 to 5 are diagrams provided to explain a touch gesture for translating a message according to an embodiment.
- a user may execute a messaging application on the user terminal device 100 and view a received message on the user terminal device 100 , and then write a message and transmit it to the counterpart user terminal device.
- a messaging application may distinctively display each of the separately transmitted and received messages, as illustrated in FIG. 3 .
- the messages being transmitted and received may be distinguished by message boxes 31 , 32 that encircle each of the messages.
- the user may perform a preset touch gesture with respect to a message 41 that is intended to be translated so that the message 41 is translated.
- a target language of the translation may be set or modified by an option menu provided from a messaging application or the user terminal device 100 .
- the message 41 When the user performs a touch and drag with respect to the message 41 in a direction from left to right, the message 41 may be translated into a preset national language.
- a unit of translation may correspond to a message box unit of distinguishing the messages being transmitted and received.
- FIG. 4 b illustrates a screen displaying a translated message for which the tough gesture is inputted.
- a preset target language is English
- the user terminal device 100 may transmit the message 41 “ ?” for which the gesture is inputted to the translation service providing server 300 , and receive a translated message 43 “Where shall we meet?” from the translation service providing server 300 .
- the user terminal device 100 may transmit target language setting information (English) together with the message for which the gesture is inputted.
- the user terminal device 100 may replace the message 41 for which touch gesture is inputted into the translated message 43 and display the translated message 43 as illustrated in FIG. 4 b.
- the user terminal device 100 may additionally display the translated message “Where shall we meet?” together with the source message “ ?” for which touch gesture is inputted. For example, the user terminal device 100 may divide a region of the message box for which touch gesture is inputted or additionally generate another message box under the message box, and display the translated message “Where shall we meet?” in the divided message box region or the additionally generated message box.
- the processor 140 may convert the translated message into voice and output the converted voice through the speaker.
- the communication unit 110 may transmit the translated message to a text to speech (TTS) server for converting text into voice, and receive the converted voice signal of the translated message from the TTS server.
- TTS text to speech
- the translated message may be un-translated into the source message as illustrated in FIG. 5 .
- the processor 140 may control such that the source message 52 before translation is displayed instead of the translated message 51 for which the touch gesture is inputted.
- the touch gesture to un-translate the translated message 51 back into the source message 52 may be a gesture in an opposite direction to the touch gesture performed to translate the source message 52 .
- the touch gesture for translating the received source message before translation may be a touch and drag in a first direction
- the touch gesture for un-translating the translated message back into the source message before translation may be a touch and drag in a second direction opposite the first direction.
- FIG. 6 is a diagram provided to explain a touch gesture for translating a message according to another embodiment.
- the user terminal device 100 may display an instruction statement to inform about an operation to be performed according to touch and drag (e.g., “View translated text 61 ”) upon inputting of the touch and drag.
- touch and drag e.g., “View translated text 61 ”
- the instruction statement disappears and a translated text for the message may be displayed.
- an instruction statement such as “View source text 62 ” may be displayed while a touch and drag is being inputted in a direction from right to left with respect to the translated message.
- the instruction statement may disappear and the source message of the translated message may be displayed.
- FIG. 7 is a diagram provided to explain a method for displaying pronunciation information of a message according to an embodiment.
- the processor 140 may control such that pronunciation information of any of the messages displayed on the display 120 is displayed. Specifically, when the user performs a preset touch gesture for a message having pronunciation information, the processor 140 may display the pronunciation information such as phonetic alphabets of corresponding message. In an example, the pronunciation information may be additional information displayed together with corresponding message.
- the processor 140 may display phonetic alphabets of such Chinese language message.
- a region of a message box 71 for which gesture is inputted may be divided in response to the pinch-out gesture, and a Chinese language message 72 along with the phonetic alphabets 73 of the Chinese language message may be displayed in each of the divided regions.
- Chinese pronunciation information may be information stored in the user terminal device 100 .
- the processor 140 may transmit a message for which gesture is inputted to the translation service providing server 300 , and receive and display pronunciation information of the message from the translation service providing server 300 .
- the user terminal device 100 may further include the speaker such that, when preset touch gesture is inputted for the pronunciation information, the processor 140 may convert the pronunciation information for which touch gesture is inputted into voice and output the converted voice through the speaker. For example, when a user double-touches the region displaying the pronunciation information 73 of the Chinese language message, the speaker may output voice according to pronunciation information.
- FIGS. 8 and 9 are diagrams provided to explain a motion gesture for translating all of displayed messages according to an embodiment.
- a user may have a conversation with a single user or a plurality of users using a mobile messenger.
- the mobile messenger includes commercial messenger applications such as Kakao Talk, Line, WhatsApp, or the like.
- FIG. 8 illustrates an embodiment in which a user transmits and receives messages at real time among a plurality of users including “Mike” who uses English language and “Michiko” who uses Japanese language.
- real-time translation with respect to the messages may be performed in response to a preset gesture performed for each of the messages.
- a target language of the translation is set to be Korean
- both the English language message 81 and the Japanese language message 82 may be translated into Korean language messages.
- the users may actively communicate without having a pause in their conversation because screen with messages in translated languages is available to be viewed on a corresponding messaging application.
- the user may perform a preset motion gesture on the user terminal device 100 to view each of the translated messages of all the messages displayed on the display 120 .
- the processor 140 may translate the messages in foreign language 81 , 82 for which translation is available, among the messages displayed on the display 120 in preset languages, and replace the messages 81 , 82 with the translated target messages 91 , 92 and display the translated target messages 91 , 92 .
- the processor 140 may transmit all the displayed foreign-language messages 81 , 82 to the translation service providing server 300 , and receive and display corresponding translated messages 91 , 92 .
- the user may translate all of displayed messages into preset national languages by simply performing one motion gesture without having to individually perform a touch gesture with respect to each of the messages.
- FIG. 10 is a diagram provided to explain a touch gesture for translating a message posted on the social network service according to an embodiment.
- real-time translation may be performed with respect to a message such as article posted on a mobile page or comments thereof provided through the social network service (SNS).
- SNS includes online services such as Facebook or Twitter for building up a relation network of internet users on online.
- the SNS may be executed on the user terminal device 100 using an SNS providing application as a platform.
- FIG. 10 a illustrates screen in which the user terminal device 100 connects to SNS.
- SNS may be configured with basic platform including a posted message 1010 and comment type messages 1020 - 1040 thereof.
- the posted message 1010 and comment messages 1020 - 1040 may be distinctively displayed on a message unit basis, in which case each of the messages may be translated on the message unit basis.
- the user may perform a touch and drag gesture with respect to one (e.g., message 1040 ) of the comment messages 1020 - 1040 , and the message 1040 for which touch and drag gesture is inputted may be replaced by a translated message 1050 in a preset target language and displayed.
- one e.g., message 1040
- the message 1040 for which touch and drag gesture is inputted may be replaced by a translated message 1050 in a preset target language and displayed.
- FIG. 11 is a block diagram illustrating a constitution of a user terminal device in detail according to another embodiment.
- the user terminal device 100 ′ according to another embodiment includes a communication unit 110 , a display 120 , a sensing unit 130 , a processor 140 , a storage 150 , an image processor 160 , an audio processor 170 , an audio outputter 180 and a user interface 190 .
- a communication unit 110 a communication unit 110
- the display 120 includes a display 120 , a sensing unit 130 , a processor 140 , a storage 150 , an image processor 160 , an audio processor 170 , an audio outputter 180 and a user interface 190 .
- the processor 140 includes RAM 141 , ROM 142 , a graphic processor 143 (CPU 144 , first to nth interfaces 145 - 1 to 145 - n ) and a bus 146 .
- RAM 141 , ROM 142 , and the graphic processor 143 may be connected to each other via the bus 146 .
- the first to nth interfaces 145 - 1 to 145 - n are connected to the elements described above.
- One of the interfaces may be a network interface connected to an external device through network.
- the CPU 144 may access the storage 140 and perform booting by using O/S stored in the storage 140 . Further, the CPU 144 may perform various operations by using various programs, contents and data stored in the storage 140 .
- the RAM 141 stores instruction sets for system booting. Upon powering-on in response to input of turn-on command, the CPU 144 copies RAM 141 stored in the storage 150 onto the RAM 141 according to the instructions stored in the ROM 142 , and executes the O/S to boot the system. When the booting is completed, the CPU 144 copies various application programs stored in the storage 150 onto the RAM 141 , and executes the application program copied to the RAM 141 to perform various operations.
- the graphic processor 143 may generate a screen including various objects such as icons, images, texts or the like, using an arithmetic unit (not illustrated) or a renderer (not illustrated).
- the arithmetic unit calculates attribute values such as coordinate values, forms, sizes, colors or the like according to layouts of screens.
- the renderer may generate various layouts of screens including objects based on the attribute values calculated at the arithmetic unit.
- the operation of the processor 140 described above may be performed by implementing the programs stored in the storage 150 .
- the storage 150 may store O/S (operating system) software module for driving the user terminal device 100 ′ and various multimedia contents.
- O/S operating system
- the storage 150 may store a base module for processing signals delivered from each hardware included in the user terminal device 100 ′, a storage module for managing database (DB) or registry, a graphic processing module for generating layouts of screens, a security module or the like.
- the storage 150 may store programs such as communication module, translation module or the like, which are necessary for implementation of the translation service according to an embodiment.
- the processor 140 may perform communication with the counterpart user terminal device 200 , the messaging service providing server 300 , the translation service providing server 300 or the like by using the communication module.
- the image processor 160 is configured to process various images such as decoding, scaling, noise filtering, frame rate converting, resolution converting or the like with respect to contents.
- the audio processor 170 is configured to process audio data. Specifically, the audio processor 170 may process the pronunciation information for which touch gesture is inputted to convert it into voice data, and deliver the converted voice data to the audio outputter 180 .
- the audio outputter 180 is configured to output audio data processed in the audio processor 170 .
- the audio outputter 180 may output the converted voice data through a receiver or a speaker.
- the user interface 190 is configured to sense user interaction for controlling overall operation of the user terminal device 100 ′.
- FIG. 12 is a flowchart provided to explain a control method of a user terminal device according to an embodiment.
- the user terminal device may control such that a message transmitted and received in a communication with an external device is displayed, at S 1210 .
- the messages being transmitted and received may be distinctively displayed on a message unit basis.
- a gesture for the user terminal device may be sensed.
- the gesture inputted for a message may be a touch and drag in a first direction.
- a touch gesture inputted for the translated message of the message may be a touch and drag in a second direction opposite the first direction.
- the user terminal device may provide the translation service for at least one of the displayed messages.
- translation service for the displayed messages may be provided on a message unit basis.
- it in response to sensing a touch gesture inputted for at least one of the display messages, it may be controlled such that the message for which touch gesture may be inputted is translated and the corresponding translated message in a preset language may be displayed.
- it in response to sensing a touch gesture inputted for the translated message, it may be controlled such that the translated message for which touch gesture is inputted may be un-translated back into the source message before translation.
- it may be controlled such that the message for which touch gesture is inputted may be replaced by the translated message and displayed. Further, it may be controlled such that the message for which touch gesture is inputted may be displayed together with the translated message.
- pronunciation information with respect to the translated message for which touch gesture is inputted may be displayed.
- messages transmitted and received among user terminal devices may be instantly translated. Accordingly, communication between users using different languages from each other may be performed more actively.
- control method of the user terminal device may be implemented as a program and stored in various recording media.
- a computer program processible with various processors for implementing various control methods described above may be stored in recording media and used.
- a non-transitory computer readable recording medium may be provided, storing therein a program for performing operations of displaying a message transmitted and received in a communication with an external device, sensing a gesture inputted for the user terminal device, and providing translation service for at least one of the displayed messages.
- the non-transitory computer readable medium is a medium capable of storing data semi-permanently and being readable by a device, rather than a medium such as register, cash, and memory that stores the data for a brief period of time.
- a non-transitory computer readable medium such as CD, DVD, hard disk, Blu-ray disk, USB, memory card, ROM, and so on.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Acoustics & Sound (AREA)
- Information Transfer Between Computers (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
- The present disclosure relates to a user terminal device and a method for controlling the same, and more particularly, to a user terminal device for providing translation service and a method for controlling the same.
- With distribution of smart phones and development of the information and communication technology, instant message transmitting applications such as mobile messenger, social network service (SNS) or the like are expanded and users thereof are exponentially increased. With the services such as instant message transmitting applications and so on are actively implemented, even the general users need free communication with foreigners using different languages.
- However, inconvenience is experienced due to language barrier when users of different languages are conversing with each other on the message transmitting applications. Conventionally, in order to translate messages transmitted and received in foreign languages, translation is performed by using separately installed translation applications or web pages providing translation service. However, these methods require works such as installing and implementing the translation applications, accessing web pages, or the like. Further, because each foreign language message should be copied and pasted in order to perform translation, problems such as user inconvenience and difficulty of continuing smooth conversation may occur.
- Accordingly, in order to solve inconvenience mentioned above, a form of translation solution suitable for transmitted and received messages is required.
- Accordingly, an object of the present disclosure is to provide a user terminal device for translating messages transmitted and received between user terminal devices and providing a result more conveniently and a method for controlling the same.
- In order to accomplish the above-mentioned objects, the present invention provides a user terminal device for providing translation service, including a communication unit configured to perform communication with an external device, a display configured to display a message transmitted and received in a communication with the external device, a sensing unit configured to sense a gesture for the user terminal device, and a processor configured to provide the translation service for at least one of the displayed messages when a preset gesture is sensed.
- Further, the processor may distinctively display the messages being transmitted and received on a message unit basis, and provide the translation service for the displayed messages on the message unit basis.
- Further, in response to a touch gesture for at least one of the displayed messages being sensed, the processor may control such that a translated message of the message for which touch gesture is inputted is displayed in a preset language.
- Further, in response to a touch gesture for the translated message being sensed, the processor may control such that the translated message for which the touch gesture is inputted is displayed in a source language before the translation.
- Further, the touch gesture for the message may be a touch and drag in a first direction, and the touch gesture for the translated message may be a touch and drag in a second direction opposite the first direction.
- Further, the processor may control such that, the message for which the touch gesture is inputted is replaced by the translated message and displayed, or the translated message is displayed together with the message for which the touch gesture is inputted.
- Further, in response to a touch gesture for the translated message being input, the processor may control such that pronunciation information with respect to the translated message for which touch gesture is inputted may be displayed.
- Further, in an embodiment, the user terminal device may further include a speaker and, in response to a touch gesture with respect to the pronunciation information being input, the processor may converts the pronunciation information for which the touch gesture is inputted into voice and outputs the converted voice through the speaker.
- Further, in response to a motion gesture for the user terminal device being sensed, the processor may control such that all of the displayed messages may be translated into a preset language and displayed.
- Further, the communication unit may perform communication with an external server for providing translation service, and the processor may controls such that at least one of the displayed message is transmitted to the external server, and a translated message of the at least one message received from the external server may be displayed.
- Meanwhile, a control method of a user terminal device for providing translation service according to an embodiment is provided, which may include displaying a message transmitted and received in a communication with an external device, sensing a gesture for the user terminal device, and providing the translation service for at least one of the displayed messages.
- Further, the displaying may include distinctively displaying the transmitted and received message on a message unit basis, and the providing the translation service may include providing the translation service for the displayed message on a message unit basis.
- Further, in response to a touch gesture for at least one of the displayed messages being sensed, the providing the translation service may include displaying a translated message of the message for which touch gesture is inputted, in a preset language.
- Further, in response to a touch gesture for the translated message being sensed, the providing the translation service may include displaying the translated message for which the touch gesture is inputted in a source language before translation.
- Further, the touch gesture inputted for the message may be a touch and drag in a first direction, and the touch gesture inputted for the translated message may be a touch and drag in a second direction opposite the first direction.
- Further, the providing the translation service may include replacing the message for which the touch gesture is inputted by the translated message and displaying a result, or displaying the translated message together with the message for which the touch gesture is inputted.
- Further, in response to sensing a touch gesture inputted for the translated message, pronunciation information with respect to the translated message for which touch gesture is inputted may be displayed.
- Further, in response to inputting of a touch gesture for the pronunciation information, the pronunciation information for which the touch gesture is inputted may be converted into voice, and the converted voice may be outputted through the speaker.
- Further, in response to sensing a motion gesture inputted for the user terminal device, all of the displayed messages may be translated into a preset target language and displayed.
- Further, the providing translation service may include transmitting at least one of the displayed messages to the external server, receiving a translated message with respect to at least one message from the external server, and displaying the received translated message.
- According to the above various embodiments, a message transmitted and received between user terminal devices can be instantly translated, and communication between users using different languages from each other can be performed further facilitated.
- The above and other objects, features and advantages of the present disclosure will become more apparent to those of ordinary skill in the art by describing in detail exemplary embodiments thereof with reference to the accompanying drawings, in which:
-
FIG. 1 is a block diagram briefly illustrating a constitution of a user terminal device according to an embodiment; -
FIG. 2 is a diagram illustrating a system in which a user terminal device in a communication with an external server performs translation according to an embodiment; -
FIGS. 3 to 5 are diagrams provided to explain a touch gesture for translating messages according to an embodiment; -
FIG. 6 is a diagram provided to explain a touch gesture for translating messages according to another embodiment; -
FIG. 7 is a diagram provided to explain a method for displaying pronunciation information of a message according to an embodiment; -
FIGS. 8 and 9 are diagrams provided to explain a motion gesture for translating all of displayed messages according to an embodiment; -
FIG. 10 is a diagram provided to explain a touch gesture for translating messages posted on the social network service according to an embodiment; -
FIG. 11 is a block diagram illustrating a constitution of a user terminal device in detail according to another embodiment; and -
FIG. 12 is a flowchart provided to explain a method of a user terminal device according to an embodiment. -
FIG. 1 is a block diagram briefly illustrating a constitution of a user terminal device according to an embodiment. - Referring to
FIG. 1 , theuser terminal device 100 according to an embodiment includes acommunication unit 110, adisplay 120, asensing unit 130 and aprocessor 140. - The
communication unit 110 is configured to perform communication with various types of external devices according to various types of communication methods. In an example, the external devices may include at least one among a messagingservice providing server 200, a translationservice providing server 300 and a counterpart user terminal device. - The
communication unit 110 may transmit a message written on theuser terminal device 100 or receive a message from the counterpart user terminal device or the messagingservice providing server 200 providing the message transmitting and receiving services, in a communication with the counterpart user terminal device or the messagingservice providing server 200. The messagingservice providing server 200 refers to a server for providing service to relay message transmission and reception with respect to the counterpart user terminal device. - Further, the
communication unit 110 may perform communication with the translationservice providing server 300. Thecommunication unit 110 may generate a translation request message to request translation of a message selected by a user into a language according to preset translation options, and transmit the generated translation request message to the translationservice providing server 300. Thecommunication unit 110 may receive the translated message of the selected message from the translationservice providing server 300. - The
communication unit 110 may include Wi-Fi chip, Bluetooth chip, wireless communication chip, NFC chip or the like. Thecontroller 140 may perform communication with the external devices described above by using thecommunication unit 110. - Specifically, Wi-Fi chip and Bluetooth chip perform communication according to Wi-Fi method and Bluetooth method, respectively. When Wi-Fi chip or Bluetooth chip is used, various connection information, such as SSID and session key, may be first transmitted and received, so that communication may be connected by using the connection information and various information may be transmitted and received. The wireless communication chip refers to a chip for performing communication according to various communication standards, such as IEEE, ZigBee, 3rd Generation (3G), 3rd Generation Partnership Project (3GPP), Long Term Evolution (LTE), or the like. The NFC chip refers to a chip that operates according to Near Field Communication (NFC) utilizing 13.56 MHz bandwidth among various RF-ID frequency bandwidths such as 135 kHz, 13.56 MHz, 433 MHz, 860-960 MHz, 2.45 GHz, or the like.
- The
display 120 may provide various content screens. In an example, content screen may include various contents such as image, video, and text, application running screen including various contents, graphic user interface (GUI) screen, or the like. According to an embodiment, thedisplay 120 may display a message transmitted to, or received from an external device, and a translated message of the transmitted and received message for the notice of a user. - A method for implementing the
display 120 is not strictly limited. For example, the display may be implemented as various forms of displays, such as liquid crystal display (LCD), organic light emitting diodes (OLED) display, active-matrix organic light-emitting diode (AM-OLED), plasma display panel (PDP), or the like. Thedisplay 120 may additionally include additional configurations depending on the method for implementing the same. For example, for a liquid display type ofdisplay 120, thedisplay 120 may include a LCD display panel (not illustrated), a backlight unit (not illustrated) for providing light, and a panel drive substrate (not illustrated) for driving a panel (not illustrated). Specifically, according to an embodiment, thedisplay 120 may be preferably combined with a touch sensing unit of thesensing unit 130 to be thus provided as a touch screen. - The
sensing unit 130 may sense various user interactions. Thesensing unit 130 may be configured to include a motion gesture sensing unit (not illustrated) and a touch gesture sensing unit (not illustrated). - The motion gesture sensing unit may include at least one of an acceleration sensor and a gyro sensor, which can sense a motion of the
user terminal device 100. - Further, the touch gesture sensing unit may include a touch sensor. The touch gesture sensing unit may sense a touch input of a user, using the touch sensor attached on a back side of the display panel. The
processor 140 may obtain information such as touch coordinates, touch time or the like from the touch gesture sensor to determine the type of the sensed touch input (e.g., tap gesture, double tap gesture, panning gesture, flick gesture, touch and drag gesture, and so on). Further, the touch gesture sensing unit may directly determine the type of the touch input, using the obtained touch coordinates and touch time. - The
processor 140 is configured to control overall operation of theuser terminal device 100. - The
processor 140 may provide the translation service for at least one message among the messages displayed on thedisplay 120, when a preset gesture is sensed by thesensing unit 130. Specifically, theprocessor 140 may transmit at least one message of the displayed messages to the external server for providing translation service, and when a translated message is received from the external device, control thedisplay 120 to display the received translated message. - Specifically, the
processor 140 may distinctively display the messages being transmitted and received on a message unit basis, and provide the translation service for the displayed messages on the message unit basis. In other words, theprocessor 140 may provide the translation service for a message selected by a user by a unit of a message box that includes word balloon, comment window or the like, based on which separately transmitted and received messages are divided and displayed. - According to an embodiment, in response to sensing a touch gesture inputted for at least one of the displayed messages, the
processor 140 may control such that the message for which touch gesture is inputted is translated and the corresponding translated message in a preset language is displayed. In an example, the touch gesture for the message may be a touch and drag in one direction. Theprocessor 140 may control thedisplay 120 to replace the message for which the touch gesture is inputted with a translated message, or display the translated message together with the message for which the touch gesture is inputted. - According to another embodiment, in response to sensing a motion gesture inputted to the
user terminal device 100, theprocessor 140 may control thedisplay 120 to translate all of displayed messages into a preset language and display the result. In an example, the motion gesture may be various motions of theuser terminal device 100 including rotating movement, linear movement, reciprocal movement such as shaking, and so on. -
FIG. 2 is a diagram illustrating a system in which a user terminal device in a communication with an external server performs translation. - A network 20 may include a messaging
service providing server 200 and a translationservice providing server 300. - The network 20 may be a single network or a combination of networks, which may wirelessly connect the translation
service providing server 300, theuser terminal device 100, and the messagingservice providing server 200 for mutual communication of message-related data. - The
user terminal device 100 according to an embodiment may be generally implemented as a small-sized device such as a smartphone, and so on, and accordingly, has a limit in storing data in theuser terminal device 100 in view of a memory capacity. Therefore, theuser terminal device 100 according to an embodiment may be provided with the translation data from the translationservice providing server 300 via thecommunication unit 110. - The translation
service providing server 300 may receive a selected message from theuser terminal device 100 and perform translation of the received message. Specifically, the translationservice providing server 300 may perform translation of a source message to be translated based on the translation data included in a translation DB loaded therein and transmit a translated message to theuser terminal device 100. In an example, the translation DB may store data for performing translation according to various national languages. Theuser terminal device 100 may transmit language setting information in which a target national language of the translation is set, when transmitting a source message to be translated to the translationservice providing server 300. For example, a user may set at the user terminal device 100 a Korean language as a target language into which a message is to be translated, in which case, the language setting information may be transmitted together when a source message to be translated is transmitted. The translationservice providing server 300 may perform translation according to the target language of the translation based on the received language setting information. - The messaging
service providing server 200 is a mobile carrier server, which provides the messaging service. The messagingservice providing server 200 may include at least one among a server for relaying transmission of messages, such as short messaging service, multimedia messaging service or the like, a server for relaying transmission of messages by mobile messenger service, and a server for providing social network service. - The
user terminal device 100 may transmit and receive a message with the counterpart user terminal device through the messagingservice providing server 200 or transmit a message to a server for providing social network service. -
FIGS. 3 to 5 are diagrams provided to explain a touch gesture for translating a message according to an embodiment. - As illustrated in
FIG. 3 , a user may execute a messaging application on theuser terminal device 100 and view a received message on theuser terminal device 100, and then write a message and transmit it to the counterpart user terminal device. - Generally, a messaging application may distinctively display each of the separately transmitted and received messages, as illustrated in
FIG. 3 . The messages being transmitted and received may be distinguished bymessage boxes - As illustrated in
FIG. 4a , the user may perform a preset touch gesture with respect to amessage 41 that is intended to be translated so that themessage 41 is translated. In an example, a target language of the translation may be set or modified by an option menu provided from a messaging application or theuser terminal device 100. - When the user performs a touch and drag with respect to the
message 41 in a direction from left to right, themessage 41 may be translated into a preset national language. In other words, a unit of translation may correspond to a message box unit of distinguishing the messages being transmitted and received. -
FIG. 4b illustrates a screen displaying a translated message for which the tough gesture is inputted. When a preset target language is English, theuser terminal device 100 may transmit themessage 41 “?” for which the gesture is inputted to the translationservice providing server 300, and receive a translatedmessage 43 “Where shall we meet?” from the translationservice providing server 300. In an example, theuser terminal device 100 may transmit target language setting information (English) together with the message for which the gesture is inputted. - The
user terminal device 100 may replace themessage 41 for which touch gesture is inputted into the translatedmessage 43 and display the translatedmessage 43 as illustrated inFIG. 4 b. - According to another embodiment, the
user terminal device 100 may additionally display the translated message “Where shall we meet?” together with the source message “?” for which touch gesture is inputted. For example, theuser terminal device 100 may divide a region of the message box for which touch gesture is inputted or additionally generate another message box under the message box, and display the translated message “Where shall we meet?” in the divided message box region or the additionally generated message box. - According to another embodiment, when the
user terminal device 100 additionally includes a speaker and the user performs a preset touch gesture with respect to the message for which the touch gesture is inputted or with respect to the translated message, theprocessor 140 may convert the translated message into voice and output the converted voice through the speaker. In this case, thecommunication unit 110 may transmit the translated message to a text to speech (TTS) server for converting text into voice, and receive the converted voice signal of the translated message from the TTS server. - Meanwhile, according to an embodiment, the translated message may be un-translated into the source message as illustrated in
FIG. 5 . Specifically, in response to sensing a preset touch gesture inputted for the translatedmessage 51, theprocessor 140 may control such that thesource message 52 before translation is displayed instead of the translatedmessage 51 for which the touch gesture is inputted. - In an example, the touch gesture to un-translate the translated
message 51 back into thesource message 52 may be a gesture in an opposite direction to the touch gesture performed to translate thesource message 52. For example, the touch gesture for translating the received source message before translation may be a touch and drag in a first direction, and the touch gesture for un-translating the translated message back into the source message before translation may be a touch and drag in a second direction opposite the first direction. -
FIG. 6 is a diagram provided to explain a touch gesture for translating a message according to another embodiment. - According to another embodiment, when a preset gesture, e.g., a touch and drag in a direction from left to right in
FIG. 6a , is inputted for thesource message 41 to be translated, theuser terminal device 100 may display an instruction statement to inform about an operation to be performed according to touch and drag (e.g., “View translatedtext 61”) upon inputting of the touch and drag. In an example, when touch and drag in a direction from left to right is not complete, translation of the corresponding message may not be performed. When the touch and drag is complete, the instruction statement disappears and a translated text for the message may be displayed. - Likewise, as illustrated in
FIG. 6b , an instruction statement such as “Viewsource text 62” may be displayed while a touch and drag is being inputted in a direction from right to left with respect to the translated message. When touch and drag is complete, the instruction statement may disappear and the source message of the translated message may be displayed. -
FIG. 7 is a diagram provided to explain a method for displaying pronunciation information of a message according to an embodiment. - Referring to
FIG. 7 , theprocessor 140 may control such that pronunciation information of any of the messages displayed on thedisplay 120 is displayed. Specifically, when the user performs a preset touch gesture for a message having pronunciation information, theprocessor 140 may display the pronunciation information such as phonetic alphabets of corresponding message. In an example, the pronunciation information may be additional information displayed together with corresponding message. - For example, as illustrated in
FIG. 7a , when the user performs pinch-out gesture with respect to a message in Chinese language, theprocessor 140 may display phonetic alphabets of such Chinese language message. In an example, as illustrated inFIG. 7b , a region of amessage box 71 for which gesture is inputted may be divided in response to the pinch-out gesture, and aChinese language message 72 along with the phonetic alphabets 73 of the Chinese language message may be displayed in each of the divided regions. - In an example, Chinese pronunciation information may be information stored in the
user terminal device 100. Alternatively, theprocessor 140 may transmit a message for which gesture is inputted to the translationservice providing server 300, and receive and display pronunciation information of the message from the translationservice providing server 300. - In an example, the
user terminal device 100 may further include the speaker such that, when preset touch gesture is inputted for the pronunciation information, theprocessor 140 may convert the pronunciation information for which touch gesture is inputted into voice and output the converted voice through the speaker. For example, when a user double-touches the region displaying the pronunciation information 73 of the Chinese language message, the speaker may output voice according to pronunciation information. -
FIGS. 8 and 9 are diagrams provided to explain a motion gesture for translating all of displayed messages according to an embodiment. - Referring to
FIG. 8 , a user may have a conversation with a single user or a plurality of users using a mobile messenger. In an example, the mobile messenger includes commercial messenger applications such as Kakao Talk, Line, WhatsApp, or the like. -
FIG. 8 illustrates an embodiment in which a user transmits and receives messages at real time among a plurality of users including “Mike” who uses English language and “Michiko” who uses Japanese language. In an example, when each of the users transmit a message in different languages from each other, real-time translation with respect to the messages may be performed in response to a preset gesture performed for each of the messages. In an example, when a target language of the translation is set to be Korean, both theEnglish language message 81 and theJapanese language message 82 may be translated into Korean language messages. - Accordingly, without having to perform a separate application for translation such as translation application, the users may actively communicate without having a pause in their conversation because screen with messages in translated languages is available to be viewed on a corresponding messaging application.
- Meanwhile, the user may perform a preset motion gesture on the
user terminal device 100 to view each of the translated messages of all the messages displayed on thedisplay 120. For example, as illustrated inFIG. 9 , when the user performs a motion of grabbing and shaking theuser terminal device 100, theprocessor 140 may translate the messages inforeign language display 120 in preset languages, and replace themessages target messages target messages processor 140 may transmit all the displayed foreign-language messages service providing server 300, and receive and display corresponding translatedmessages - As a result, the user may translate all of displayed messages into preset national languages by simply performing one motion gesture without having to individually perform a touch gesture with respect to each of the messages.
-
FIG. 10 is a diagram provided to explain a touch gesture for translating a message posted on the social network service according to an embodiment. - Referring to
FIG. 10 , real-time translation may be performed with respect to a message such as article posted on a mobile page or comments thereof provided through the social network service (SNS). In an example, SNS includes online services such as Facebook or Twitter for building up a relation network of internet users on online. The SNS may be executed on theuser terminal device 100 using an SNS providing application as a platform. -
FIG. 10a illustrates screen in which theuser terminal device 100 connects to SNS. As illustrated inFIG. 10a , SNS may be configured with basic platform including a postedmessage 1010 and comment type messages 1020-1040 thereof. In an example, the postedmessage 1010 and comment messages 1020-1040 may be distinctively displayed on a message unit basis, in which case each of the messages may be translated on the message unit basis. - For example, as illustrated in
FIG. 10b , the user may perform a touch and drag gesture with respect to one (e.g., message 1040) of the comment messages 1020-1040, and themessage 1040 for which touch and drag gesture is inputted may be replaced by a translatedmessage 1050 in a preset target language and displayed. -
FIG. 11 is a block diagram illustrating a constitution of a user terminal device in detail according to another embodiment. As illustrated inFIG. 11 , theuser terminal device 100′ according to another embodiment includes acommunication unit 110, adisplay 120, asensing unit 130, aprocessor 140, astorage 150, animage processor 160, anaudio processor 170, anaudio outputter 180 and a user interface 190. In the following description, elements or operations overlapping with those described above with reference toFIG. 1 will not be redundantly described for the sake of brevity. - The
processor 140 includesRAM 141,ROM 142, a graphic processor 143 (CPU 144, first to nth interfaces 145-1 to 145-n) and abus 146. In an example,RAM 141,ROM 142, and the graphic processor 143 (CPU 144, first to nth interfaces 145-1 to 145-n) may be connected to each other via thebus 146. - The first to nth interfaces 145-1 to 145-n are connected to the elements described above. One of the interfaces may be a network interface connected to an external device through network.
- The
CPU 144 may access thestorage 140 and perform booting by using O/S stored in thestorage 140. Further, theCPU 144 may perform various operations by using various programs, contents and data stored in thestorage 140. - The
RAM 141 stores instruction sets for system booting. Upon powering-on in response to input of turn-on command, theCPU 144copies RAM 141 stored in thestorage 150 onto theRAM 141 according to the instructions stored in theROM 142, and executes the O/S to boot the system. When the booting is completed, theCPU 144 copies various application programs stored in thestorage 150 onto theRAM 141, and executes the application program copied to theRAM 141 to perform various operations. - The
graphic processor 143 may generate a screen including various objects such as icons, images, texts or the like, using an arithmetic unit (not illustrated) or a renderer (not illustrated). The arithmetic unit calculates attribute values such as coordinate values, forms, sizes, colors or the like according to layouts of screens. The renderer may generate various layouts of screens including objects based on the attribute values calculated at the arithmetic unit. - Meanwhile, the operation of the
processor 140 described above may be performed by implementing the programs stored in thestorage 150. - The
storage 150 may store O/S (operating system) software module for driving theuser terminal device 100′ and various multimedia contents. - Specifically, the
storage 150 may store a base module for processing signals delivered from each hardware included in theuser terminal device 100′, a storage module for managing database (DB) or registry, a graphic processing module for generating layouts of screens, a security module or the like. Specifically, thestorage 150 may store programs such as communication module, translation module or the like, which are necessary for implementation of the translation service according to an embodiment. - The
processor 140 may perform communication with the counterpartuser terminal device 200, the messagingservice providing server 300, the translationservice providing server 300 or the like by using the communication module. - The
image processor 160 is configured to process various images such as decoding, scaling, noise filtering, frame rate converting, resolution converting or the like with respect to contents. - The
audio processor 170 is configured to process audio data. Specifically, theaudio processor 170 may process the pronunciation information for which touch gesture is inputted to convert it into voice data, and deliver the converted voice data to theaudio outputter 180. - The
audio outputter 180 is configured to output audio data processed in theaudio processor 170. Theaudio outputter 180 may output the converted voice data through a receiver or a speaker. - The user interface 190 is configured to sense user interaction for controlling overall operation of the
user terminal device 100′. -
FIG. 12 is a flowchart provided to explain a control method of a user terminal device according to an embodiment. - First, the user terminal device may control such that a message transmitted and received in a communication with an external device is displayed, at S1210. In an example, the messages being transmitted and received may be distinctively displayed on a message unit basis.
- At S1220, a gesture for the user terminal device may be sensed. In an example, the gesture inputted for a message may be a touch and drag in a first direction. Further, a touch gesture inputted for the translated message of the message may be a touch and drag in a second direction opposite the first direction.
- At S1230, when gesture is sensed, the user terminal device may provide the translation service for at least one of the displayed messages. In an example, translation service for the displayed messages may be provided on a message unit basis. Further, in response to sensing a touch gesture inputted for at least one of the display messages, it may be controlled such that the message for which touch gesture may be inputted is translated and the corresponding translated message in a preset language may be displayed. Further, in response to sensing a touch gesture inputted for the translated message, it may be controlled such that the translated message for which touch gesture is inputted may be un-translated back into the source message before translation.
- In an example, it may be controlled such that the message for which touch gesture is inputted may be replaced by the translated message and displayed. Further, it may be controlled such that the message for which touch gesture is inputted may be displayed together with the translated message.
- Further, in response to sensing a touch gesture inputted for the translated message, pronunciation information with respect to the translated message for which touch gesture is inputted may be displayed.
- According to various embodiments of the present disclosure described above, messages transmitted and received among user terminal devices may be instantly translated. Accordingly, communication between users using different languages from each other may be performed more actively.
- The control method of the user terminal device according to the various embodiments described above may be implemented as a program and stored in various recording media. In other words, a computer program processible with various processors for implementing various control methods described above may be stored in recording media and used.
- For example, a non-transitory computer readable recording medium may be provided, storing therein a program for performing operations of displaying a message transmitted and received in a communication with an external device, sensing a gesture inputted for the user terminal device, and providing translation service for at least one of the displayed messages.
- The non-transitory computer readable medium is a medium capable of storing data semi-permanently and being readable by a device, rather than a medium such as register, cash, and memory that stores the data for a brief period of time. In particular, the various applications or programs described above may be stored and provided on a non-transitory computer readable medium such as CD, DVD, hard disk, Blu-ray disk, USB, memory card, ROM, and so on.
- Further, while the present disclosure has been described in detail above, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the disclosure, are given by way of illustration only, since various changes and modifications within the scope of the disclosure will become apparent to those skilled in the art from this detailed description.
Claims (15)
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR10-2015-0108233 | 2015-07-30 | ||
KR1020150108233A KR20170014589A (en) | 2015-07-30 | 2015-07-30 | User terminal apparatus for providing translation service and control method thereof |
PCT/KR2016/006585 WO2017018665A1 (en) | 2015-07-30 | 2016-06-22 | User terminal device for providing translation service, and method for controlling same |
Publications (1)
Publication Number | Publication Date |
---|---|
US20180150458A1 true US20180150458A1 (en) | 2018-05-31 |
Family
ID=57884609
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/572,400 Abandoned US20180150458A1 (en) | 2015-07-30 | 2016-06-22 | User terminal device for providing translation service, and method for controlling same |
Country Status (4)
Country | Link |
---|---|
US (1) | US20180150458A1 (en) |
KR (1) | KR20170014589A (en) |
CN (1) | CN107851096A (en) |
WO (1) | WO2017018665A1 (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3407550A1 (en) * | 2017-05-26 | 2018-11-28 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for sending messages and mobile terminal |
US20190205397A1 (en) * | 2017-01-17 | 2019-07-04 | Loveland Co., Ltd. | Multilingual communication system and multilingual communication provision method |
US20220036875A1 (en) * | 2018-11-27 | 2022-02-03 | Inventio Ag | Method and device for outputting an audible voice message in an elevator system |
US20220261148A1 (en) * | 2019-11-08 | 2022-08-18 | Vivo Mobile Communication Co., Ltd. | Message processing method and electronic device |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101979600B1 (en) * | 2017-07-27 | 2019-05-17 | 이병두 | Method and computer program for providing messaging service of expressing emotion on text messgage |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080288241A1 (en) * | 2005-11-14 | 2008-11-20 | Fumitaka Noda | Multi Language Exchange System |
US20100125410A1 (en) * | 2008-11-17 | 2010-05-20 | Mary Anne Hicks | Methods and Apparatuses for Providing Enhanced Navigation Services |
US20120092233A1 (en) * | 2008-11-18 | 2012-04-19 | Sharp Kabushiki Kaisha | Display control apparatus and display control method |
US20140081620A1 (en) * | 2012-09-18 | 2014-03-20 | Abbyy Software Ltd. | Swiping Action for Displaying a Translation of a Textual Image |
US20140125580A1 (en) * | 2012-11-02 | 2014-05-08 | Samsung Electronics Co., Ltd. | Method and device for providing information regarding an object |
US20140180670A1 (en) * | 2012-12-21 | 2014-06-26 | Maria Osipova | General Dictionary for All Languages |
US20140278441A1 (en) * | 2013-03-15 | 2014-09-18 | Qualcomm Incorporated | Systems and methods for switching processing modes using gestures |
US20140297256A1 (en) * | 2013-03-15 | 2014-10-02 | Translate Abroad, Inc. | Systems and methods for determining and displaying multi-line foreign language translations in real time on mobile devices |
US20140313143A1 (en) * | 2013-04-18 | 2014-10-23 | Hojae JUNG | Mobile terminal and control method thereof |
US20150019240A1 (en) * | 2013-07-12 | 2015-01-15 | Inventec (Pudong) Technology Corporation | System for translating target words by gesture and method thereof |
US9275046B2 (en) * | 2013-03-15 | 2016-03-01 | Translate Abroad, Inc. | Systems and methods for displaying foreign character sets and their translations in real time on resource-constrained mobile devices |
US20170169812A1 (en) * | 2015-12-15 | 2017-06-15 | Facebook, Inc. | Providing intelligent transcriptions of sound messages in a messaging application |
Family Cites Families (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101498028B1 (en) * | 2008-04-29 | 2015-03-03 | 엘지전자 주식회사 | Terminal and method for controlling the same |
CN102737238A (en) * | 2011-04-01 | 2012-10-17 | 洛阳磊石软件科技有限公司 | Gesture motion-based character recognition system and character recognition method, and application thereof |
KR20130071958A (en) * | 2011-12-21 | 2013-07-01 | 엔에이치엔(주) | System and method for providing interpretation or translation of user message by instant messaging application |
KR101375166B1 (en) * | 2012-05-14 | 2014-03-20 | 전남대학교산학협력단 | System and control method for character make-up |
CN104298491B (en) * | 2013-07-18 | 2019-10-08 | 腾讯科技(深圳)有限公司 | Message treatment method and device |
KR101421621B1 (en) * | 2013-07-30 | 2014-07-22 | (주)블루랩스 | Smartphone terminal with language translation and language translation system comprising the same |
-
2015
- 2015-07-30 KR KR1020150108233A patent/KR20170014589A/en not_active Application Discontinuation
-
2016
- 2016-06-22 US US15/572,400 patent/US20180150458A1/en not_active Abandoned
- 2016-06-22 CN CN201680043189.XA patent/CN107851096A/en active Pending
- 2016-06-22 WO PCT/KR2016/006585 patent/WO2017018665A1/en active Application Filing
Patent Citations (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20080288241A1 (en) * | 2005-11-14 | 2008-11-20 | Fumitaka Noda | Multi Language Exchange System |
US20100125410A1 (en) * | 2008-11-17 | 2010-05-20 | Mary Anne Hicks | Methods and Apparatuses for Providing Enhanced Navigation Services |
US20120092233A1 (en) * | 2008-11-18 | 2012-04-19 | Sharp Kabushiki Kaisha | Display control apparatus and display control method |
US20140081620A1 (en) * | 2012-09-18 | 2014-03-20 | Abbyy Software Ltd. | Swiping Action for Displaying a Translation of a Textual Image |
US9836128B2 (en) * | 2012-11-02 | 2017-12-05 | Samsung Electronics Co., Ltd. | Method and device for providing information regarding an object |
US20140125580A1 (en) * | 2012-11-02 | 2014-05-08 | Samsung Electronics Co., Ltd. | Method and device for providing information regarding an object |
US20140180670A1 (en) * | 2012-12-21 | 2014-06-26 | Maria Osipova | General Dictionary for All Languages |
US20140278441A1 (en) * | 2013-03-15 | 2014-09-18 | Qualcomm Incorporated | Systems and methods for switching processing modes using gestures |
US9275046B2 (en) * | 2013-03-15 | 2016-03-01 | Translate Abroad, Inc. | Systems and methods for displaying foreign character sets and their translations in real time on resource-constrained mobile devices |
US20140297256A1 (en) * | 2013-03-15 | 2014-10-02 | Translate Abroad, Inc. | Systems and methods for determining and displaying multi-line foreign language translations in real time on mobile devices |
US20140313143A1 (en) * | 2013-04-18 | 2014-10-23 | Hojae JUNG | Mobile terminal and control method thereof |
US9268463B2 (en) * | 2013-04-18 | 2016-02-23 | Lg Electronics Inc. | Mobile terminal and control method thereof |
US20150019240A1 (en) * | 2013-07-12 | 2015-01-15 | Inventec (Pudong) Technology Corporation | System for translating target words by gesture and method thereof |
US20170169812A1 (en) * | 2015-12-15 | 2017-06-15 | Facebook, Inc. | Providing intelligent transcriptions of sound messages in a messaging application |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190205397A1 (en) * | 2017-01-17 | 2019-07-04 | Loveland Co., Ltd. | Multilingual communication system and multilingual communication provision method |
US11030421B2 (en) * | 2017-01-17 | 2021-06-08 | Loveland Co., Ltd. | Multilingual communication system and multilingual communication provision method |
EP3407550A1 (en) * | 2017-05-26 | 2018-11-28 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Method for sending messages and mobile terminal |
US20220036875A1 (en) * | 2018-11-27 | 2022-02-03 | Inventio Ag | Method and device for outputting an audible voice message in an elevator system |
US20220261148A1 (en) * | 2019-11-08 | 2022-08-18 | Vivo Mobile Communication Co., Ltd. | Message processing method and electronic device |
US11861158B2 (en) * | 2019-11-08 | 2024-01-02 | Vivo Mobile Communication Co., Ltd. | Message processing method and electronic device |
Also Published As
Publication number | Publication date |
---|---|
WO2017018665A1 (en) | 2017-02-02 |
CN107851096A (en) | 2018-03-27 |
KR20170014589A (en) | 2017-02-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11113083B2 (en) | Notification interaction in a touchscreen user interface | |
US10976773B2 (en) | User terminal device and displaying method thereof | |
CN108139852B (en) | Integrating content in non-browser applications | |
US20180150458A1 (en) | User terminal device for providing translation service, and method for controlling same | |
KR102045585B1 (en) | Adaptive input language switching | |
US10158609B2 (en) | User terminal device, communication system and control method therefor | |
EP3207458B1 (en) | Input signal emulation | |
EP3873073A1 (en) | Method for connecting mobile terminal and external display and apparatus implementing the same | |
US10359901B2 (en) | Method and apparatus for providing intelligent service using inputted character in a user device | |
US20130111360A1 (en) | Accessed Location of User Interface | |
US10223061B2 (en) | Display redistribution between a primary display and a secondary display | |
US11829588B2 (en) | Method, apparatus, and system for generating resource value transfer request | |
JP6439266B2 (en) | Text input method and apparatus in electronic device with touch screen | |
US20150304251A1 (en) | Direct Manipulation of Object Size in User Interface | |
US8830165B1 (en) | User interface | |
CN105379236A (en) | User experience mode transitioning | |
US10547711B2 (en) | Using off-screen user interface data during remote sessions | |
CN113934349B (en) | Interaction method, interaction device, electronic equipment and storage medium | |
KR20150087665A (en) | Operating Method For Handwriting Data and Electronic Device supporting the same | |
US11822768B2 (en) | Electronic apparatus and method for controlling machine reading comprehension based guide user interface | |
KR102108412B1 (en) | Method for providing search service on chatting based on messaging service, and device therefor | |
CN108780400B (en) | Data processing method and electronic equipment | |
US20230117037A1 (en) | Comment sharing method, apparatus and electronic device | |
CN109683726B (en) | Character input method, character input device, electronic equipment and storage medium | |
US11163378B2 (en) | Electronic device and operating method therefor |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YOON, YOON-JIN;REEL/FRAME:044056/0444 Effective date: 20171107 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |