WO2022191425A1 - Dispositif électronique pour appliquer un effet visuel à un texte de dialogue et son procédé de commande - Google Patents

Dispositif électronique pour appliquer un effet visuel à un texte de dialogue et son procédé de commande Download PDF

Info

Publication number
WO2022191425A1
WO2022191425A1 PCT/KR2022/001280 KR2022001280W WO2022191425A1 WO 2022191425 A1 WO2022191425 A1 WO 2022191425A1 KR 2022001280 W KR2022001280 W KR 2022001280W WO 2022191425 A1 WO2022191425 A1 WO 2022191425A1
Authority
WO
WIPO (PCT)
Prior art keywords
electronic device
text
visual effect
target application
keyword
Prior art date
Application number
PCT/KR2022/001280
Other languages
English (en)
Korean (ko)
Inventor
김은선
오영학
주민지
유현우
정윤성
박민규
이봄이
이원제
이은경
문재희
박경대
Original Assignee
삼성전자주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 삼성전자주식회사 filed Critical 삼성전자주식회사
Publication of WO2022191425A1 publication Critical patent/WO2022191425A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/103Formatting, i.e. changing of presentation of documents
    • G06F40/109Font handling; Temporal or kinetic typography
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/22Indexing; Data structures therefor; Storage structures
    • G06F16/2228Indexing structures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04817Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • G06T11/001Texturing; Colouring; Generation of texture or colour
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/22Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the display of characters or indicia using display control signals derived from coded signals representing the characters or indicia, e.g. with a character-code memory
    • G09G5/24Generation of individual character patterns
    • G09G5/26Generation of individual character patterns for modifying the character dimensions, e.g. double width, double height

Definitions

  • the following disclosure relates to an electronic device for applying a visual effect to dialog text and a method for controlling the same.
  • An artificial intelligence system may perform an operation requested by a user, and may provide information or a service according to the requested operation.
  • the AI assistant 's ability to understand the user's utterances and depend on accuracy in performing tasks according to the utterance intention.
  • the AI assistant can interpret the user input to infer the user's intention and perform tasks according to the inferred intention.
  • the artificial intelligence assistant may provide a dialog in text, such as a greeting, an answer to a voice command, a proactive suggestion, and a question, on the display of the electronic device.
  • an electronic device that applies a visual effect according to the meaning of a text dialog and applies a visual effect according to a function of a related application or service may be provided.
  • an electronic device includes: a display module for outputting a dialog in text through a screen; a memory in which computer-executable instructions are stored; and a processor that accesses the memory and executes the instructions, wherein the instructions determine a characteristic of a keyword determined based on the text, and determine a visual effect in consideration of the characteristic, wherein the keyword and When a related target application exists, the visual effect is determined by further considering the characteristics of the target application, and the visual effect is applied when the text is output.
  • an operating method of an electronic device includes: when outputting a dialog in text through a screen, identifying a characteristic of a keyword determined based on the text; determining a visual effect in consideration of the characteristic, and determining the visual effect by further considering the characteristic of the target application when a target application related to the keyword exists; and applying the visual effect when outputting the text.
  • the recording medium on which a program for controlling the operation of the electronic device when the electronic device outputs a text dialog through a screen, the keywords determined based on the text characterization; determining a visual effect in consideration of the characteristic, and determining the visual effect by further considering the characteristic of the target application when a target application related to the keyword exists; and a program for applying the visual effect when outputting the text.
  • a visual effect in consideration of a keyword characteristic or a characteristic of a related application may be provided.
  • FIG. 1 is a block diagram of an electronic device in a network environment, according to various embodiments of the present disclosure
  • FIG. 2 is a block diagram illustrating an integrated intelligence system, according to an embodiment.
  • FIG. 3 is a diagram illustrating a user terminal displaying a screen for processing a voice input received through an intelligent app, according to an embodiment.
  • FIG. 4 is a diagram illustrating a form in which relation information between a concept and an action is stored in a database, according to an embodiment.
  • FIG. 5 is a block diagram illustrating an electronic device according to various embodiments.
  • FIG. 6 is a view for explaining a process in which a visual effect is applied when a dialog in text is output according to various embodiments of the present disclosure
  • FIG. 7 is a view for explaining a process of applying a visual effect in consideration of the characteristics of the application when there is an application related to a dialog according to various embodiments of the present disclosure
  • FIG. 8 is a diagram for explaining a process in which a visual effect is applied when there are a plurality of applications related to a dialog, according to various embodiments of the present disclosure
  • FIG. 9 is a diagram for describing a process in which a visual effect is applied when one application among a plurality of applications is recommended based on a distance, according to various embodiments of the present disclosure
  • FIG. 10 is a diagram for describing a process in which a visual effect is applied differently when a dialog in text is output before and after an operation is executed, according to various embodiments of the present disclosure
  • FIG. 11 is a view for explaining a process in which a visual effect is applied to a background screen when a dialog in text is output, according to various embodiments of the present disclosure
  • FIG. 12 is a diagram for explaining a process of applying different visual effects according to meanings even for the same keyword, according to various embodiments of the present disclosure
  • FIG. 13 is a view for explaining a process of changing a visual expression according to a screen size, according to various embodiments of the present disclosure
  • FIG. 14 is a diagram for explaining a process of executing an operation related to a keyword or changing an option, according to various embodiments of the present disclosure
  • 15 is a view for explaining various visualization expression methods according to various embodiments of the present disclosure.
  • 16 is a diagram for explaining a process of switching a screen for outputting a text of a dialog, according to various embodiments of the present disclosure
  • 17 is a diagram for describing mapping between a dialog and a visual effect, according to various embodiments of the present disclosure.
  • FIG. 18 is a flowchart illustrating a method of operating an electronic device according to various embodiments of the present disclosure
  • FIG. 1 is a block diagram of an electronic device in a network environment, according to various embodiments of the present disclosure
  • FIG. 1 is a block diagram of an electronic device 101 in a network environment 100, according to various embodiments.
  • the electronic device 101 communicates with the electronic device 102 through a first network 198 (eg, a short-range wireless communication network) or a second network 199 . It may communicate with at least one of the electronic device 104 and the server 108 through (eg, a long-distance wireless communication network). According to an embodiment, the electronic device 101 may communicate with the electronic device 104 through the server 108 .
  • the electronic device 101 includes a processor 120 , a memory 130 , an input module 150 , a sound output module 155 , a display module 160 , an audio module 170 , and a sensor module ( 176), interface 177, connection terminal 178, haptic module 179, camera module 180, power management module 188, battery 189, communication module 190, subscriber identification module 196 , or an antenna module 197 .
  • at least one of these components eg, the connection terminal 178
  • some of these components are integrated into one component (eg, display module 160 ). can be
  • the processor 120 for example, executes software (eg, a program 140) to execute at least one other component (eg, a hardware or software component) of the electronic device 101 connected to the processor 120. It can control and perform various data processing or operations. According to one embodiment, as at least part of data processing or operation, the processor 120 converts commands or data received from other components (eg, the sensor module 176 or the communication module 190 ) to the volatile memory 132 . may be stored in , process commands or data stored in the volatile memory 132 , and store the result data in the non-volatile memory 134 .
  • software eg, a program 140
  • the processor 120 converts commands or data received from other components (eg, the sensor module 176 or the communication module 190 ) to the volatile memory 132 .
  • the volatile memory 132 may be stored in , process commands or data stored in the volatile memory 132 , and store the result data in the non-volatile memory 134 .
  • the processor 120 is a main processor 121 (eg, a central processing unit or an application processor) or a secondary processor 123 (eg, a graphic processing unit, a neural network processing unit) a neural processing unit (NPU), an image signal processor, a sensor hub processor, or a communication processor).
  • a main processor 121 eg, a central processing unit or an application processor
  • a secondary processor 123 eg, a graphic processing unit, a neural network processing unit
  • NPU neural processing unit
  • an image signal processor e.g., a sensor hub processor, or a communication processor.
  • the secondary processor 123 may, for example, act on behalf of the main processor 121 while the main processor 121 is in an inactive (eg, sleep) state, or when the main processor 121 is active (eg, executing an application). ), together with the main processor 121, at least one of the components of the electronic device 101 (eg, the display module 160, the sensor module 176, or the communication module 190) It is possible to control at least some of the related functions or states.
  • the coprocessor 123 eg, an image signal processor or a communication processor
  • may be implemented as part of another functionally related component eg, the camera module 180 or the communication module 190. have.
  • the auxiliary processor 123 may include a hardware structure specialized for processing an artificial intelligence model.
  • Artificial intelligence models can be created through machine learning. Such learning may be performed, for example, in the electronic device 101 itself on which the artificial intelligence model is performed, or may be performed through a separate server (eg, the server 108).
  • the learning algorithm may include, for example, supervised learning, unsupervised learning, semi-supervised learning, or reinforcement learning, but in the above example not limited
  • the artificial intelligence model may include a plurality of artificial neural network layers.
  • Artificial neural networks include deep neural networks (DNNs), convolutional neural networks (CNNs), recurrent neural networks (RNNs), restricted boltzmann machines (RBMs), deep belief networks (DBNs), bidirectional recurrent deep neural networks (BRDNNs), It may be one of deep Q-networks or a combination of two or more of the above, but is not limited to the above example.
  • the artificial intelligence model may include, in addition to, or alternatively, a software structure in addition to the hardware structure.
  • the memory 130 may store various data used by at least one component (eg, the processor 120 or the sensor module 176 ) of the electronic device 101 .
  • the data may include, for example, input data or output data for software (eg, the program 140 ) and instructions related thereto.
  • the memory 130 may include a volatile memory 132 or a non-volatile memory 134 .
  • the program 140 may be stored as software in the memory 130 , and may include, for example, an operating system 142 , middleware 144 , or an application 146 .
  • the input module 150 may receive a command or data to be used by a component (eg, the processor 120 ) of the electronic device 101 from the outside (eg, a user) of the electronic device 101 .
  • the input module 150 may include, for example, a microphone, a mouse, a keyboard, a key (eg, a button), or a digital pen (eg, a stylus pen).
  • the sound output module 155 may output a sound signal to the outside of the electronic device 101 .
  • the sound output module 155 may include, for example, a speaker or a receiver.
  • the speaker can be used for general purposes such as multimedia playback or recording playback.
  • the receiver can be used to receive incoming calls. According to one embodiment, the receiver may be implemented separately from or as part of the speaker.
  • the display module 160 may visually provide information to the outside (eg, a user) of the electronic device 101 .
  • the display module 160 may include, for example, a control circuit for controlling a display, a hologram device, or a projector and a corresponding device.
  • the display module 160 may include a touch sensor configured to sense a touch or a pressure sensor configured to measure the intensity of a force generated by the touch.
  • the audio module 170 may convert a sound into an electric signal or, conversely, convert an electric signal into a sound. According to an embodiment, the audio module 170 acquires a sound through the input module 150 or an external electronic device (eg, a sound output module 155 ) directly or wirelessly connected to the electronic device 101 .
  • the electronic device 102) eg, a speaker or headphones
  • the sensor module 176 detects an operating state (eg, power or temperature) of the electronic device 101 or an external environmental state (eg, a user state), and generates an electrical signal or data value corresponding to the sensed state. can do.
  • the sensor module 176 may include, for example, a gesture sensor, a gyro sensor, a barometric pressure sensor, a magnetic sensor, an acceleration sensor, a grip sensor, a proximity sensor, a color sensor, an IR (infrared) sensor, a biometric sensor, It may include a temperature sensor, a humidity sensor, a Hall sensor, or an illuminance sensor.
  • the interface 177 may support one or more specified protocols that may be used by the electronic device 101 to directly or wirelessly connect with an external electronic device (eg, the electronic device 102 ).
  • the interface 177 may include, for example, a high definition multimedia interface (HDMI), a universal serial bus (USB) interface, an SD card interface, or an audio interface.
  • the connection terminal 178 may include a connector through which the electronic device 101 can be physically connected to an external electronic device (eg, the electronic device 102 ).
  • the connection terminal 178 may include, for example, an HDMI connector, a USB connector, an SD card connector, or an audio connector (eg, a headphone connector).
  • the haptic module 179 may convert an electrical signal into a mechanical stimulus (eg, vibration or movement) or an electrical stimulus that the user can perceive through tactile or kinesthetic sense.
  • the haptic module 179 may include, for example, a motor, a piezoelectric element, or an electrical stimulation device.
  • the camera module 180 may capture still images and moving images. According to an embodiment, the camera module 180 may include one or more lenses, image sensors, image signal processors, or flashes.
  • the power management module 188 may manage power supplied to the electronic device 101 .
  • the power management module 188 may be implemented as, for example, at least a part of a power management integrated circuit (PMIC).
  • PMIC power management integrated circuit
  • the battery 189 may supply power to at least one component of the electronic device 101 .
  • battery 189 may include, for example, a non-rechargeable primary cell, a rechargeable secondary cell, or a fuel cell.
  • the communication module 190 is a direct (eg, wired) communication channel or a wireless communication channel between the electronic device 101 and an external electronic device (eg, the electronic device 102, the electronic device 104, or the server 108). It can support establishment and communication performance through the established communication channel.
  • the communication module 190 may include one or more communication processors that operate independently of the processor 120 (eg, an application processor) and support direct (eg, wired) communication or wireless communication.
  • the communication module 190 is a wireless communication module 192 (eg, a cellular communication module, a short-range wireless communication module, or a global navigation satellite system (GNSS) communication module) or a wired communication module 194 (eg, : It may include a local area network (LAN) communication module, or a power line communication module).
  • a wireless communication module 192 eg, a cellular communication module, a short-range wireless communication module, or a global navigation satellite system (GNSS) communication module
  • GNSS global navigation satellite system
  • wired communication module 194 eg, : It may include a local area network (LAN) communication module, or a power line communication module.
  • a corresponding communication module among these communication modules is a first network 198 (eg, a short-range communication network such as Bluetooth, wireless fidelity (WiFi) direct, or infrared data association (IrDA)) or a second network 199 (eg, legacy It may communicate with the external electronic device 104 through a cellular network, a 5G network, a next-generation communication network, the Internet, or a computer network (eg, a telecommunication network such as a LAN or a WAN).
  • a first network 198 eg, a short-range communication network such as Bluetooth, wireless fidelity (WiFi) direct, or infrared data association (IrDA)
  • a second network 199 eg, legacy It may communicate with the external electronic device 104 through a cellular network, a 5G network, a next-generation communication network, the Internet, or a computer network (eg, a telecommunication network such as a LAN or a WAN).
  • a telecommunication network
  • the wireless communication module 192 uses subscriber information (eg, International Mobile Subscriber Identifier (IMSI)) stored in the subscriber identification module 196 within a communication network such as the first network 198 or the second network 199 .
  • subscriber information eg, International Mobile Subscriber Identifier (IMSI)
  • IMSI International Mobile Subscriber Identifier
  • the electronic device 101 may be identified or authenticated.
  • the wireless communication module 192 may support a 5G network after a 4G network and a next-generation communication technology, for example, a new radio access technology (NR).
  • NR access technology includes high-speed transmission of high-capacity data (eMBB (enhanced mobile broadband)), minimization of terminal power and access to multiple terminals (mMTC (massive machine type communications)), or high reliability and low latency (URLLC (ultra-reliable and low-latency) -latency communications)).
  • eMBB enhanced mobile broadband
  • mMTC massive machine type communications
  • URLLC ultra-reliable and low-latency
  • the wireless communication module 192 may support a high frequency band (eg, mmWave band) to achieve a high data rate, for example.
  • a high frequency band eg, mmWave band
  • the wireless communication module 192 uses various techniques for securing performance in a high-frequency band, for example, beamforming, massive multiple-input and multiple-output (MIMO), all-dimensional multiplexing. It may support technologies such as full dimensional MIMO (FD-MIMO), an array antenna, analog beam-forming, or a large scale antenna.
  • the wireless communication module 192 may support various requirements defined in the electronic device 101 , an external electronic device (eg, the electronic device 104 ), or a network system (eg, the second network 199 ).
  • the wireless communication module 192 may include a peak data rate (eg, 20 Gbps or more) for realizing eMBB, loss coverage (eg, 164 dB or less) for realizing mMTC, or U-plane latency for realizing URLLC ( Example: Downlink (DL) and uplink (UL) each 0.5 ms or less, or round trip 1 ms or less) can be supported.
  • a peak data rate eg, 20 Gbps or more
  • loss coverage eg, 164 dB or less
  • U-plane latency for realizing URLLC
  • the antenna module 197 may transmit or receive a signal or power to the outside (eg, an external electronic device).
  • the antenna module 197 may include an antenna including a conductor formed on a substrate (eg, a PCB) or a radiator formed of a conductive pattern.
  • the antenna module 197 may include a plurality of antennas (eg, an array antenna). In this case, at least one antenna suitable for a communication method used in a communication network such as the first network 198 or the second network 199 is connected from the plurality of antennas by, for example, the communication module 190 . can be selected. A signal or power may be transmitted or received between the communication module 190 and an external electronic device through the selected at least one antenna.
  • other components eg, a radio frequency integrated circuit (RFIC)
  • RFIC radio frequency integrated circuit
  • the antenna module 197 may form a mmWave antenna module.
  • the mmWave antenna module comprises a printed circuit board, an RFIC disposed on or adjacent to a first side (eg, bottom side) of the printed circuit board and capable of supporting a designated high frequency band (eg, mmWave band); and a plurality of antennas (eg, an array antenna) disposed on or adjacent to a second side (eg, top or side) of the printed circuit board and capable of transmitting or receiving signals of the designated high frequency band. can do.
  • peripheral devices eg, a bus, general purpose input and output (GPIO), serial peripheral interface (SPI), or mobile industry processor interface (MIPI)
  • GPIO general purpose input and output
  • SPI serial peripheral interface
  • MIPI mobile industry processor interface
  • the command or data may be transmitted or received between the electronic device 101 and the external electronic device 104 through the server 108 connected to the second network 199 .
  • Each of the external electronic devices 102 or 104 may be the same as or different from the electronic device 101 .
  • all or part of the operations performed by the electronic device 101 may be executed by one or more external electronic devices 102 , 104 , or 108 .
  • the electronic device 101 may perform the function or service itself instead of executing the function or service itself.
  • one or more external electronic devices may be requested to perform at least a part of the function or the service.
  • One or more external electronic devices that have received the request may execute at least a part of the requested function or service, or an additional function or service related to the request, and transmit a result of the execution to the electronic device 101 .
  • the electronic device 101 may process the result as it is or additionally and provide it as at least a part of a response to the request.
  • cloud computing, distributed computing, mobile edge computing (MEC), or client-server computing technology may be used.
  • the electronic device 101 may provide an ultra-low latency service using, for example, distributed computing or mobile edge computing.
  • the external electronic device 104 may include an Internet of things (IoT) device.
  • the server 108 may be an intelligent server using machine learning and/or neural networks.
  • the external electronic device 104 or the server 108 may be included in the second network 199 .
  • the electronic device 101 may be applied to an intelligent service (eg, smart home, smart city, smart car, or health care) based on 5G communication technology and IoT-related technology.
  • FIG. 2 is a block diagram illustrating an integrated intelligence system according to an embodiment.
  • the integrated intelligent system 20 may include an electronic device 101 , an intelligent server 200 , and a service server 300 .
  • the electronic device 101 may be a terminal device (or electronic device) connectable to the Internet, for example, a mobile phone, a smart phone, a personal digital assistant (PDA), a notebook computer, a TV, a white home appliance, It may be a wearable device, an HMD, or a smart speaker.
  • a terminal device or electronic device connectable to the Internet
  • PDA personal digital assistant
  • TV TV
  • white home appliance It may be a wearable device, an HMD, or a smart speaker.
  • the electronic device 101 includes an interface 177 , a microphone 150 - 1 , a speaker 155 - 1 , a display module 160 , a memory 130 , or a processor 120 . can do.
  • the components listed above may be operatively or electrically connected to each other.
  • the microphone 150 - 1 may be included in an input module (eg, the input module 150 of FIG. 1 ).
  • the speaker 155 - 1 may be included in a sound output module (eg, the sound output module 155 of FIG. 1 ).
  • the interface 177 may be configured to transmit/receive data by being connected to an external device.
  • the microphone 150 - 1 may receive a sound (eg, a user's utterance) and convert it into an electrical signal.
  • the speaker 155 - 1 may output an electrical signal as a sound (eg, voice).
  • the display module 160 may be configured to display an image or a video.
  • the display module 160 may also display a graphic user interface (GUI) of an executed app (or an application program).
  • GUI graphic user interface
  • the memory 130 may store the client module 151 , a software development kit (SDK) 153 , and a plurality of apps 146 .
  • the client module 151 and the SDK 153 may constitute a framework (or solution program) for performing general functions.
  • the client module 151 or the SDK 153 may configure a framework for processing a voice input.
  • the plurality of apps 146 may be a program for performing a specified function.
  • the plurality of apps 155 may include a first app 146 - 1 and a second app 146 - 2 .
  • each of the plurality of apps 155 may include a plurality of operations for performing a specified function.
  • the apps may include an alarm app, a message app, and/or a schedule app.
  • the plurality of apps 155 may be executed by the processor 120 to sequentially execute at least some of the plurality of operations.
  • the processor 120 may control the overall operation of the electronic device 101 .
  • the processor 120 may be electrically connected to the interface 177 , the microphone 150 - 1 , the speaker 155 - 1 , and the display module 160 to perform a specified operation.
  • the processor 120 may also execute a program stored in the memory 130 to perform a designated function.
  • the processor 120 may execute at least one of the client module 151 and the SDK 153 to perform the following operation for processing a voice input.
  • the processor 120 may control the operation of the plurality of apps 146 through, for example, the SDK 153 .
  • the following operations described as operations of the client module 151 or the SDK 153 may be operations by the execution of the processor 120 .
  • the client module 151 may receive a voice input.
  • the client module 151 may receive a voice signal corresponding to the user's utterance sensed through the microphone 150 - 1 .
  • the client module 151 may transmit the received voice input to the intelligent server 200 .
  • the client module 151 may transmit status information of the electronic device 101 to the intelligent server 200 together with the received voice input.
  • the state information may be, for example, execution state information of an app.
  • the client module 151 may receive a result corresponding to the received voice input. For example, when the intelligent server 200 can calculate a result corresponding to the received voice input, the client module 151 may receive a result corresponding to the received voice input. The client module 151 may display the received result on the display module 160 .
  • the client module 151 may receive a plan corresponding to the received voice input.
  • the client module 151 may display a result of executing a plurality of operations of the app according to the plan on the display module 160 .
  • the client module 151 may, for example, sequentially display execution results of a plurality of operations on the display.
  • the electronic device 101 may display only partial results of executing a plurality of operations (eg, results of the last operation) on the display.
  • the client module 151 may receive a request for obtaining information necessary for calculating a result corresponding to a voice input from the intelligent server 200 . According to an embodiment, the client module 151 may transmit the necessary information to the intelligent server 200 in response to the request.
  • the client module 151 may transmit result information of executing a plurality of operations according to a plan to the intelligent server 200 .
  • the intelligent server 200 may confirm that the received voice input has been correctly processed using the result information.
  • the client module 151 may include a voice recognition module. According to an embodiment, the client module 151 may recognize a voice input performing a limited function through the voice recognition module. For example, the client module 151 may execute an intelligent app for processing a voice input for performing an organic operation through a specified input (eg, wake up!).
  • a voice recognition module may recognize a voice input performing a limited function through the voice recognition module.
  • the client module 151 may execute an intelligent app for processing a voice input for performing an organic operation through a specified input (eg, wake up!).
  • the intelligent server 200 may receive information related to a user's voice input from the electronic device 101 through a communication network. According to an embodiment, the intelligent server 200 may change data related to the received voice input into text data. According to an embodiment, the intelligent server 200 may generate a plan for performing a task corresponding to the user's voice input based on the text data.
  • the plan may be generated by an artificial intelligent (AI) system.
  • the artificial intelligence system may be a rule-based system, a neural network-based system (eg, a feedforward neural network (FNN)), a recurrent neural network (RNN) ))) can also be Alternatively, it may be a combination of the above or other artificial intelligence systems.
  • the plan may be selected from a set of predefined plans or may be generated in real time in response to a user request. For example, the artificial intelligence system may select at least one plan from among a plurality of predefined plans.
  • the intelligent server 200 may transmit a result according to the generated plan to the electronic device 101 or transmit the generated plan to the electronic device 101 .
  • the electronic device 101 may display a result according to the plan on the display.
  • the electronic device 101 may display the result of executing the operation according to the plan on the display.
  • the intelligent server 200 of an embodiment includes a front end 210, a natural language platform 220, a capsule DB 230, an execution engine 240, It may include an end user interface 250 , a management platform 260 , a big data platform 270 , or an analytics platform 280 .
  • the front end 210 may receive a voice input received from the electronic device 101 .
  • the front end 210 may transmit a response corresponding to the voice input.
  • the natural language platform 220 includes an automatic speech recognition module (ASR module) 221 , a natural language understanding module (NLU module) 223 , a planner module ( planner module 225 , a natural language generator module (NLG module) 227 , or a text to speech module (TTS module) 229 .
  • ASR module automatic speech recognition module
  • NLU module natural language understanding module
  • planner module planner module 225
  • NLG module natural language generator module
  • TTS module text to speech module
  • the automatic voice recognition module 221 may convert a voice input received from the electronic device 101 into text data.
  • the natural language understanding module 223 may recognize the user's intention by using text data of the voice input. For example, the natural language understanding module 223 may determine the user's intention by performing syntactic analysis or semantic analysis.
  • the natural language understanding module 223 recognizes the meaning of a word extracted from a voice input using a linguistic feature (eg, a grammatical element) of a morpheme or phrase, and matches the meaning of the identified word with the intention of the user. You can decide your intentions.
  • the planner module 225 may generate a plan using the intent and parameters determined by the natural language understanding module 223 .
  • the planner module 225 may determine a plurality of domains required to perform a task based on the determined intention.
  • the planner module 225 may determine a plurality of operations included in each of the plurality of domains determined based on the intention.
  • the planner module 225 may determine a parameter required to execute the determined plurality of operations or a result value output by the execution of the plurality of operations.
  • the parameter and the result value may be defined as a concept of a specified format (or class).
  • the plan may include a plurality of actions and a plurality of concepts determined by the user's intention.
  • the planner module 225 may determine the relationship between the plurality of operations and the plurality of concepts in stages (or hierarchically). For example, the planner module 225 may determine the execution order of the plurality of operations determined based on the user's intention based on the plurality of concepts. In other words, the planner module 225 may determine the execution order of the plurality of operations based on parameters required for execution of the plurality of operations and results output by the execution of the plurality of operations. Accordingly, the planner module 225 may generate a plan including a plurality of operations and related information (eg, an ontology) between a plurality of concepts. The planner module 225 may generate a plan using information stored in the capsule database 230 in which a set of relationships between concepts and operations is stored.
  • the planner module 225 may generate a plan using information stored in the capsule database 230 in which a set of relationships between concepts and operations is stored.
  • the natural language generation module 227 may change the specified information into a text form.
  • the information changed to the text form may be in the form of natural language utterance.
  • the text-to-speech conversion module 229 may change information in a text format into information in a voice format.
  • some or all of the functions of the natural language platform 220 may also be implemented in the electronic device 101 .
  • the capsule database 230 may store information on relationships between a plurality of concepts and operations corresponding to a plurality of domains.
  • a capsule may include a plurality of action objects (action objects or action information) and concept objects (concept objects or concept information) included in the plan.
  • the capsule database 230 may store a plurality of capsules in the form of a concept action network (CAN).
  • the plurality of capsules may be stored in a function registry included in the capsule database 230 .
  • the capsule database 230 may include a strategy registry in which strategy information necessary for determining a plan corresponding to a voice input is stored.
  • the strategy information may include reference information for determining one plan when there are a plurality of plans corresponding to the voice input.
  • the capsule database 230 may include a follow up registry in which information on a subsequent operation for suggesting a subsequent operation to the user in a specified situation is stored.
  • the subsequent operation may include, for example, a subsequent utterance.
  • the capsule database 230 may include a layout registry that stores layout information of information output through the electronic device 101 .
  • the capsule database 230 may include a vocabulary registry in which vocabulary information included in the capsule information is stored.
  • the capsule database 230 may include a dialog registry (dialog registry) in which dialog (or interaction) information with the user is stored.
  • the capsule database 230 may update a stored object through a developer tool.
  • the developer tool may include, for example, a function editor for updating an action object or a concept object.
  • the developer tool may include a vocabulary editor for updating the vocabulary.
  • the developer tool may include a strategy editor for creating and registering strategies for determining plans.
  • the developer tool may include a dialog editor that creates a conversation with the user.
  • the developer tool can include a follow up editor that can edit subsequent utterances that activate follow-up goals and provide hints. The subsequent goal may be determined based on a currently set goal, a user's preference, or an environmental condition.
  • the capsule database 230 may be implemented in the electronic device 101 as well.
  • the execution engine 240 may calculate a result using the generated plan.
  • the end user interface 250 may transmit the calculated result to the electronic device 101 . Accordingly, the electronic device 101 may receive the result and provide the received result to the user.
  • the management platform 260 may manage information used in the intelligent server 200 .
  • the big data platform 270 according to an embodiment may collect user data.
  • the analysis platform 280 of an embodiment may manage the quality of service (QoS) of the intelligent server 200 . For example, the analysis platform 280 may manage the components and processing speed (or efficiency) of the intelligent server 200 .
  • QoS quality of service
  • the service server 300 may provide a specified service (eg, food order or hotel reservation) to the electronic device 101 .
  • the service server 300 may be a server operated by a third party.
  • the service server 300 of an embodiment may provide information for generating a plan corresponding to the received voice input to the intelligent server 200 .
  • the provided information may be stored in the capsule database 230 .
  • the service server 300 may provide result information according to the plan to the intelligent server 200 .
  • the electronic device 101 may provide various intelligent services to the user in response to a user input.
  • the user input may include, for example, an input through a physical button, a touch input, or a voice input.
  • the electronic device 101 may provide a voice recognition service through an intelligent app (or a voice recognition app) stored therein.
  • the electronic device 101 may recognize a user utterance or a voice input received through the microphone, and provide a service corresponding to the recognized voice input to the user. .
  • the electronic device 101 may perform a specified operation alone or together with the intelligent server and/or the service server, based on the received voice input. For example, the electronic device 101 may execute an app corresponding to the received voice input and perform a specified operation through the executed app.
  • the user terminal when the electronic device 101 provides a service together with the intelligent server 200 and/or the service server, the user terminal detects a user's utterance using the microphone 150-1, , a signal (or voice data) corresponding to the sensed user's utterance may be generated. The user terminal may transmit the voice data to the intelligent server 200 using the interface 177 .
  • the intelligent server 200 In response to the voice input received from the electronic device 101 , the intelligent server 200 according to an embodiment is a plan for performing a task corresponding to the voice input, or performs an operation according to the plan. results can be generated.
  • the plan may include, for example, a plurality of actions for performing a task corresponding to a user's voice input, and a plurality of concepts related to the plurality of actions.
  • the concept may define parameters input to the execution of the plurality of operations or result values output by the execution of the plurality of operations.
  • the plan may include a plurality of actions and association information between a plurality of concepts.
  • the electronic device 101 may receive the response using the interface 177 .
  • the electronic device 101 outputs a voice signal generated inside the electronic device 101 to the outside using the speaker 155 - 1 , or generates a voice signal generated inside the electronic device 101 using the display module 160 . Images can be output externally.
  • FIG. 3 is a diagram illustrating a screen on which an electronic device processes a voice input received through an intelligent app according to various embodiments of the present disclosure
  • the electronic device 101 may execute an intelligent app to process a user input through an intelligent server (eg, the intelligent server 200 of FIG. 2 ).
  • an intelligent server eg, the intelligent server 200 of FIG. 2 .
  • the electronic device 101 processes the voice input.
  • You can run intelligent apps for The electronic device 101 may, for example, run the intelligent app in a state in which the schedule app is running.
  • the electronic device 101 may display an object (eg, an icon) 311 corresponding to an intelligent app on a display (eg, the display module 160 of FIG. 1 ).
  • the electronic device 101 may receive a voice input by a user's utterance. For example, the electronic device 101 may receive a voice input saying “Tell me about this week’s schedule!”.
  • the electronic device 101 may display a user interface (UI) 313 (eg, an input window) of an intelligent app in which text data of the received voice input is displayed on the display.
  • UI user interface
  • the electronic device 101 may display a result corresponding to the received voice input on the display.
  • the electronic device 101 may receive a plan corresponding to the received user input, and display 'this week's schedule' on the display according to the plan.
  • FIG. 4 is a diagram illustrating a form in which relation information between a concept and an action is stored in a database, according to an embodiment.
  • the capsule database (eg, the capsule database 230 of FIG. 2 ) of the intelligent server may store the capsule in the form of a concept action network (CAN).
  • the capsule database may store an operation for processing a task corresponding to a user's voice input and parameters necessary for the operation in the form of a concept action network (CAN).
  • the capsule database may store a plurality of capsules (capsule(A) 401, capsule(B) 404) corresponding to each of a plurality of domains (eg, applications).
  • one capsule eg, capsule(A) 401
  • one domain eg, location (geo), application
  • at least one service provider eg, CP 1 402 or CP 2 403
  • one capsule may include at least one operation 410 and at least one concept 420 for performing a specified function.
  • the natural language platform may generate a plan for performing a task corresponding to the received voice input using the capsule stored in the capsule database.
  • the planner module of the natural language platform eg, the planner module 225 of FIG. 2
  • the electronic device may have various types of devices.
  • the electronic device may include, for example, a portable communication device (eg, a smart phone), a computer device, a portable multimedia device, a portable medical device, a camera, a wearable device, or a home appliance device.
  • a portable communication device eg, a smart phone
  • a computer device e.g., a smart phone
  • a portable multimedia device e.g., a portable medical device
  • a camera e.g., a portable medical device
  • a camera e.g., a portable medical device
  • a camera e.g., a portable medical device
  • a wearable device e.g., a smart bracelet
  • a home appliance device e.g., a home appliance
  • a or B at least one of A and B, “at least one of A or B”, “A, B or C”, “at least one of A, B and C”, and "A
  • Each of the phrases “at least one or two of , B, or C” may include any one of the items listed together in the corresponding one of the phrases, or all possible combinations thereof.
  • Terms such as “first”, “second”, or “first” or “second” may simply be used to distinguish an element from other elements in question, and may refer elements to other aspects (e.g., importance or order) is not limited.
  • one (eg, first) component is “coupled” or “connected” to another (eg, second) component, with or without the terms “functionally” or “communicatively”.
  • module used in various embodiments of this document may include a unit implemented in hardware, software, or firmware, and is interchangeable with terms such as, for example, logic, logic block, component, or circuit.
  • a module may be an integrally formed part or a minimum unit or a part of the part that performs one or more functions.
  • the module may be implemented in the form of an application-specific integrated circuit (ASIC).
  • ASIC application-specific integrated circuit
  • Various embodiments of the present document include one or more instructions stored in a storage medium (eg, internal memory 136 or external memory 138) readable by a machine (eg, electronic device 101).
  • a machine eg, electronic device 101
  • the processor eg, the processor 120
  • the device eg, the electronic device 101
  • the one or more instructions may include code generated by a compiler or code executable by an interpreter.
  • the device-readable storage medium may be provided in the form of a non-transitory storage medium.
  • 'non-transitory' only means that the storage medium is a tangible device and does not contain a signal (eg, electromagnetic wave), and this term refers to the case where data is semi-permanently stored in the storage medium and It does not distinguish between temporary storage cases.
  • a signal eg, electromagnetic wave
  • the method according to various embodiments disclosed in this document may be provided in a computer program product (computer program product).
  • Computer program products may be traded between sellers and buyers as commodities.
  • the computer program product is distributed in the form of a machine-readable storage medium (eg compact disc read only memory (CD-ROM)), or through an application store (eg Play StoreTM) or on two user devices ( It can be distributed (eg downloaded or uploaded) directly, online between smartphones (eg: smartphones).
  • a portion of the computer program product may be temporarily stored or temporarily created in a machine-readable storage medium such as a memory of a server of a manufacturer, a server of an application store, or a relay server.
  • each component eg, a module or a program of the above-described components may include a singular or a plurality of entities, and some of the plurality of entities may be separately disposed in other components. have.
  • one or more components or operations among the above-described corresponding components may be omitted, or one or more other components or operations may be added.
  • a plurality of components eg, a module or a program
  • the integrated component may perform one or more functions of each component of the plurality of components identically or similarly to those performed by the corresponding component among the plurality of components prior to the integration. .
  • operations performed by a module, program, or other component are executed sequentially, in parallel, repeatedly, or heuristically, or one or more of the operations are executed in a different order, or omitted. , or one or more other operations may be added.
  • FIG. 5 is a block diagram illustrating an electronic device according to various embodiments.
  • the electronic device 101 may include a display module 160 , a memory 130 in which computer-executable instructions are stored, and a processor 120 that accesses the memory and executes the instructions.
  • the commands stored in the memory 130 determine the characteristics of the keyword determined based on the text, and determine the visual effect in consideration of the characteristics, but the keyword-related target application exists.
  • the visual effect may be determined by further considering the characteristics of the target application, and the visual effect may be applied when the text is output.
  • the instructions may be configured to identify a keyword determined based on text and identify a target application related to the keyword.
  • the text of the dialog may be generated in a natural language platform (eg, the natural language platform 220 of FIG. 2 ) of the intelligent server (eg, the intelligent server 200 of FIG. 2 ).
  • the natural language platform 220 includes an automatic speech recognition module (eg, the automatic speech recognition module 221 of FIG. 2 ), a natural language understanding module (eg, the natural language understanding module 223 of FIG. 2 ), and a planner.
  • a module eg, the planner module 225 of FIG. 2
  • a natural language generation module eg, the natural language generation module 227 of FIG. 2
  • a text-to-speech module eg, the text-to-speech module 229 of FIG. 2
  • a dialog output as text through the electronic device 101 through the natural language understanding module, the planner module, and the natural language generating module, a keyword and a target application in the corresponding dialog may be determined.
  • the dialog may be a greeting from an AI assistant, an answer to a user's voice command, a proactive suggestion, or a question.
  • the planner module may determine a plurality of domains (eg, the application 146 of FIG. 1 ) using the intention and parameters determined by the natural language understanding module, and perform a plurality of operations included in each of the plurality of domains. can decide
  • a plurality of capsules eg, the capsule 401 of FIG. 4
  • a capsule database eg, the capsule database 230 of FIG. 2 ).
  • the instructions may be configured to determine a visual effect by identifying a keyword and/or a characteristic of a target application.
  • a characteristic of the keyword may be, for example, small talk, information, or suggestion.
  • the target application may be, for example, an application related to the setting of the electronic device 101 or a third party application.
  • the visual effect may include a text font, characteristics such as text color or size, icons and templates, text animation, and the like.
  • the visual effect in consideration of the characteristic of the target application includes a logo of the target application, a color of the logo, and the target application. At least one of icons associated with the .
  • the target application is a coffee-related application
  • a logo of a corresponding company a representative color of the logo, a coffee icon, etc. may be included in a visual effect in consideration of the characteristics of the target application.
  • the commands determine the logo of the target application, extract a representative color from the logo, and icon of a product related to a service provided by the target application can be configured to determine
  • the electronic device 101 may further include a database 510 , and the commands refer to the database 510 to provide a visual effect corresponding to a characteristic of a keyword and a visual effect corresponding to a characteristic of a target application.
  • the database 510 may include an icon and a template 512 , a text size or color-related feature 514 , a text font 516 , and a text animation 518 .
  • the icon and template 512 may include icons related to the characteristics of keywords, for example, a coffee cup in relation to coffee, dumbbells in case of exercise, and the like.
  • text features 514 may include various text sizes, colors, and the like.
  • the text font 516 may include various typefaces.
  • the text animation 518 may include an animation yPosition bounce, an animation yPosition wave, and the like in a vertical direction.
  • the display module 160 may include a touch sensor configured to sense a touch or a pressure sensor configured to measure the intensity of a force generated by the touch. According to an embodiment, the display module 160 may detect a touch through the screen. According to various embodiments, the size of the screen may vary according to the type of the electronic device 101 . According to various embodiments, the electronic device 101 may be a smartphone, a PC, a tablet, a notebook computer, a smart watch, a car equipped with an AVN (audio video navigation) function, etc., and the size of the screen is It may vary depending on the electronic device.
  • AVN audio video navigation
  • FIG. 6 is a view for explaining a process in which a visual effect is applied when a dialog in text is output according to various embodiments of the present disclosure
  • the text to be output on the screen of the electronic device may be determined as 'Shall I order a large coffee with a shot added today?' (673), and a logo color of the target application may be determined as a visual effect.
  • the screen of the electronic device may change in order of 610, 630, 650, and 670.
  • screens of the electronic device eg, 610, 630, 650, 670 in FIG. 6, 710, 730, 750 in FIG. 7, 810, 830, 850, 870 in FIG. 910, 940, 970 of FIG. 9, 1010, 1030, 1050, 1070 of FIG. 11, 1110, 1140, 1170, 1210, 1230, 1250, 1270, 1410, 1430, 1450, 1470 of FIG. , 1510 , 1530 , 1550 , 1570 of FIG. 15 , and 1610 , 1630 , 1650 of FIG. 16 ) are illustrated as screens of the smart watch, but are not limited thereto.
  • the electronic device may be a smartphone, a PC, a tablet, a notebook computer, a smart watch, or the like, and the size of the screen may vary depending on the electronic device.
  • a visual effect may be determined by extracting a feature from a keyword of the determined text 673 or extracting a feature from a target application related to the text 673 .
  • 'coffee' may be determined as a keyword in the text 673 of 'Should I order a large coffee with a shot added today?', and a coffee cup icon 635 may be output as a visual effect at 630 .
  • the color of the corresponding icon 635 may be a representative color of the logo of the target application.
  • the color of the text representing the time provided on the basic screen in 650 may be changed to the color of the logo of the target application ( 655 ).
  • the coffee cup 635 of 630 is changed to a shape that is poured into the text 08 and 30 indicating the current time, and the color of the text 08 and 30 indicating the current time is changed (655) to the color of the logo of the target application.
  • a dialog 673 in a text format may be provided at 670 .
  • the keyword 676 and the target application of the dialog may be predetermined by a natural language platform (eg, the natural language platform 220 of FIG. 2 ) as described with reference to FIG. 5 , and a visual effect in consideration of the keyword characteristics and the characteristics of the target application can be applied.
  • the logo color of the target application may be applied to the keyword 676 to appear.
  • the logo color of the target application may be stored in icons and templates (eg, icons and templates 512 in FIG. 5) of a database (eg, database 510 in FIG. 5), It may be extracted from the target application logo to determine the visual effect.
  • FIG. 7 is a diagram for explaining a process of applying a visual effect in consideration of the characteristics of a dialog when an application related to a dialog exists, according to various embodiments of the present disclosure
  • the keyword of the dialog output as text may be 'coffee' or 'should I order', and among them, a target application related to the keyword 'coffee' may be searched.
  • instructions stored in the memory (eg, the memory 130 of FIG. 5 ) of the electronic device (eg, the electronic device of FIG. 5 ) are stored in the memory (eg, the memory 130 of FIG. 1 ) based on the determined text.
  • a target application may be configured to search for a target application related to a keyword among applications stored in (eg, the application 146 of FIG. 1 ).
  • a color may be extracted from the logo of the target application, and the color may be utilized as a visual effect.
  • a coffee mug icon 715 associated with 'coffee' which is a product related to a target application or keyword determined based on the dialog, may be output in 710 .
  • the color of the target application logo may be applied to the coffee cup icon (S715).
  • the color of the target application logo may be applied to the keyword 'Should I order' in step 730 (733).
  • the target application logo may be output together with the text of the dialog ( 736 ).
  • the target application when the user selects (eg, tap or touch) 738 the keyword 'Should I order', the target application may be linked and the following operation may be executed.
  • the user may select the target application logo 736 and the following operation may be executed.
  • in addition to tactile feedback such as a user's selection (eg, tap or touch), a user's voice response (eg, "Yes"), a gesture (eg, , swipe) or a user input of various methods may execute the following operation.
  • an order may be entered through a target application according to a user input at 730 , and text of a dialog related to the order may appear.
  • the location of a plurality of stores related to the target application may be searched based on the user's location, and an order may be placed at a point close to the user's location.
  • the text of the dialog stating that an order is made at 750 may be output, and a visual effect may be applied to the keyword 755 as in 730 .
  • the color of the target application logo may be applied to the keyword 755 .
  • FIG. 8 is a diagram for explaining a process in which a visual effect is applied when there are a plurality of applications related to a dialog, according to various embodiments of the present disclosure
  • a process in which a visual effect is applied when there are a plurality of target applications will be described at 810 , 830 , 850 , and 870 .
  • text may be output and the screen may be switched to 830 .
  • in 830 there may be a plurality of target applications related to the keyword 832 of the dialog text, and logos 834 and 836 of the plurality of target applications may be output together.
  • the user taps (or touches) the logo 834 of a specific target application 838 , the next operation through the target application may be executed.
  • the next operation may be executed in response to the user's voice response (eg, "Yes") in addition to the user's tap.
  • the operation when the operation is executed, a text indicating that the operation is executed through the corresponding application is output as shown in 850 , and the color of the corresponding application logo may be applied to the keyword 855 .
  • the artificial intelligence assistant may recommend one of a plurality of target applications.
  • the AI assistant may recommend a single target application based on location or based on a specific event (eg, a sale, new product launch, event, etc.).
  • a specific event eg, a sale, new product launch, event, etc.
  • the logo 874 of the recommended target application may appear larger among the logos 874 and 876 of the two target applications.
  • the color of the recommended target application logo may be applied to the keyword 872 .
  • FIG. 9 is a diagram for describing a process in which a visual effect is applied when one application among a plurality of applications is recommended based on a distance, according to various embodiments of the present disclosure
  • a process of applying a visual effect is illustrated at 910 , 940 , and 970 .
  • text for recommending a target application based on the distance is output.
  • a visual effect applied to the keyword 915 may vary according to the distance.
  • a visual effect such as a color or brightness of the keyword 945 displayed on the screen may increase as the distance increases.
  • visual effects such as color or brightness of the keyword 945 displayed on the screen may become lighter.
  • the font size of the keyword 975 may increase as the distance increases.
  • the font size of the keyword 975 may decrease.
  • the AI assistant can determine whether the distance is close based on global positioning system (GPS) information.
  • FIG. 10 is a diagram for describing a process in which a visual effect is applied differently when a dialog in text is output before and after an operation is executed, according to various embodiments of the present disclosure
  • steps 1010, 1030, 1050, and 1070 when the artificial intelligence assistant outputs the text of a dialog and the user responds to a certain action, a process in which visual effects are differently applied to the same keyword will be described in steps 1010, 1030, 1050, and 1070.
  • a visual effect in consideration of the characteristics of the keyword may be applied to the keyword 1015 of the text.
  • a logo 1017 eg, an IoT-related application icon
  • a related operation eg, turning on a boiler
  • the user may tap the target application logo 1017 to execute a related operation.
  • a related operation may be executed by the user's voice response (eg, "Yes").
  • a visual effect may be applied to the keyword 1035 .
  • different visual effects may be applied to the keyword 1015 before the operation execution and the keyword 1035 after the operation execution even if the keywords 'boilers' 1015 and 1035 are the same in 1010 and 1030 .
  • a red color may be applied to the keyword 1015 before the operation is executed, and a blue color may be applied to the keyword 1035 after the operation is executed.
  • a process in which different visual effects are applied to the same keyword is illustrated in steps 1050 and 1070 similar to those described above.
  • a visual effect may be applied to the keyword 1055 , and a logo 1057 (eg, an alarm application icon) of a target application related to the keyword 1055 may appear.
  • a logo 1057 eg, an alarm application icon
  • a related action eg, 7:10 am alarm
  • a visual effect when a dialog in text is output after a related operation is executed in 1070 , a visual effect may be applied to the keyword 1075 .
  • different visual effects may be applied to the keyword 1055 before the action execution and the keyword 1075 after the action execution even if the keywords '7:10 alarm' 1055 and 1075 are the same in 1050 and 1070.
  • a visual effect of a red color may be applied to the keyword 1055 before the operation is executed
  • a visual effect of a blue color may be applied to the keyword 1075 after the operation is executed.
  • FIG. 11 is a view for explaining a process in which a visual effect is applied to a background screen when a dialog in text is output, according to various embodiments of the present disclosure
  • FIG. 11 when a dialog in text is output, various embodiments in which a visual effect is applied not only to a keyword but also to a background are illustrated through 1110 , 1140 , and 1170 .
  • a visual effect of rain falling according to the characteristic of the keyword 'rain' 1113 may be applied to the background 1117 .
  • the background 1177 is different from the background 1147 of the case 1143 that is not related to the 'light' in the case (1173) associated with 'light'. ), a visual effect can be applied with a dark color.
  • FIG. 12 is a diagram for explaining a process of applying different visual effects according to meanings even for the same keyword, according to various embodiments of the present disclosure
  • 'should I turn on' may be included in the keyword of the text dialog in 1210 and 1250 ( 1216 and 1256 ).
  • the meaning may vary according to other keywords 1213 and 1253 in the text.
  • the keyword 1216 of 'should I turn on' is used with the keyword 1213 of 'smartphone do not disturb mode', so that a specific function (eg, the electronic device 101 of FIG. 1 ) of the electronic device (eg, the electronic device 101 of FIG. 1 ) is used. It means to turn on the smart phone do not disturb mode), and in 1250, the keyword 'turn on' (1256) is used with the keyword 'air conditioner' (1253) to turn on the air conditioner, which is an IoT device.
  • a visual effect of thickening the blurred keyword may be applied.
  • the keyword 1236 of 'Should I turn it on' in 1230 may be darker than the keyword 1216 of 'Should I turn it on' in 1210 .
  • the keyword 1256 'Should I turn on' used to mean turning on the air conditioner a visual effect in which text is blown in the wind may be applied.
  • the keyword 1276 of 'should I turn it on' in 1270 may be a keyword to which the effect of shaking in the wind is applied to the keyword 1256 of 'should I turn on' in 1250 .
  • FIG. 13 is a view for explaining a process of changing a visual expression according to a screen size, according to various embodiments of the present disclosure
  • the electronic device may be a notebook 1310 , smartphones 1330 and 1350 , and a smart watch 1370 , respectively, as illustrated in 1313 , 1333 , 1353 , and 1373 , respectively.
  • Screen sizes may vary.
  • the commands stored in the memory may be configured to be output by changing the length according to the screen size when outputting a dialog in text.
  • text 1317 in the form of a relatively complete sentence may be output on the screen 1313 when the electronic device is a notebook computer.
  • a relatively abbreviated text 1337 may be output on the screen 1333 when the electronic device is a smartphone with a horizontal folding axis.
  • texts 1357 and 1377 changed based on the screen size are output on the screen 1353 when the electronic device is a smartphone with a vertical folding axis or the screen 1373 when the electronic device is a smart watch.
  • an auxiliary visualization representation may be used to reduce the text length.
  • the text may be imaged or the text may be replaced with an icon corresponding to the text.
  • the length of the text may be reduced by changing the verb to the noun form or leaving only the key keyword.
  • FIG. 14 is a diagram for explaining a process of executing an operation related to a keyword or changing an option, according to various embodiments of the present disclosure
  • an option may be changed or a related operation may be executed by a user input with respect to a keyword of the text of a dialog.
  • a dialog in text is output in steps 1410 and 1450, a user input is possible for a keyword.
  • a user interface for changing a size option as shown in 1430 may be provided.
  • an order may be entered through the target application, and text indicating that the order is completed may be provided as shown in 1470. have.
  • the text 'Should I start 10 lunges?' may be output (not shown).
  • a user interface for changing the number of times may be provided (not shown).
  • a screen eg, a stopwatch
  • a target application, an exercise application, is executed and that an exercise is started after a predetermined time may be provided (not shown).
  • 15 is a view for explaining various visualization expression methods according to various embodiments of the present disclosure.
  • a different type of visual expression may be applied to keywords that are auxiliary information according to keywords in the text of a dialog.
  • a blue-based color of the image indicating that it is cold may be applied to a portion of 'it's chilly' among the text keywords.
  • a blurred visual effect may be applied to the auxiliary information 1515 that the temperature is 14 degrees Celsius, and may be disposed on the periphery of the screen.
  • keywords 1535 of a minimum of 6 degrees and a maximum of 14 degrees which are auxiliary information for information that today is colder than yesterday, may appear small in the periphery of the screen.
  • a blurry visual effect may be applied to the keyword 1555 'be strong a little more'.
  • a blurry visual effect may be applied to the keyword 1575 'Shall I turn it on?'.
  • various animation effects may be applied to keywords. For example, if text such as "How about a glass of cold beer when you go home?" is output, a visual effect of blue color is applied to the keyword 'a cold beer', and the text 'Han' and ' A visual effect such as an animation in which the text 'Gun' collides as if toasting each other may be applied (not shown).
  • 16 is a diagram for explaining a process of switching a screen for outputting a text of a dialog, according to various embodiments of the present disclosure
  • a process of switching between the screen on which the text of the dialog is output and the basic screen of the electronic device is illustrated.
  • text 1613 is output, and visual effects such as color, font, size, and animation may be applied to a keyword (eg, '7640 steps of 1610 ).
  • a basic screen eg, a screen indicating time
  • the electronic device may appear darker than the text 1613 as a background ( 1617 ).
  • the basic screen of the electronic device may come forward, and the text of the dialog may appear blurry as a background ( 1630 ).
  • the text of the dialog appears again, and the basic screen of the electronic device may appear blurry as a background ( 1650 ).
  • 17 is a diagram for describing mapping between a dialog and a visual effect, according to various embodiments of the present disclosure.
  • an example 1710 of a dialog text output on a screen of an electronic device eg, the electronic device 101 of FIG. 1
  • a color 1740 and an animation effect 1770 corresponding thereto are mapped. have.
  • the dialog may be a small talk, information, or suggestion.
  • the dialog when the dialog is a daily conversation, information, and proposal, a color pool, a font pool, and an animation pool that are mapped respectively are determined, and can be arbitrarily determined within the pool.
  • the visual effect eg, the database 510 of FIG. 5
  • the database eg, the database 510 of FIG. 5
  • icons and templates 512 may be mapped.
  • FIG. 18 is a flowchart illustrating a method of operating an electronic device according to various embodiments of the present disclosure
  • FIG. 18 portions overlapping with those described in FIGS. 1 to 17 will be omitted and described.
  • a characteristic of a keyword determined based on a text dialog output through a screen is identified.
  • a dialog in text may be generated in a natural language platform (eg, the natural language platform 220 of FIG. 2 ), and the keyword determined based on the dialog may be generated by an intelligent server (eg, the intelligent server 200 of FIG. 2 ) )) may be determined based on a capsule (eg, capsule A 401 of FIG. 4 ) stored in a capsule database (eg, capsule database 230 of FIG. 2 ).
  • characteristics of the keyword may be small talk, information, and suggestion.
  • a visual effect may be determined in consideration of the characteristics of the keyword.
  • visual effects include text fonts, text color or size characteristics, icons and templates, text animations (eg, icons and templates 512 of FIG. 5 , text features 514 , and text fonts 516 ). , text animation 518), and the like.
  • the visual effect determination process may vary depending on whether the target application exists.
  • the target application may be, for example, an application related to the setting of the electronic device 101 and/or a third party application.
  • a visual effect may be determined by further considering the characteristics of the target application.
  • the visual effect when the target application does not exist, the visual effect may be determined in consideration of only the characteristics of the keyword.
  • the visual effect in consideration of the characteristics of the target application may include at least one of a logo of the target application, a color of the logo, and an icon associated with the target application.
  • a logo of the target application may be included in a visual effect in consideration of the characteristics of the target application.
  • a color of the logo may be included in a visual effect in consideration of the characteristics of the target application.
  • an icon associated with the target application may be included in a visual effect in consideration of the characteristics of the target application.
  • a logo of the target application is determined, a representative color is extracted from the logo, and an icon of a product related to a service provided by the target application is determined.
  • a visual effect corresponding to a characteristic of a keyword and a visual effect corresponding to a characteristic of a target application may be determined with reference to a database (eg, the database 510 of FIG. 5 ).
  • the database includes icons and templates, features related to the size or color of text, text fonts, and text animations (eg, icons and templates 512, text features 514, and text fonts 516 in FIG. 5). , text animation 518) may be included.
  • a color pool, a font pool, and an animation pool each mapped according to a dialog are determined, and a visual effect may be arbitrarily determined within the corresponding pool.
  • the visual effect eg, the database 510 of FIG. 5
  • the database eg, the database 510 of FIG. 5
  • Icons and templates 512 may be mapped (eg, the table of FIG. 17 ).
  • the electronic device eg, the electronic device 101 of FIG. 1
  • the electronic device is a display module that outputs a dialog in text through a screen (eg, 1313 , 1333 , 1353 , 1373 of FIG. 13 ) (eg, the display module 160 of FIG. 1 ), a memory in which computer-executable instructions are stored (eg, the memory 130 of FIG. 1 ) and accessing the memory to execute the commands and a processor (eg, the processor 120 of FIG. 1 ).
  • the commands determine the characteristics of the keyword determined based on the text, and determine the visual effect in consideration of the characteristic, but if there is a target application related to the keyword, the visual effect is further taken into consideration in consideration of the characteristics of the target application It may be configured to determine and apply a visual effect when outputting text.
  • the visual effect further considering the characteristics of the target application may include at least one of a logo of the target application, a color of the logo, and an icon associated with the target application.
  • the commands when there are a plurality of target applications and one is recommended, the commands may be configured to output logos of the plurality of target applications, but to output a larger logo of the target application to be recommended.
  • the commands may be configured to differently output a visual effect before and after a user input for a keyword.
  • the commands may be configured to change and output text according to the size of the screen.
  • the electronic device (eg, the electronic device 101 of FIG. 1 ) stores a visual effect corresponding to a characteristic of a keyword and a visual effect corresponding to a characteristic of a target application (eg, a database). Further comprising database 510 of FIG. 5 , the instructions may be configured to determine a visual effect with reference to the database.
  • the text and keywords may be determined by a natural language platform (eg, the natural language platform 220 of FIG. 2 ).
  • a characteristic of a keyword determined based on the text is identified. action to do; an operation of determining a visual effect in consideration of the characteristic, but further considering the characteristic of the target application when a target application related to the keyword exists; and applying a visual effect when outputting text.
  • the visual effect further considering the characteristics of the target application may include at least one of a logo of the target application, a color of the logo, or a product icon associated with the target application.
  • the target application to be recommended may include an operation of outputting a larger logo of .
  • a method of operating an electronic device may include an operation of differently outputting visual effects before and after a user input for a keyword.
  • a method of operating an electronic device may include an operation of changing and outputting text according to a size of a screen.
  • the method of operating an electronic device includes a database that stores a visual effect corresponding to a characteristic of a keyword and a visual effect corresponding to a characteristic of a target application. It may include an operation of determining a visual effect with reference to (eg, the database 510 of FIG. 5 ).
  • the text and keywords may be determined by a natural language platform (eg, the natural language platform 220 of FIG. 2 ).
  • the electronic device on a recording medium in which a program for controlling the operation of the electronic device (eg, the electronic device 101 of FIG. 1 ) is recorded, the electronic device outputs a text dialog through a screen.
  • the operation of identifying the characteristics of the keyword determined based on the text; an operation of determining a visual effect in consideration of the characteristic, but further considering the characteristic of the target application when a target application related to the keyword exists; and a program for applying a visual effect when outputting text may be recorded.
  • the visual effect further considering the characteristics of the target application may include at least one of a logo of the target application, a color of the logo, or a product icon associated with the target application.
  • a program for controlling the operation of the electronic device eg, the electronic device 101 of FIG. 1
  • the electronic device recommends a plurality of target applications
  • a program for outputting logos of a plurality of target applications and outputting a larger logo of the target application to be recommended may be recorded.
  • the electronic device in a recording medium in which a program for controlling the operation of the electronic device (eg, the electronic device 101 of FIG. 1 ) is recorded, the electronic device may change the visual effect before and after the user input for the keyword differently.
  • a program that causes an operation to output may be recorded.
  • the electronic device in a recording medium in which a program for controlling the operation of the electronic device (eg, the electronic device 101 of FIG. 1 ) is recorded, the electronic device outputs the changed text according to the size of the screen.
  • a program that causes the operation to be recorded may be recorded.
  • the electronic device in a recording medium in which a program for controlling the operation of the electronic device (eg, the electronic device 101 of FIG. 1 ) is recorded, the electronic device includes a visual effect corresponding to the characteristic of a keyword and a target application
  • a program for determining the visual effect may be recorded with reference to a database (eg, the database 510 of FIG. 5 ) storing visual effects corresponding to the characteristics of .

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Human Computer Interaction (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Computer Hardware Design (AREA)
  • Electrotherapy Devices (AREA)
  • Ultra Sonic Daignosis Equipment (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Un dispositif électronique selon un mode de réalisation peut comprendre : un module d'affichage qui affiche un dialogue texte à travers un écran ; une mémoire dans laquelle sont stockées des instructions exécutables par ordinateur ; et un processeur qui accède à la mémoire pour exécuter les instructions, lorsqu'un dialogue texte est affiché à travers le module d'affichage, les instructions sont conçues pour : identifier les caractéristiques d'un mot-clé déterminé sur la base d'un texte ; déterminer un effet visuel en tenant compte des caractéristiques et, lorsqu'il existe une application cible associée au mot-clé, considérer en outre les caractéristiques de l'application cible dans la détermination de l'effet visuel ; et appliquer l'effet visuel lorsque le texte est affiché. Divers autres modes de réalisation sont possibles.
PCT/KR2022/001280 2021-03-11 2022-01-25 Dispositif électronique pour appliquer un effet visuel à un texte de dialogue et son procédé de commande WO2022191425A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020210032089A KR20220127600A (ko) 2021-03-11 2021-03-11 다이얼로그 텍스트에 시각적 효과를 적용하는 전자 장치 및 이의 제어 방법
KR10-2021-0032089 2021-03-11

Publications (1)

Publication Number Publication Date
WO2022191425A1 true WO2022191425A1 (fr) 2022-09-15

Family

ID=83226906

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2022/001280 WO2022191425A1 (fr) 2021-03-11 2022-01-25 Dispositif électronique pour appliquer un effet visuel à un texte de dialogue et son procédé de commande

Country Status (2)

Country Link
KR (1) KR20220127600A (fr)
WO (1) WO2022191425A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150277681A1 (en) * 2014-03-31 2015-10-01 Monticello Enterprises LLC System and method for providing a single input field having multiple processing possibilities
KR20190104197A (ko) * 2017-04-18 2019-09-06 구글 엘엘씨 문맥형 음성-구동 딥 북마킹
KR20190134975A (ko) * 2018-05-01 2019-12-05 삼성전자주식회사 인공지능 시스템의 앱들 또는 스킬들의 리스트를 표시하는 증강 현실 장치 및 동작 방법
KR20200057426A (ko) * 2018-11-16 2020-05-26 삼성전자주식회사 음성 인식 기반 이미지를 표시하는 전자 장치
KR102116309B1 (ko) * 2018-12-17 2020-05-28 주식회사 인공지능연구원 가상 캐릭터와 텍스트의 동기화 애니메이션 출력 시스템

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150277681A1 (en) * 2014-03-31 2015-10-01 Monticello Enterprises LLC System and method for providing a single input field having multiple processing possibilities
KR20190104197A (ko) * 2017-04-18 2019-09-06 구글 엘엘씨 문맥형 음성-구동 딥 북마킹
KR20190134975A (ko) * 2018-05-01 2019-12-05 삼성전자주식회사 인공지능 시스템의 앱들 또는 스킬들의 리스트를 표시하는 증강 현실 장치 및 동작 방법
KR20200057426A (ko) * 2018-11-16 2020-05-26 삼성전자주식회사 음성 인식 기반 이미지를 표시하는 전자 장치
KR102116309B1 (ko) * 2018-12-17 2020-05-28 주식회사 인공지능연구원 가상 캐릭터와 텍스트의 동기화 애니메이션 출력 시스템

Also Published As

Publication number Publication date
KR20220127600A (ko) 2022-09-20

Similar Documents

Publication Publication Date Title
WO2021075736A1 (fr) Dispositif électronique et procédé associé de partage de commande vocale
WO2022010157A1 (fr) Procédé permettant de fournir un écran dans un service de secrétaire virtuel à intelligence artificielle, et dispositif de terminal d'utilisateur et serveur pour le prendre en charge
WO2020180000A1 (fr) Procédé d'expansion de langues utilisées dans un modèle de reconnaissance vocale et dispositif électronique comprenant un modèle de reconnaissance vocale
WO2023113502A1 (fr) Dispositif électronique et procédé de recommandation de commande vocale associé
WO2020101389A1 (fr) Dispositif électronique d'affichage d'une image fondée sur la reconnaissance vocale
WO2022191395A1 (fr) Appareil de traitement d'une instruction utilisateur et son procédé de fonctionnement
WO2022139420A1 (fr) Dispositif électronique et procédé de partage d'informations d'exécution d'un dispositif électronique concernant une entrée d'utilisateur avec continuité
WO2022131566A1 (fr) Dispositif électronique et procédé de fonctionnement de dispositif électronique
WO2022191425A1 (fr) Dispositif électronique pour appliquer un effet visuel à un texte de dialogue et son procédé de commande
WO2020171545A1 (fr) Dispositif électronique et système de traitement de saisie d'utilisateur et procédé associé
WO2022177165A1 (fr) Dispositif électronique et procédé permettant d'analyser un résultat de reconnaissance vocale
WO2022182038A1 (fr) Dispositif et procédé de traitement de commande vocale
WO2022025448A1 (fr) Dispositif électronique et son procédé de fonctionnement
WO2022196994A1 (fr) Dispositif électronique comprenant un module de conversion de texte en parole personnalisé, et son procédé de commande
WO2023008798A1 (fr) Dispositif électronique de gestion de réponses inappropriées et son procédé de fonctionnement
WO2023158076A1 (fr) Dispositif électronique et son procédé de traitement d'énoncé
WO2023048379A1 (fr) Serveur et dispositif électronique pour traiter un énoncé d'utilisateur, et son procédé de fonctionnement
WO2023106862A1 (fr) Dispositif électronique et procédé de fonctionnement d'un dispositif électronique
WO2023017975A1 (fr) Dispositif électronique permettant de délivrer en sortie un résultat de traitement de commande vocale à la suite d'un changement d'état et son procédé de fonctionnement
WO2022131805A1 (fr) Procédé de fourniture de réponse à une entrée vocale et dispositif électronique pour le prendre en charge
WO2024029845A1 (fr) Dispositif électronique et son procédé de reconnaissance vocale
WO2023177079A1 (fr) Serveur et dispositif électronique permettant de traiter une parole d'utilisateur sur la base d'un vecteur synthétique, et procédé de fonctionnement associé
WO2023043025A1 (fr) Procédé de traitement d'un énoncé continu incomplet, serveur et procédé de réalisation de dispositif électronique
WO2024043670A1 (fr) Procédé d'analyse de la parole d'un utilisateur, et dispositif électronique prenant celui-ci en charge
WO2024058597A1 (fr) Dispositif électronique et procédé de traitement d'énoncé d'utilisateur

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22767322

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22767322

Country of ref document: EP

Kind code of ref document: A1