WO2018135803A1 - Voice input processing method and electronic device for supporting the same - Google Patents
Voice input processing method and electronic device for supporting the same Download PDFInfo
- Publication number
- WO2018135803A1 WO2018135803A1 PCT/KR2018/000540 KR2018000540W WO2018135803A1 WO 2018135803 A1 WO2018135803 A1 WO 2018135803A1 KR 2018000540 W KR2018000540 W KR 2018000540W WO 2018135803 A1 WO2018135803 A1 WO 2018135803A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- electronic device
- voice input
- user
- indication
- processor
- Prior art date
Links
- 238000003672 processing method Methods 0.000 title claims description 8
- 238000004891 communication Methods 0.000 claims abstract description 72
- 230000000007 visual effect Effects 0.000 claims abstract description 28
- 230000002618 waking effect Effects 0.000 claims abstract description 17
- 238000000034 method Methods 0.000 claims description 82
- 230000006870 function Effects 0.000 claims description 68
- 238000013507 mapping Methods 0.000 claims description 33
- 230000008569 process Effects 0.000 description 21
- 238000012545 processing Methods 0.000 description 13
- 238000010586 diagram Methods 0.000 description 10
- 230000004044 response Effects 0.000 description 9
- 238000001228 spectrum Methods 0.000 description 9
- 230000001413 cellular effect Effects 0.000 description 8
- 238000005286 illumination Methods 0.000 description 7
- 230000035945 sensitivity Effects 0.000 description 7
- 230000008859 change Effects 0.000 description 6
- 230000014509 gene expression Effects 0.000 description 5
- 230000000694 effects Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 108010076504 Protein Sorting Signals Proteins 0.000 description 2
- 239000008280 blood Substances 0.000 description 2
- 210000004369 blood Anatomy 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 238000002591 computed tomography Methods 0.000 description 2
- 230000036541 health Effects 0.000 description 2
- 235000012736 patent blue V Nutrition 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- WQZGKKKJIJFFOK-GASJEMHNSA-N Glucose Natural products OC[C@H]1OC(O)[C@H](O)[C@@H](O)[C@@H]1O WQZGKKKJIJFFOK-GASJEMHNSA-N 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 238000002583 angiography Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000036772 blood pressure Effects 0.000 description 1
- 230000036760 body temperature Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000005611 electricity Effects 0.000 description 1
- 238000002567 electromyography Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000446 fuel Substances 0.000 description 1
- 239000011521 glass Substances 0.000 description 1
- 239000008103 glucose Substances 0.000 description 1
- 230000006698 induction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000002595 magnetic resonance imaging Methods 0.000 description 1
- 238000001646 magnetic resonance method Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012567 pattern recognition method Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000005406 washing Methods 0.000 description 1
- 229910052724 xenon Inorganic materials 0.000 description 1
- FHNFHKCVQCLJFQ-UHFFFAOYSA-N xenon atom Chemical compound [Xe] FHNFHKCVQCLJFQ-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3228—Monitoring task completion, e.g. by use of idle timers, stop commands or wait commands
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/26—Power supply means, e.g. regulation thereof
- G06F1/32—Means for saving power
- G06F1/3203—Power management, i.e. event-based initiation of a power-saving mode
- G06F1/3206—Monitoring of events, devices or parameters that trigger a change in power modality
- G06F1/3231—Monitoring the presence, absence or movement of users
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/30—Authentication, i.e. establishing the identity or authorisation of security principals
- G06F21/31—User authentication
- G06F21/32—User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/083—Recognition networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
- G10L15/32—Multiple recognisers used in sequence or in parallel; Score combination systems therefor, e.g. voting systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L2015/088—Word spotting
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/221—Announcement of recognition results
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/225—Feedback of the input speech
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Definitions
- the present disclosure generally relates to a voice input processing method and an electronic device for supporting the same.
- electronic devices have been developed to include a microphone and provide a voice interface, so that the devices may perform functions such as purchasing products or operating smart home systems via voice input received through the microphone. In doing so, the electronic device may perform voice recognition for the received voice input. If a specified word (e.g., a wake-up word) for waking up the electronic device is included in the result of the voice recognition, the electronic device may wake up itself or a module associated with performing a specified function.
- a specified word e.g., a wake-up word
- the plurality of electronic devices may be woken up irrespective of intention of the user.
- the user may unnecessarily receive the same response from multiple devices or may receive error information (e.g., a voice, a screen indication, or the like) regarding the voice recognition from some devices.
- error information e.g., a voice, a screen indication, or the like
- an aspect of the present disclosure is to provide a voice input processing method for waking up one of a plurality of electronic devices if there are the plurality of electronic devices which are woken up by the same specified word.
- the electronic device for supporting the same is also described.
- an electronic device includes a microphone configured to receive a voice input, a communication circuitry configured to communicate with an external electronic device, an indicator configured to provide at least one visual indication, and a processor configured to be electrically connected with the microphone, the communication circuitry, and the indicator, and a memory configured to be electrically connected with the processor.
- the memory stores instructions, when executed, cause the processor to receive a first voice input through the microphone, perform a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, display a first visual indication through the indicator, receive a second voice input through the microphone, perform a second voice recognition for the second voice input, and if a second specified word corresponding to the first visual indication is included in a result of the second voice recognition, wake up the electronic device.
- a voice input processing method of an electronic device includes receiving a first voice input through a microphone, performing a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, displaying an indication through the indicator, receiving a second voice input through the microphone, performing a second voice recognition for the second voice input, and if a second specified word corresponding to the indication is included in a result of the second voice recognition, waking up the electronic device.
- the plurality of electronic devices may provide a method for determining intention of the user (e.g., a method for selecting one electronic device to provide a response).
- FIG. 1 is a block diagram illustrating an electronic device that processes voice inputs according to an embodiment.
- FIG. 2 is a drawing illustrating a system that processes voice inputs according to an embodiment.
- FIG. 3 is a drawing illustrating a method for providing an indication according to an embodiment.
- FIG. 4 is a flowchart illustrating an operation method of an electronic device that processes voice inputs according to an embodiment.
- FIG. 5 is a flowchart illustrating another operation method of an electronic device that processes voice inputs according to an embodiment.
- FIG. 6 is a drawing illustrating a method for determining an indication according to an embodiment.
- FIG. 7 is a drawing illustrating another method for determining an indication according to an embodiment.
- FIG. 8 is a flowchart illustrating an operation method of an electronic device for determining an indication according to an embodiment.
- FIG. 9 is a flowchart illustrating a method for displaying an indication determined through an external electronic device at an electronic device according to an embodiment.
- FIG. 10 is a drawing illustrating a method for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
- FIG. 11 is a signal sequence diagram illustrating a system for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
- FIG. 12 is a drawing illustrating a method for determining whether to wake up an electronic device through adjustment of receive sensitivity of a voice input according to an embodiment.
- FIG. 13 is a flowchart illustrating an operation method of an electronic device for adjusting the received sensitivity of a voice input according to an embodiment.
- FIG. 14 is a drawing illustrating a method for determining whether to wake up an electronic device based on usage history information according to an embodiment.
- FIG. 15 is a flowchart illustrating an operation method of an electronic device for determining whether to wake up an electronic device based on usage history information according to an embodiment.
- FIG. 16 is a drawing illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
- FIG. 17 is a flowchart illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
- FIG. 18 is a block diagram illustrating an electronic device 1801 in a network environment 1800 according to an embodiment.
- FIG. 19 is a block diagram illustrating an electronic device 1901 according to one embodiment.
- FIG. 20 is a block diagram illustrating a program module 2010 according to an embodiment.
- the expressions “A or B,” or “at least one of A and/or B” may indicate A and B, A, or B.
- the expression “A or B” or “at least one of A and/or B” may indicate (1) at least one A, (2) at least one B, or (3) both at least one A and at least one B.
- first may refer to modifying various different elements of various embodiments of the present disclosure, but are not intended to limit the elements.
- a first user device and “a second user device” may indicate different users regardless of order or importance.
- a first component may be referred to as a second component and vice versa without departing from the scope of the present disclosure.
- a component for example, a first component
- another component for example, a second component
- the component may be directly connected to the other component or connected through another component (for example, a third component).
- a component for example, a first component
- another component for example, a third component
- the expression “a device configured to” in some situations may indicate that the device and another device or part are “capable of.”
- the expression "a processor configured to perform A, B, and C” may indicate a dedicated processor (for example, an embedded processor) for performing a corresponding operation or a general purpose processor (for example, a central processing unit (CPU) or application processor (AP)) for performing corresponding operations by executing at least one software program stored in a memory device.
- a dedicated processor for example, an embedded processor
- a general purpose processor for example, a central processing unit (CPU) or application processor (AP)
- An electronic device may be a smartphone, a tablet personal computer (PC), a mobile phone, a video telephone, an electronic book reader, a desktop PC, a laptop PC, a netbook computer, a workstation, a server, a personal digital assistant (PDA), a portable multimedia player (PMP), a Motion Picture Experts Group (MPEG-1 or MPEG-2) Audio Layer 3 (MP3) player, a mobile medical device, a camera, a wearable device, etc.
- PDA personal digital assistant
- PMP portable multimedia player
- MPEG-1 or MPEG-2 Motion Picture Experts Group Audio Layer 3
- the wearable device may be an accessory-type device (e.g., a watch, a ring, a bracelet, an anklet, a necklace, glasses, a contact lens, a head-mounted device (HMD)), a textile- or clothing-integrated-type device (e.g., an electronic apparel), a body-attached-type device (e.g., a skin pad or a tattoo), or a bio-implantable-type device (e.g., an implantable circuit)
- HMD head-mounted device
- a textile- or clothing-integrated-type device e.g., an electronic apparel
- a body-attached-type device e.g., a skin pad or a tattoo
- a bio-implantable-type device e.g., an implantable circuit
- the electronic device may be a home appliance.
- the smart home appliance may be a television (TV), a digital video/versatile disc (DVD) player, an audio, a refrigerator, an air conditioner, a cleaner, an oven, a microwave oven, a washing machine, an air cleaner, a set-top box, a home automation control panel, a security control panel, a television (TV) box (e.g., Samsung HomeSyncTM, Apple TVTM, or Google TVTM), a game console (e.g., XboxTM or PlayStationTM), an electronic dictionary, an electronic key, a camcorder, an electronic picture frame, etc.
- TV television
- DVD digital video/versatile disc
- the electronic device may be various medical devices (e.g., various portable medical measurement devices (e.g., a blood glucose measuring device, a heart rate measuring device, a blood pressure measuring device, a body temperature measuring device, or the like), a magnetic resonance angiography (MRA) device, a magnetic resonance imaging (MRI) device, a computed tomography (CT) device, a scanner, an ultrasonic device, or the like), a navigation device, a global navigation satellite system (GNSS), an event data recorder (EDR), a flight data recorder (FDR), a vehicle infotainment device, electronic equipment for vessels (e.g., a navigation system, a gyrocompass, or the like), avionics, a security device, a head unit for a vehicle, an industrial or home robot, an automatic teller machine (ATM), a point of sales (POS) device of a store, or an Internet of things (IoT) device (e.g., a light bulb, various portable medical measurement
- the electronic device may be at least one of a part of furniture or a building/structure, an electronic board, an electronic signature receiving device, a projector, or a measuring instrument (e.g., a water meter, an electricity meter, a gas meter, a wave meter, or the like).
- the electronic device may be one or more combinations of the above-mentioned devices.
- the electronic device according to some various embodiments of the present disclosure may be a flexible device.
- the electronic device according to an embodiment of the present disclosure is not limited to the above-mentioned devices, and may include new electronic devices as new technology is developed.
- the term “user” used herein may refer to a person who uses an electronic device or may refer to a device (e.g., an artificial intelligence electronic device) that uses an electronic device.
- FIG. 1 is a block diagram illustrating an electronic device that processes voice inputs according to an embodiment.
- the electronic device 100 may perform voice recognition for a voice input received through a microphone 120. If a first specified word (e.g., a first wake-up word) for waking up the electronic device 100 is included in the result of the voice recognition, the electronic device 100 may wake up itself or at least one module associated with performing a function included in the electronic device 100. Meanwhile, if there is another external electronic device which can be woken up by the first wake-up word, the electronic device 100 may provide an indication to the user that is different from the indication provided in the external electronic device. This way, both the electronic device 100 and the external electronic device are not woken up simultaneously when the first specified word is spoken.
- a first specified word e.g., a first wake-up word
- the user may verify that the electronic device 100, which the user wants to wake up, has not woken up.
- the user may then speak a second specified word (e.g., a second wake-up word) corresponding to the indication provided from the electronic device 100 to wake up the electronic device 100.
- a second specified word e.g., a second wake-up word
- the electronic device 100 for performing the above-mentioned function may include a processor 110, the microphone 120, a speaker 130, an indicator 140 (or a user interface), a memory 150, and a communication circuitry 160.
- the elements of the electronic device 100 are not limited thereto.
- the electronic device 100 may fail to include at least one of the above-mentioned elements and may further include at least one other element.
- the electronic device 100 may further include a proximity sensor which measures a proximity state and/or distance of an object (e.g. the user).
- at least one of the above-mentioned elements may be located and fixed within a housing.
- the housing may include a front surface, a rear surface, and a side surface which surrounds at least part of a space between the front surface and the rear surface.
- the processor 110 may execute arithmetic operations or data processing to control at least one other element of the electronic device 100.
- the processor 110 may drive, for example, an operating system (OS) or an application program to control a plurality of hardware or software elements of the electronic device 100.
- the processor 110 may include one or more of a central processing unit (CPU), an application processor (AP), or a communication processor (CP).
- the processor 110 may be implemented as a system on chip (SoC).
- the processor 110 may include a microprocessor or any suitable type of processing circuitry, such as one or more general-purpose processors (e.g., ARM-based processors), a Digital Signal Processor (DSP), a Programmable Logic Device (PLD), an Application-Specific Integrated Circuit (ASIC), a Field-Programmable Gate Array (FPGA), a Graphical Processing Unit (GPU), a video card controller, etc.
- general-purpose processors e.g., ARM-based processors
- DSP Digital Signal Processor
- PLD Programmable Logic Device
- ASIC Application-Specific Integrated Circuit
- FPGA Field-Programmable Gate Array
- GPU Graphical Processing Unit
- the processor 110 may analyze a voice input received through the microphone 120. According to an embodiment, the processor 110 may perform voice recognition for the received voice input.
- the processor 110 may analyze voice data corresponding to a voice input received through the microphone 120 and may convert the voice data into text. For example, if a user speaks the word “HI”, the microphone 120 may convert vibration energy from the user’s voice into an electronic signal and may transmit the converted electric signal to the processor 110.
- the processor 110 may convert the transmitted electric signal into frequency spectrum information and may compare the converted frequency spectrum information with a plurality of frequency spectrum information corresponding to various characters stored in the memory 150 to determine that characters (e.g., “H” and “I”) were spoken.
- the processor 110 may perform voiceprint identification for a received voice input.
- the processor 110 may perform voiceprint identification in substantially the same manner as voice recognition.
- the processor 110 may use the frequency spectrum information converted from the user’s voice input. For example, if the user speaks the word “HI”, voice recognition may only determine that characters (e.g., “H” and “I”) were included in the input, whereas voiceprint identification may determine that a specific user spoke “HI.”
- the processor 110 may wake up at least one other element included in the electronic device 100.
- the processor 110 may supply power to one or more hardware elements and/or may load a software element (e.g., a program module) into the memory 150 so that the software element can be executed.
- a software element e.g., a program module
- the processor 110 may determine whether there are one or more external electronic devices which also can be woken up by the specified word. For example, if a signal corresponding to the specified word is received from the external electronic device, the processor 110 may determine that the external electronic device is able to be woken up by the specified word.
- the signal corresponding to the wake-up word may be a signal transmitted from the external electronic device and may include, for example, a notification that the external electronic device recognizes the specified word and is in a state of waiting for wake-up. In this case, the processor 110 may not wake up elements in the electronic device 100 may provide (e.g. display) an indication through the indicator 140.
- the processor 110 may provide an indication that is different from the indication provided from the external electronic device such that the user can distinguish the electronic device 100 from the external electronic device.
- the indication may be, for example, a color displayed by the indicator 140.
- the indication may include identification information assigned to each electronic device, such as a nickname. A display object corresponding to the identification information may be displayed on the display.
- the processor 110 may wake up at least one other element included in the electronic device 100. This way, the processor 110 may replace the specified word (e.g., a first specified word) for waking up the electronic device 100 with another specified word (e.g., a second specified word) after providing the indication.
- the specified word e.g., a first specified word
- another specified word e.g., a second specified word
- the processor 110 may manage the types of the indications and how the indications correspond to different electronic devices.
- the processor 110 may store information (e.g., a table) of mapping a first indication to the identification information of a first electronic device and mapping a second indication to the identification information of a second electronic device.
- red indication may be assigned to the first electronic device and blue indication may be assigned to the second electronic device.
- the processor 110 may receive information regarding the assigned indication via the communication circuitry 160 from an external electronic device and may provide the indication based on the received information about the indication. This way, the processor 110 may not directly manage the mapping of the indications to the various electronic devices.
- the electronic device 100 may be used as a device (e.g., a server device, a hub device, or a master device) which manages the mapping information or may be used as a device (e.g., a slave device or a client device) which receives the mapping information from another device which manages the mapping information.
- the processor 110 may display the indication assigned to the electronic device 100 and may transmit information about indications assigned to other external electronic devices to the other external electronic devices via the communication circuitry 160.
- the processor 110 may determine whether to wake up the microphone 120 based on sensing information obtained by a sensor module (e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like) included in the electronic device 100. For example, using the sensor module, the processor 110 may determine whether the user or a body part of the user is close to (or in contact with) the electronic device 100. If so, the processor 110 may wake up the microphone 120 to receive voice input of the user. Further, the processor 110 may inform another external electronic device that the user is close to (or in contact with) the electronic device 100. In this case, the external electronic device may deactivate its microphone so that the external electronic device does not erroneously receive the voice input from the user.
- a sensor module e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like
- the microphone in the external electronic device may be only deactivated for a specified time period. For example, after the specified time period elapses, the microphone may be reactivated.
- the processor 110 may recognize a face or an operation of the user through a camera or the like. For example, using the camera, the processor 110 may determine whether the user looks at the electronic device 100, and may wake up the microphone 120 only in the electronic device 100 that the user looked at.
- the processor 110 may adjust the received sensitivity of the voice input received through the microphone 120. For example, the processor 110 may adjust a threshold of received intensity (or receive strength or a volume level) of the voice input for voice recognition. According to an embodiment, if the received intensity is a first level and if received intensity of the voice input received by another external electronic device is at a second level lower than the first level, the processor 110 may set the threshold of received intensity for the microphone 120 and the received intensity for the microphone included in the external electronic device to values between the first level and the second level.
- the electronic device 100 may be able to process the voice input from the user while the external electronic device may not process the voice input.
- the processor 110 may provide an indication depending on the threshold of received intensity to guide the user. For example, the electronic device with the highest threshold may display an indication of a first color (e.g., a green color) to inform the user that it is capable of performing voice recognition at the current volume. An electronic device with the lowest received intensity may display an indication of a second color (e.g., a red color) to inform the user that it cannot perform voice recognition at the current volume.
- a first color e.g., a green color
- a second color e.g., a red color
- the processor 110 may determine whether to wake up the electronic device 100 based on usage history information of the user. For example, the processor 110 may identify the user through voiceprint identification and may identify the history in which the identified user uses the electronic device 100. For example, the history information may include the frequency of use, the number of use, the difference in the frequency of use/the number of use with another user, etc. The processor 110 can then use this history information to determine whether to wake up the electronic device 100. For example, if the electronic device 100 is an electronic device frequently used by a first user, the processor 110 may not wake up the electronic device 100 when voice input from a second user is received. In this case, the processor 110 may wake up the electronic device 100 only if voice input from the first user is received.
- the processor 110 may determine whether to wake up the electronic device 100, based on settings information set by the user. For example, if the user sets the electronic device 100 as an electronic device frequently used by him or her, the processor 110 may wake up the electronic device 100 when voice input from the user is received.
- the processor 110 may determine whether to wake up the electronic device 100 based on a location of a personal device (e.g., a smart watch, a smartphone, or the like) of the user. For example, in case of a wearable electronic device, the processor 110 may determine whether the user wears the wearable electronic device. If determining that the user wears the wearable electronic device, the processor 110 may wake up the electronic device 100 adjacent to the wearable electronic device. The processor 110 may determine a location relationship between the personal device of the user and the electronic device 100 using an illumination sensor included in the electronic device 100 or the degree of noise measured through the microphone 120.
- a personal device e.g., a smart watch, a smartphone, or the like
- the processor 110 may determine whether the user wears the wearable electronic device. If determining that the user wears the wearable electronic device, the processor 110 may wake up the electronic device 100 adjacent to the wearable electronic device.
- the processor 110 may determine a location relationship between the personal device of the user and the electronic device 100 using an illumination sensor included in the
- the electronic device 100 may determine that the personal device of the user is adjacent to the electronic device 100. According to an embodiment, the electronic device 100 may compare the measured values (e.g., the illumination value and the value indicating the degree of noise) with values measured by other external electronic devices and may select the most similar device to the personal device of the user as the device closest to the user.
- the measured values e.g., the illumination value and the value indicating the degree of noise
- the location relationship between the personal device of the user and the electronic device 100 may be determined through various indoor positioning methods, such as ones using communication beacons (e.g., Bluetooth low energy (BLE), wireless-fidelity (Wi-Fi), or the like), fingerprinting, visible light communication using light-fidelity (Li-Fi), multilateration using a received signal strength indicator (RSSI) and a round trip time (RTT), etc.
- BLE Bluetooth low energy
- Wi-Fi wireless-fidelity
- RTSI received signal strength indicator
- RTT round trip time
- the BLE device may transmit an advertising packet to peripheral electronic devices and the electronic device 100 which receives the packet may calculate its location through a cell ID or multilateration using the device ID of the BLE device or the distance between the electronic device 100 and the BLE device.
- the packet may include data such as a device ID, a service place name, and a transmit signal strength.
- the electronic device 100 may calculate the distance value between the BLE device and the electronic device 100 using an RSSI signal value of the BLE device.
- the method for calculating the distance using the RSSI signal value may include calculating the distance using strength (power) loss of the RSSI signal or path loss between the electronic device 100 and the BLE device.
- the fingerprint method may be a pattern recognition method for measuring strength of signals received based on Wi-Fi, Bluetooth, or the like.
- the method may involve knowing several reference locations in advance and configuring a radio map. When a reference location on the radio map is determined to have a signal strength characteristic most similar to the strength of the signal received by the electronic device 100, the reference location may be determined to be the location of the electronic device 100.
- the visible light communication method may involve devices capable of transmitting light signals that are so brief that they cannot be visibly detected by humans.
- the light signals may be encoded with unique information (e.g. location information) so that by receiving the light signal and decoding it, the electronic device 100 may obtain location information.
- unique information e.g. location information
- the multilateration method using the RSSI and the RTT may be a method for converting an RSSI of a Wi-Fi signal to a distance using signal propagation modeling and calculating a location of the electronic device 100 by triangulating the electronic device 100 from a plurality of Wi-Fi access points (APs). Distance between a Wi-Fi AP and the electronic device 100 may be determined based on a delay time interval from when a pulse is transmitted from the Wi-Fi AP to when a response pulse is received by the Wi-Fi AP from the electronic device 100.
- the processor 110 may wake-up the electronic device 100 depending on whether an instruction corresponding to a function performable by the electronic device 100 is included in a result of voice recognition of a received voice input. For example, the processor 110 may wake up the electronic device 100 only if the electronic device 100 is capable of performing the requested function in the voice input. In another embodiment, although a plurality of electronic devices recognize a specified word (e.g. the wake word), only one electronic device may respond to the specified word and the other electronic devices may remain in the sleep state. Further, the electronic devices may be able to perform the same function (e.g. play music). If an electronic device most suitable for performing the function is determined, the other electronic devices may not need to wake up.
- a specified word e.g. the wake word
- only one electronic device may respond to the specified word and the other electronic devices may remain in the sleep state.
- the electronic devices may be able to perform the same function (e.g. play music). If an electronic device most suitable for performing the function is determined, the other electronic devices may not need to wake up
- Factors in determining which device is most suitable may include the ability to process the instruction from the user, the distance from the user, the volume of the voice input from the user, etc.
- an apparatus for responding to the specified word and analyzing the instruction may be selected because, among the plurality of electronic devices, it is the device closest to the user.
- a server device connected with the plurality of electronic devices may analyze the instruction.
- the server device may determine an electronic device capable of performing a function corresponding to the instruction among the plurality of electronic devices and may transmit a related signal such that the determined electronic device performs the function.
- the microphone 120 may receive various audio inputs generated outside the electronic device, such as voice inputs from the user.
- the microphone 120 may be one microphone located in the electronic device 100, or a plurality of microphones may be located in the electronic device 100.
- the speaker 130 may output audio.
- the speaker 130 may output a synthetic voice output corresponding to an indication.
- the speaker 130 may output a synthetic voice output corresponding to the result of performing a function.
- the indicator 140 may be exposed on part of the housing and may provide (or display) a visual indication.
- the indicator 140 may display a color.
- the indicator 140 may include, for example, an LED or a display.
- the memory 150 may include a volatile and/or nonvolatile memory.
- the memory 150 may store instructions or data associated with at least one other element of the electronic device 100.
- the memory 150 may store software and/or a program.
- the program may include an application.
- the application may be a set of a series of programs (or instructions) for performing at least one specified function.
- the memory 150 may include mapping information that maps a type of an indication to a state of an application. Further, the memory 150 may store frequency spectrum information corresponding to various characters and frequency spectrum information of the user’s voice so that the electronic device 100 may perform voice recognition and voiceprinting.
- the memory 150 may include, for example, an internal memory or an external memory.
- the communication circuitry 160 may establish communication between the electronic device 100 and an external electronic device.
- the communication circuitry 160 may be connected to, for example, a network through wireless communication or wired communication.
- At least some of the elements of the electronic device 100 may exist outside the housing of the electronic device 100 and may be electrically connected with the processor 110.
- at least part of the microphone 120, the speaker 130, or the indicator 140 may exist outside the housing of the electronic device 100.
- an electronic device may include a housing, a user interface (e.g., the indicator 140) configured to be exposed through the housing and provide a visual indication, a first microphone (e.g., the microphone 120), a speaker (e.g., the speaker 130), a communication circuitry (e.g., the communication circuitry 160), a processor (e.g., the processor 110) configured to be electronically connected with the user interface, the first microphone, the speaker, and the communication circuitry, and a memory (e.g., the memory 150) configured to be electrically connected with the processor.
- a user interface e.g., the indicator 140
- a first microphone e.g., the microphone 120
- a speaker e.g., the speaker 130
- a communication circuitry e.g., the communication circuitry 160
- a processor e.g., the processor 110
- a memory e.g., the memory 150
- the memory may store instructions, when executed, cause the processor to receive a first user voice input for waking up the electronic device through the first microphone, determine whether the electronic device is in a state for performing a selected operation based at least in part on the first user voice input, if the electronic device is in the state for performing the operation, provide a first indication using the user interface, receive a user input for selecting the electronic device to perform the operation, receive a second user voice input associated with the operation through the first microphone, transmit data associated with the operation to an external server through the communication circuitry, receive a response from the external server via the communication circuitry, and provide the response using at least one of the user interface and the speaker, and if the electronic device is not in the state for performing the operation, provide a second indication different from the first indication using the user interface and set the electronic device to a waiting state.
- the instructions may cause the processor to receive information via the communication circuitry from at least one external electronic device, and determine whether the electronic device is in the state for performing the selected operation based at least in part on the information.
- the at least one external electronic device may include a second microphone and obtain at least part of the first user voice input through the second microphone in a state where the at least one external electronic device is close to the electronic device.
- the first indication may include a first color
- the second indication may include a second color
- an electronic device may include a microphone (e.g., the microphone 120) configured to receive a voice input, a communication circuitry (e.g., the communication circuitry 160) configured to communicate with an external electronic device, an indicator (e.g., the indicator 140) configured to provide at least one visual indication, and a processor (e.g., the processor 110) configured to be electrically connected with the microphone, the communication circuitry, and the indicator, and a memory (e.g., the memory 150) configured to be electrically connected with the processor.
- a microphone e.g., the microphone 120
- a communication circuitry e.g., the communication circuitry 160
- an indicator e.g., the indicator 140
- a processor e.g., the processor 110
- a memory e.g., the memory 150
- the memory may store instructions, when executed, cause the processor to receive a first voice input through the microphone, perform a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, display a first visual indication through the indicator, receive a second voice input through the microphone, perform a second voice recognition for the second voice input, and if a second specified word corresponding to the first visual indication is included in a result of the second voice recognition, wake up the electronic device.
- the indicator may include at least one of a light emitting diode (LED) or a display, and the first visual indication may be at least one color displayed by the indicator.
- LED light emitting diode
- the memory may further include mapping information that maps a type of the at least one visual indication to an assignment state of the at least one visual indication, and the instructions may cause the processor to determine the first visual indication based on the mapping information.
- the instructions may cause the processor to transmit information about a second visual indication different from the first visual indication to the external electronic device if a signal corresponding to the first specified word is received from the external electronic device located near the electronic device.
- the signal may include at least one of information regarding when the external electronic device received the first voice input and a first volume level of the first voice input received by the external electronic device.
- the instructions may cause the processor to adjust a threshold of a volume level for the second voice recognition based on the first volume level of the first voice input included in the signal and a second volume level of the first voice input received through the microphone.
- the electronic device may further include a speaker, and the instructions may cause the processor to output a voice output corresponding to the first visual indication through the speaker.
- the electronic device may further include a sensor module, and the instructions may cause the processor to determine at least one of whether a body part of a user is close to the electronic device and whether the body part of the user is in contact with the electronic device, based on sensing information obtained through the sensor module, and determine whether to wake up the microphone based on the determination that the body part of the user is close to the electronic device or the determination that the body part of the user is in contact with the electronic device.
- the memory may include usage history information of a user for the electronic device, and the instructions may cause the processor to perform voiceprint identification for the first voice input to identify the user, and determine whether to wake up the electronic device based on the usage history information of the user.
- the instructions may cause the processor to determine whether to wake up the electronic device based on whether an instruction corresponding to a function performable by the electronic device is included in the result of the second voice recognition.
- FIG. 2 is a drawing illustrating a system that processes voice inputs according to an embodiment.
- a user 200 may call an electronic device (e.g., a first electronic device 210) by speaking a message including a specified word (e.g., “AAA”).
- a specified word e.g., “AAA”
- other external electronic devices e.g., a second electronic device 230 and a third electronic device 250
- the other external electronic devices may also process the voice input from the user, even though the user does not intend to user the other electronic devices.
- a short range e.g. 20 m
- the first electronic device 210 may perform voice recognition of the voice input from the user 200 received through the first microphone 211. Further, if the specified word 203a is included in the result of the voice recognition, in operation 205, the first electronic device 210 may determine whether another external electronic device can be woken up by the specified word 203a. For example, if receiving a signal corresponding to the specified word 203a from the other external electronic device, the first electronic device 210 may determine that the other external electronic device can be woken up by the specified word 203a. In this operation, each of the second electronic device 230 and the third electronic device 250 may receive the voice input of the user 200 through the second microphone 231 and the third microphone 251 and perform voice recognition. If the specified word 203a is included in the result of the voice recognition, each of the second electronic device 230 and the third electronic device 250 may transmit the signal corresponding to the specified word 203a to the electronic device 210. The signal may include identification information of each electronic device.
- the first electronic device 210 may assign different indications to all three electronic devices capable of being woken up by the specified word 203a. For example, the first electronic device 210 may assign indications of a first color, a second color, and a third color to the first electronic device 210, the second electronic device 230, and the third electronic device 250, respectively.
- the first electronic device 210 may display the assigned indication of the first color through the first indicator 213.
- the first electronic device 210 may transmit information about the indication of the second color and information about the indication of the third color to the second electronic device 230 and the third electronic device 250, respectively.
- the second electronic device 230 and the third electronic device 250 may display the indication of the second color and the indication of the third color through the second indicator 233 and the third indicator 253, respectively.
- the electronic devices capable of being woken up by the same specified word 203a may select an indication themselves.
- the first electronic device may select the indication of the first color and may transmit information about the first color to the second electronic device 230.
- the second electronic device 230 may select the indication of the second color different from the first color and may transmit information about the first color and the second color to the third electronic device 250.
- the third electronic device 250 may select the indication of the third color different from the first color and the second color.
- the method or order in which the electronic devices e.g., the first electronic device 210, the second electronic device 230, or the third electronic device 250) transmit their color information to other electronic devices is not limited thereto. Various methods may be used to determine the method or order.
- FIG. 3 is a drawing illustrating a method for providing an indication according to an embodiment.
- an electronic device 310 may receive a first voice input 301 of the user 300 through the microphone 311. If a first specified word 301a is included in the result of voice recognition for the received first voice input 301, the electronic device 310 may display an indication assigned to the electronic device 310 through an indicator 313.
- the indication may include, for example, a color assigned for the electronic device 310.
- the assignment of the indication may be performed based on mapping information that maps the type of indication and the assignment state of the indication. This mapping information may be stored in a database managed by the electronic device 310. In some embodiments, the database may be managed by a separate server device, a hub device, or a master device, and the electronic device 310 may receive information about the indication via a communication circuitry and may display the corresponding indication through the indicator 313.
- the electronic device 310 displays a specified indication through the indicator 313, the user 300 may see the displayed indication (see reference numeral 303). In some embodiments, the electronic device 310 may output the specified indication as voice output through a voice output device (e.g., a speaker) so that the user may hear the indication. If the user 300 speaks a second voice input 305 including a new second specified word 305a corresponding to the indication (i.e. either the displayed indication or the voiced indication), the electronic device 310 may determine that the second specified word 305a is included in the result of voice recognition for the second voice input 305. The electronic device 310 may then wake up at least one of its elements. The second voice input 305 may include the second specified word 305a and an instruction 305b.
- a voice output device e.g., a speaker
- the electronic device 310 may wake up at least one element for performing a function corresponding to the instruction 305b.
- the electronic device 310 may provide audio output for informing the user 300 that the electronic device 310 is woken up and/or audio output 307 for informing the user 300 of the result of the performance of the function.
- the electronic device 310 may also display an indication for informing the user that the electronic device 310 is woken up, through the indicator 313.
- FIG. 4 is a flowchart illustrating an operation method of an electronic device that processes voice inputs according to an embodiment.
- an electronic device may receive a first voice input through a microphone (e.g., a microphone 120 of FIG. 1) and may perform voice recognition for the received first voice input.
- the first voice input may include a specified word (e.g. a wake-up word) for waking up the electronic device.
- the electronic device may determine whether a first specified word is included in the result of the voice recognition. For example, the processor 110 may determine whether the first specified word is included in the text converted from the first voice input.
- the electronic device may display an indication.
- the electronic device e.g., the processor 110
- the electronic device may display an indication it selected for itself irrespective of the mapping information.
- the electronic device e.g., the communication circuitry 160 of FIG. 1
- the electronic device e.g., the processor 110
- the electronic device may maintain the waiting state (i.e. the sleep state).
- the electronic device may receive a second voice input through the microphone (e.g., the microphone 120) and may perform voice recognition for the received second voice input.
- the second voice input may be one spoken by the user after he or she has seen the displayed indication.
- the second voice input may include the second specified word corresponding to the indication.
- the electronic device may determine whether the second specified word is included in the result of the voice recognition. For example, the processor 110 may determine whether the second specified word is included in the text converted from the second voice input.
- the electronic device may wake up at least one element in the electronic device.
- the electronic device e.g., the processor 110
- the electronic device e.g., the processor 110
- the electronic device may display an indication different from the indication displayed in step 430 and/or may maintain the waiting state.
- FIG. 5 is a flowchart illustrating another operation method of an electronic device that processes voice inputs according to an embodiment.
- an electronic device may receive a first voice input and may perform voice recognition for the first voice input.
- the first voice may include a specified word (or a wake-up word) for waking up the electronic device. Further, the specified word may be used as the wake-up word of another external electronic device as well as the electronic device.
- the electronic device may determine whether the first specified word is included in the result of the voice recognition.
- the processor 110 may determine whether the first specified word is included in text converted from the first voice input.
- the electronic device may determine whether a signal corresponding to the first specified word is received from the external electronic device. For example, the external electronic device capable of being woken up by the first specified word may transmit the signal corresponding to the first specified word. In this case, the electronic device may receive the transmitted signal via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1) and may verify that there is another external electronic device capable of being woken up by the first specified word.
- a communication circuitry e.g., a communication circuitry 160 of FIG.
- the electronic device may display an indication distinguishing the electronic device from the external electronic device through an indicator.
- the processor 110 may specify an indication to distinguish itself from the external electronic device based on mapping information mapping the types of indications to the assignment states of the indications.
- the electronic device may then display the specified indication through an indicator (e.g., an indicator 140 of FIG. 1).
- the electronic device may receive a second voice input through the microphone (e.g., the microphone 120 of FIG. 1) and may perform voice recognition for the second voice input.
- the second voice input may be spoken by the user has seen the displayed indication, and the second voice input may include a second specified word corresponding to the indication. For example, if a red indication is displayed through the indicator, the second voice input may be “Hi, red” where “red” corresponds to the indication and is the second specified word.
- the user may speak an instruction including the second specified word, for example, “Red, how is the weather today?”
- the electronic device may determine whether the second specified word is included in the result of the voice recognition. For example, the electronic device may determine whether the second specified word is included in the text converted from the second voice input.
- the electronic device may wake up at least one element of the electronic device.
- the electronic device may wake itself up to perform a function corresponding to the instruction.
- the electronic device may return to operation 550. For example, the electronic device may repeatedly perform operations 550 and 560 until the second voice input including the second specified word is received. In some embodiments, if the second voice input including in the second specified word is not received for a specified time period, the electronic device may skip operations 550 to 570 and may change to a waiting state or an inactive (sleep) state.
- FIG. 6 is a drawing illustrating a method for determining an indication according to an embodiment.
- a first electronic device 610 may receive a voice input through a first microphone 611. If a specified word is included in a result of voice recognition for the received voice input, the first electronic device 610 may provide (or display) a first indication through the first indicator 613. According to an embodiment, the first electronic device 610 may manage mapping information mapping the types of indications to the assignment states of the indications in a database 615 stored in memory (e.g., the memory 150 of FIG. 1). Thus, the first electronic device 610 may determine the first indication based on the mapping information 617.
- the mapping information 617 may include, for example, information associated with an assigning order 617a of electronic devices and types 617b of indications.
- the types 617b of the indications may be color, unique nicknames of the electronic devices, etc. Further, each electronic device (e.g., the first electronic device 610, the second electronic device 630, or a third electronic device 650) may display its corresponding indication through an indicator (e.g., the first indicator 613, a second indicator 633, or a third indicator 653) or may output voice information corresponding to the indication through its speaker. In some embodiments, the types 617b of the indications may depend on how each electronic device (e.g., the first electronic device 610, the second electronic device 630, or the third electronic device 650) outputs the indications.
- the indication may be set to a specific color. But if the electronic device outputs the indication on a display, the indication may be set to a display object that includes a specific color or a nickname of the device. And if the electronic device outputs an indication through a speaker, the indication may be set to voice information corresponding to a specific color or a nickname.
- the first electronic device 610 may determine indications of other external electronic devices and may transmit the determined indications to the external electronic devices.
- the first electronic device 610 may function as a master device.
- the second electronic device 630 and the third electronic device 650 may receive voice inputs from the user through a second microphone 631 and a third microphone 651, respectively. If a specified word is included in the result of the voice recognition for the received voice input, the second electronic device 630 and the third electronic device 650 may transmit a signal corresponding to the specified word to the electronic device 610.
- the first electronic device 610 may receive the signal transmitted from the second electronic device 630 and the third electronic device 650 and may sequentially assign the second indication and the third indication to the second electronic device 630 and the third electronic device 650, respectively, based on an order where the signal is received. Further, the first electronic device 610 may transmit information about the second indication and information about the third indication to the second electronic device 630 and the third electronic device 650, respectively, via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1). Receiving the information about the second indication and the information about the third indication, the second electronic device 630 and the third electronic device 650 may provide (or display) the second indication and the third indication through the second indicator 633 and the third indicator 653, respectively.
- a communication circuitry e.g., a communication circuitry 160 of FIG.
- the types 617b of the indications assigned to the electronic devices may not duplicate each other. Further, if one electronic device is selected (or woken up) by a user, the type 617b of the assigned indication may be reset.
- the indications are spoken but there is some overlap in the indications, then these indications may not be used. For example, spoken indications “blue” and “sky blue” overlap with the word “blue” and may cause confusion. Thus, they may not be used as indications.
- the indications are displayed colors, to avoid confusion, only colors that highly contrast with each other may be used. For example, the various electronic devices may be use blue and sky blue as two different indications.
- the types 617b of the indications assigned to the electronic devices may be differently specified according to performance of the electronic devices, a temporal order of the received voice input, volume levels of the received voice inputs, etc.
- the indication may be assigned sequentially from an electronic device with the best performance.
- the indication may be assigned sequentially from an electronic device having the best response time to a voice input.
- the indication may be assigned sequentially from an electronic device with the highest volume of the received voice input.
- the indication may be assigned from an electronic device which is always powered on.
- the types 617b of the indications assigned to the electronic devices may be differently specified according to settings specified by the user.
- FIG. 7 is a drawing illustrating another method for determining an indication according to an embodiment.
- each of the plurality of electronic devices may function as a slave or client device.
- a server device 710 or a hub device which determines the types of indications for the plurality of electronic devices and transmits the information about the determined indications to the plurality of electronic devices.
- the server device 710 may manage, for example, mapping information 713 that maps the types of indications to the assignment states of the indications in a database 711.
- the mapping information 713 may include, for example, information associated with an assigning order 713a of the electronic devices and types 713b of indications.
- the first electronic device 730, the second electronic device 750, and the third electronic device 770 may receive voice inputs through a first microphone 731, a second microphone 751, and a third microphone 771, respectively. If the specified word is included in the result of the voice recognition for the received voice inputs, the first electronic device 730, the second electronic device 750, and the third electronic device 770 may transmit signals corresponding to the specified word to the server device 710.
- the server device 710 may receive the signals transmitted from at least one of the electronic devices (e.g., the first electronic device 730, the second electronic device 750, and the third electronic device 770) and may transmit information based on the mapping information 713 to the electronic devices.
- the server device 710 may receive signals transmitted from the first electronic device 730, the second electronic device 750, and the third electronic device 770 and may assign a first indication, a second indication, and a third indication to the first electronic device 730, the second electronic device 750, and the third electronic device 770 sequentially based on the mapping information 713 in an order in which the signals are received.
- the server device 710 may receive signals transmitted form the first electronic device 730, the second electronic device 750, and the third electronic device 770 and may assign the first indication, the second indication, and the third indication to the first electronic device 730, the second electronic device 750, and the third electronic device 770, respectively, based on assignment history.
- the server device 710 may transmit information about the first indication, information about the second indication, and information about the third indication to the first electronic device 730, the second electronic device 750, and the third electronic device 770, respectively. Respectively receiving the information about the first indication, the information about the second indication, and the information about the third indication, the first electronic device 730, the second electronic device 750, and the third electronic device 770 may provide (or display) the first indication, the second indication, and the third indication through the first indicator 733, the second indication 753, and the third indication 773, respectively.
- FIG. 8 is a flowchart illustrating an operation method of an electronic device for determining an indication according to an embodiment.
- an electronic device may receive a signal corresponding to a specified word via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1) from at least one external electronic device.
- the electronic device may be a server device, a hub device, or a master device.
- the signal corresponding to the specified word may include information relating to the fact that the external electronic device recognizes the specified word and may wake up based on the recognition.
- the signal corresponding to the specified word may include an indication for when the external electronic device received the voice input containing the specified word, received strength of the voice input, etc.
- the electronic device may determine an indication for the external electronic device.
- the electronic device may manage mapping information mapping the types of indications to the assignment states of the indications in a database stored in memory (e.g., a memory 150 of FIG. 1).
- the electronic device may determine (i.e. assign) the indication for the external electronic device based on the mapping information stored in the database.
- the electronic device may transmit information about the determined indication to the external electronic device via the communication circuitry (e.g., the communication circuitry 160). Receiving the information about the indication, the external electronic device may provide (or display) the indication through its indicator.
- FIG. 9 is a flowchart illustrating a method for displaying an indication determined through an external electronic device at an electronic device according to an embodiment.
- an electronic device e.g., an electronic device 100 or a processor 110 of FIG. 1
- a microphone e.g., the microphone 120 of FIG. 1
- the processor 110 may analyze voice data corresponding to the voice input received through the microphone and may convert the voice data into text.
- the electronic device may determine whether a specified word is included in the result of the voice recognition.
- the processor 110 may determine whether the specified word is included in the converted text.
- the electronic device may transmit a signal corresponding to the specified word to an external electronic device (e.g., a server device, a hub device, or a master device).
- the electronic device may send out the signal corresponding to the specified word via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1).
- the electronic device may transmit the signal and may change to a state of waiting for wake-up. For example, the electronic device may change to a state of waiting for a response to the signal.
- the electronic device may receive information about an indication via the communication circuitry (e.g., the communication circuitry 160) from the external electronic device. If receiving the indication, in operation 990, the electronic device (e.g., the electronic device 100 or the processor 110) may provide (or display) the indication through an indicator (e.g., the indicator 140 of FIG. 1).
- the electronic device e.g., the electronic device 100 or the processor 110 may provide (or display) the indication through an indicator (e.g., the indicator 140 of FIG. 1).
- FIG. 10 is a drawing illustrating a method for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
- electronic devices may determine whether to wake up microphones (e.g., a first microphone 1011, a second microphone 1031, and a third microphone 1051) included in the electronic devices depending on whether a body part of the user 1000 (e.g., a finger) is close to (or in contact with) the electronic devices.
- the electronic devices may determine whether the body part 1001 is close to (or in contact with) the electronic devices based on sensing information obtained by a sensor module (e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like).
- That electronic device may wake up its microphones to receive a voice input from the user 1000. Further, that electronic device may transmit a notification to another external electronic device indicating that it has detected the body part 1001. In this case, after receiving the notification, the external electronic device may deactivate its own microphone so that it does not receive the voice input of the user 1000.
- the body part 1001 of the user 1000 is close to (or in contact with) the first electronic device 1010.
- the first electronic device 1010 may wake up the first microphone 1011 and may transmit a notification to the second electronic device 1030 and the third electronic device 1050, where the notification indicates that the body part 1001 is close to (in contact with) the first electronic device 1010.
- the second electronic device 1030 and the third electronic device 1050 may deactivate the second microphone 1031 and the third microphone 1051, respectively.
- the microphone may be deactivated for a specified time period. After the specified time period elapses, the microphone may be reactivated. For example, after the specified time period elapses, the first electronic device 1010 may deactivate the first microphone 1011, and the second electronic device 1030 and the third electronic device 1050 may wake up the second microphone 1031 and the third microphone 1051, respectively. In another example, the first electronic device 1010 may maintain the wake-up state of the microphone 1011, and the second electronic device 1030 and the third electronic device 1050 may wake up the second microphone 1031 and the third microphone 1051, respectively.
- the electronic device which receives an additional selection e.g., a finger touch, proximity, or the like
- the electronic device which receives an additional selection may maintain a wake-up state (e.g., a state of waiting for receiving an additional voice) and the electronic device which does not receive a selection of the user 1000 may change to an inactive state.
- FIG. 11 is a signal sequence diagram illustrating a system for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
- a user 1000 may select (e.g., be close to or come into contact with) one of a plurality of electronic devices (e.g., a first electronic device 1110, a second electronic device 1130, and a third electronic device 1150) and may speak a specified word (e.g., a wake-up word).
- a specified word e.g., a wake-up word
- the order of the user operations may be reversed. In other words, the user may speak the specified word (e.g., the wake-up word) and then may select one of the plurality of electronic devices.
- the user 1100 may be close to (or come into contact with) one electronic device (e.g., the first electronic device 1110).
- the user 1100 may touch the one electronic device using a part of his or her body (e.g., a finger).
- the first electronic device 1110 may recognize the proximity (or contact) of the user 1100. According to an embodiment, the first electronic device 1110 may receive the proximity (or contact) of the user 1100 based on sensing information obtained by a sensor module (e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like).
- a sensor module e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like.
- the first electronic device 1110 may inform external electronic devices (e.g., the second electronic device 1130 and the third electronic device 1150) of the proximity (or contact) state of the user 1100. Further, in operation 1115, the first electronic device 1110 may wake up a microphone (e.g., a microphone 120 of FIG. 1) included in the first electronic device 1110. If the microphone included in the first electronic device 1110 is already woken up, the first electronic device 1110 may maintain the wake-up state of the microphone. In operations 1131 and 1151, the second electronic device 1130 and the third electronic device 1150 may receive the proximity (or contact) state of the user 1100.
- a microphone e.g., a microphone 120 of FIG.
- the second electronic device 1130 and the third electronic device 1150 may deactivate microphones included in the second electronic device 1130 and the third electronic device 1150, respectively, so that the second electronic device 1130 and the third electronic device 1150 are prevented from detecting the voice input from the user.
- the user 1100 may utter (or speak) a voice input, which may include an instruction corresponding to a function to be performed by the first electronic device 1110.
- the first electronic device 1110 having an awake microphone, may receive the voice input through the microphone and may perform voice recognition for the received voice input.
- the second electronic device 1130 and the third electronic device 1150 may not detect the voice input because their microphones are deactivated.
- the microphones of the second electronic device 1130 and the third electronic device 1150 may be deactivated for a specified time period or until they receive a notification that the proximity state of the user 1100 has changed (e.g., the user is no longer close to or in contact with the first electronic device 1110).
- the first electronic device 1110 may wake up at least some of its modules related to the function specified by the user to perform that function.
- FIG. 12 is a drawing illustrating a method for determining whether to wake up an electronic device through adjustment of received sensitivity of a voice input according to an embodiment.
- an electronic device may adjust receive sensitivity of a voice input received through a microphone (e.g., a first microphone 1211 or a second microphone 1231).
- the electronic device may adjust a threshold 1205 of the received intensity of the voice input.
- the threshold 1205 may indicate, for example, a minimum value for preventing errors in voice recognition.
- each electronic device may have a different threshold.
- the electronic devices may have different thresholds depending on the performances of the electronic devices.
- the threshold 1205 of the received intensity may be adjusted to a value between the first level 1201 and the second level 1203. For example, if the first level 1201 is higher than the second level 1203, the threshold 1205 may be adjusted to be lower than the first level 1201 and be higher than the second level 1203.
- voice inputs may be processed by the first electronic device 1210 but not the second electronic device 1230. As a result, only the first electronic device 1210 may be woken up in response to the voice input from the user.
- the electronic device may change the type of indication shown in the indicator (e.g., the first indicator 1213 or the second indicator 1233) according to the received intensity.
- the first electronic device 1210 in which the received intensity is the first level 1201 may display an indication of a first color
- the second electronic device 1230 in which the received intensity is the second level may display an indication of a second color different from the first color.
- a user may verify each indication (e.g., a first indication or a second indication) so that he or she knows which electronic device is capable of performing voice recognition.
- the threshold 1205 may be set to correspond to the received intensity of a voice input including a first specified word.
- the electronic device may not process the voice input.
- FIG. 13 is a flowchart illustrating an operation method of an electronic device for adjusting the received sensitivity of a voice input according to an embodiment.
- an electronic device may receive a voice input of a user through a microphone (e.g., a microphone 120 of FIG. 1) and may analyze the received intensity of the voice input. Further, in operation 1330, the electronic device (e.g., the electronic device 100 or the processor 110) may receive information regarding the received intensity of a voice input from the external electronic device via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1).
- a communication circuitry e.g., a communication circuitry 160 of FIG.
- the electronic device may adjust the received sensitivity of the voice input based on the received voice input and the information received from the external electronic device.
- the electronic device may adjust the threshold of received intensity of a voice input. For example, if the received intensity of a voice input received by the electronic device is a first level and if received intensity of a voice received by the external electronic device is a second level, the electronic device may adjust the threshold to a value between the first level and the second level.
- the electronic device e.g., the electronic device 100 or the processor 110
- FIG. 14 is a drawing illustrating a method for determining whether to wake up an electronic device based on usage history information according to an embodiment.
- an electronic device may determine whether to wake up the electronic device based on usage history information of user A 1400.
- the electronic device e.g., the first electronic device 1410 or the second electronic device 1430
- may perform voiceprint identification for a voice input received through a microphone e.g., a first microphone 1411 or a second microphone 1431 to identify user A 1400.
- the electronic device may verify history where the identified user A 1400 uses the electronic device based on usage history information (e.g., first usage history information 1415 or second usage history information 1435) stored in a database (e.g., a first database 1413 or a second database 1433) and may determine whether to wake up the electronic device based on the usage history information.
- usage history information may include, for example, information associated with the frequency of use by user A 1400, the number of use by user A 1400, the difference in the frequency of user/the number of use with another user, etc.
- the usage history information may include information about the number of times where user A 1400 selects that specific electronic device when two or more electronic devices (e.g., the first electronic device 1410 or the second electronic device 1430) use the specified word (e.g., a wake-up word) to wake up.
- the specified word e.g., a wake-up word
- the first electronic device 1410 and the second electronic device 1430 may receive the voice input with the same voice intensity at substantially and almost the same time.
- the first electronic device 1410 and the second electronic device 1430 may verify usage history information of user A 1400 to determine whether user A 1400 is a user who frequently use the first electronic device 1410 or the second electronic device 1430. If determining that user A 1400 is the user who frequently use the first electronic device 1410, the first electronic device 1410 may wake up.
- FIG. 15 is a flowchart illustrating an operation method of an electronic device for determining whether to wake up an electronic device based on usage history information according to an embodiment.
- an electronic device may receive a voice input of a user through a microphone (e.g., the microphone 120 of FIG. 1) and may perform voice identification for the received voice.
- the electronic device e.g., the electronic device 100 or the processor 110
- the processor 110 may convert vibration energy of the voice input an electric signal, may convert the electric signal into frequency spectrum information.
- the electronic device may then compare the converted frequency spectrum information with a plurality of frequency spectrum information corresponding to different voiceprints of different users, thus identifying the user.
- the electronic device may verify usage history information of the user.
- the usage history information of the user may be managed through a database (e.g., the database 615 of FIG. 6).
- the usage history information of the user may include, for example, information associated with the frequency of use by the user, the number of use by the user, the difference in the frequency of use/the number of use with another user, etc.
- the usage history information may also include information on the number of times the user has selected a particular device.
- the electronic device may determine whether to wake up the electronic device based on the usage history information. For example, if the number of times that the user has used the electronic device is larger than the number of times that another user has used the electronic device, or if the user more recently uses the electronic device than another user, the electronic device may be woken up.
- FIG. 16 is a drawing illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
- an electronic device may receive a voice input of a user 1600 through its microphone and may perform voice recognition for the received voice. Further, the electronic device may determine wake-up of the electronic device based on an instruction included in the voice input that corresponds to a function performable by the electronic device. For example, the electronic device may be woken up only if it is possible for the electronic device to perform the included function.
- a plurality of electronic devices e.g., the first electronic device 1610 and the second electronic device 1630
- the first electronic device 1610 may respond to the specified word 1601a and the other electronic device (e.g., the second electronic device 1630) may not wake up.
- the first electronic device 1610 determines that it may perform the specified operation 1603a but the second electronic device 1630 determines that it cannot perform the operation 1603a. If there are a plurality of electronic devices capable of performing the same function, the electronic device that is closer to the user may be designated to perform the function.
- the first electronic device 1610 and the second electronic device 1630 may be connected to hub devices, respectively.
- the first electronic device 1610 may be connected to a first hub device
- the second electronic device 1630 may be connected to a second hub device.
- the first and second electronic devices 1610 and 1630 may be controlled by the first and second hub devices, respectively.
- the first and second hub devices may receive a voice input spoken from the user 1600 and may analyze the received voice input. The first and second hub devices then control the first and second electronic devices 1610 and 1630 based on the analyzed result.
- FIG. 17 is a flowchart illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
- an electronic device e.g., an electronic device 100 or a processor 110 of FIG. 1 may receive a voice input of a user and may perform voice recognition for the received voice input.
- the electronic device may determine whether an instruction corresponding to a function performable by the electronic device is included in the result of performing the voice recognition.
- the electronic device e.g., the electronic device 100 or the processor
- the electronic device may wake up at least one element necessary for performing the function. If the instruction corresponds to a function that cannot be performed by the electronic device, the electronic device may not wake up.
- an electronic device for performing the function may be selected based on the location of the user, the distance between the user and each electronic device, or the like.
- a voice input processing method of an electronic device may include receiving a first voice input through a microphone, performing a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, displaying an indication through the indicator, receiving a second voice input through the microphone, performing a second voice recognition for the second voice input, and if a second specified word corresponding to the indication is included in a result of the second voice recognition, waking up the electronic device.
- the method may further include verifying mapping information that maps a type of at least one indication to an assignment state of the at least one indication which are stored in a memory of the electronic device, and determining the indication based on the mapping information.
- the method may further include transmitting information about an another indication different from the indication to the external electronic device if a signal corresponding to the first specified word is received from an external electronic device located near the electronic device.
- the method may further include adjusting a threshold of a volume level for the second voice recognition based on a first volume level of the first voice input received in the external electronic device included in the signal and a second volume level of the first voice input received through the microphone.
- the method may further include determining at least one of whether a body part of a user is close to the electronic device and whether the body part of the user is in contact with the electronic device, based on sensing information obtained through a sensor module of the electronic device, and determining whether to wake up the microphone based on the determination that the body part of the user is close to the electronic device or the determination that the body part of the user is in contact with the electronic device.
- the method may further include performing voiceprint identification for the first voice input to identify a user, and determining whether to wake up the electronic device based on usage history information of the user, the usage history information being stored in a memory of the electronic device.
- the method may further include determining whether to wake up the electronic device based on whether an instruction corresponding to a function performable by the electronic device is included in the result of the second voice recognition.
- FIG. 18 is a block diagram illustrating an electronic device 1801 in a network environment 1800 according to an embodiment.
- the electronic device 1801 shown in FIG. 18 may be the same or similar structure to an electronic device 100 shown in FIG. 1.
- an electronic device 1801 may include a bus 1810, a processor 1820, a memory 1830, an input/output interface 1850, a display 1860, and a communication interface 1870. According to an embodiment, the electronic device 1801 may not include at least one of the elements or may further include other element(s).
- the bus 1810 may interconnect the elements 1820 to 1870 and may include a circuit for conveying communications (e.g., a control message or data) among the elements.
- the processor 1820 may include one or more of a central processing unit (CPU), an application processor (AP), or a communication processor (CP).
- the processor 1820 may perform an arithmetic operation or data processing associated with control and/or communication of at least one other elements of the electronic device 1801.
- the memory 1830 may include a volatile and/or nonvolatile memory.
- the memory 1830 may store instructions or data associated with at least one other element(s) of the electronic device 1801.
- the memory 1830 may store software and/or a program 1840.
- the program 1840 may include, for example, a kernel 1841, a middleware 1843, an application programming interface (API) 1845, and/or an application program (or “an application”) 1847. At least a part of the kernel 1841, the middleware 1843, or the API 1845 may be referred to as an “operating system (OS)”.
- OS operating system
- the kernel 1841 may control or manage system resources (e.g., the bus 1810, the processor 1820, the memory 1830, and the like) that are used to execute operations or functions implemented in other programs (e.g., the middleware 1843, the API 1845, and the application program 1847). Furthermore, the kernel 1841 may provide an interface that allows the middleware 1843, the API 1845, or the application program 1847 to access discrete elements of the electronic device 1801 so as to control or manage system resources.
- system resources e.g., the bus 1810, the processor 1820, the memory 1830, and the like
- other programs e.g., the middleware 1843, the API 1845, and the application program 1847.
- the kernel 1841 may provide an interface that allows the middleware 1843, the API 1845, or the application program 1847 to access discrete elements of the electronic device 1801 so as to control or manage system resources.
- the middleware 1843 may perform, for example, a mediation role such that the API 1845 or the application program 1847 communicates with the kernel 1841 to exchange data. Furthermore, the middleware 1843 may process one or more task requests received from the application program 1847 according to a priority. For example, the middleware 1843 may assign the priority, which makes it possible to use a system resource (e.g., the bus 1810, the processor 1820, the memory 1830, or the like) of the electronic device 1801, to at least one of the application program 1847 and may process the one or more task requests.
- a system resource e.g., the bus 1810, the processor 1820, the memory 1830, or the like
- the API 1845 may be, for example, an interface through which the application program 1847 controls a function provided by the kernel 1841 or the middleware 1843, and may include, for example, at least one interface or function (e.g., an instruction) for a file control, a window control, image processing, a character control, or the like.
- interface or function e.g., an instruction
- the input/output interface 1850 may transmit, for example, an instruction or data, input from a user or another external device, to other element(s) of the electronic device 1801 or may output an instruction or data, received from other element(s) of the electronic device 1801, to a user or another external device.
- the display 1860 may include, for example, a liquid crystal display (LCD), a light-emitting diode (LED) display, an organic LED (OLED) display, a microelectromechanical systems (MEMS) display, or an electronic paper display.
- the display 1860 may display, for example, various contents (e.g., a text, an image, a video, an icon, a symbol, and the like) to a user.
- the display 1860 may include a touch screen and may receive, for example, a touch, gesture, proximity, or hovering input using an electronic pen or a part of a user’s body.
- the communication interface 1870 may establish communication between the electronic device 1801 and an external device (e.g., the first electronic device 1802, the second electronic device 1804, or the server 1806).
- the communication interface 1870 may be connected to the network 1862 over wireless communication or wired communication to communicate with the external device (e.g., the second electronic device 1804 or the server 1806).
- the wireless communication may include, for example, cellular communication which uses at least one of long-term evolution (LTE), LTE Advanced (LTE-A), Code Division Multiple Access (CDMA), Wideband CDMA (WCDMA), Universal Mobile Telecommunications System (UMTS), Wireless Broadband (WiBro), Global System for Mobile Communications (GSM), or the like.
- the wireless communication may be, for example, an element 1864 of FIG. 18.
- the wireless communication may include at least one of wireless fidelity (Wi-Fi), light fidelity (Li-Fi), Bluetooth, Bluetooth low energy (BLE), ZigBee, near field communication (NFC), magnetic secure transmission (MST), radio frequency (RF), or a body area network (BAN).
- the wireless communication may include a global navigation satellite system (GNSS).
- GNSS global navigation satellite system
- the GNSS may be, for example, a global positioning system (GPS), a global navigation satellite system (Glonass), a Beidou navigation satellite system (hereinafter referred to as “Beidou”), or an European global satellite-based navigation system (hereinafter referred to as “Galileo”).
- GPS global positioning system
- GNSS Beidou navigation satellite system
- Galileo European global satellite-based navigation system
- the wired communication may include at least one of, for example, a universal serial bus (USB), a high definition multimedia interface (HDMI), a recommended standard-232 (RS-232), a plain old telephone service (POTS), or the like.
- the network 1862 may include at least one of telecommunications networks, for example, a computer network (e.g., LAN or WAN), an Internet, or a telephone network.
- Each of the first and second electronic devices 1802 and 1804 may be a device of which the type is different from or the same as that of the electronic device 1801. According to various embodiments, all or a portion of operations that the electronic device 1801 will perform may be executed by another or plural electronic devices (e.g., the first electronic device 1802, the second electronic device 1804 or the server 1806). According to an embodiment, in the case where the electronic device 1801 executes any function or service automatically or in response to a request, the electronic device 1801 may not perform the function or the service internally, but, alternatively or additionally, it may request at least a portion of a function associated with the electronic device 1801 at other electronic device (e.g., the electronic device 1802 or 1804 or the server 1806).
- the other electronic device may execute the requested function or additional function and may transmit the execution result to the electronic device 1801.
- the electronic device 1801 may provide the requested function or service using the received result or may additionally process the received result to provide the requested function or service.
- cloud computing distributed computing, or client-server computing may be used.
- FIG. 19 is a block diagram illustrating an electronic device 1901 according to one embodiment.
- an electronic device 1901 may include, for example, all or a part of the electronic device 100 illustrated in FIG. 1 or the electronic device 1801 illustrated in FIG. 18.
- the electronic device 1901 may include one or more processors (e.g., an application processor (AP)) 1910, a communication module 1920, a subscriber identification module 1924, a memory 1930, a sensor module 1940, an input device 1950, a display 1960, an interface 1970, an audio module 1980, a camera module 1991, a power management module 1995, a battery 1096, an indicator 1997, and a motor 1998.
- processors e.g., an application processor (AP) 1910, a communication module 1920, a subscriber identification module 1924, a memory 1930, a sensor module 1940, an input device 1950, a display 1960, an interface 1970, an audio module 1980, a camera module 1991, a power management module 1995, a battery 1096, an indicator 1997, and a motor 1998.
- AP application processor
- the processor 1910 may drive, for example, an operating system (OS) or an application to control a plurality of hardware or software elements connected to the processor 1910 and may process and compute a variety of data.
- the processor 1910 may be implemented with a System on Chip (SoC).
- SoC System on Chip
- the processor 1910 may further include a graphic processing unit (GPU) and/or an image signal processor.
- the processor 1910 may include at least a part (e.g., a cellular module 1921) of the elements illustrated in FIG. 19.
- the processor 1910 may load an instruction or data, which is received from at least one of other elements (e.g., a nonvolatile memory), into a volatile memory and process the loaded instruction or data.
- the processor 1910 may store result data in the nonvolatile memory.
- the communication module 1920 may be configured the same as or similar to the communication interface 1870 of FIG. 18.
- the communication module 1920 may include the cellular module 1921, a Wi-Fi module 1923, a Bluetooth (BT) module 1925, a GNSS module 1927, a near field communication (NFC) module 1928, and a radio frequency (RF) module 1929.
- BT Bluetooth
- NFC near field communication
- RF radio frequency
- the cellular module 1921 may provide, for example, voice communication, video communication, a character service, an Internet service, or the like over a communication network. According to an embodiment, the cellular module 1921 may perform discrimination and authentication of the electronic device 1901 within a communication network by using the subscriber identification module (e.g., a SIM card) 1924. According to an embodiment, the cellular module 1921 may perform at least a portion of functions that the processor 1910 provides. According to an embodiment, the cellular module 1921 may include a communication processor (CP).
- CP communication processor
- At least some (e.g., two or more) of the cellular module 1921, the Wi-Fi module 1923, the BT module 1925, the GNSS module 1927, or the NFC module 1928 may be included within one Integrated Circuit (IC) or an IC package.
- IC Integrated Circuit
- the RF module 1929 may transmit and receive a communication signal (e.g., an RF signal).
- the RF module 1929 may include a transceiver, a power amplifier module (PAM), a frequency filter, a low noise amplifier (LNA), an antenna, or the like.
- PAM power amplifier module
- LNA low noise amplifier
- at least one of the cellular module 1921, the Wi-Fi module 1923, the BT module 1925, the GNSS module 1927, or the NFC module 1928 may transmit and receive an RF signal through a separate RF module.
- the subscriber identification module 1924 may include, for example, a card and/or embedded SIM that includes a subscriber identification module and may include unique identify information (e.g., integrated circuit card identifier (ICCID)) or subscriber information (e.g., international mobile subscriber identity (IMSI)).
- ICCID integrated circuit card identifier
- IMSI international mobile subscriber identity
- the memory 1930 may include an internal memory 1932 or an external memory 1934.
- the internal memory 1932 may include at least one of, for example, a volatile memory (e.g., a dynamic random access memory (DRAM), a static RAM (SRAM), a synchronous DRAM (SDRAM), or the like), a nonvolatile memory (e.g., a one-time programmable read only memory (OTPROM), a programmable ROM (PROM), an erasable and programmable ROM (EPROM), an electrically erasable and programmable ROM (EEPROM), a mask ROM, a flash ROM, a flash memory, a hard drive, or a solid state drive (SSD)).
- a volatile memory e.g., a dynamic random access memory (DRAM), a static RAM (SRAM), a synchronous DRAM (SDRAM), or the like
- a nonvolatile memory e.g., a one-time programmable read only memory (OTPROM),
- the external memory 1934 may include a flash drive such as compact flash (CF), secure digital (SD), micro secure digital (Micro-SD), mini secure digital (Mini-SD), extreme digital (xD), a multimedia card (MMC), a memory stick, or the like.
- the external memory 1934 may be operatively and/or physically connected to the electronic device 1901 through various interfaces.
- the sensor module 1940 may measure, for example, a physical quantity or may detect an operation state of the electronic device 1901.
- the sensor module 1940 may convert the measured or detected information to an electric signal.
- the sensor module 1940 may include at least one of a gesture sensor 1940A, a gyro sensor 1940B, a barometric pressure sensor 1940C, a magnetic sensor 1940D, an acceleration sensor 1940E, a grip sensor 1940F, the proximity sensor 1940G, a color sensor 1940H (e.g., red, green, blue (RGB) sensor), a biometric sensor 1940I, a temperature/humidity sensor 1940J, an illuminance sensor 1940K, or an UV sensor 1940M.
- a gesture sensor 1940A e.g., a gyro sensor 1940B
- a barometric pressure sensor 1940C e.g., a MEMS acceleration sensor 1940E
- a grip sensor 1940F e.g., the proximity sensor 1940G
- a color sensor 1940H e.g., red, green
- the sensor module 1940 may further include, for example, an E-nose sensor, an electromyography (EMG) sensor, an electroencephalogram (EEG) sensor, an electrocardiogram (ECG) sensor, an infrared (IR) sensor, an iris sensor, and/or a fingerprint sensor.
- the sensor module 1940 may further include a control circuit for controlling at least one or more sensors included therein.
- the electronic device 1901 may further include a processor that is a part of the processor 1910 or independent of the processor 1910 and is configured to control the sensor module 1940. The processor may control the sensor module 1940 while the processor 1910 remains at a sleep state.
- the input device 1950 may include, for example, a touch panel 1952, a (digital) pen sensor 1954, a key 1956, or an ultrasonic input unit 1958.
- the touch panel 1952 may use at least one of capacitive, resistive, infrared and ultrasonic detecting methods.
- the touch panel 1952 may further include a control circuit.
- the touch panel 1952 may further include a tactile layer to provide a tactile reaction to a user.
- the (digital) pen sensor 1954 may be, for example, a part of a touch panel or may include an additional sheet for recognition.
- the key 1956 may include, for example, a physical button, an optical key, or a keypad.
- the ultrasonic input device 1958 may detect (or sense) an ultrasonic signal, which is generated from an input device, through a microphone (e.g., a microphone 1988) and may check data corresponding to the detected ultrasonic signal.
- the display 1960 may include a panel 1962, a hologram device 1964, a projector 1966, and/or a control circuit for controlling the panel 1962, the hologram device 1964, or the projector 1966.
- the panel 1962 may be implemented, for example, to be flexible, transparent or wearable.
- the panel 1962 and the touch panel 1952 may be integrated into a single module.
- the panel 1962 may include a pressure sensor (or force sensor) that measures the intensity of touch pressure by a user.
- the pressure sensor may be implemented integrally with the touch panel 1952, or may be implemented as at least one sensor separately from the touch panel 1952.
- the hologram device 1964 may display a stereoscopic image in a space using a light interference phenomenon.
- the projector 1966 may project light onto a screen so as to display an image.
- the screen may be arranged in the inside or the outside of the electronic device 1901.
- the interface 1970 may include, for example, a high-definition multimedia interface (HDMI) 1972, a universal serial bus (USB) 1974, an optical interface 1976, or a D-subminiature (D-sub) 1978.
- the interface 1970 may be included, for example, in a communication interface 1870 illustrated in FIG. 18. Additionally or generally, the interface 1970 may include, for example, a mobile high definition link (MHL) interface, a SD card/multi-media card (MMC) interface, or an infrared data association (IrDA) standard interface.
- MHL mobile high definition link
- MMC SD card/multi-media card
- IrDA infrared data association
- the audio module 1980 may convert, for example, a sound and an electric signal in dual directions. At least a part of the audio module 1980 may be included, for example, in an input/output interface 1850 illustrated in FIG. 18. The audio module 1980 may process, for example, sound information that is input or output through a speaker 1982, a receiver 1984, an earphone 1986, or the microphone 1988.
- the camera module 1991 may capture a still image or a video.
- the camera module 1991 may include at least one or more image sensors (e.g., a front sensor or a rear sensor), a lens, an image signal processor (ISP), or a flash (e.g., an LED or a xenon lamp).
- image sensors e.g., a front sensor or a rear sensor
- ISP image signal processor
- flash e.g., an LED or a xenon lamp
- the power management module 1995 may manage, for example, power of the electronic device 1901.
- a power management integrated circuit (PMIC), a charger IC, or a battery or fuel gauge may be included in the power management module 1995.
- the PMIC may have a wired charging method and/or a wireless charging method.
- the wireless charging method may include, for example, a magnetic resonance method, a magnetic induction method or an electromagnetic method and may further include an additional circuit, for example, a coil loop, a resonant circuit, a rectifier, or the like.
- the battery gauge may measure, for example, a remaining capacity of the battery 1996 and a voltage, current or temperature thereof while the battery is charged.
- the battery 1996 may include, for example, a rechargeable battery and/or a solar battery.
- the indicator 1997 may display a specific state of the electronic device 1901 or a part thereof (e.g., the processor 1910), such as a booting state, a message state, a charging state, and the like.
- the motor 1998 may convert an electrical signal into a mechanical vibration and may generate the following effects: vibration, haptic, and the like.
- the electronic device 1901 may include, for example, a mobile TV supporting device (e.g., a GPU) capable of processing media data according to the standards of digital multimedia broadcasting (DMB), digital video broadcasting (DVB), MediaFLO TM , or the like.
- a mobile TV supporting device e.g., a GPU
- DMB digital multimedia broadcasting
- DVD digital video broadcasting
- MediaFLO TM MediaFLO TM
- Each of the above-mentioned elements of the electronic device according to various embodiments of the present disclosure may be configured with one or more components, and the names of the elements may be changed according to the type of the electronic device.
- the electronic device e.g., the electronic device 1901
- some of the elements of the electronic device according to various embodiments may be combined with each other so as to form one entity, so that the functions of the elements may be performed in the same manner as before the combination.
- FIG. 20 is a block diagram illustrating a program module 2010 according to an embodiment.
- a program module 2010 may include an operating system (OS) to control resources associated with an electronic device (e.g., an electronic device 1801 of FIG. 18), and/or diverse applications (e.g., an application program 1847 of FIG. 18) driven on the OS.
- the OS may be, for example, Android TM , iOS TM , Windows TM , Symbian TM , Tizen TM , or Bada TM .
- the program module 2010 may include a kernel 2020 (e.g., a kernel 1841 of FIG. 18), a middleware 2030 (e.g., a middleware 1843 of FIG. 18), an application programming interface (API) 2060 (e.g., an API 1845 of FIG. 18), and/or an application 2070 (e.g., an application program 1847 of FIG. 18).
- a kernel 2020 e.g., a kernel 1841 of FIG. 18
- a middleware 2030 e.g., a middleware 1843 of FIG. 18
- API application programming interface
- an application 2070 e.g., an application program 1847 of FIG. 18
- At least a portion of the program module 2010 may be preloaded on an electronic device or may be downloadable from an external electronic device (e.g., a first electronic device 1802, a second electronic device 1804, a server 1806, or the like of FIG. 18).
- the kernel 2020 may include, for example, a system resource manager 2021 or a device driver 2023.
- the system resource manager 2021 may control, allocate, or retrieve system resources.
- the system resource manager 2021 may include a process managing unit, a memory managing unit, a file system managing unit, or the like.
- the device driver 2023 may include, for example, a display driver, a camera driver, a Bluetooth driver, a shared memory driver, a USB driver, a keypad driver, a Wi-Fi driver, an audio driver, or an inter-process communication (IPC) driver.
- IPC inter-process communication
- the middleware 2030 may provide, for example, a function that the application 2070 needs in common, or may provide diverse functions to the application 2070 through the API 2060 to allow the application 2070 to efficiently use limited system resources of the electronic device.
- the middleware 2030 may include at least one of a runtime library 2035, an application manager 2041, a window manager 2042, a multimedia manager 2043, a resource manager 2044, a power manager 2045, a database manager 2046, a package manager 2047, a connectivity manager 2048, a notification manager 2049, a location manager 2050, a graphic manager 2051, or a security manager 2052.
- the runtime library 2035 may include, for example, a library module that is used by a compiler to add a new function through a programming language while the application 2070 is being executed.
- the runtime library 2035 may perform input/output management, memory management, or capacities about arithmetic functions.
- the application manager 2041 may manage, for example, a life cycle of at least one application of the application 2070.
- the window manager 2042 may manage a graphic user interface (GUI) resource that is used in a screen.
- GUI graphic user interface
- the multimedia manager 2043 may identify a format necessary for playing diverse media files, and may perform encoding or decoding of media files by using a codec suitable for the format.
- the resource manager 2044 may manage resources such as a storage space, memory, or source code of at least one application of the application 2070.
- the power manager 2045 may operate, for example, with a basic input/output system (BIOS) to manage a battery or power, and may provide power information for an operation of an electronic device.
- the database manager 2046 may generate, search for, or modify database that is to be used in at least one application of the application 2070.
- the package manager 2047 may install or update an application that is distributed in the form of package file.
- the connectivity manager 2048 may manage, for example, wireless connection.
- the notification manager 2049 may display or notify an event such as arrival message, appointment, or proximity notification in a mode that does not disturb a user.
- the location manager 2050 may manage location information about an electronic device.
- the graphic manager 2051 may manage a graphic effect that is provided to a user, or manage a user interface relevant thereto.
- the security manager 2052 may provide a general security function necessary for system security, user authentication, or the like.
- the middleware 2030 may include a telephony manager for managing a voice or video call function of the electronic device or a middleware module that combines diverse functions of the above-described elements.
- the middleware 2030 may provide a module specialized to each OS kind to provide differentiated functions. Additionally, the middleware 2030 may dynamically remove a part of the preexisting elements or may add new elements thereto.
- the API 2060 may be, for example, a set of programming functions and may be provided with a configuration that is variable depending on an OS. For example, in the case where an OS is the android or the iOS, it may provide one API set per platform. In the case where an OS is the Tizen, it may provide two or more API sets per platform.
- the application 2070 may include, for example, applications such as a home 2071, a dialer 2072, an SMS/MMS 2073, an instant message (IM) 2074, a browser 2075, a camera 2076, an alarm 2077, a contact 2078, a voice dial 2079, an e-mail 2080, a calendar 2081, a media player 2082, an album 2083, a watch 2084, health care (e.g., measuring an exercise quantity, blood sugar, or the like) or offering of environment information (e.g., information of barometric pressure, humidity, temperature, or the like).
- IM instant message
- the application 2070 may include an information exchanging application to support information exchange between an electronic device and an external electronic device.
- the information exchanging application may include, for example, a notification relay application for transmitting specific information to an external electronic device, or a device management application for managing the external electronic device.
- the notification relay application may include a function of transmitting notification information, which arise from other applications, to an external electronic device or may receive, for example, notification information from an external electronic device and provide the notification information to a user.
- the device management application may install, delete, or update for example, a function (e.g., turn-on/turn-off of an external electronic device itself (or a part of components) or adjustment of brightness (or resolution) of a display) of the external electronic device which communicates with the electronic device, and an application running in the external electronic device.
- a function e.g., turn-on/turn-off of an external electronic device itself (or a part of components) or adjustment of brightness (or resolution) of a display
- the application 2070 may include an application (e.g., a health care application of a mobile medical device) that is assigned in accordance with an attribute of an external electronic device. According to an embodiment, the application 2070 may include an application that is received from an external electronic device.
- an application e.g., a health care application of a mobile medical device
- the application 2070 may include an application that is received from an external electronic device.
- At least a portion of the program module 2010 may be implemented by software, firmware, hardware (e.g., the processor 1910), or a combination (e.g., execution) of two or more thereof, and may include modules, programs, routines, sets of instructions, processes, or the like for performing one or more functions.
- module used herein may represent, for example, a unit including one of hardware, software and firmware or a combination thereof.
- the term “module” may be interchangeably used with the terms “unit”, “logic”, “logical block”, “component” and “circuit”.
- the “module” may be implemented mechanically or electronically.
- the “module” may include at least one of an application-specific integrated circuit (ASIC) chip, a field-programmable gate array (FPGA), and a programmable-logic device for performing some operations, which are known or will be developed.
- ASIC application-specific integrated circuit
- FPGA field-programmable gate array
- At least a part of devices (e.g., modules or functions thereof) or methods (e.g., operations) according to various embodiments of the present disclosure may be implemented as instructions stored in a computer-readable storage medium in the form of a program module.
- the instructions are performed by a processor (e.g., the processor 1820), the processor may perform functions corresponding to the instructions.
- the computer-readable storage medium may be, for example, the memory 1830.
- a computer-readable recording medium may include a hard disk, a floppy disk, a magnetic medium (e.g., a magnetic tape), an optical medium (e.g., CD-ROM, digital versatile disc (DVD)), a magneto-optical medium (e.g., a floptical disk), or a hardware device (e.g., a ROM, a RAM, a flash memory, or the like).
- the program instructions may include machine language codes generated by compilers and high-level language codes that can be executed by computers using interpreters.
- the above-mentioned hardware device may be configured to be operated as one or more software modules for performing operations of various embodiments of the present disclosure and vice versa.
- a module or a program module according to various embodiments of the present disclosure may include at least one of the above-mentioned elements, or some elements may be omitted or other additional elements may be added. Operations performed by the module, the program module or other elements according to various embodiments of the present disclosure may be performed in a sequential, parallel, iterative or heuristic way. Furthermore, some operations may be performed in another order or may be omitted, or other operations may be added.
- Certain aspects of the above-described embodiments of the present disclosure can be implemented in hardware, firmware or via the execution of software or computer code that can be stored in a recording medium such as a CD ROM, a Digital Versatile Disc (DVD), a magnetic tape, a RAM, a floppy disk, a hard disk, or a magneto-optical disk or computer code downloaded over a network originally stored on a remote recording medium or a non-transitory machine readable medium and to be stored on a local recording medium, so that the methods described herein can be rendered via such software that is stored on the recording medium using a general purpose computer, or a special processor or in programmable or dedicated hardware, such as an ASIC or FPGA.
- a recording medium such as a CD ROM, a Digital Versatile Disc (DVD), a magnetic tape, a RAM, a floppy disk, a hard disk, or a magneto-optical disk or computer code downloaded over a network originally stored on a remote recording medium or a non
- the computer, the processor, microprocessor controller or the programmable hardware include memory components, e.g., RAM, ROM, Flash, etc. that may store or receive software or computer code that when accessed and executed by the computer, processor or hardware implement the processing methods described herein.
- memory components e.g., RAM, ROM, Flash, etc. that may store or receive software or computer code that when accessed and executed by the computer, processor or hardware implement the processing methods described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Acoustics & Sound (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Computer Security & Cryptography (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- User Interface Of Digital Computer (AREA)
- Telephone Function (AREA)
Abstract
An electronic device is provided. The electronic device includes a microphone, a communication circuitry, an indicator configured to provide at least one visual indication, and a processor configured to be electrically connected with the microphone, the communication circuitry, and the indicator, and a memory. The memory stores instructions, when executed, cause the processor to receive a first voice input through the microphone, perform a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, display a first visual indication through the indicator, receive a second voice input through the microphone, perform a second voice recognition for the second voice input, and if a second specified word corresponding to the first visual indication is included in a result of the second voice recognition, wake up the electronic device.
Description
The present disclosure generally relates to a voice input processing method and an electronic device for supporting the same.
Recently, electronic devices have been developed to include a microphone and provide a voice interface, so that the devices may perform functions such as purchasing products or operating smart home systems via voice input received through the microphone. In doing so, the electronic device may perform voice recognition for the received voice input. If a specified word (e.g., a wake-up word) for waking up the electronic device is included in the result of the voice recognition, the electronic device may wake up itself or a module associated with performing a specified function.
However, if there are a plurality of electronic devices which are woken up by the same specified word (e.g., “Hi, Galaxy”) and are located within a predetermined distance (e.g., about 20 m), when the user speaks the specified word, the plurality of electronic devices may be woken up irrespective of intention of the user.
If the plurality of electronic devices are woken up by the same specified word, the user may unnecessarily receive the same response from multiple devices or may receive error information (e.g., a voice, a screen indication, or the like) regarding the voice recognition from some devices.
Aspects of the present disclosure are to address at least the above-mentioned problems and/or disadvantages and to provide at least the advantages described below. Accordingly, an aspect of the present disclosure is to provide a voice input processing method for waking up one of a plurality of electronic devices if there are the plurality of electronic devices which are woken up by the same specified word. The electronic device for supporting the same is also described.
In accordance with an aspect of the present disclosure, an electronic device includes a microphone configured to receive a voice input, a communication circuitry configured to communicate with an external electronic device, an indicator configured to provide at least one visual indication, and a processor configured to be electrically connected with the microphone, the communication circuitry, and the indicator, and a memory configured to be electrically connected with the processor. The memory stores instructions, when executed, cause the processor to receive a first voice input through the microphone, perform a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, display a first visual indication through the indicator, receive a second voice input through the microphone, perform a second voice recognition for the second voice input, and if a second specified word corresponding to the first visual indication is included in a result of the second voice recognition, wake up the electronic device.
In accordance with another aspect of the present disclosure, a voice input processing method of an electronic device includes receiving a first voice input through a microphone, performing a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, displaying an indication through the indicator, receiving a second voice input through the microphone, performing a second voice recognition for the second voice input, and if a second specified word corresponding to the indication is included in a result of the second voice recognition, waking up the electronic device.
According to embodiments disclosed in the present disclosure, as a plurality of electronic devices which are woken up by the same specified word display different indications, only the electronic device a user wants may be woken up.
According to embodiments disclosed in the present disclosure, the plurality of electronic devices may provide a method for determining intention of the user (e.g., a method for selecting one electronic device to provide a response).
In addition, various effects or advantages directly or indirectly ascertained through the present disclosure may be provided.
Other aspects, advantages, and salient features of the disclosure will become apparent to those skilled in the art from the following detailed description, which, taken in conjunction with the annexed drawings, discloses various embodiments of the present disclosure.
The above and other aspects, features, and advantages of certain embodiments of the present disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
FIG. 1 is a block diagram illustrating an electronic device that processes voice inputs according to an embodiment.
FIG. 2 is a drawing illustrating a system that processes voice inputs according to an embodiment.
FIG. 3 is a drawing illustrating a method for providing an indication according to an embodiment.
FIG. 4 is a flowchart illustrating an operation method of an electronic device that processes voice inputs according to an embodiment.
FIG. 5 is a flowchart illustrating another operation method of an electronic device that processes voice inputs according to an embodiment.
FIG. 6 is a drawing illustrating a method for determining an indication according to an embodiment.
FIG. 7 is a drawing illustrating another method for determining an indication according to an embodiment.
FIG. 8 is a flowchart illustrating an operation method of an electronic device for determining an indication according to an embodiment.
FIG. 9 is a flowchart illustrating a method for displaying an indication determined through an external electronic device at an electronic device according to an embodiment.
FIG. 10 is a drawing illustrating a method for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
FIG. 11 is a signal sequence diagram illustrating a system for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
FIG. 12 is a drawing illustrating a method for determining whether to wake up an electronic device through adjustment of receive sensitivity of a voice input according to an embodiment.
FIG. 13 is a flowchart illustrating an operation method of an electronic device for adjusting the received sensitivity of a voice input according to an embodiment.
FIG. 14 is a drawing illustrating a method for determining whether to wake up an electronic device based on usage history information according to an embodiment.
FIG. 15 is a flowchart illustrating an operation method of an electronic device for determining whether to wake up an electronic device based on usage history information according to an embodiment.
FIG. 16 is a drawing illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
FIG. 17 is a flowchart illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
FIG. 18 is a block diagram illustrating an electronic device 1801 in a network environment 1800 according to an embodiment.
FIG. 19 is a block diagram illustrating an electronic device 1901 according to one embodiment.
FIG. 20 is a block diagram illustrating a program module 2010 according to an embodiment.
Throughout the drawings, it should be noted that like reference numbers are used to depict the same or similar elements, features, and structures.
Hereinafter, various embodiments of the present disclosure are disclosed with reference to the accompanying drawings. However, the present disclosure is not intended to be limited by the various embodiments of the present disclosure. Instead, it is intended that the present disclosure covers all modifications, equivalents, and/or alternatives provided they come within the scope of the appended claims and their equivalents.
The terms and words used in the following description and claims are not limited to their dictionary meanings, but, are merely used by the inventor(s) to enable a clear and consistent understanding of the present disclosure. Accordingly, it should be apparent to those skilled in the art that the following description of various embodiments of the present disclosure is provided for illustration purpose only and not for the purpose of limiting the present disclosure as defined by the appended claims and their equivalents.
It is to be understood that the singular forms “a,” “an,” and “the” include plural referents unless the context clearly dictates otherwise. Thus, for example, reference to “a component surface” includes reference to one or more of such surfaces.
The term “include,” “comprise,” and “have”, or “may include,” or “may comprise” and “may have” used herein indicates disclosed functions, operations, or existence of elements but does not exclude other functions, operations or elements.
For example, the expressions “A or B,” or “at least one of A and/or B” may indicate A and B, A, or B. For instance, the expression “A or B” or “at least one of A and/or B” may indicate (1) at least one A, (2) at least one B, or (3) both at least one A and at least one B.
The terms such as “1st,” “2nd,” “first,” “second,” and the like used herein may refer to modifying various different elements of various embodiments of the present disclosure, but are not intended to limit the elements. For instance, “a first user device” and “a second user device” may indicate different users regardless of order or importance. For example, a first component may be referred to as a second component and vice versa without departing from the scope of the present disclosure.
In various embodiments of the present disclosure, it is intended that when a component (for example, a first component) is referred to as being “operatively or communicatively coupled with/to” or “connected to” another component (for example, a second component), the component may be directly connected to the other component or connected through another component (for example, a third component). In various embodiments of the present disclosure, it is intended that when a component (for example, a first component) is referred to as being “directly connected to” or “directly accessed” another component (for example, a second component), another component (for example, a third component) does not exist between the component (for example, the first component) and the other component (for example, the second component).
The expression “configured to” used in various embodiments of the present disclosure may be interchangeably used with "suitable for," "having the capacity to," "designed to," "adapted to," "made to," or "capable of" according to the situation, for example. The term “configured to” may not necessarily indicate “specifically designed to” in terms of hardware. Instead, the expression “a device configured to” in some situations may indicate that the device and another device or part are “capable of.” For example, the expression "a processor configured to perform A, B, and C” may indicate a dedicated processor (for example, an embedded processor) for performing a corresponding operation or a general purpose processor (for example, a central processing unit (CPU) or application processor (AP)) for performing corresponding operations by executing at least one software program stored in a memory device.
Terms used in various embodiments of the present disclosure are used to describe certain embodiments of the present disclosure, but are not intended to limit the scope of other embodiments. All terms used herein may have the same meanings that are generally understood by a person skilled in the art. In any case, even the terms explicitly defined in the present specification are not intended to be interpreted as excluding embodiments of the present disclosure.
An electronic device according to various embodiments of the present disclosure may be a smartphone, a tablet personal computer (PC), a mobile phone, a video telephone, an electronic book reader, a desktop PC, a laptop PC, a netbook computer, a workstation, a server, a personal digital assistant (PDA), a portable multimedia player (PMP), a Motion Picture Experts Group (MPEG-1 or MPEG-2) Audio Layer 3 (MP3) player, a mobile medical device, a camera, a wearable device, etc. The wearable device may be an accessory-type device (e.g., a watch, a ring, a bracelet, an anklet, a necklace, glasses, a contact lens, a head-mounted device (HMD)), a textile- or clothing-integrated-type device (e.g., an electronic apparel), a body-attached-type device (e.g., a skin pad or a tattoo), or a bio-implantable-type device (e.g., an implantable circuit)
In another embodiment, the electronic device may be a home appliance. The smart home appliance may be a television (TV), a digital video/versatile disc (DVD) player, an audio, a refrigerator, an air conditioner, a cleaner, an oven, a microwave oven, a washing machine, an air cleaner, a set-top box, a home automation control panel, a security control panel, a television (TV) box (e.g., Samsung HomeSync™, Apple TV™, or Google TV™), a game console (e.g., Xbox™ or PlayStation™), an electronic dictionary, an electronic key, a camcorder, an electronic picture frame, etc.
In another embodiment, the electronic device may be various medical devices (e.g., various portable medical measurement devices (e.g., a blood glucose measuring device, a heart rate measuring device, a blood pressure measuring device, a body temperature measuring device, or the like), a magnetic resonance angiography (MRA) device, a magnetic resonance imaging (MRI) device, a computed tomography (CT) device, a scanner, an ultrasonic device, or the like), a navigation device, a global navigation satellite system (GNSS), an event data recorder (EDR), a flight data recorder (FDR), a vehicle infotainment device, electronic equipment for vessels (e.g., a navigation system, a gyrocompass, or the like), avionics, a security device, a head unit for a vehicle, an industrial or home robot, an automatic teller machine (ATM), a point of sales (POS) device of a store, or an Internet of things (IoT) device (e.g., a light bulb, various sensors, an electric or gas meter, a sprinkler, a fire alarm, a thermostat, a streetlamp, a toaster, exercise equipment, a hot water tank, a heater, a boiler, or the like).
In yet another embodiment, the electronic device may be at least one of a part of furniture or a building/structure, an electronic board, an electronic signature receiving device, a projector, or a measuring instrument (e.g., a water meter, an electricity meter, a gas meter, a wave meter, or the like). The electronic device may be one or more combinations of the above-mentioned devices. The electronic device according to some various embodiments of the present disclosure may be a flexible device. The electronic device according to an embodiment of the present disclosure is not limited to the above-mentioned devices, and may include new electronic devices as new technology is developed.
Hereinafter, the electronic devices according to various embodiments of the present disclosure will be described in more detail with reference to the accompanying drawings. The term “user” used herein may refer to a person who uses an electronic device or may refer to a device (e.g., an artificial intelligence electronic device) that uses an electronic device.
FIG. 1 is a block diagram illustrating an electronic device that processes voice inputs according to an embodiment.
The electronic device 100 may perform voice recognition for a voice input received through a microphone 120. If a first specified word (e.g., a first wake-up word) for waking up the electronic device 100 is included in the result of the voice recognition, the electronic device 100 may wake up itself or at least one module associated with performing a function included in the electronic device 100. Meanwhile, if there is another external electronic device which can be woken up by the first wake-up word, the electronic device 100 may provide an indication to the user that is different from the indication provided in the external electronic device. This way, both the electronic device 100 and the external electronic device are not woken up simultaneously when the first specified word is spoken. Thus, using the indication provided by the electronic device 100, the user may verify that the electronic device 100, which the user wants to wake up, has not woken up. The user may then speak a second specified word (e.g., a second wake-up word) corresponding to the indication provided from the electronic device 100 to wake up the electronic device 100.
Referring to FIG. 1, the electronic device 100 for performing the above-mentioned function may include a processor 110, the microphone 120, a speaker 130, an indicator 140 (or a user interface), a memory 150, and a communication circuitry 160. The elements of the electronic device 100 are not limited thereto. According to various embodiments, the electronic device 100 may fail to include at least one of the above-mentioned elements and may further include at least one other element. According to an embodiment, the electronic device 100 may further include a proximity sensor which measures a proximity state and/or distance of an object (e.g. the user). Further, at least one of the above-mentioned elements may be located and fixed within a housing. The housing may include a front surface, a rear surface, and a side surface which surrounds at least part of a space between the front surface and the rear surface.
The processor 110 may execute arithmetic operations or data processing to control at least one other element of the electronic device 100. The processor 110 may drive, for example, an operating system (OS) or an application program to control a plurality of hardware or software elements of the electronic device 100. The processor 110 may include one or more of a central processing unit (CPU), an application processor (AP), or a communication processor (CP). According to an embodiment, the processor 110 may be implemented as a system on chip (SoC). The processor 110 may include a microprocessor or any suitable type of processing circuitry, such as one or more general-purpose processors (e.g., ARM-based processors), a Digital Signal Processor (DSP), a Programmable Logic Device (PLD), an Application-Specific Integrated Circuit (ASIC), a Field-Programmable Gate Array (FPGA), a Graphical Processing Unit (GPU), a video card controller, etc. In addition, it would be recognized that when a general purpose computer accesses code for implementing the processing shown herein, the execution of the code transforms the general purpose computer into a special purpose computer for executing the processing shown herein. Certain of the functions and steps provided in the Figures may be implemented in hardware, software or a combination of both and may be performed in whole or in part within the programmed instructions of a computer.
The processor 110 may analyze a voice input received through the microphone 120. According to an embodiment, the processor 110 may perform voice recognition for the received voice input. The processor 110 may analyze voice data corresponding to a voice input received through the microphone 120 and may convert the voice data into text. For example, if a user speaks the word “HI”, the microphone 120 may convert vibration energy from the user’s voice into an electronic signal and may transmit the converted electric signal to the processor 110. The processor 110 may convert the transmitted electric signal into frequency spectrum information and may compare the converted frequency spectrum information with a plurality of frequency spectrum information corresponding to various characters stored in the memory 150 to determine that characters (e.g., “H” and “I”) were spoken.
According to an embodiment, the processor 110 may perform voiceprint identification for a received voice input. The processor 110 may perform voiceprint identification in substantially the same manner as voice recognition. Herein, in case of the voiceprint identification, the processor 110 may use the frequency spectrum information converted from the user’s voice input. For example, if the user speaks the word “HI”, voice recognition may only determine that characters (e.g., “H” and “I”) were included in the input, whereas voiceprint identification may determine that a specific user spoke “HI.”
According to an embodiment, if specified word or words are included in the result of the voice recognition, the processor 110 may wake up at least one other element included in the electronic device 100. For example, the processor 110 may supply power to one or more hardware elements and/or may load a software element (e.g., a program module) into the memory 150 so that the software element can be executed.
According to an embodiment, the processor 110 may determine whether there are one or more external electronic devices which also can be woken up by the specified word. For example, if a signal corresponding to the specified word is received from the external electronic device, the processor 110 may determine that the external electronic device is able to be woken up by the specified word. The signal corresponding to the wake-up word may be a signal transmitted from the external electronic device and may include, for example, a notification that the external electronic device recognizes the specified word and is in a state of waiting for wake-up. In this case, the processor 110 may not wake up elements in the electronic device 100 may provide (e.g. display) an indication through the indicator 140. For example, the processor 110 may provide an indication that is different from the indication provided from the external electronic device such that the user can distinguish the electronic device 100 from the external electronic device. The indication may be, for example, a color displayed by the indicator 140. For example, if the external electronic device displays a red color, the electronic device 100 may display another color, for example, blue or green. In some embodiments, if the indicator 140 includes a display, the indication may include identification information assigned to each electronic device, such as a nickname. A display object corresponding to the identification information may be displayed on the display.
According to an embodiment, after providing an indication through the indicator 140, if a voice input corresponding to another specified word (e.g., a second specified word) is received, the processor 110 may wake up at least one other element included in the electronic device 100. This way, the processor 110 may replace the specified word (e.g., a first specified word) for waking up the electronic device 100 with another specified word (e.g., a second specified word) after providing the indication.
According to an embodiment, the processor 110 may manage the types of the indications and how the indications correspond to different electronic devices. For example, the processor 110 may store information (e.g., a table) of mapping a first indication to the identification information of a first electronic device and mapping a second indication to the identification information of a second electronic device. For example, red indication may be assigned to the first electronic device and blue indication may be assigned to the second electronic device. In some embodiments, the processor 110 may receive information regarding the assigned indication via the communication circuitry 160 from an external electronic device and may provide the indication based on the received information about the indication. This way, the processor 110 may not directly manage the mapping of the indications to the various electronic devices. For example, the electronic device 100 may be used as a device (e.g., a server device, a hub device, or a master device) which manages the mapping information or may be used as a device (e.g., a slave device or a client device) which receives the mapping information from another device which manages the mapping information. If the electronic device 100 is used as the master device, the processor 110 may display the indication assigned to the electronic device 100 and may transmit information about indications assigned to other external electronic devices to the other external electronic devices via the communication circuitry 160.
According to an embodiment, the processor 110 may determine whether to wake up the microphone 120 based on sensing information obtained by a sensor module (e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like) included in the electronic device 100. For example, using the sensor module, the processor 110 may determine whether the user or a body part of the user is close to (or in contact with) the electronic device 100. If so, the processor 110 may wake up the microphone 120 to receive voice input of the user. Further, the processor 110 may inform another external electronic device that the user is close to (or in contact with) the electronic device 100. In this case, the external electronic device may deactivate its microphone so that the external electronic device does not erroneously receive the voice input from the user. In some embodiments, the microphone in the external electronic device may be only deactivated for a specified time period. For example, after the specified time period elapses, the microphone may be reactivated. In some embodiments, the processor 110 may recognize a face or an operation of the user through a camera or the like. For example, using the camera, the processor 110 may determine whether the user looks at the electronic device 100, and may wake up the microphone 120 only in the electronic device 100 that the user looked at.
According to an embodiment, the processor 110 may adjust the received sensitivity of the voice input received through the microphone 120. For example, the processor 110 may adjust a threshold of received intensity (or receive strength or a volume level) of the voice input for voice recognition. According to an embodiment, if the received intensity is a first level and if received intensity of the voice input received by another external electronic device is at a second level lower than the first level, the processor 110 may set the threshold of received intensity for the microphone 120 and the received intensity for the microphone included in the external electronic device to values between the first level and the second level. This way, when the voice input is above the threshold of received intensity for the microphone 120 but below the threshold of received intensity for the microphone included in the external electronic device, the electronic device 100 may be able to process the voice input from the user while the external electronic device may not process the voice input. In some embodiments, the processor 110 may provide an indication depending on the threshold of received intensity to guide the user. For example, the electronic device with the highest threshold may display an indication of a first color (e.g., a green color) to inform the user that it is capable of performing voice recognition at the current volume. An electronic device with the lowest received intensity may display an indication of a second color (e.g., a red color) to inform the user that it cannot perform voice recognition at the current volume.
According to an embodiment, the processor 110 may determine whether to wake up the electronic device 100 based on usage history information of the user. For example, the processor 110 may identify the user through voiceprint identification and may identify the history in which the identified user uses the electronic device 100. For example, the history information may include the frequency of use, the number of use, the difference in the frequency of use/the number of use with another user, etc. The processor 110 can then use this history information to determine whether to wake up the electronic device 100. For example, if the electronic device 100 is an electronic device frequently used by a first user, the processor 110 may not wake up the electronic device 100 when voice input from a second user is received. In this case, the processor 110 may wake up the electronic device 100 only if voice input from the first user is received. In some embodiments, the processor 110 may determine whether to wake up the electronic device 100, based on settings information set by the user. For example, if the user sets the electronic device 100 as an electronic device frequently used by him or her, the processor 110 may wake up the electronic device 100 when voice input from the user is received.
According to an embodiment, the processor 110 may determine whether to wake up the electronic device 100 based on a location of a personal device (e.g., a smart watch, a smartphone, or the like) of the user. For example, in case of a wearable electronic device, the processor 110 may determine whether the user wears the wearable electronic device. If determining that the user wears the wearable electronic device, the processor 110 may wake up the electronic device 100 adjacent to the wearable electronic device. The processor 110 may determine a location relationship between the personal device of the user and the electronic device 100 using an illumination sensor included in the electronic device 100 or the degree of noise measured through the microphone 120. For example, if illumination values measured by the illumination sensor or values indicating the degree of noise measured through the microphone 120 are the same or similar to values measured by the personal device of the user, the electronic device 100 may determine that the personal device of the user is adjacent to the electronic device 100. According to an embodiment, the electronic device 100 may compare the measured values (e.g., the illumination value and the value indicating the degree of noise) with values measured by other external electronic devices and may select the most similar device to the personal device of the user as the device closest to the user.
Further, the location relationship between the personal device of the user and the electronic device 100 may be determined through various indoor positioning methods, such as ones using communication beacons (e.g., Bluetooth low energy (BLE), wireless-fidelity (Wi-Fi), or the like), fingerprinting, visible light communication using light-fidelity (Li-Fi), multilateration using a received signal strength indicator (RSSI) and a round trip time (RTT), etc. In the indoor positioning method using a BLE device, the BLE device may transmit an advertising packet to peripheral electronic devices and the electronic device 100 which receives the packet may calculate its location through a cell ID or multilateration using the device ID of the BLE device or the distance between the electronic device 100 and the BLE device. The packet may include data such as a device ID, a service place name, and a transmit signal strength. The electronic device 100 may calculate the distance value between the BLE device and the electronic device 100 using an RSSI signal value of the BLE device. The method for calculating the distance using the RSSI signal value may include calculating the distance using strength (power) loss of the RSSI signal or path loss between the electronic device 100 and the BLE device.
The fingerprint method may be a pattern recognition method for measuring strength of signals received based on Wi-Fi, Bluetooth, or the like. The method may involve knowing several reference locations in advance and configuring a radio map. When a reference location on the radio map is determined to have a signal strength characteristic most similar to the strength of the signal received by the electronic device 100, the reference location may be determined to be the location of the electronic device 100.
The visible light communication method may involve devices capable of transmitting light signals that are so brief that they cannot be visibly detected by humans. The light signals may be encoded with unique information (e.g. location information) so that by receiving the light signal and decoding it, the electronic device 100 may obtain location information.
The multilateration method using the RSSI and the RTT may be a method for converting an RSSI of a Wi-Fi signal to a distance using signal propagation modeling and calculating a location of the electronic device 100 by triangulating the electronic device 100 from a plurality of Wi-Fi access points (APs). Distance between a Wi-Fi AP and the electronic device 100 may be determined based on a delay time interval from when a pulse is transmitted from the Wi-Fi AP to when a response pulse is received by the Wi-Fi AP from the electronic device 100.
According to an embodiment, the processor 110 may wake-up the electronic device 100 depending on whether an instruction corresponding to a function performable by the electronic device 100 is included in a result of voice recognition of a received voice input. For example, the processor 110 may wake up the electronic device 100 only if the electronic device 100 is capable of performing the requested function in the voice input. In another embodiment, although a plurality of electronic devices recognize a specified word (e.g. the wake word), only one electronic device may respond to the specified word and the other electronic devices may remain in the sleep state. Further, the electronic devices may be able to perform the same function (e.g. play music). If an electronic device most suitable for performing the function is determined, the other electronic devices may not need to wake up. Factors in determining which device is most suitable may include the ability to process the instruction from the user, the distance from the user, the volume of the voice input from the user, etc. According to an embodiment, an apparatus for responding to the specified word and analyzing the instruction may be selected because, among the plurality of electronic devices, it is the device closest to the user. Alternatively, a server device connected with the plurality of electronic devices may analyze the instruction. In this case, the server device may determine an electronic device capable of performing a function corresponding to the instruction among the plurality of electronic devices and may transmit a related signal such that the determined electronic device performs the function.
The microphone 120 may receive various audio inputs generated outside the electronic device, such as voice inputs from the user. The microphone 120 may be one microphone located in the electronic device 100, or a plurality of microphones may be located in the electronic device 100.
The speaker 130 may output audio. For example, the speaker 130 may output a synthetic voice output corresponding to an indication. For another example, the speaker 130 may output a synthetic voice output corresponding to the result of performing a function.
The indicator 140 may be exposed on part of the housing and may provide (or display) a visual indication. For example, the indicator 140 may display a color. The indicator 140 may include, for example, an LED or a display.
The memory 150 may include a volatile and/or nonvolatile memory. For example, the memory 150 may store instructions or data associated with at least one other element of the electronic device 100. According to an embodiment, the memory 150 may store software and/or a program. The program may include an application. The application may be a set of a series of programs (or instructions) for performing at least one specified function.
According to an embodiment, the memory 150 may include mapping information that maps a type of an indication to a state of an application. Further, the memory 150 may store frequency spectrum information corresponding to various characters and frequency spectrum information of the user’s voice so that the electronic device 100 may perform voice recognition and voiceprinting. The memory 150 may include, for example, an internal memory or an external memory.
The communication circuitry 160 may establish communication between the electronic device 100 and an external electronic device. The communication circuitry 160 may be connected to, for example, a network through wireless communication or wired communication.
According to an embodiment, at least some of the elements of the electronic device 100 may exist outside the housing of the electronic device 100 and may be electrically connected with the processor 110. For example, at least part of the microphone 120, the speaker 130, or the indicator 140 may exist outside the housing of the electronic device 100.
As described above, according to one embodiment, an electronic device (e.g., the electronic device 100) may include a housing, a user interface (e.g., the indicator 140) configured to be exposed through the housing and provide a visual indication, a first microphone (e.g., the microphone 120), a speaker (e.g., the speaker 130), a communication circuitry (e.g., the communication circuitry 160), a processor (e.g., the processor 110) configured to be electronically connected with the user interface, the first microphone, the speaker, and the communication circuitry, and a memory (e.g., the memory 150) configured to be electrically connected with the processor. The memory may store instructions, when executed, cause the processor to receive a first user voice input for waking up the electronic device through the first microphone, determine whether the electronic device is in a state for performing a selected operation based at least in part on the first user voice input, if the electronic device is in the state for performing the operation, provide a first indication using the user interface, receive a user input for selecting the electronic device to perform the operation, receive a second user voice input associated with the operation through the first microphone, transmit data associated with the operation to an external server through the communication circuitry, receive a response from the external server via the communication circuitry, and provide the response using at least one of the user interface and the speaker, and if the electronic device is not in the state for performing the operation, provide a second indication different from the first indication using the user interface and set the electronic device to a waiting state.
According to one embodiment, the instructions may cause the processor to receive information via the communication circuitry from at least one external electronic device, and determine whether the electronic device is in the state for performing the selected operation based at least in part on the information. And the at least one external electronic device may include a second microphone and obtain at least part of the first user voice input through the second microphone in a state where the at least one external electronic device is close to the electronic device.
According to one embodiment, the first indication may include a first color, and the second indication may include a second color.
As described above, according to one embodiment, an electronic device (e.g., the electronic device 100) may include a microphone (e.g., the microphone 120) configured to receive a voice input, a communication circuitry (e.g., the communication circuitry 160) configured to communicate with an external electronic device, an indicator (e.g., the indicator 140) configured to provide at least one visual indication, and a processor (e.g., the processor 110) configured to be electrically connected with the microphone, the communication circuitry, and the indicator, and a memory (e.g., the memory 150) configured to be electrically connected with the processor. The memory may store instructions, when executed, cause the processor to receive a first voice input through the microphone, perform a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, display a first visual indication through the indicator, receive a second voice input through the microphone, perform a second voice recognition for the second voice input, and if a second specified word corresponding to the first visual indication is included in a result of the second voice recognition, wake up the electronic device.
According to one embodiment, the indicator may include at least one of a light emitting diode (LED) or a display, and the first visual indication may be at least one color displayed by the indicator.
According to one embodiment, the memory may further include mapping information that maps a type of the at least one visual indication to an assignment state of the at least one visual indication, and the instructions may cause the processor to determine the first visual indication based on the mapping information.
According to one embodiment, the instructions may cause the processor to transmit information about a second visual indication different from the first visual indication to the external electronic device if a signal corresponding to the first specified word is received from the external electronic device located near the electronic device.
According to one embodiment, the signal may include at least one of information regarding when the external electronic device received the first voice input and a first volume level of the first voice input received by the external electronic device.
According to one embodiment, the instructions may cause the processor to adjust a threshold of a volume level for the second voice recognition based on the first volume level of the first voice input included in the signal and a second volume level of the first voice input received through the microphone.
According to one embodiment, the electronic device may further include a speaker, and the instructions may cause the processor to output a voice output corresponding to the first visual indication through the speaker.
According to one embodiment, the electronic device may further include a sensor module, and the instructions may cause the processor to determine at least one of whether a body part of a user is close to the electronic device and whether the body part of the user is in contact with the electronic device, based on sensing information obtained through the sensor module, and determine whether to wake up the microphone based on the determination that the body part of the user is close to the electronic device or the determination that the body part of the user is in contact with the electronic device.
According to one embodiment, the memory may include usage history information of a user for the electronic device, and the instructions may cause the processor to perform voiceprint identification for the first voice input to identify the user, and determine whether to wake up the electronic device based on the usage history information of the user.
According to one embodiment, the instructions may cause the processor to determine whether to wake up the electronic device based on whether an instruction corresponding to a function performable by the electronic device is included in the result of the second voice recognition.
FIG. 2 is a drawing illustrating a system that processes voice inputs according to an embodiment.
Referring to FIG. 2, in operation 201, a user 200 may call an electronic device (e.g., a first electronic device 210) by speaking a message including a specified word (e.g., “AAA”). However, if there are other external electronic devices (e.g., a second electronic device 230 and a third electronic device 250) which use the specified word as a wake-up word as well as the first electronic device 210 within a short range (e.g., 20 m), the other external electronic devices may also process the voice input from the user, even though the user does not intend to user the other electronic devices. Thus, there may be a need for processing so that only the first electronic device 210 wakes up.
In operation 203, the first electronic device 210 may perform voice recognition of the voice input from the user 200 received through the first microphone 211. Further, if the specified word 203a is included in the result of the voice recognition, in operation 205, the first electronic device 210 may determine whether another external electronic device can be woken up by the specified word 203a. For example, if receiving a signal corresponding to the specified word 203a from the other external electronic device, the first electronic device 210 may determine that the other external electronic device can be woken up by the specified word 203a. In this operation, each of the second electronic device 230 and the third electronic device 250 may receive the voice input of the user 200 through the second microphone 231 and the third microphone 251 and perform voice recognition. If the specified word 203a is included in the result of the voice recognition, each of the second electronic device 230 and the third electronic device 250 may transmit the signal corresponding to the specified word 203a to the electronic device 210. The signal may include identification information of each electronic device.
Receiving the signal, the first electronic device 210 may assign different indications to all three electronic devices capable of being woken up by the specified word 203a. For example, the first electronic device 210 may assign indications of a first color, a second color, and a third color to the first electronic device 210, the second electronic device 230, and the third electronic device 250, respectively.
In operation 207, the first electronic device 210 may display the assigned indication of the first color through the first indicator 213. In some embodiments, the first electronic device 210 may transmit information about the indication of the second color and information about the indication of the third color to the second electronic device 230 and the third electronic device 250, respectively. In this case, the second electronic device 230 and the third electronic device 250 may display the indication of the second color and the indication of the third color through the second indicator 233 and the third indicator 253, respectively.
According to another embodiment, the electronic devices (e.g., the first electronic device 210, the second electronic device 230, and the third electronic device 250) capable of being woken up by the same specified word 203a may select an indication themselves. For example, the first electronic device may select the indication of the first color and may transmit information about the first color to the second electronic device 230. After receiving the information about the first color, the second electronic device 230 may select the indication of the second color different from the first color and may transmit information about the first color and the second color to the third electronic device 250. After receiving the information about the first color and the second color, the third electronic device 250 may select the indication of the third color different from the first color and the second color. The method or order in which the electronic devices (e.g., the first electronic device 210, the second electronic device 230, or the third electronic device 250) transmit their color information to other electronic devices is not limited thereto. Various methods may be used to determine the method or order.
FIG. 3 is a drawing illustrating a method for providing an indication according to an embodiment.
Referring to FIG. 3, an electronic device 310 may receive a first voice input 301 of the user 300 through the microphone 311. If a first specified word 301a is included in the result of voice recognition for the received first voice input 301, the electronic device 310 may display an indication assigned to the electronic device 310 through an indicator 313. The indication may include, for example, a color assigned for the electronic device 310. According to an embodiment, the assignment of the indication may be performed based on mapping information that maps the type of indication and the assignment state of the indication. This mapping information may be stored in a database managed by the electronic device 310. In some embodiments, the database may be managed by a separate server device, a hub device, or a master device, and the electronic device 310 may receive information about the indication via a communication circuitry and may display the corresponding indication through the indicator 313.
If the electronic device 310 displays a specified indication through the indicator 313, the user 300 may see the displayed indication (see reference numeral 303). In some embodiments, the electronic device 310 may output the specified indication as voice output through a voice output device (e.g., a speaker) so that the user may hear the indication. If the user 300 speaks a second voice input 305 including a new second specified word 305a corresponding to the indication (i.e. either the displayed indication or the voiced indication), the electronic device 310 may determine that the second specified word 305a is included in the result of voice recognition for the second voice input 305. The electronic device 310 may then wake up at least one of its elements. The second voice input 305 may include the second specified word 305a and an instruction 305b. In this case, the electronic device 310 may wake up at least one element for performing a function corresponding to the instruction 305b. In some embodiments, the electronic device 310 may provide audio output for informing the user 300 that the electronic device 310 is woken up and/or audio output 307 for informing the user 300 of the result of the performance of the function. The electronic device 310 may also display an indication for informing the user that the electronic device 310 is woken up, through the indicator 313.
FIG. 4 is a flowchart illustrating an operation method of an electronic device that processes voice inputs according to an embodiment.
Referring to FIG. 4, in operation 410, an electronic device (e.g., a processor 110 of FIG. 1) may receive a first voice input through a microphone (e.g., a microphone 120 of FIG. 1) and may perform voice recognition for the received first voice input. The first voice input may include a specified word (e.g. a wake-up word) for waking up the electronic device.
In operation 420, the electronic device (e.g., the processor 110) may determine whether a first specified word is included in the result of the voice recognition. For example, the processor 110 may determine whether the first specified word is included in the text converted from the first voice input.
If the first specified word is included in the result of the voice recognition, in operation 430, the electronic device (e.g., the indicator 140 of FIG. 1) may display an indication. According to an embodiment, the electronic device (e.g., the processor 110) may specify an indication to distinguish the electronic device from an external electronic device based on mapping information mapping the types of indications to the assignment states of the indications. In some embodiments, the electronic device may display an indication it selected for itself irrespective of the mapping information. In this case, the electronic device (e.g., the communication circuitry 160 of FIG. 1) may transmit information about the selected indication to the external electronic device. Receiving the information about the indication, the external electronic device may select an indication that is different from the indication of the electronic device.
If the first specified word is not included in the result of performing the voice recognition in operation 420, the electronic device (e.g., the processor 110) according to an embodiment may maintain the waiting state (i.e. the sleep state).
In operation 440, the electronic device (e.g., the processor 110) may receive a second voice input through the microphone (e.g., the microphone 120) and may perform voice recognition for the received second voice input. The second voice input may be one spoken by the user after he or she has seen the displayed indication. The second voice input may include the second specified word corresponding to the indication.
In operation 450, the electronic device (e.g., the processor 110) may determine whether the second specified word is included in the result of the voice recognition. For example, the processor 110 may determine whether the second specified word is included in the text converted from the second voice input.
If the second specified word is included in the result of the voice recognition, in operation 460, the electronic device (e.g., the processor 110) may wake up at least one element in the electronic device. For example, the electronic device (e.g., the processor 110) may change from the waiting state to an active state where the electronic device may perform a function.
If the second specified word is not included in the result of the voice recognition in operation 450, the electronic device (e.g., the processor 110) according to an embodiment may display an indication different from the indication displayed in step 430 and/or may maintain the waiting state.
FIG. 5 is a flowchart illustrating another operation method of an electronic device that processes voice inputs according to an embodiment.
Referring to FIG. 5, in operation 510, an electronic device (e.g., an electronic device 100 or a processor 110 of FIG. 1) may receive a first voice input and may perform voice recognition for the first voice input. The first voice may include a specified word (or a wake-up word) for waking up the electronic device. Further, the specified word may be used as the wake-up word of another external electronic device as well as the electronic device.
In operation 520, the electronic device may determine whether the first specified word is included in the result of the voice recognition. For example, the processor 110 may determine whether the first specified word is included in text converted from the first voice input.
If the first specified word is included in the result of the voice recognition, in operation 530, the electronic device (e.g., the processor 110) may determine whether a signal corresponding to the first specified word is received from the external electronic device. For example, the external electronic device capable of being woken up by the first specified word may transmit the signal corresponding to the first specified word. In this case, the electronic device may receive the transmitted signal via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1) and may verify that there is another external electronic device capable of being woken up by the first specified word.
If receiving the signal corresponding to the first specified word from the external electronic device, in operation 540, the electronic device may display an indication distinguishing the electronic device from the external electronic device through an indicator. For example, the processor 110 may specify an indication to distinguish itself from the external electronic device based on mapping information mapping the types of indications to the assignment states of the indications. The electronic device may then display the specified indication through an indicator (e.g., an indicator 140 of FIG. 1).
In operation 550, the electronic device may receive a second voice input through the microphone (e.g., the microphone 120 of FIG. 1) and may perform voice recognition for the second voice input. The second voice input may be spoken by the user has seen the displayed indication, and the second voice input may include a second specified word corresponding to the indication. For example, if a red indication is displayed through the indicator, the second voice input may be “Hi, red” where “red” corresponds to the indication and is the second specified word. For another example, the user may speak an instruction including the second specified word, for example, “Red, how is the weather today?”
In operation 560, the electronic device may determine whether the second specified word is included in the result of the voice recognition. For example, the electronic device may determine whether the second specified word is included in the text converted from the second voice input.
If the second specified word is included in the result of performing the voice recognition, in operation 570, the electronic device may wake up at least one element of the electronic device. Alternatively, if the second specified word and an instruction are included in the result of the voice recognition, the electronic device may wake itself up to perform a function corresponding to the instruction.
If the second specified word is not included in the result of performing the voice recognition, the electronic device may return to operation 550. For example, the electronic device may repeatedly perform operations 550 and 560 until the second voice input including the second specified word is received. In some embodiments, if the second voice input including in the second specified word is not received for a specified time period, the electronic device may skip operations 550 to 570 and may change to a waiting state or an inactive (sleep) state.
FIG. 6 is a drawing illustrating a method for determining an indication according to an embodiment.
Referring to FIG. 6, a first electronic device 610 may receive a voice input through a first microphone 611. If a specified word is included in a result of voice recognition for the received voice input, the first electronic device 610 may provide (or display) a first indication through the first indicator 613. According to an embodiment, the first electronic device 610 may manage mapping information mapping the types of indications to the assignment states of the indications in a database 615 stored in memory (e.g., the memory 150 of FIG. 1). Thus, the first electronic device 610 may determine the first indication based on the mapping information 617. The mapping information 617 may include, for example, information associated with an assigning order 617a of electronic devices and types 617b of indications.
According to an embodiment, the types 617b of the indications may be color, unique nicknames of the electronic devices, etc. Further, each electronic device (e.g., the first electronic device 610, the second electronic device 630, or a third electronic device 650) may display its corresponding indication through an indicator (e.g., the first indicator 613, a second indicator 633, or a third indicator 653) or may output voice information corresponding to the indication through its speaker. In some embodiments, the types 617b of the indications may depend on how each electronic device (e.g., the first electronic device 610, the second electronic device 630, or the third electronic device 650) outputs the indications. For example, if one electronic device outputs the indication through an LED, the indication may be set to a specific color. But if the electronic device outputs the indication on a display, the indication may be set to a display object that includes a specific color or a nickname of the device. And if the electronic device outputs an indication through a speaker, the indication may be set to voice information corresponding to a specific color or a nickname.
According to an embodiment, the first electronic device 610 may determine indications of other external electronic devices and may transmit the determined indications to the external electronic devices. In this embodiment, the first electronic device 610 may function as a master device. The second electronic device 630 and the third electronic device 650 may receive voice inputs from the user through a second microphone 631 and a third microphone 651, respectively. If a specified word is included in the result of the voice recognition for the received voice input, the second electronic device 630 and the third electronic device 650 may transmit a signal corresponding to the specified word to the electronic device 610. In this case, the first electronic device 610 may receive the signal transmitted from the second electronic device 630 and the third electronic device 650 and may sequentially assign the second indication and the third indication to the second electronic device 630 and the third electronic device 650, respectively, based on an order where the signal is received. Further, the first electronic device 610 may transmit information about the second indication and information about the third indication to the second electronic device 630 and the third electronic device 650, respectively, via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1). Receiving the information about the second indication and the information about the third indication, the second electronic device 630 and the third electronic device 650 may provide (or display) the second indication and the third indication through the second indicator 633 and the third indicator 653, respectively.
According to an embodiment, the types 617b of the indications assigned to the electronic devices may not duplicate each other. Further, if one electronic device is selected (or woken up) by a user, the type 617b of the assigned indication may be reset. In some embodiments, if the indications are spoken but there is some overlap in the indications, then these indications may not be used. For example, spoken indications “blue” and “sky blue” overlap with the word “blue” and may cause confusion. Thus, they may not be used as indications. Further, when the indications are displayed colors, to avoid confusion, only colors that highly contrast with each other may be used. For example, the various electronic devices may be use blue and sky blue as two different indications.
According to an embodiment, the types 617b of the indications assigned to the electronic devices may be differently specified according to performance of the electronic devices, a temporal order of the received voice input, volume levels of the received voice inputs, etc. For example, the indication may be assigned sequentially from an electronic device with the best performance. Alternatively, the indication may be assigned sequentially from an electronic device having the best response time to a voice input. Alternatively, the indication may be assigned sequentially from an electronic device with the highest volume of the received voice input. Alternatively, the indication may be assigned from an electronic device which is always powered on. In some embodiments, the types 617b of the indications assigned to the electronic devices may be differently specified according to settings specified by the user.
FIG. 7 is a drawing illustrating another method for determining an indication according to an embodiment.
Referring to FIG. 7, if there are a plurality of electronic devices (e.g., a first electronic device 730, a second electronic device 750, and a third electronic device 770) capable of being called (woken up) by one specified word (e.g., a wake-up word) and if they receive voice inputs including the specified word, the electronic devices may transmit signals corresponding to the specified word. In this embodiment, each of the plurality of electronic devices may function as a slave or client device. In this case, there may be a need for a server device 710 (or a hub device) which determines the types of indications for the plurality of electronic devices and transmits the information about the determined indications to the plurality of electronic devices.
The server device 710 may manage, for example, mapping information 713 that maps the types of indications to the assignment states of the indications in a database 711. The mapping information 713 may include, for example, information associated with an assigning order 713a of the electronic devices and types 713b of indications.
The first electronic device 730, the second electronic device 750, and the third electronic device 770 may receive voice inputs through a first microphone 731, a second microphone 751, and a third microphone 771, respectively. If the specified word is included in the result of the voice recognition for the received voice inputs, the first electronic device 730, the second electronic device 750, and the third electronic device 770 may transmit signals corresponding to the specified word to the server device 710. The server device 710 may receive the signals transmitted from at least one of the electronic devices (e.g., the first electronic device 730, the second electronic device 750, and the third electronic device 770) and may transmit information based on the mapping information 713 to the electronic devices. For example, the server device 710 may receive signals transmitted from the first electronic device 730, the second electronic device 750, and the third electronic device 770 and may assign a first indication, a second indication, and a third indication to the first electronic device 730, the second electronic device 750, and the third electronic device 770 sequentially based on the mapping information 713 in an order in which the signals are received. For another example, the server device 710 may receive signals transmitted form the first electronic device 730, the second electronic device 750, and the third electronic device 770 and may assign the first indication, the second indication, and the third indication to the first electronic device 730, the second electronic device 750, and the third electronic device 770, respectively, based on assignment history. The server device 710 may transmit information about the first indication, information about the second indication, and information about the third indication to the first electronic device 730, the second electronic device 750, and the third electronic device 770, respectively. Respectively receiving the information about the first indication, the information about the second indication, and the information about the third indication, the first electronic device 730, the second electronic device 750, and the third electronic device 770 may provide (or display) the first indication, the second indication, and the third indication through the first indicator 733, the second indication 753, and the third indication 773, respectively.
FIG. 8 is a flowchart illustrating an operation method of an electronic device for determining an indication according to an embodiment.
According to an embodiment, referring to FIG. 8, in operation 810, an electronic device (e.g., an electronic device 100 or a processor 110 of FIG. 1) may receive a signal corresponding to a specified word via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1) from at least one external electronic device. The electronic device may be a server device, a hub device, or a master device. The signal corresponding to the specified word may include information relating to the fact that the external electronic device recognizes the specified word and may wake up based on the recognition. The signal corresponding to the specified word may include an indication for when the external electronic device received the voice input containing the specified word, received strength of the voice input, etc.
In operation 830, the electronic device (e.g., the electronic device 100 or the processor 110) may determine an indication for the external electronic device. According to an embodiment, the electronic device may manage mapping information mapping the types of indications to the assignment states of the indications in a database stored in memory (e.g., a memory 150 of FIG. 1). Thus, the electronic device may determine (i.e. assign) the indication for the external electronic device based on the mapping information stored in the database.
In operation 850, the electronic device (e.g., the electronic device 100 or the processor 110) may transmit information about the determined indication to the external electronic device via the communication circuitry (e.g., the communication circuitry 160). Receiving the information about the indication, the external electronic device may provide (or display) the indication through its indicator.
FIG. 9 is a flowchart illustrating a method for displaying an indication determined through an external electronic device at an electronic device according to an embodiment.
According to an embodiment, referring to FIG. 9, in operation 910, an electronic device (e.g., an electronic device 100 or a processor 110 of FIG. 1) (e.g., the slave device) may receive a voice input through a microphone (e.g., the microphone 120 of FIG. 1) and may perform voice recognition for the received voice input. According to an embodiment, the processor 110 may analyze voice data corresponding to the voice input received through the microphone and may convert the voice data into text.
In operation 930, the electronic device (e.g., the electronic device 100 or the processor 110) may determine whether a specified word is included in the result of the voice recognition. For example, the processor 110 may determine whether the specified word is included in the converted text.
If the specified word is included in the result of the voice recognition, in operation 950, the electronic device (e.g., the electronic device 100 or the processor 110) may transmit a signal corresponding to the specified word to an external electronic device (e.g., a server device, a hub device, or a master device). In some embodiments, the electronic device may send out the signal corresponding to the specified word via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1). Further, the electronic device may transmit the signal and may change to a state of waiting for wake-up. For example, the electronic device may change to a state of waiting for a response to the signal.
In operation 970, the electronic device (e.g., the electronic device 100 or the processor 110) may receive information about an indication via the communication circuitry (e.g., the communication circuitry 160) from the external electronic device. If receiving the indication, in operation 990, the electronic device (e.g., the electronic device 100 or the processor 110) may provide (or display) the indication through an indicator (e.g., the indicator 140 of FIG. 1).
FIG. 10 is a drawing illustrating a method for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
Referring to FIG. 10, electronic devices (e.g., a first electronic device 1010, a second electronic device 1030, and a third electronic device 1050) may determine whether to wake up microphones (e.g., a first microphone 1011, a second microphone 1031, and a third microphone 1051) included in the electronic devices depending on whether a body part of the user 1000 (e.g., a finger) is close to (or in contact with) the electronic devices. For example, the electronic devices may determine whether the body part 1001 is close to (or in contact with) the electronic devices based on sensing information obtained by a sensor module (e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like). If the body part 1001 is close to (or in contact with) one of the electronic devices, that electronic device may wake up its microphones to receive a voice input from the user 1000. Further, that electronic device may transmit a notification to another external electronic device indicating that it has detected the body part 1001. In this case, after receiving the notification, the external electronic device may deactivate its own microphone so that it does not receive the voice input of the user 1000. In the example shown in FIG. 10, the body part 1001 of the user 1000 is close to (or in contact with) the first electronic device 1010. The first electronic device 1010 may wake up the first microphone 1011 and may transmit a notification to the second electronic device 1030 and the third electronic device 1050, where the notification indicates that the body part 1001 is close to (in contact with) the first electronic device 1010. After receiving the notification, the second electronic device 1030 and the third electronic device 1050 may deactivate the second microphone 1031 and the third microphone 1051, respectively.
According to an embodiment, the microphone may be deactivated for a specified time period. After the specified time period elapses, the microphone may be reactivated. For example, after the specified time period elapses, the first electronic device 1010 may deactivate the first microphone 1011, and the second electronic device 1030 and the third electronic device 1050 may wake up the second microphone 1031 and the third microphone 1051, respectively. In another example, the first electronic device 1010 may maintain the wake-up state of the microphone 1011, and the second electronic device 1030 and the third electronic device 1050 may wake up the second microphone 1031 and the third microphone 1051, respectively.
According to an embodiment, if a plurality of electronic devices receive a voice of the user 1000, the electronic device which receives an additional selection (e.g., a finger touch, proximity, or the like) of the user 1000 may maintain a wake-up state (e.g., a state of waiting for receiving an additional voice) and the electronic device which does not receive a selection of the user 1000 may change to an inactive state.
FIG. 11 is a signal sequence diagram illustrating a system for determining whether to wake up an electronic device based on a proximity sensor according to an embodiment.
According to an embodiment, a user 1000 may select (e.g., be close to or come into contact with) one of a plurality of electronic devices (e.g., a first electronic device 1110, a second electronic device 1130, and a third electronic device 1150) and may speak a specified word (e.g., a wake-up word). The order of the user operations may be reversed. In other words, the user may speak the specified word (e.g., the wake-up word) and then may select one of the plurality of electronic devices.
Referring to FIG. 11, in this example, in operation 1101, the user 1100 may be close to (or come into contact with) one electronic device (e.g., the first electronic device 1110). For example, the user 1100 may touch the one electronic device using a part of his or her body (e.g., a finger).
In operation 1111, the first electronic device 1110 may recognize the proximity (or contact) of the user 1100. According to an embodiment, the first electronic device 1110 may receive the proximity (or contact) of the user 1100 based on sensing information obtained by a sensor module (e.g., a proximity sensor, an illumination sensor, a sound pressure sensor, or the like).
Recognizing the proximity (or contact) of the user 1100, in operation 1113, the first electronic device 1110 may inform external electronic devices (e.g., the second electronic device 1130 and the third electronic device 1150) of the proximity (or contact) state of the user 1100. Further, in operation 1115, the first electronic device 1110 may wake up a microphone (e.g., a microphone 120 of FIG. 1) included in the first electronic device 1110. If the microphone included in the first electronic device 1110 is already woken up, the first electronic device 1110 may maintain the wake-up state of the microphone. In operations 1131 and 1151, the second electronic device 1130 and the third electronic device 1150 may receive the proximity (or contact) state of the user 1100. Receiving the proximity (or contact) of the user 1100, in operations 1133 and 1153, the second electronic device 1130 and the third electronic device 1150 may deactivate microphones included in the second electronic device 1130 and the third electronic device 1150, respectively, so that the second electronic device 1130 and the third electronic device 1150 are prevented from detecting the voice input from the user.
In operation 1103, the user 1100 may utter (or speak) a voice input, which may include an instruction corresponding to a function to be performed by the first electronic device 1110. In operation 1117, the first electronic device 1110, having an awake microphone, may receive the voice input through the microphone and may perform voice recognition for the received voice input. As explained above, the second electronic device 1130 and the third electronic device 1150 may not detect the voice input because their microphones are deactivated. The microphones of the second electronic device 1130 and the third electronic device 1150 may be deactivated for a specified time period or until they receive a notification that the proximity state of the user 1100 has changed (e.g., the user is no longer close to or in contact with the first electronic device 1110).
In operation 1119, the first electronic device 1110 may wake up at least some of its modules related to the function specified by the user to perform that function.
FIG. 12 is a drawing illustrating a method for determining whether to wake up an electronic device through adjustment of received sensitivity of a voice input according to an embodiment.
Referring to FIG. 12, an electronic device (e.g., a first electronic device 1210 or a second electronic device 1230) may adjust receive sensitivity of a voice input received through a microphone (e.g., a first microphone 1211 or a second microphone 1231). For example, the electronic device may adjust a threshold 1205 of the received intensity of the voice input. According to an embodiment, after the adjustment, only one of the plurality of electronic devices may recognize the voice input of the user. The threshold 1205 may indicate, for example, a minimum value for preventing errors in voice recognition. Further, each electronic device may have a different threshold. For example, the electronic devices may have different thresholds depending on the performances of the electronic devices.
As shown in FIG. 12, if the received intensity of the first microphone 1211 is a first level (v1) 1201 and if the received intensity of the second microphone 1231 is a second level (v2) 1203, the threshold 1205 of the received intensity may be adjusted to a value between the first level 1201 and the second level 1203. For example, if the first level 1201 is higher than the second level 1203, the threshold 1205 may be adjusted to be lower than the first level 1201 and be higher than the second level 1203. Thus, voice inputs may be processed by the first electronic device 1210 but not the second electronic device 1230. As a result, only the first electronic device 1210 may be woken up in response to the voice input from the user.
According to an embodiment, the electronic device (e.g., the first electronic device 1210 or the second electronic device 1230) may change the type of indication shown in the indicator (e.g., the first indicator 1213 or the second indicator 1233) according to the received intensity. For example, the first electronic device 1210 in which the received intensity is the first level 1201 may display an indication of a first color, and the second electronic device 1230 in which the received intensity is the second level may display an indication of a second color different from the first color. Thus, a user may verify each indication (e.g., a first indication or a second indication) so that he or she knows which electronic device is capable of performing voice recognition.
According to an embodiment, the threshold 1205 may be set to correspond to the received intensity of a voice input including a first specified word. Thus, if the user is distant from an electronic device or changes the direction where he directs his voice input, since the received intensity of the voice input does not meet the threshold 1205, the electronic device may not process the voice input.
FIG. 13 is a flowchart illustrating an operation method of an electronic device for adjusting the received sensitivity of a voice input according to an embodiment.
Referring to FIG. 13, in operation 1310, an electronic device (e.g., an electronic device 100 or a processor 110 of FIG. 1) may receive a voice input of a user through a microphone (e.g., a microphone 120 of FIG. 1) and may analyze the received intensity of the voice input. Further, in operation 1330, the electronic device (e.g., the electronic device 100 or the processor 110) may receive information regarding the received intensity of a voice input from the external electronic device via a communication circuitry (e.g., a communication circuitry 160 of FIG. 1).
In operation 1350, the electronic device (e.g., the electronic device 100 or the processor 110) may adjust the received sensitivity of the voice input based on the received voice input and the information received from the external electronic device. According to an embodiment, in doing so, the electronic device may adjust the threshold of received intensity of a voice input. For example, if the received intensity of a voice input received by the electronic device is a first level and if received intensity of a voice received by the external electronic device is a second level, the electronic device may adjust the threshold to a value between the first level and the second level. The electronic device (e.g., the electronic device 100 or the processor 110) may perform voice recognition depending on whether the intensity of the received voice input exceeds the threshold.
FIG. 14 is a drawing illustrating a method for determining whether to wake up an electronic device based on usage history information according to an embodiment.
Referring to FIG. 14, an electronic device (e.g., a first electronic device 1410 or a second electronic device 1430) may determine whether to wake up the electronic device based on usage history information of user A 1400. According to an embodiment, the electronic device (e.g., the first electronic device 1410 or the second electronic device 1430) may perform voiceprint identification for a voice input received through a microphone (e.g., a first microphone 1411 or a second microphone 1431) to identify user A 1400. Further, the electronic device may verify history where the identified user A 1400 uses the electronic device based on usage history information (e.g., first usage history information 1415 or second usage history information 1435) stored in a database (e.g., a first database 1413 or a second database 1433) and may determine whether to wake up the electronic device based on the usage history information. The usage history information may include, for example, information associated with the frequency of use by user A 1400, the number of use by user A 1400, the difference in the frequency of user/the number of use with another user, etc. For another example, the usage history information may include information about the number of times where user A 1400 selects that specific electronic device when two or more electronic devices (e.g., the first electronic device 1410 or the second electronic device 1430) use the specified word (e.g., a wake-up word) to wake up.
As shown in the drawing, if user A 1400 speaks a voice input, only the first electronic device 1410, which is frequently or recently used by user A 1400, may be woken up. The second electronic device 1430, which is frequently or recently used by user B, may not wake up. For example, in a state where the first electronic device 1410 and the second electronic device 1430 are located within a specified distance (e.g., about 20 m), if user A 1400 is located between the first electronic device 1410 and the second electronic device 1430 and speaks a voice input including a specified word, the first electronic device 1410 and the second electronic device 1430 may receive the voice input with the same voice intensity at substantially and almost the same time. In this case, the first electronic device 1410 and the second electronic device 1430 may verify usage history information of user A 1400 to determine whether user A 1400 is a user who frequently use the first electronic device 1410 or the second electronic device 1430. If determining that user A 1400 is the user who frequently use the first electronic device 1410, the first electronic device 1410 may wake up.
FIG. 15 is a flowchart illustrating an operation method of an electronic device for determining whether to wake up an electronic device based on usage history information according to an embodiment.
Referring to FIG. 15, in operation 1510, an electronic device (e.g., an electronic device 100 or a processor 110 of FIG. 1) may receive a voice input of a user through a microphone (e.g., the microphone 120 of FIG. 1) and may perform voice identification for the received voice. According to an embodiment, the electronic device (e.g., the electronic device 100 or the processor 110) may analyze voice data corresponding to the received voice input and may convert the voice data into text. For example, the processor 110 may convert vibration energy of the voice input an electric signal, may convert the electric signal into frequency spectrum information. The electronic device may then compare the converted frequency spectrum information with a plurality of frequency spectrum information corresponding to different voiceprints of different users, thus identifying the user.
In operation 1530, the electronic device may verify usage history information of the user. According to an embodiment, the usage history information of the user may be managed through a database (e.g., the database 615 of FIG. 6). The usage history information of the user may include, for example, information associated with the frequency of use by the user, the number of use by the user, the difference in the frequency of use/the number of use with another user, etc. The usage history information may also include information on the number of times the user has selected a particular device.
In operation 1550, the electronic device (e.g., the electronic device 100 or the processor 110) may determine whether to wake up the electronic device based on the usage history information. For example, if the number of times that the user has used the electronic device is larger than the number of times that another user has used the electronic device, or if the user more recently uses the electronic device than another user, the electronic device may be woken up.
FIG. 16 is a drawing illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
Referring to FIG. 16, an electronic device (e.g., a first electronic device 1610 or a second electronic device 1630) may receive a voice input of a user 1600 through its microphone and may perform voice recognition for the received voice. Further, the electronic device may determine wake-up of the electronic device based on an instruction included in the voice input that corresponds to a function performable by the electronic device. For example, the electronic device may be woken up only if it is possible for the electronic device to perform the included function.
According to an embodiment, in operation 1601, although a plurality of electronic devices (e.g., the first electronic device 1610 and the second electronic device 1630) recognize a specified word 1601a included in a first voice of the user 1600, only one electronic device (e.g., the first electronic device 1610) may respond to the specified word 1601a and the other electronic device (e.g., the second electronic device 1630) may not wake up. This is because in operation 1603, the first electronic device 1610 determines that it may perform the specified operation 1603a but the second electronic device 1630 determines that it cannot perform the operation 1603a. If there are a plurality of electronic devices capable of performing the same function, the electronic device that is closer to the user may be designated to perform the function.
As shown in the example of FIG. 16, as the user 1600 speaks “Turn on the light,” only the first electronic device 1610 (e.g., a light) is woken up to perform the function and the second electronic device 1630 (e.g., a refrigerator) is not woken up. In some embodiments, the first electronic device 1610 and the second electronic device 1630 may be connected to hub devices, respectively. For example, the first electronic device 1610 may be connected to a first hub device, and the second electronic device 1630 may be connected to a second hub device. In this case, the first and second electronic devices 1610 and 1630 may be controlled by the first and second hub devices, respectively. The first and second hub devices may receive a voice input spoken from the user 1600 and may analyze the received voice input. The first and second hub devices then control the first and second electronic devices 1610 and 1630 based on the analyzed result.
FIG. 17 is a flowchart illustrating a method for determining whether to wake up an electronic device based on a performable function according to an embodiment.
Referring to FIG. 17, in operation 1710, an electronic device (e.g., an electronic device 100 or a processor 110 of FIG. 1) may receive a voice input of a user and may perform voice recognition for the received voice input.
In operation 1730, the electronic device (e.g., the electronic device 100 or the processor) may determine whether an instruction corresponding to a function performable by the electronic device is included in the result of performing the voice recognition.
If the instruction corresponding to the function is included, in operation 1750, the electronic device (e.g., the electronic device 100 or the processor) may wake up at least one element necessary for performing the function. If the instruction corresponds to a function that cannot be performed by the electronic device, the electronic device may not wake up.
According to an embodiment, if there are a plurality of electronic devices capable of performing the function, an electronic device for performing the function may be selected based on the location of the user, the distance between the user and each electronic device, or the like.
As described above, according to one embodiment, a voice input processing method of an electronic device may include receiving a first voice input through a microphone, performing a first voice recognition for the first voice input, if a first specified word for waking up the electronic device is included in a result of the first voice recognition, displaying an indication through the indicator, receiving a second voice input through the microphone, performing a second voice recognition for the second voice input, and if a second specified word corresponding to the indication is included in a result of the second voice recognition, waking up the electronic device.
According to one embodiment, the method may further include verifying mapping information that maps a type of at least one indication to an assignment state of the at least one indication which are stored in a memory of the electronic device, and determining the indication based on the mapping information.
According to one embodiment, the method may further include transmitting information about an another indication different from the indication to the external electronic device if a signal corresponding to the first specified word is received from an external electronic device located near the electronic device.
According to one embodiment, the method may further include adjusting a threshold of a volume level for the second voice recognition based on a first volume level of the first voice input received in the external electronic device included in the signal and a second volume level of the first voice input received through the microphone.
According to one embodiment, the method may further include determining at least one of whether a body part of a user is close to the electronic device and whether the body part of the user is in contact with the electronic device, based on sensing information obtained through a sensor module of the electronic device, and determining whether to wake up the microphone based on the determination that the body part of the user is close to the electronic device or the determination that the body part of the user is in contact with the electronic device.
According to one embodiment, the method may further include performing voiceprint identification for the first voice input to identify a user, and determining whether to wake up the electronic device based on usage history information of the user, the usage history information being stored in a memory of the electronic device.
According to one embodiment, the method may further include determining whether to wake up the electronic device based on whether an instruction corresponding to a function performable by the electronic device is included in the result of the second voice recognition.
FIG. 18 is a block diagram illustrating an electronic device 1801 in a network environment 1800 according to an embodiment. The electronic device 1801 shown in FIG. 18 may be the same or similar structure to an electronic device 100 shown in FIG. 1.
Referring to FIG. 18, an electronic device 1801 may include a bus 1810, a processor 1820, a memory 1830, an input/output interface 1850, a display 1860, and a communication interface 1870. According to an embodiment, the electronic device 1801 may not include at least one of the elements or may further include other element(s). The bus 1810 may interconnect the elements 1820 to 1870 and may include a circuit for conveying communications (e.g., a control message or data) among the elements.
The processor 1820 (e.g., a processor 110 of FIG. 1) may include one or more of a central processing unit (CPU), an application processor (AP), or a communication processor (CP). For example, the processor 1820 may perform an arithmetic operation or data processing associated with control and/or communication of at least one other elements of the electronic device 1801.
The memory 1830 (e.g., a memory 150 of FIG. 1) may include a volatile and/or nonvolatile memory. For example, the memory 1830 may store instructions or data associated with at least one other element(s) of the electronic device 1801. According to an embodiment, the memory 1830 may store software and/or a program 1840. The program 1840 may include, for example, a kernel 1841, a middleware 1843, an application programming interface (API) 1845, and/or an application program (or “an application”) 1847. At least a part of the kernel 1841, the middleware 1843, or the API 1845 may be referred to as an “operating system (OS)”.
For example, the kernel 1841 may control or manage system resources (e.g., the bus 1810, the processor 1820, the memory 1830, and the like) that are used to execute operations or functions implemented in other programs (e.g., the middleware 1843, the API 1845, and the application program 1847). Furthermore, the kernel 1841 may provide an interface that allows the middleware 1843, the API 1845, or the application program 1847 to access discrete elements of the electronic device 1801 so as to control or manage system resources.
The middleware 1843 may perform, for example, a mediation role such that the API 1845 or the application program 1847 communicates with the kernel 1841 to exchange data. Furthermore, the middleware 1843 may process one or more task requests received from the application program 1847 according to a priority. For example, the middleware 1843 may assign the priority, which makes it possible to use a system resource (e.g., the bus 1810, the processor 1820, the memory 1830, or the like) of the electronic device 1801, to at least one of the application program 1847 and may process the one or more task requests.
The API 1845 may be, for example, an interface through which the application program 1847 controls a function provided by the kernel 1841 or the middleware 1843, and may include, for example, at least one interface or function (e.g., an instruction) for a file control, a window control, image processing, a character control, or the like.
The input/output interface 1850 (e.g., a microphone 120, a speaker 130, or an indicator 140 of FIG. 1) may transmit, for example, an instruction or data, input from a user or another external device, to other element(s) of the electronic device 1801 or may output an instruction or data, received from other element(s) of the electronic device 1801, to a user or another external device.
The display 1860 may include, for example, a liquid crystal display (LCD), a light-emitting diode (LED) display, an organic LED (OLED) display, a microelectromechanical systems (MEMS) display, or an electronic paper display. The display 1860 may display, for example, various contents (e.g., a text, an image, a video, an icon, a symbol, and the like) to a user. The display 1860 may include a touch screen and may receive, for example, a touch, gesture, proximity, or hovering input using an electronic pen or a part of a user’s body.
For example, the communication interface 1870 (e.g., a communication circuitry 160 of FIG. 1) may establish communication between the electronic device 1801 and an external device (e.g., the first electronic device 1802, the second electronic device 1804, or the server 1806). For example, the communication interface 1870 may be connected to the network 1862 over wireless communication or wired communication to communicate with the external device (e.g., the second electronic device 1804 or the server 1806).
The wireless communication may include, for example, cellular communication which uses at least one of long-term evolution (LTE), LTE Advanced (LTE-A), Code Division Multiple Access (CDMA), Wideband CDMA (WCDMA), Universal Mobile Telecommunications System (UMTS), Wireless Broadband (WiBro), Global System for Mobile Communications (GSM), or the like. According to an embodiment, the wireless communication may be, for example, an element 1864 of FIG. 18. As shown, the wireless communication may include at least one of wireless fidelity (Wi-Fi), light fidelity (Li-Fi), Bluetooth, Bluetooth low energy (BLE), ZigBee, near field communication (NFC), magnetic secure transmission (MST), radio frequency (RF), or a body area network (BAN). According to an embodiment, the wireless communication may include a global navigation satellite system (GNSS). The GNSS may be, for example, a global positioning system (GPS), a global navigation satellite system (Glonass), a Beidou navigation satellite system (hereinafter referred to as “Beidou”), or an European global satellite-based navigation system (hereinafter referred to as “Galileo”). Hereinafter, in this disclosure, “GPS” and “GNSS” may be interchangeably used. The wired communication may include at least one of, for example, a universal serial bus (USB), a high definition multimedia interface (HDMI), a recommended standard-232 (RS-232), a plain old telephone service (POTS), or the like. The network 1862 may include at least one of telecommunications networks, for example, a computer network (e.g., LAN or WAN), an Internet, or a telephone network.
Each of the first and second electronic devices 1802 and 1804 may be a device of which the type is different from or the same as that of the electronic device 1801. According to various embodiments, all or a portion of operations that the electronic device 1801 will perform may be executed by another or plural electronic devices (e.g., the first electronic device 1802, the second electronic device 1804 or the server 1806). According to an embodiment, in the case where the electronic device 1801 executes any function or service automatically or in response to a request, the electronic device 1801 may not perform the function or the service internally, but, alternatively or additionally, it may request at least a portion of a function associated with the electronic device 1801 at other electronic device (e.g., the electronic device 1802 or 1804 or the server 1806). The other electronic device may execute the requested function or additional function and may transmit the execution result to the electronic device 1801. The electronic device 1801 may provide the requested function or service using the received result or may additionally process the received result to provide the requested function or service. To this end, for example, cloud computing, distributed computing, or client-server computing may be used.
FIG. 19 is a block diagram illustrating an electronic device 1901 according to one embodiment.
Referring to FIG. 19, an electronic device 1901 may include, for example, all or a part of the electronic device 100 illustrated in FIG. 1 or the electronic device 1801 illustrated in FIG. 18. The electronic device 1901 may include one or more processors (e.g., an application processor (AP)) 1910, a communication module 1920, a subscriber identification module 1924, a memory 1930, a sensor module 1940, an input device 1950, a display 1960, an interface 1970, an audio module 1980, a camera module 1991, a power management module 1995, a battery 1096, an indicator 1997, and a motor 1998.
The processor 1910 may drive, for example, an operating system (OS) or an application to control a plurality of hardware or software elements connected to the processor 1910 and may process and compute a variety of data. For example, the processor 1910 may be implemented with a System on Chip (SoC). According to an embodiment, the processor 1910 may further include a graphic processing unit (GPU) and/or an image signal processor. The processor 1910 may include at least a part (e.g., a cellular module 1921) of the elements illustrated in FIG. 19. The processor 1910 may load an instruction or data, which is received from at least one of other elements (e.g., a nonvolatile memory), into a volatile memory and process the loaded instruction or data. The processor 1910 may store result data in the nonvolatile memory.
The communication module 1920 may be configured the same as or similar to the communication interface 1870 of FIG. 18. The communication module 1920 may include the cellular module 1921, a Wi-Fi module 1923, a Bluetooth (BT) module 1925, a GNSS module 1927, a near field communication (NFC) module 1928, and a radio frequency (RF) module 1929.
The cellular module 1921 may provide, for example, voice communication, video communication, a character service, an Internet service, or the like over a communication network. According to an embodiment, the cellular module 1921 may perform discrimination and authentication of the electronic device 1901 within a communication network by using the subscriber identification module (e.g., a SIM card) 1924. According to an embodiment, the cellular module 1921 may perform at least a portion of functions that the processor 1910 provides. According to an embodiment, the cellular module 1921 may include a communication processor (CP). In some embodiments, at least some (e.g., two or more) of the cellular module 1921, the Wi-Fi module 1923, the BT module 1925, the GNSS module 1927, or the NFC module 1928 may be included within one Integrated Circuit (IC) or an IC package.
For example, the RF module 1929 may transmit and receive a communication signal (e.g., an RF signal). For example, the RF module 1929 may include a transceiver, a power amplifier module (PAM), a frequency filter, a low noise amplifier (LNA), an antenna, or the like. According to another embodiment, at least one of the cellular module 1921, the Wi-Fi module 1923, the BT module 1925, the GNSS module 1927, or the NFC module 1928 may transmit and receive an RF signal through a separate RF module.
The subscriber identification module 1924 may include, for example, a card and/or embedded SIM that includes a subscriber identification module and may include unique identify information (e.g., integrated circuit card identifier (ICCID)) or subscriber information (e.g., international mobile subscriber identity (IMSI)).
The memory 1930 (e.g., the memory 1830 of FIG. 18) may include an internal memory 1932 or an external memory 1934. For example, the internal memory 1932 may include at least one of, for example, a volatile memory (e.g., a dynamic random access memory (DRAM), a static RAM (SRAM), a synchronous DRAM (SDRAM), or the like), a nonvolatile memory (e.g., a one-time programmable read only memory (OTPROM), a programmable ROM (PROM), an erasable and programmable ROM (EPROM), an electrically erasable and programmable ROM (EEPROM), a mask ROM, a flash ROM, a flash memory, a hard drive, or a solid state drive (SSD)). The external memory 1934 may include a flash drive such as compact flash (CF), secure digital (SD), micro secure digital (Micro-SD), mini secure digital (Mini-SD), extreme digital (xD), a multimedia card (MMC), a memory stick, or the like. The external memory 1934 may be operatively and/or physically connected to the electronic device 1901 through various interfaces.
The sensor module 1940 may measure, for example, a physical quantity or may detect an operation state of the electronic device 1901. The sensor module 1940 may convert the measured or detected information to an electric signal. For example, the sensor module 1940 may include at least one of a gesture sensor 1940A, a gyro sensor 1940B, a barometric pressure sensor 1940C, a magnetic sensor 1940D, an acceleration sensor 1940E, a grip sensor 1940F, the proximity sensor 1940G, a color sensor 1940H (e.g., red, green, blue (RGB) sensor), a biometric sensor 1940I, a temperature/humidity sensor 1940J, an illuminance sensor 1940K, or an UV sensor 1940M. Although not illustrated, additionally or alternatively, the sensor module 1940 may further include, for example, an E-nose sensor, an electromyography (EMG) sensor, an electroencephalogram (EEG) sensor, an electrocardiogram (ECG) sensor, an infrared (IR) sensor, an iris sensor, and/or a fingerprint sensor. The sensor module 1940 may further include a control circuit for controlling at least one or more sensors included therein. According to an embodiment, the electronic device 1901 may further include a processor that is a part of the processor 1910 or independent of the processor 1910 and is configured to control the sensor module 1940. The processor may control the sensor module 1940 while the processor 1910 remains at a sleep state.
The input device 1950 may include, for example, a touch panel 1952, a (digital) pen sensor 1954, a key 1956, or an ultrasonic input unit 1958. For example, the touch panel 1952 may use at least one of capacitive, resistive, infrared and ultrasonic detecting methods. Also, the touch panel 1952 may further include a control circuit. The touch panel 1952 may further include a tactile layer to provide a tactile reaction to a user. The (digital) pen sensor 1954 may be, for example, a part of a touch panel or may include an additional sheet for recognition. The key 1956 may include, for example, a physical button, an optical key, or a keypad. The ultrasonic input device 1958 may detect (or sense) an ultrasonic signal, which is generated from an input device, through a microphone (e.g., a microphone 1988) and may check data corresponding to the detected ultrasonic signal.
The display 1960 (e.g., a display 1860 of FIG. 18) may include a panel 1962, a hologram device 1964, a projector 1966, and/or a control circuit for controlling the panel 1962, the hologram device 1964, or the projector 1966. The panel 1962 may be implemented, for example, to be flexible, transparent or wearable. The panel 1962 and the touch panel 1952 may be integrated into a single module. According to an embodiment, the panel 1962 may include a pressure sensor (or force sensor) that measures the intensity of touch pressure by a user. The pressure sensor may be implemented integrally with the touch panel 1952, or may be implemented as at least one sensor separately from the touch panel 1952. The hologram device 1964 may display a stereoscopic image in a space using a light interference phenomenon. The projector 1966 may project light onto a screen so as to display an image. For example, the screen may be arranged in the inside or the outside of the electronic device 1901.
The interface 1970 may include, for example, a high-definition multimedia interface (HDMI) 1972, a universal serial bus (USB) 1974, an optical interface 1976, or a D-subminiature (D-sub) 1978. The interface 1970 may be included, for example, in a communication interface 1870 illustrated in FIG. 18. Additionally or generally, the interface 1970 may include, for example, a mobile high definition link (MHL) interface, a SD card/multi-media card (MMC) interface, or an infrared data association (IrDA) standard interface.
The audio module 1980 may convert, for example, a sound and an electric signal in dual directions. At least a part of the audio module 1980 may be included, for example, in an input/output interface 1850 illustrated in FIG. 18. The audio module 1980 may process, for example, sound information that is input or output through a speaker 1982, a receiver 1984, an earphone 1986, or the microphone 1988.
For example, the camera module 1991 may capture a still image or a video. According to an embodiment, the camera module 1991 may include at least one or more image sensors (e.g., a front sensor or a rear sensor), a lens, an image signal processor (ISP), or a flash (e.g., an LED or a xenon lamp).
The power management module 1995 may manage, for example, power of the electronic device 1901. According to an embodiment, a power management integrated circuit (PMIC), a charger IC, or a battery or fuel gauge may be included in the power management module 1995. The PMIC may have a wired charging method and/or a wireless charging method. The wireless charging method may include, for example, a magnetic resonance method, a magnetic induction method or an electromagnetic method and may further include an additional circuit, for example, a coil loop, a resonant circuit, a rectifier, or the like. The battery gauge may measure, for example, a remaining capacity of the battery 1996 and a voltage, current or temperature thereof while the battery is charged. The battery 1996 may include, for example, a rechargeable battery and/or a solar battery.
The indicator 1997 may display a specific state of the electronic device 1901 or a part thereof (e.g., the processor 1910), such as a booting state, a message state, a charging state, and the like. The motor 1998 may convert an electrical signal into a mechanical vibration and may generate the following effects: vibration, haptic, and the like.
The electronic device 1901 may include, for example, a mobile TV supporting device (e.g., a GPU) capable of processing media data according to the standards of digital multimedia broadcasting (DMB), digital video broadcasting (DVB), MediaFLOTM, or the like. Each of the above-mentioned elements of the electronic device according to various embodiments of the present disclosure may be configured with one or more components, and the names of the elements may be changed according to the type of the electronic device. In various embodiments, the electronic device (e.g., the electronic device 1901) may include at least one of the above-mentioned elements, and some elements may be omitted or other additional elements may be added. Furthermore, some of the elements of the electronic device according to various embodiments may be combined with each other so as to form one entity, so that the functions of the elements may be performed in the same manner as before the combination.
FIG. 20 is a block diagram illustrating a program module 2010 according to an embodiment.
According to an embodiment, a program module 2010 (e.g., a program 1840 of FIG. 18) may include an operating system (OS) to control resources associated with an electronic device (e.g., an electronic device 1801 of FIG. 18), and/or diverse applications (e.g., an application program 1847 of FIG. 18) driven on the OS. The OS may be, for example, AndroidTM, iOSTM, WindowsTM, SymbianTM, TizenTM, or BadaTM.
Referring to FIG. 20, the program module 2010 may include a kernel 2020 (e.g., a kernel 1841 of FIG. 18), a middleware 2030 (e.g., a middleware 1843 of FIG. 18), an application programming interface (API) 2060 (e.g., an API 1845 of FIG. 18), and/or an application 2070 (e.g., an application program 1847 of FIG. 18). At least a portion of the program module 2010 may be preloaded on an electronic device or may be downloadable from an external electronic device (e.g., a first electronic device 1802, a second electronic device 1804, a server 1806, or the like of FIG. 18).
The kernel 2020 may include, for example, a system resource manager 2021 or a device driver 2023. The system resource manager 2021 may control, allocate, or retrieve system resources. According to an embodiment, the system resource manager 2021 may include a process managing unit, a memory managing unit, a file system managing unit, or the like. The device driver 2023 may include, for example, a display driver, a camera driver, a Bluetooth driver, a shared memory driver, a USB driver, a keypad driver, a Wi-Fi driver, an audio driver, or an inter-process communication (IPC) driver.
The middleware 2030 may provide, for example, a function that the application 2070 needs in common, or may provide diverse functions to the application 2070 through the API 2060 to allow the application 2070 to efficiently use limited system resources of the electronic device. According to an embodiment, the middleware 2030 may include at least one of a runtime library 2035, an application manager 2041, a window manager 2042, a multimedia manager 2043, a resource manager 2044, a power manager 2045, a database manager 2046, a package manager 2047, a connectivity manager 2048, a notification manager 2049, a location manager 2050, a graphic manager 2051, or a security manager 2052.
The runtime library 2035 may include, for example, a library module that is used by a compiler to add a new function through a programming language while the application 2070 is being executed. The runtime library 2035 may perform input/output management, memory management, or capacities about arithmetic functions.
The application manager 2041 may manage, for example, a life cycle of at least one application of the application 2070. The window manager 2042 may manage a graphic user interface (GUI) resource that is used in a screen. The multimedia manager 2043 may identify a format necessary for playing diverse media files, and may perform encoding or decoding of media files by using a codec suitable for the format. The resource manager 2044 may manage resources such as a storage space, memory, or source code of at least one application of the application 2070.
The power manager 2045 may operate, for example, with a basic input/output system (BIOS) to manage a battery or power, and may provide power information for an operation of an electronic device. The database manager 2046 may generate, search for, or modify database that is to be used in at least one application of the application 2070. The package manager 2047 may install or update an application that is distributed in the form of package file. The connectivity manager 2048 may manage, for example, wireless connection.
The notification manager 2049 may display or notify an event such as arrival message, appointment, or proximity notification in a mode that does not disturb a user. The location manager 2050 may manage location information about an electronic device. The graphic manager 2051 may manage a graphic effect that is provided to a user, or manage a user interface relevant thereto. The security manager 2052 may provide a general security function necessary for system security, user authentication, or the like.
According to an embodiment, the middleware 2030 may include a telephony manager for managing a voice or video call function of the electronic device or a middleware module that combines diverse functions of the above-described elements. According to an embodiment, the middleware 2030 may provide a module specialized to each OS kind to provide differentiated functions. Additionally, the middleware 2030 may dynamically remove a part of the preexisting elements or may add new elements thereto.
The API 2060 may be, for example, a set of programming functions and may be provided with a configuration that is variable depending on an OS. For example, in the case where an OS is the android or the iOS, it may provide one API set per platform. In the case where an OS is the Tizen, it may provide two or more API sets per platform.
The application 2070 may include, for example, applications such as a home 2071, a dialer 2072, an SMS/MMS 2073, an instant message (IM) 2074, a browser 2075, a camera 2076, an alarm 2077, a contact 2078, a voice dial 2079, an e-mail 2080, a calendar 2081, a media player 2082, an album 2083, a watch 2084, health care (e.g., measuring an exercise quantity, blood sugar, or the like) or offering of environment information (e.g., information of barometric pressure, humidity, temperature, or the like).
According to an embodiment, the application 2070 may include an information exchanging application to support information exchange between an electronic device and an external electronic device. The information exchanging application may include, for example, a notification relay application for transmitting specific information to an external electronic device, or a device management application for managing the external electronic device. For example, the notification relay application may include a function of transmitting notification information, which arise from other applications, to an external electronic device or may receive, for example, notification information from an external electronic device and provide the notification information to a user. The device management application may install, delete, or update for example, a function (e.g., turn-on/turn-off of an external electronic device itself (or a part of components) or adjustment of brightness (or resolution) of a display) of the external electronic device which communicates with the electronic device, and an application running in the external electronic device.
According to an embodiment, the application 2070 may include an application (e.g., a health care application of a mobile medical device) that is assigned in accordance with an attribute of an external electronic device. According to an embodiment, the application 2070 may include an application that is received from an external electronic device.
At least a portion of the program module 2010 may be implemented by software, firmware, hardware (e.g., the processor 1910), or a combination (e.g., execution) of two or more thereof, and may include modules, programs, routines, sets of instructions, processes, or the like for performing one or more functions.
The term “module” used herein may represent, for example, a unit including one of hardware, software and firmware or a combination thereof. The term “module” may be interchangeably used with the terms “unit”, “logic”, “logical block”, “component” and “circuit”. The “module” may be implemented mechanically or electronically. For example, the “module” may include at least one of an application-specific integrated circuit (ASIC) chip, a field-programmable gate array (FPGA), and a programmable-logic device for performing some operations, which are known or will be developed.
At least a part of devices (e.g., modules or functions thereof) or methods (e.g., operations) according to various embodiments of the present disclosure may be implemented as instructions stored in a computer-readable storage medium in the form of a program module. In the case where the instructions are performed by a processor (e.g., the processor 1820), the processor may perform functions corresponding to the instructions. The computer-readable storage medium may be, for example, the memory 1830.
A computer-readable recording medium may include a hard disk, a floppy disk, a magnetic medium (e.g., a magnetic tape), an optical medium (e.g., CD-ROM, digital versatile disc (DVD)), a magneto-optical medium (e.g., a floptical disk), or a hardware device (e.g., a ROM, a RAM, a flash memory, or the like). The program instructions may include machine language codes generated by compilers and high-level language codes that can be executed by computers using interpreters. The above-mentioned hardware device may be configured to be operated as one or more software modules for performing operations of various embodiments of the present disclosure and vice versa.
A module or a program module according to various embodiments of the present disclosure may include at least one of the above-mentioned elements, or some elements may be omitted or other additional elements may be added. Operations performed by the module, the program module or other elements according to various embodiments of the present disclosure may be performed in a sequential, parallel, iterative or heuristic way. Furthermore, some operations may be performed in another order or may be omitted, or other operations may be added.
While the present disclosure has been shown and described with reference to various embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the scope of the present disclosure as defined by the appended claims and their equivalents.
Certain aspects of the above-described embodiments of the present disclosure can be implemented in hardware, firmware or via the execution of software or computer code that can be stored in a recording medium such as a CD ROM, a Digital Versatile Disc (DVD), a magnetic tape, a RAM, a floppy disk, a hard disk, or a magneto-optical disk or computer code downloaded over a network originally stored on a remote recording medium or a non-transitory machine readable medium and to be stored on a local recording medium, so that the methods described herein can be rendered via such software that is stored on the recording medium using a general purpose computer, or a special processor or in programmable or dedicated hardware, such as an ASIC or FPGA. As would be understood in the art, the computer, the processor, microprocessor controller or the programmable hardware include memory components, e.g., RAM, ROM, Flash, etc. that may store or receive software or computer code that when accessed and executed by the computer, processor or hardware implement the processing methods described herein.
Claims (15)
- An electronic device, comprising:a microphone configured to receive a voice input;a communication circuitry configured to communicate with an external electronic device;an indicator configured to provide at least one visual indication; anda processor configured to be electrically connected with the microphone, the communication circuitry, and the indicator; anda memory configured to be electrically connected with the processor,wherein the memory stores instructions, when executed, cause the processor to:receive a first voice input through the microphone;perform a first voice recognition for the first voice input;if a first specified word for waking up the electronic device is included in a result of the first voice recognition, display a first visual indication through the indicator;receive a second voice input through the microphone;perform a second voice recognition for the second voice input; andif a second specified word corresponding to the first visual indication is included in a result of the second voice recognition, wake up the electronic device.
- The electronic device of claim 1, wherein the indicator comprises at least one of a light emitting diode (LED) or a display, andwherein the first visual indication is at least one color displayed by the indicator.
- The electronic device of claim 1, wherein the memory further include mapping information that maps a type of the at least one visual indication to an assignment state of the at least one visual indication, andwherein the instructions cause the processor to:determine the first visual indication based on the mapping information.
- The electronic device of claim 1, wherein the instructions cause the processor to:if a signal corresponding to the first specified word is received from the external electronic device located near the electronic device, transmit information about a second visual indication different from the first visual indication to the external electronic device.
- The electronic device of claim 4, wherein the signal comprises at least one of information regarding when the external electronic device received the first voice input and a first volume level of the first voice input received by the external electronic device.
- The electronic device of claim 5, wherein the instructions cause the processor to:adjust a threshold of a volume level for the second voice recognition based on the first volume level of the first voice input included in the signal and a second volume level of the first voice input received through the microphone.
- The electronic device of claim 1, further comprising:a speaker,wherein the instructions cause the processor to:output a voice output corresponding to the first visual indication through the speaker.
- The electronic device of claim 1, further comprising:a sensor module,wherein the instructions cause the processor to:determine at least one of whether a body part of a user is close to the electronic device and whether the body part of the user is in contact with the electronic device, based on sensing information obtained through the sensor module; anddetermine whether to wake up the microphone based on the determination that the body part of the user is close to the electronic device or the determination that the body part of the user is in contact with the electronic device.
- The electronic device of claim 1, wherein the memory includes usage history information of a user for the electronic device, andwherein the instructions cause the processor to:perform voiceprint identification for the first voice input to identify the user; anddetermine whether to wake up the electronic device based on the usage history information of the user.
- The electronic device of claim 1, wherein the instructions cause the processor to:determine whether to wake up the electronic device based on whether an instruction corresponding to a function performable by the electronic device is included in the result of the second voice recognition.
- A voice input processing method of an electronic device, the method comprising:receiving a first voice input through a microphone;performing a first voice recognition for the first voice input;if a first specified word for waking up the electronic device is included in a result of the first voice recognition, displaying an indication through the indicator;receiving a second voice input through the microphone;performing a second voice recognition for the second voice input; andif a second specified word corresponding to the indication is included in a result of the second voice recognition, waking up the electronic device.
- The method of claim 11, further comprising:verifying mapping information that maps a type of at least one indication to an assignment state of the at least one indication which are stored in a memory of the electronic device; anddetermining the indication based on the mapping information.
- The method of claim 11, further comprising:if a signal corresponding to the first specified word is received from an external electronic device located near the electronic device, transmitting information about an another indication different from the indication to the external electronic device.
- The method of claim 13, further comprising:adjusting a threshold of a volume level for the second voice recognition based on a first volume level of the first voice input received in the external electronic device included in the signal and a second volume level of the first voice input received through the microphone.
- The method of claim 11, further comprising:performing voiceprint identification for the first voice input to identify a user; anddetermining whether to wake up the electronic device based on usage history information of the user, the usage history information being stored in a memory of the electronic device.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP18741894.2A EP3523798A4 (en) | 2017-01-20 | 2018-01-11 | Voice input processing method and electronic device for supporting the same |
CN201880007265.0A CN110192248B (en) | 2017-01-20 | 2018-01-11 | Voice input processing method and electronic device for supporting the same |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020170009635A KR20180085931A (en) | 2017-01-20 | 2017-01-20 | Voice input processing method and electronic device supporting the same |
KR10-2017-0009635 | 2017-01-20 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2018135803A1 true WO2018135803A1 (en) | 2018-07-26 |
Family
ID=62906590
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2018/000540 WO2018135803A1 (en) | 2017-01-20 | 2018-01-11 | Voice input processing method and electronic device for supporting the same |
Country Status (5)
Country | Link |
---|---|
US (2) | US10832670B2 (en) |
EP (1) | EP3523798A4 (en) |
KR (1) | KR20180085931A (en) |
CN (1) | CN110192248B (en) |
WO (1) | WO2018135803A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10489111B2 (en) | 2017-03-28 | 2019-11-26 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
WO2020224346A1 (en) * | 2019-05-09 | 2020-11-12 | 北京京东尚科信息技术有限公司 | Control device and operation method therefor, and speech interaction device and operation method therefor |
Families Citing this family (119)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9185062B1 (en) | 2014-05-31 | 2015-11-10 | Apple Inc. | Message user interfaces for capture and transmittal of media and location content |
US9772817B2 (en) | 2016-02-22 | 2017-09-26 | Sonos, Inc. | Room-corrected voice detection |
US10095470B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Audio response playback |
US10509626B2 (en) | 2016-02-22 | 2019-12-17 | Sonos, Inc | Handling of loss of pairing between networked devices |
US9965247B2 (en) | 2016-02-22 | 2018-05-08 | Sonos, Inc. | Voice controlled media playback system based on user profile |
US10264030B2 (en) | 2016-02-22 | 2019-04-16 | Sonos, Inc. | Networked microphone device control |
US9947316B2 (en) | 2016-02-22 | 2018-04-17 | Sonos, Inc. | Voice control of a media playback system |
US9978390B2 (en) | 2016-06-09 | 2018-05-22 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US10134399B2 (en) | 2016-07-15 | 2018-11-20 | Sonos, Inc. | Contextualization of voice inputs |
US10115400B2 (en) | 2016-08-05 | 2018-10-30 | Sonos, Inc. | Multiple voice services |
US9942678B1 (en) | 2016-09-27 | 2018-04-10 | Sonos, Inc. | Audio playback settings for voice interaction |
US10181323B2 (en) | 2016-10-19 | 2019-01-15 | Sonos, Inc. | Arbitration-based voice recognition |
CN111343060B (en) | 2017-05-16 | 2022-02-11 | 苹果公司 | Method and interface for home media control |
CN107564532A (en) * | 2017-07-05 | 2018-01-09 | 百度在线网络技术(北京)有限公司 | Awakening method, device, equipment and the computer-readable recording medium of electronic equipment |
US10475449B2 (en) | 2017-08-07 | 2019-11-12 | Sonos, Inc. | Wake-word detection suppression |
US10048930B1 (en) | 2017-09-08 | 2018-08-14 | Sonos, Inc. | Dynamic computation of system response volume |
US10531157B1 (en) * | 2017-09-21 | 2020-01-07 | Amazon Technologies, Inc. | Presentation and management of audio and visual content across devices |
US10446165B2 (en) | 2017-09-27 | 2019-10-15 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US10621981B2 (en) | 2017-09-28 | 2020-04-14 | Sonos, Inc. | Tone interference cancellation |
US10051366B1 (en) | 2017-09-28 | 2018-08-14 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US10482868B2 (en) | 2017-09-28 | 2019-11-19 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US10466962B2 (en) | 2017-09-29 | 2019-11-05 | Sonos, Inc. | Media playback system with voice assistance |
US11343614B2 (en) | 2018-01-31 | 2022-05-24 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US10600408B1 (en) * | 2018-03-23 | 2020-03-24 | Amazon Technologies, Inc. | Content output management based on speech quality |
DK201870364A1 (en) | 2018-05-07 | 2019-12-03 | Apple Inc. | Multi-participant live communication user interface |
EP3769510A1 (en) | 2018-05-07 | 2021-01-27 | Apple Inc. | User interfaces for viewing live video feeds and recorded video |
US11175880B2 (en) | 2018-05-10 | 2021-11-16 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US10959029B2 (en) | 2018-05-25 | 2021-03-23 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
US10681460B2 (en) | 2018-06-28 | 2020-06-09 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
KR20200017292A (en) * | 2018-08-08 | 2020-02-18 | 삼성전자주식회사 | The Method for Recognizing Voice and the Electronic Device supporting the same |
CN112334977B (en) * | 2018-08-14 | 2024-05-17 | 华为技术有限公司 | Voice recognition method, wearable device and system |
US10461710B1 (en) | 2018-08-28 | 2019-10-29 | Sonos, Inc. | Media playback system with maximum volume setting |
US11076035B2 (en) | 2018-08-28 | 2021-07-27 | Sonos, Inc. | Do not disturb feature for audio notifications |
NO20181210A1 (en) | 2018-08-31 | 2020-03-02 | Elliptic Laboratories As | Voice assistant |
US10587430B1 (en) | 2018-09-14 | 2020-03-10 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US11258671B1 (en) * | 2018-09-18 | 2022-02-22 | Amazon Technologies, Inc. | Functionality management for devices |
US11315553B2 (en) * | 2018-09-20 | 2022-04-26 | Samsung Electronics Co., Ltd. | Electronic device and method for providing or obtaining data for training thereof |
US11024331B2 (en) | 2018-09-21 | 2021-06-01 | Sonos, Inc. | Voice detection optimization using sound metadata |
US10978062B1 (en) * | 2018-09-27 | 2021-04-13 | Amazon Technologies, Inc. | Voice-controlled device switching between modes based on speech input |
US11128792B2 (en) | 2018-09-28 | 2021-09-21 | Apple Inc. | Capturing and displaying images with multiple focal planes |
US11100923B2 (en) | 2018-09-28 | 2021-08-24 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US10692518B2 (en) | 2018-09-29 | 2020-06-23 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
KR102606789B1 (en) * | 2018-10-01 | 2023-11-28 | 삼성전자주식회사 | The Method for Controlling a plurality of Voice Recognizing Device and the Electronic Device supporting the same |
KR102393774B1 (en) | 2018-10-08 | 2022-05-03 | 에스케이텔레콤 주식회사 | Voice recognition apparatus and control method for the same |
US11899519B2 (en) | 2018-10-23 | 2024-02-13 | Sonos, Inc. | Multiple stage network microphone device with reduced power consumption and processing load |
EP3644096A1 (en) * | 2018-10-23 | 2020-04-29 | HERE Global B.V. | Radio maps for location and/or positioning services |
EP3654249A1 (en) | 2018-11-15 | 2020-05-20 | Snips | Dilated convolutions and gating for efficient keyword spotting |
JP1640131S (en) * | 2018-11-27 | 2019-09-02 | 360° image display | |
US11183183B2 (en) | 2018-12-07 | 2021-11-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11132989B2 (en) | 2018-12-13 | 2021-09-28 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US10602268B1 (en) | 2018-12-20 | 2020-03-24 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
CN109326291A (en) * | 2018-12-24 | 2019-02-12 | 深圳创维-Rgb电子有限公司 | Television mode control method, TV and storage medium |
CN111369988A (en) * | 2018-12-26 | 2020-07-03 | 华为终端有限公司 | Voice awakening method and electronic equipment |
US11315556B2 (en) | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
US10867604B2 (en) | 2019-02-08 | 2020-12-15 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
WO2020185497A1 (en) * | 2019-03-08 | 2020-09-17 | Ball Michael Robert | Security system |
EP3709194A1 (en) | 2019-03-15 | 2020-09-16 | Spotify AB | Ensemble-based data comparison |
KR102624327B1 (en) * | 2019-03-15 | 2024-01-15 | 삼성전자주식회사 | Method for location inference of IoT device, server and electronic device supporting the same |
US10943598B2 (en) * | 2019-03-18 | 2021-03-09 | Rovi Guides, Inc. | Method and apparatus for determining periods of excessive noise for receiving smart speaker voice commands |
KR102716781B1 (en) * | 2019-04-02 | 2024-10-15 | 삼성전자주식회사 | Electronic device and Method for controlling the electronic device thereof |
US11120794B2 (en) | 2019-05-03 | 2021-09-14 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
CN111913590A (en) * | 2019-05-07 | 2020-11-10 | 北京搜狗科技发展有限公司 | Input method, device and equipment |
US11363071B2 (en) | 2019-05-31 | 2022-06-14 | Apple Inc. | User interfaces for managing a local network |
US10996917B2 (en) | 2019-05-31 | 2021-05-04 | Apple Inc. | User interfaces for audio media control |
US10904029B2 (en) | 2019-05-31 | 2021-01-26 | Apple Inc. | User interfaces for managing controllable external devices |
US11620103B2 (en) | 2019-05-31 | 2023-04-04 | Apple Inc. | User interfaces for audio media control |
US10586540B1 (en) | 2019-06-12 | 2020-03-10 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US11200894B2 (en) | 2019-06-12 | 2021-12-14 | Sonos, Inc. | Network microphone device with command keyword eventing |
US11361756B2 (en) | 2019-06-12 | 2022-06-14 | Sonos, Inc. | Conditional wake word eventing based on environment |
US11810575B2 (en) * | 2019-06-12 | 2023-11-07 | Lg Electronics Inc. | Artificial intelligence robot for providing voice recognition function and method of operating the same |
KR102246661B1 (en) * | 2019-06-25 | 2021-05-03 | 엘지전자 주식회사 | Method and apparatus for selecting voice enable device |
WO2021002493A1 (en) * | 2019-07-01 | 2021-01-07 | 엘지전자 주식회사 | Intelligent gateway device, and control system comprising same |
CN110290468B (en) * | 2019-07-04 | 2020-09-22 | 英华达(上海)科技有限公司 | Virtual sound insulation communication method, device, system, electronic device and storage medium |
US11138975B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US10871943B1 (en) | 2019-07-31 | 2020-12-22 | Sonos, Inc. | Noise classification for event detection |
US11138969B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US11094319B2 (en) | 2019-08-30 | 2021-08-17 | Spotify Ab | Systems and methods for generating a cleaned version of ambient sound |
CN110660390B (en) * | 2019-09-17 | 2022-05-03 | 百度在线网络技术(北京)有限公司 | Intelligent device wake-up method, intelligent device and computer readable storage medium |
WO2021075774A1 (en) * | 2019-10-16 | 2021-04-22 | Samsung Electronics Co., Ltd. | Method for controlling iot device and electronic device therefor |
US11189286B2 (en) | 2019-10-22 | 2021-11-30 | Sonos, Inc. | VAS toggle based on device orientation |
US10785414B1 (en) * | 2019-11-04 | 2020-09-22 | Facebook Technologies, Llc | Synchronized visual indicators for multimedia device actions |
JP7426619B2 (en) * | 2019-11-05 | 2024-02-02 | パナソニックIpマネジメント株式会社 | Control system and control method |
WO2021118269A1 (en) * | 2019-12-12 | 2021-06-17 | Samsung Electronics Co., Ltd. | Electronic device and method for controlling electronic device |
US11200900B2 (en) | 2019-12-20 | 2021-12-14 | Sonos, Inc. | Offline voice control |
CN111192581A (en) * | 2020-01-07 | 2020-05-22 | 百度在线网络技术(北京)有限公司 | Voice wake-up method, device and storage medium |
US11562740B2 (en) | 2020-01-07 | 2023-01-24 | Sonos, Inc. | Voice verification for media playback |
US11556307B2 (en) | 2020-01-31 | 2023-01-17 | Sonos, Inc. | Local voice data processing |
US11308958B2 (en) | 2020-02-07 | 2022-04-19 | Sonos, Inc. | Localized wakeword verification |
CN111192591B (en) * | 2020-02-10 | 2022-12-13 | 百度在线网络技术(北京)有限公司 | Awakening method and device of intelligent equipment, intelligent sound box and storage medium |
US11328722B2 (en) * | 2020-02-11 | 2022-05-10 | Spotify Ab | Systems and methods for generating a singular voice audio stream |
US11308959B2 (en) | 2020-02-11 | 2022-04-19 | Spotify Ab | Dynamic adjustment of wake word acceptance tolerance thresholds in voice-controlled devices |
EP4115277A1 (en) * | 2020-03-05 | 2023-01-11 | Telefonaktiebolaget LM ERICSSON (PUBL) | Controlling a set of devices by a voice-controlled device |
US11308962B2 (en) * | 2020-05-20 | 2022-04-19 | Sonos, Inc. | Input detection windowing |
US11482224B2 (en) | 2020-05-20 | 2022-10-25 | Sonos, Inc. | Command keywords with input detection windowing |
US11727919B2 (en) | 2020-05-20 | 2023-08-15 | Sonos, Inc. | Memory allocation for keyword spotting engines |
CN113900714A (en) * | 2020-06-22 | 2022-01-07 | 阿里巴巴集团控股有限公司 | Electronic equipment and nearby awakening method and device thereof |
US11395252B2 (en) | 2020-07-23 | 2022-07-19 | Skyhook Wireless, Inc. | Crowd sourced RTT-based positioning |
US11698771B2 (en) | 2020-08-25 | 2023-07-11 | Sonos, Inc. | Vocal guidance engines for playback devices |
US11392291B2 (en) | 2020-09-25 | 2022-07-19 | Apple Inc. | Methods and interfaces for media control with dynamic feedback |
WO2022102888A1 (en) * | 2020-11-11 | 2022-05-19 | Samsung Electronics Co., Ltd. | Methods and systems for identifying devices and positions of devices in an iot environment |
US11984123B2 (en) | 2020-11-12 | 2024-05-14 | Sonos, Inc. | Network device interaction by range |
JP2022129867A (en) * | 2021-02-25 | 2022-09-06 | パナソニックIpマネジメント株式会社 | Method for activating voice interface, computer program, and control device |
CN112908330B (en) * | 2021-03-04 | 2022-08-09 | 深圳市云希谷科技有限公司 | Voice wake-up method and device for terminal equipment and computer readable storage medium |
CN115086096A (en) * | 2021-03-15 | 2022-09-20 | Oppo广东移动通信有限公司 | Method, apparatus, device and storage medium for responding control voice |
US11915708B2 (en) | 2021-03-18 | 2024-02-27 | Samsung Electronics Co., Ltd. | Methods and systems for invoking a user-intended internet of things (IoT) device from a plurality of IoT devices |
WO2022196896A1 (en) * | 2021-03-18 | 2022-09-22 | Samsung Electronics Co., Ltd. | Methods and systems for invoking a user-intended internet of things (iot) device from a plurality of iot devices |
US20220358917A1 (en) * | 2021-04-21 | 2022-11-10 | Meta Platforms, Inc. | Multi-device Mediation for Assistant Systems |
US11861315B2 (en) | 2021-04-21 | 2024-01-02 | Meta Platforms, Inc. | Continuous learning for natural-language understanding models for assistant systems |
US12118790B2 (en) * | 2021-04-21 | 2024-10-15 | Meta Platforms, Inc. | Auto-capture of interesting moments by assistant systems |
US11907605B2 (en) | 2021-05-15 | 2024-02-20 | Apple Inc. | Shared-content session user interfaces |
US11893214B2 (en) | 2021-05-15 | 2024-02-06 | Apple Inc. | Real-time communication user interface |
US20220368548A1 (en) | 2021-05-15 | 2022-11-17 | Apple Inc. | Shared-content session user interfaces |
CN113380246A (en) * | 2021-06-08 | 2021-09-10 | 阿波罗智联(北京)科技有限公司 | Instruction execution method, related device and computer program product |
US11770600B2 (en) | 2021-09-24 | 2023-09-26 | Apple Inc. | Wide angle video conference |
US20240265921A1 (en) * | 2021-09-30 | 2024-08-08 | Sonos, Inc. | Conflict management for wake-word detection processes |
KR20230086500A (en) * | 2021-12-08 | 2023-06-15 | 엘지전자 주식회사 | Artificial intelligence device |
WO2023163427A1 (en) * | 2022-02-28 | 2023-08-31 | 삼성전자주식회사 | Method and device for adjusting volume of electronic device |
US20230418618A1 (en) * | 2022-06-24 | 2023-12-28 | Evga Corporation | Computer intelligent switching system and method thereof |
CN117746849A (en) * | 2022-09-14 | 2024-03-22 | 荣耀终端有限公司 | Voice interaction method, device and terminal |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140207452A1 (en) * | 2013-01-24 | 2014-07-24 | Microsoft Corporation | Visual feedback for speech recognition system |
US20140343946A1 (en) * | 2013-05-14 | 2014-11-20 | Rawles Llc | Storing State Information From Network-Based User Devices |
US20150046157A1 (en) * | 2012-03-16 | 2015-02-12 | Nuance Communications, Inc. | User Dedicated Automatic Speech Recognition |
US20150088518A1 (en) * | 2012-03-08 | 2015-03-26 | Lg Electronics Inc. | Apparatus and method for multiple device voice control |
US9361885B2 (en) | 2013-03-12 | 2016-06-07 | Nuance Communications, Inc. | Methods and apparatus for detecting a voice command |
US20160300571A1 (en) * | 2014-10-09 | 2016-10-13 | Google Inc. | Hotword detection on multiple devices |
Family Cites Families (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8068881B2 (en) | 2002-08-09 | 2011-11-29 | Avon Associates, Inc. | Voice controlled multimedia and communications system |
US8498425B2 (en) | 2008-08-13 | 2013-07-30 | Onvocal Inc | Wearable headset with self-contained vocal feedback and vocal command |
US10321541B2 (en) * | 2011-03-11 | 2019-06-11 | Ilumi Solutions, Inc. | LED lighting device |
US8340975B1 (en) * | 2011-10-04 | 2012-12-25 | Theodore Alfred Rosenberger | Interactive speech recognition device and system for hands-free building control |
US9786294B1 (en) * | 2012-07-30 | 2017-10-10 | Amazon Technologies, Inc. | Visual indication of an operational state |
US9779757B1 (en) * | 2012-07-30 | 2017-10-03 | Amazon Technologies, Inc. | Visual indication of an operational state |
WO2014064324A1 (en) | 2012-10-26 | 2014-05-01 | Nokia Corporation | Multi-device speech recognition |
US9275637B1 (en) * | 2012-11-06 | 2016-03-01 | Amazon Technologies, Inc. | Wake word evaluation |
US9275642B2 (en) * | 2012-11-13 | 2016-03-01 | Unified Computer Intelligence Corporation | Voice-operated internet-ready ubiquitous computing device and method thereof |
US9047857B1 (en) * | 2012-12-19 | 2015-06-02 | Rawles Llc | Voice commands for transitioning between device states |
US9466286B1 (en) * | 2013-01-16 | 2016-10-11 | Amazong Technologies, Inc. | Transitioning an electronic device between device states |
DE112014000709B4 (en) * | 2013-02-07 | 2021-12-30 | Apple Inc. | METHOD AND DEVICE FOR OPERATING A VOICE TRIGGER FOR A DIGITAL ASSISTANT |
JP2014153663A (en) | 2013-02-13 | 2014-08-25 | Sony Corp | Voice recognition device, voice recognition method and program |
US10133546B2 (en) * | 2013-03-14 | 2018-11-20 | Amazon Technologies, Inc. | Providing content on multiple devices |
US20140376747A1 (en) * | 2013-06-20 | 2014-12-25 | Qmotion Incorporated | Voice control of lights and motorized window coverings |
US10147441B1 (en) * | 2013-12-19 | 2018-12-04 | Amazon Technologies, Inc. | Voice controlled system |
JP6282516B2 (en) | 2014-04-08 | 2018-02-21 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | Multi-device voice operation system, voice operation method, and program |
US9632748B2 (en) * | 2014-06-24 | 2017-04-25 | Google Inc. | Device designation for audio input monitoring |
US10339293B2 (en) * | 2014-08-15 | 2019-07-02 | Apple Inc. | Authenticated device used to unlock another device |
US9548053B1 (en) * | 2014-09-19 | 2017-01-17 | Amazon Technologies, Inc. | Audible command filtering |
US9812128B2 (en) * | 2014-10-09 | 2017-11-07 | Google Inc. | Device leadership negotiation among voice interface devices |
KR102299330B1 (en) * | 2014-11-26 | 2021-09-08 | 삼성전자주식회사 | Method for voice recognition and an electronic device thereof |
JP6520100B2 (en) | 2014-12-15 | 2019-05-29 | オンキヨー株式会社 | Electronic device control system, terminal device, and server |
US9811312B2 (en) * | 2014-12-22 | 2017-11-07 | Intel Corporation | Connected device voice command support |
US10770067B1 (en) * | 2015-09-08 | 2020-09-08 | Amazon Technologies, Inc. | Dynamic voice search transitioning |
CN105741838B (en) | 2016-01-20 | 2019-10-15 | 百度在线网络技术(北京)有限公司 | Voice awakening method and device |
US20170256270A1 (en) * | 2016-03-02 | 2017-09-07 | Motorola Mobility Llc | Voice Recognition Accuracy in High Noise Conditions |
US10332516B2 (en) * | 2016-05-10 | 2019-06-25 | Google Llc | Media transfer among media output devices |
DK179415B1 (en) * | 2016-06-11 | 2018-06-14 | Apple Inc | Intelligent device arbitration and control |
US10339769B2 (en) * | 2016-11-18 | 2019-07-02 | Google Llc | Server-provided visual output at a voice interface device |
US20180177029A1 (en) * | 2016-12-19 | 2018-06-21 | Pilot, Inc. | Voice-controlled light bulb |
US10672387B2 (en) * | 2017-01-11 | 2020-06-02 | Google Llc | Systems and methods for recognizing user speech |
KR20180084392A (en) * | 2017-01-17 | 2018-07-25 | 삼성전자주식회사 | Electronic device and operating method thereof |
KR20190008663A (en) * | 2017-07-17 | 2019-01-25 | 삼성전자주식회사 | Voice data processing method and system supporting the same |
US10951967B2 (en) * | 2017-08-23 | 2021-03-16 | Amazon Technologies, Inc. | Voice-controlled multimedia device and universal remote |
US10461712B1 (en) * | 2017-09-25 | 2019-10-29 | Amazon Technologies, Inc. | Automatic volume leveling |
KR102419597B1 (en) * | 2017-09-29 | 2022-07-11 | 삼성전자주식회사 | Input device, electronic device, system comprising the same and control method thereof |
-
2017
- 2017-01-20 KR KR1020170009635A patent/KR20180085931A/en active Search and Examination
-
2018
- 2018-01-11 WO PCT/KR2018/000540 patent/WO2018135803A1/en unknown
- 2018-01-11 EP EP18741894.2A patent/EP3523798A4/en active Pending
- 2018-01-11 CN CN201880007265.0A patent/CN110192248B/en active Active
- 2018-01-12 US US15/869,749 patent/US10832670B2/en active Active
-
2020
- 2020-11-09 US US17/092,516 patent/US11823673B2/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150088518A1 (en) * | 2012-03-08 | 2015-03-26 | Lg Electronics Inc. | Apparatus and method for multiple device voice control |
US20150046157A1 (en) * | 2012-03-16 | 2015-02-12 | Nuance Communications, Inc. | User Dedicated Automatic Speech Recognition |
US20140207452A1 (en) * | 2013-01-24 | 2014-07-24 | Microsoft Corporation | Visual feedback for speech recognition system |
US9361885B2 (en) | 2013-03-12 | 2016-06-07 | Nuance Communications, Inc. | Methods and apparatus for detecting a voice command |
US20140343946A1 (en) * | 2013-05-14 | 2014-11-20 | Rawles Llc | Storing State Information From Network-Based User Devices |
US20160300571A1 (en) * | 2014-10-09 | 2016-10-13 | Google Inc. | Hotword detection on multiple devices |
Non-Patent Citations (1)
Title |
---|
See also references of EP3523798A4 |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10489111B2 (en) | 2017-03-28 | 2019-11-26 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
EP3382949B1 (en) * | 2017-03-28 | 2020-07-22 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
US11372619B2 (en) | 2017-03-28 | 2022-06-28 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
US11385861B2 (en) | 2017-03-28 | 2022-07-12 | Lg Electronics Inc. | Smart controlling device and method of controlling therefor |
WO2020224346A1 (en) * | 2019-05-09 | 2020-11-12 | 北京京东尚科信息技术有限公司 | Control device and operation method therefor, and speech interaction device and operation method therefor |
Also Published As
Publication number | Publication date |
---|---|
CN110192248A (en) | 2019-08-30 |
US20180211665A1 (en) | 2018-07-26 |
US11823673B2 (en) | 2023-11-21 |
US20210056971A1 (en) | 2021-02-25 |
CN110192248B (en) | 2023-05-16 |
EP3523798A4 (en) | 2019-10-30 |
US10832670B2 (en) | 2020-11-10 |
EP3523798A1 (en) | 2019-08-14 |
KR20180085931A (en) | 2018-07-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2018135803A1 (en) | Voice input processing method and electronic device for supporting the same | |
AU2017304413B2 (en) | Electronic device and method for displaying image | |
AU2015350680B2 (en) | Power control method and apparatus for reducing power consumption | |
WO2017048000A1 (en) | Method and electronic device for providing content | |
WO2016085265A1 (en) | Method and apparatus for detecting that a device is immersed in a liquid | |
WO2018021739A1 (en) | Method for providing video content and electronic device for supporting the same | |
WO2018044055A1 (en) | Method and system for managing wireless charging | |
WO2016060400A1 (en) | Method and apparatus for managing images using a voice tag | |
WO2016035901A1 (en) | Method for recognizing iris and electronic device therefor | |
WO2016129852A1 (en) | Permission control method and electronic device operating the same | |
WO2016089142A1 (en) | Device for performing communication and method thereof | |
WO2017034166A1 (en) | Method for processing sound by electronic device and electronic device thereof | |
WO2018217066A1 (en) | Electronic device for measuring biometric information and operation method thereof | |
WO2017209446A1 (en) | Electronic device and information processing system including the same | |
WO2018236150A1 (en) | Electronic device for playing contents and operating method thereof | |
WO2018236082A1 (en) | Method for determining data of barometer sensor using data obtained from motion sensor and electronic device for the same | |
WO2017052216A1 (en) | Method for providing events corresponding to touch attributes and electronic device thereof | |
WO2017052097A1 (en) | Activity information providing method and electronic device supporting the same | |
WO2018143669A1 (en) | Method and electronic device for providing health content | |
WO2017119690A1 (en) | Electronic device and method for operating same | |
WO2018164445A1 (en) | Electronic device and method for controlling application thereof | |
WO2017026726A1 (en) | Electronic device and method for commonly using the same | |
WO2017023040A1 (en) | Screen controlling method and electronic device for supporting the same | |
WO2017142359A1 (en) | Electronic device and operation method therefor | |
WO2019009520A1 (en) | Electronic device and method for providing adsorption information of foreign substance adsorbed by camera |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18741894 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2018741894 Country of ref document: EP Effective date: 20190507 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |