JP2004354721A - Controller by speech, speech control method, and speech control program - Google Patents

Controller by speech, speech control method, and speech control program Download PDF

Info

Publication number
JP2004354721A
JP2004354721A JP2003152634A JP2003152634A JP2004354721A JP 2004354721 A JP2004354721 A JP 2004354721A JP 2003152634 A JP2003152634 A JP 2003152634A JP 2003152634 A JP2003152634 A JP 2003152634A JP 2004354721 A JP2004354721 A JP 2004354721A
Authority
JP
Japan
Prior art keywords
voice
control
data
lt
gt
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP2003152634A
Other languages
Japanese (ja)
Inventor
Kazuhiro Sato
和浩 佐藤
Original Assignee
Shimizu Corp
清水建設株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shimizu Corp, 清水建設株式会社 filed Critical Shimizu Corp
Priority to JP2003152634A priority Critical patent/JP2004354721A/en
Publication of JP2004354721A publication Critical patent/JP2004354721A/en
Application status is Pending legal-status Critical

Links

Images

Abstract

<P>PROBLEM TO BE SOLVED: To provide a controller by speech which can enhance convenience and reliability, a speech control method, and a speech control program. <P>SOLUTION: The controller by speech is equipped with microphones 30<SB>1</SB>to 30<SB>n</SB>which are disposed in correspondence to lighting fixtures 20<SB>1</SB>to 20<SB>n</SB>, i.e., control objects, and output speech data O<SB>1</SB>to O<SB>n</SB>including speech commands, a switching section 43 which outputs data by making the speech data correspondent to position data (any among the position data P<SB>1</SB>to P<SB>n</SB>) relating to the control objects when the speech data is outputted from any of the microphones among the microphones 301<SB>1</SB>to 30<SB>n</SB>and a control section 45 which specifies the control objects from the lighting fixtures 20<SB>1</SB>to 20<SB>0</SB>based on the speech data and the position data and controls the control objects based on the speech command obtained from the speech data by speech recognition. <P>COPYRIGHT: (C)2005,JPO&NCIPI

Description

[0001]
TECHNICAL FIELD OF THE INVENTION
The present invention relates to a voice control device, a voice control method, and a voice control program for controlling a control target (lighting or the like) by voice recognition, and particularly relates to a voice control device capable of improving convenience and reliability. The present invention relates to a voice control method and a voice control program.
[0002]
[Prior art]
2. Description of the Related Art Conventionally, there is a voice control device that controls a control target such as lighting or air conditioning installed indoors using a voice recognition technology. In this voice control device, when a speaker speaks a voice command (for example, “turn on living room lighting”) toward a microphone as a voice command (for example, “turn on living room lighting”), voice recognition is performed, and the living room lighting is turned on. Turned on.
[0003]
[Patent Document 1]
JP-A-10-232691
[Problems to be solved by the invention]
By the way, as described above, in the conventional voice control device, since the voice command is composed of the position of the control target (for example, a living room) and the control content (for example, with lighting), the voice command ( (The number of pronunciation characters) tends to be long, and there is a problem that the convenience for the speaker is low.
[0005]
In particular, when the user is in a panic in an emergency, part of the voice command (for example, position) is often omitted, so that the voice control device does not recognize the voice command correctly and the control is not executed. There are also reliability issues.
[0006]
The present invention has been made in view of the above, and an object of the present invention is to provide a voice control device, a voice control method, and a voice control program capable of improving convenience and reliability.
[0007]
[Means for Solving the Problems]
In order to achieve the above object, the invention according to claim 1 is provided in correspondence with a plurality of control targets, and outputs a plurality of microphones that output voice data including a voice command, and any one of the plurality of microphones When voice data is output from the microphone, output means for outputting the voice data in association with position data relating to the control target, voice recognition means for recognizing the voice command from the voice data, A control unit that specifies the control target from the plurality of control targets based on the position data; and controls the control target specified by the specification unit based on the voice command recognized by the voice recognition unit. And control means for performing the control.
[0008]
According to the present invention, when audio data including an audio command is output from any one of the plurality of microphones, the audio data is output in association with position data relating to the control target, and the audio data and the position Based on the data, after specifying the control target from a plurality of control targets, based on the voice command, based on the specified control target is determined, since the position information is not required in the voice command, The convenience of the speaker and the reliability of the control can be improved.
[0009]
The invention according to claim 2 is the voice control device according to claim 1, further comprising a registration unit that registers a relative positional relationship between the plurality of control targets, wherein the identification unit includes the voice data, the position data, And specifying the control target from the plurality of control targets based on the relative positional relationship.
[0010]
According to the present invention, the relative positional relationship between a plurality of control targets is registered, and the control target is specified from the plurality of control targets based on the audio data, the position data, and the relative positional relationship. Voice commands corresponding to the positional relationship can be used, and the convenience of the speaker can be further improved.
[0011]
According to a third aspect of the present invention, in the voice control device according to the first or second aspect, when the voice data indicates the occurrence of an emergency, the position data and the emergency data indicating the occurrence of the emergency are notified to the outside. An emergency notification means is provided.
[0012]
According to the present invention, when the voice data indicates the occurrence of an emergency, the position data and the emergency data indicating the occurrence of the emergency are notified to the outside, so that it is possible to quickly respond to the emergency.
[0013]
In addition, the invention according to claim 4 is provided when a voice data is output from any one of a plurality of microphones that are provided corresponding to a plurality of control targets and output voice data including a voice command. An output step of outputting data in association with position data relating to the control target; a voice recognition step of recognizing the voice command from the voice data; and the plurality of control targets based on the voice data and the position data. A control step of controlling the control target specified in the specifying step based on the voice command recognized in the voice recognition step. .
[0014]
According to the present invention, when audio data including an audio command is output from any one of the plurality of microphones, the audio data is output in association with position data relating to the control target, and the audio data and the position Based on the data, after specifying the control target from a plurality of control targets, based on the voice command, based on the specified control target is determined, since the position information is not required in the voice command, The convenience of the speaker and the reliability of the control can be improved.
[0015]
According to a fifth aspect of the present invention, in the voice control method according to the fourth aspect, the method further includes a registration step of registering a relative positional relationship between the plurality of control targets. And specifying the control target from the plurality of control targets based on the relative positional relationship.
[0016]
According to the present invention, the relative positional relationship between a plurality of control targets is registered, and the control target is specified from the plurality of control targets based on the audio data, the position data, and the relative positional relationship. Voice commands corresponding to the positional relationship can be used, and the convenience of the speaker can be further improved.
[0017]
According to a sixth aspect of the present invention, in the voice control method according to the fourth or fifth aspect, when the voice data indicates the occurrence of an emergency, the position data and the emergency data indicating the occurrence of the emergency are notified to the outside. An emergency notification step is included.
[0018]
According to the present invention, when the voice data indicates the occurrence of an emergency, the position data and the emergency data indicating the occurrence of the emergency are notified to the outside, so that it is possible to quickly respond to the emergency.
[0019]
According to a seventh aspect of the present invention, a computer is provided in correspondence with a plurality of control targets, and when audio data is output from any one of a plurality of microphones that output audio data including an audio command. An output unit that outputs the voice data in association with position data relating to the control object; a voice recognition unit that recognizes the voice command from the voice data; and the plurality of controls based on the voice data and the position data. A voice control program for functioning as identification means for identifying the control target from an object, and control means for controlling the control target identified by the identification means based on the voice command recognized by the voice recognition means. It is.
[0020]
According to the present invention, when audio data including an audio command is output from any one of the plurality of microphones, the audio data is output in association with position data relating to the control target, and the audio data and the position Based on the data, after specifying the control target from a plurality of control targets, based on the voice command, based on the specified control target is determined, since the position information is not required in the voice command, The convenience of the speaker and the reliability of the control can be improved.
[0021]
The invention according to claim 8 is the sound control program according to claim 7, wherein the computer causes the computer to function as registration means for registering a relative positional relationship between the plurality of control targets, and The control target is specified from the plurality of control targets based on the data, the position data, and the relative positional relationship.
[0022]
According to the present invention, the relative positional relationship between a plurality of control targets is registered, and the control target is specified from the plurality of control targets based on the audio data, the position data, and the relative positional relationship. Voice commands corresponding to the positional relationship can be used, and the convenience of the speaker can be further improved.
[0023]
According to a ninth aspect of the present invention, in the voice control program according to the seventh or eighth aspect, when the voice data indicates that an emergency has occurred, the computer executes the position data and the emergency data indicating that an emergency has occurred. It is characterized by functioning as emergency notification means for notifying outside.
[0024]
According to the present invention, when the voice data indicates the occurrence of an emergency, the position data and the emergency data indicating the occurrence of the emergency are notified to the outside, so that it is possible to quickly respond to the emergency.
[0025]
BEST MODE FOR CARRYING OUT THE INVENTION
Hereinafter, an embodiment of a voice control device, a voice control method, and a voice control program according to the present invention will be described in detail with reference to the drawings.
[0026]
FIG. 1 is a block diagram showing a configuration of an embodiment according to the present invention. In this figure, a voice control device 40 installed in a house having a living room 10 1 , a kitchen 10 2 ,..., And a children's room 10 n is illustrated.
[0027]
These living room 10 1, Kitchen 10 2, ..., to each of the children's room 10 n, the luminaire 20 1, 20 2 as a control object, ..., 20 n are installed. Microphone 30 1 is installed in a living room 10 1, sound produced in the living room 10 1 (e.g., "with a lighting" as voice command) converts the to voice data O 1.
[0028]
Microphone 30 2 is installed in the kitchen 10 2, it converts the voice generated in the kitchen 10 2 to the audio data O 2. Microphone 30 3 is installed in the children's room 10 n, it converts the voice generated by the child's room 10 n to the voice data O n.
[0029]
Voice control unit 40, based on the audio data O 1 ~ O n from the microphone 30 1 to 30 n, and a voice command by the speech recognition, the position recognition of the controlled object, for controlling the lighting fixture 20 1 to 20 n Device.
[0030]
Input port 41 1 is connected to a microphone 30 1 is a port for inputting the audio data O 1. Position data output unit 42 1 is provided in correspondence to the input port 41 1 and microphone 30 1, when the voice data O 1 is input, the luminaire 20 1 (microphone 30 1) is provided The position data P 1 representing the position (living room 10 1 ) is output.
[0031]
The input port 41 2 is connected to a microphone 30 2, a port for inputting voice data O 2. Position data output unit 42 2 is provided in correspondence to the input port 41 2 and the microphone 30 2, when the audio data O 2 is input, the luminaire 20 2 (microphone 30 2) is provided position outputs the position data P 2 representing the (kitchen 10 2).
[0032]
Input port 41 n is connected to the microphone 30 n, a port for inputting voice data O n. Position data output unit 42 n are provided corresponding to the input port 41 n and microphone 30 n, when the audio data O n is input, the luminaire 20 n (microphone 30 n) is provided The position data Pn representing the position (child room 10 n ) is output.
[0033]
The switch unit 43 has a function of controlling a connection state between the input ports 41 1 to 41 n and the output port 44. Specifically, when the position data is input from any of the position data output units among the position data output units 42 1 to 42 n , the switch unit 43 activates the corresponding input port and sets the sound from the input port. The data and the position data are output from the output port 44. When the position data is not input from the position data output unit, the switch unit 43 deactivates the corresponding input port.
[0034]
Control unit 45, audio data inputted from the output port 44, the position data, based on the position / control object table 50 and the voice command table 60, controls the control target (the luminaire 20 1 ~20 n). The operation of the control unit 45 will be described later in detail.
[0035]
The position / control target table 50 is a table indicating the correspondence between the position and the control target. As shown in FIG. 2, there are provided fields of “position data”, “position”, and “control target”. “Position data” represents the position data P 1 to P n output from the position data output units 42 1 to 42 n .
[0036]
“Position” corresponds to the position data P 1 to P n and represents a specific position (living room, kitchen,..., Child room) where the control target (microphone) is installed. "Control object" corresponds to the position data P 1 to P n, representing the luminaire 20 1 to 20 n.
[0037]
Returning to FIG. 1, the voice command table 60 is a table indicating a correspondence between voice commands issued to the microphones 30 1 to 30 n and control contents. Specifically, as shown in FIG. 3, the voice command table 60 has fields of “voice command” and “control contents”.
[0038]
The “voice command” is “turn on the light” or “turn off the light” and does not include position information as in the related art. “Control content” indicates specific control content for “voice command”.
[0039]
For example, in the case of “turn on lighting”, the lighting fixture at the position searched in the position / control target table 50 (see FIG. 2) is controlled to be turned on. On the other hand, in the case of “turn off the lighting”, the lighting fixture at the position searched in the position / control target table 50 (see FIG. 2) is turned off.
[0040]
Next, the operation of the embodiment will be described with reference to the flowchart shown in FIG. At step SA1 shown in the figure, each of the position data output unit 42 1 through 42 n, determines whether the audio data O 1 ~ O n is input, in this case, the result of determination as "No", Repeat the same judgment. In this state, all of the input ports 41 1 to 41 n are inactive.
[0041]
Then, in the living room 10 1, in order to turn on the lighting equipment 20 1 speaker, corresponding to and toward the microphone 30 1 pronounced "with the lighting", "with the illumination" from the microphone 30 1 voice data O 1 is outputted.
[0042]
Thus, audio data O 1 is input to the position data output unit 42 1, the position data output unit 42 1, the result of determination in step SA1 is "Yes". At step SA2, the position data output unit 42 1 After generating the location data P 1, and outputs it to the switch unit 43.
[0043]
At step SA3, the switch unit 43, the input port 41 1 corresponding to the position data P 1 and active. Thus, in step SA4, the audio data O 1 from the position data P 1 and the input port 41 1 is output to the control unit 45 from the output port 44.
[0044]
At step SA5, the control unit 45 performs speech recognition based on the sound data O 1. In this case, the voice "turn on the light" is recognized. In step SA6, the control unit 45 refers to the voice command table 60 (see FIG. 3) to determine whether or not the voice recognized in step SA5 is a voice command. In this case, the determination result is “Yes”. ".
[0045]
If the result of the determination in step SA6 is "No", the determination in step SA1 is made as voice other than the voice command (noise or normal conversation voice).
[0046]
In step SA7, the control unit 45 retrieves the position (in this case, the living room) and the control target (in this case, the lighting fixture 20 1 ) from the position / control target table 50 shown in FIG. 2 using the position data P 1 as a key. I do.
[0047]
In step SA8, the control unit 45 uses the voice command (in this case, “turn on lighting”), which is the result of the voice recognition in step SA5, as a key to control the voice command table 60 shown in FIG. "Turn on the lighting equipment at that position").
[0048]
In step SA9, the control unit 45, based on the controlled object found in step SA7 (living room 10 1 to set up the luminaire 20 1) and control content ( "on the luminaire at the position"), controls the Execute. Thus, the luminaire 20 1 is turned on. Thereafter, the determination in step SA1 is performed.
[0049]
As described above, according to the embodiment, (one of the audio data O 1 ~ O n) audio data including voice commands from any of the microphones of the plurality of microphones 30 1 to 30 n is output If it is, the voice data, and outputs the correlated position data relating to the controlled object, based on the audio data and position data, after identifying the control object from the luminaire 20 1 to 20 n, the voice command Since the specified control target is controlled on the basis of the position, the position information is not necessary for the voice command, so that it is possible to enhance the convenience of the speaker and the reliability of the control.
[0050]
For example, in the conventional voice control device, in the case of a voice command "turn on living room lighting", the number of pronunciation characters is eleven. On the other hand, in one embodiment, in the case of the voice command “turn on the light”, the number of pronunciation characters is reduced to eight.
[0051]
Although one embodiment according to the present invention has been described in detail with reference to the drawings, a specific configuration example is not limited to this embodiment, and a design change within a range not departing from the gist of the present invention. The present invention is also included in the present invention.
[0052]
For example, in the above-described embodiment, a program for realizing the function of the voice control device 40 shown in FIG. 1 is recorded on the computer-readable recording medium 200 shown in FIG. Each function may be realized by causing the computer 100 shown in FIG.
[0053]
The computer 100 shown in FIG. 1 includes a CPU (Central Processing Unit) 110 for executing the above-described program, an input device 120 such as a keyboard and a mouse, a ROM (Read Only Memory) 130 for storing various data, an operation parameter and the like. (Random access memory) 140 that stores a program, a reading device 150 that reads a program from the recording medium 200, and an output device 160 such as a display and a printer.
[0054]
After reading the program recorded on the recording medium 200 via the reading device 150, the CPU 110 executes the program to realize the functions described above. Note that the recording medium 200 includes an optical disk, a flexible disk, a hard disk, and the like.
[0055]
In one embodiment, a combination of position information and voice recognition may be applied to security. For example, when absent, the position data output function in the position data output units 42 1 to 42 n can be used as a security sensor.
[0056]
Further, in one embodiment, when “help!” Is pronounced by voice instead of the emergency push button, the voice control device 40 notifies the outside of emergency data and position data indicating the occurrence of an emergency. Or display on a monitor. In this case, it is possible to quickly respond to an emergency.
[0057]
Further, in one embodiment, the relative positional relationship between the control targets (microphones) is registered in the position / control target table 50, and an ambiguous expression such as "turn off the lighting of the next room" is also used as a voice command. You may comprise so that it can be used. In this case, the voice command corresponding to the relative positional relationship can be used, and the convenience of the speaker can be further improved.
[0058]
Further, in one embodiment, an example of application to a house has been described.
[0059]
【The invention's effect】
As described above, according to the first, fourth, and seventh aspects of the present invention, when audio data including an audio command is output from any one of a plurality of microphones, the control is applied to the audio data. Outputting the position data relating to the target in association with each other, and specifying the control target from a plurality of control targets based on the voice data and the position data, and then controlling the specified control target based on the voice command; Therefore, since the position information is not required for the voice command, it is possible to improve the convenience of the speaker and the reliability of the control.
[0060]
According to the invention described in claims 2, 5, and 8, the relative positional relationship between a plurality of control targets is registered, and the plurality of control targets are registered based on the audio data, the position data, and the relative positional relationship. Since the control target is specified, the voice command corresponding to the relative positional relationship can be used, and the effect that the convenience of the speaker can be further improved can be achieved.
[0061]
According to the third, sixth, and ninth aspects of the present invention, when the voice data indicates the occurrence of an emergency, the position data and the emergency data indicating the occurrence of the emergency are notified to the outside. This has the effect of being able to respond quickly.
[Brief description of the drawings]
FIG. 1 is a block diagram showing a configuration of an embodiment according to the present invention.
FIG. 2 is a diagram showing a position / control target table 50 shown in FIG.
FIG. 3 is a diagram showing a voice command table 60 shown in FIG. 1;
FIG. 4 is a flowchart illustrating the operation of the same embodiment.
FIG. 5 is a block diagram showing a configuration of a modification of the same embodiment.
[Explanation of symbols]
30 1 to 30 n Microphone 40 Voice control device 41 1 to 41 n Input port 42 1 to 42 n Position data output unit 43 Switch unit 44 Output port 45 Control unit 50 Position / control target table 60 Voice command table

Claims (9)

  1. A plurality of microphones provided corresponding to the plurality of control targets and outputting voice data including voice commands,
    When audio data is output from any one of the plurality of microphones, an output unit that outputs the audio data in association with position data related to the control target,
    Voice recognition means for recognizing the voice command from the voice data,
    Specifying means for specifying the control target from the plurality of control targets based on the audio data and the position data,
    Control means for controlling the control target specified by the specifying means, based on the voice command recognized by the voice recognition means,
    A voice control device comprising:
  2. A registration unit that registers a relative positional relationship among the plurality of control targets, wherein the specifying unit specifies the control target from the plurality of control targets based on the audio data, the position data, and the relative positional relationship. The voice control device according to claim 1, wherein:
  3. 3. The voice control device according to claim 1, further comprising: an emergency notification unit that notifies the position data and the emergency data indicating the occurrence of the emergency to the outside when the voice data indicates the occurrence of an emergency. 4.
  4. When audio data is output from any of a plurality of microphones that are provided corresponding to a plurality of control targets and output audio data including an audio command, position data relating to the control target is output to the audio data. An output step of outputting in association with;
    A voice recognition step of recognizing the voice command from the voice data,
    A specifying step of specifying the control target from the plurality of control targets based on the audio data and the position data;
    A control step of controlling the control target specified in the specifying step, based on the voice command recognized in the voice recognition step,
    A voice control method comprising:
  5. A registration step of registering a relative positional relationship between the plurality of control targets, wherein the specifying step specifies the control target from the plurality of control targets based on the audio data, the position data, and the relative positional relationship. The voice control method according to claim 4, wherein:
  6. 6. The voice control method according to claim 4, further comprising an emergency notification step of notifying the position data and the emergency data indicating the occurrence of the emergency to the outside when the audio data indicates the occurrence of an emergency.
  7. Computer
    When voice data is output from any of a plurality of microphones that are provided corresponding to a plurality of control targets and output voice data including a voice command, position data relating to the control target is included in the voice data. Output means for outputting in association with
    Voice recognition means for recognizing the voice command from the voice data,
    Specifying means for specifying the control target from the plurality of control targets based on the audio data and the position data,
    Control means for controlling the control target specified by the specifying means based on the voice command recognized by the voice recognition means,
    Voice control program to function as a.
  8. The computer causes the computer to function as a registration unit that registers a relative positional relationship between the plurality of control targets. The sound control program according to claim 7, wherein a control target is specified.
  9. 9. The computer according to claim 7, wherein when the voice data indicates the occurrence of an emergency, the computer functions as emergency notification means for notifying the position data and the emergency data indicating the occurrence of the emergency to the outside. 10. Voice control program.
JP2003152634A 2003-05-29 2003-05-29 Controller by speech, speech control method, and speech control program Pending JP2004354721A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2003152634A JP2004354721A (en) 2003-05-29 2003-05-29 Controller by speech, speech control method, and speech control program

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2003152634A JP2004354721A (en) 2003-05-29 2003-05-29 Controller by speech, speech control method, and speech control program

Publications (1)

Publication Number Publication Date
JP2004354721A true JP2004354721A (en) 2004-12-16

Family

ID=34047793

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2003152634A Pending JP2004354721A (en) 2003-05-29 2003-05-29 Controller by speech, speech control method, and speech control program

Country Status (1)

Country Link
JP (1) JP2004354721A (en)

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012063582A (en) * 2010-09-16 2012-03-29 Denso Corp In-vehicle speech recognition device
WO2017147081A1 (en) * 2016-02-22 2017-08-31 Sonos, Inc. Voice control of a media playback system
US9794720B1 (en) 2016-09-22 2017-10-17 Sonos, Inc. Acoustic position measurement
US9811314B2 (en) 2016-02-22 2017-11-07 Sonos, Inc. Metadata exchange involving a networked playback system and a networked microphone system
US9942678B1 (en) 2016-09-27 2018-04-10 Sonos, Inc. Audio playback settings for voice interaction
US9947316B2 (en) 2016-02-22 2018-04-17 Sonos, Inc. Voice control of a media playback system
US9965247B2 (en) 2016-02-22 2018-05-08 Sonos, Inc. Voice controlled media playback system based on user profile
US9978390B2 (en) 2016-06-09 2018-05-22 Sonos, Inc. Dynamic player selection for audio signal processing
US10021503B2 (en) 2016-08-05 2018-07-10 Sonos, Inc. Determining direction of networked microphone device relative to audio playback device
US10051366B1 (en) 2017-09-28 2018-08-14 Sonos, Inc. Three-dimensional beam forming with a microphone array
US10075793B2 (en) 2016-09-30 2018-09-11 Sonos, Inc. Multi-orientation playback device microphones
US10097939B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Compensation for speaker nonlinearities
US10095470B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Audio response playback
US10115400B2 (en) 2016-08-05 2018-10-30 Sonos, Inc. Multiple voice services
US10134399B2 (en) 2016-07-15 2018-11-20 Sonos, Inc. Contextualization of voice inputs
JP2018194810A (en) * 2017-05-15 2018-12-06 ネイバー コーポレーションNAVER Corporation Device controlling method and electronic apparatus
US10152969B2 (en) 2016-07-15 2018-12-11 Sonos, Inc. Voice detection by multiple devices
US10181323B2 (en) 2016-10-19 2019-01-15 Sonos, Inc. Arbitration-based voice recognition
US10264030B2 (en) 2016-02-22 2019-04-16 Sonos, Inc. Networked microphone device control
US10446165B2 (en) 2017-09-27 2019-10-15 Sonos, Inc. Robust short-time fourier transform acoustic echo cancellation during audio playback
US10445057B2 (en) 2017-09-08 2019-10-15 Sonos, Inc. Dynamic computation of system response volume
US10466962B2 (en) 2017-09-29 2019-11-05 Sonos, Inc. Media playback system with voice assistance
US10475449B2 (en) 2017-08-07 2019-11-12 Sonos, Inc. Wake-word detection suppression
US10482868B2 (en) 2017-09-28 2019-11-19 Sonos, Inc. Multi-channel acoustic echo cancellation
US10573321B1 (en) 2018-09-25 2020-02-25 Sonos, Inc. Voice detection optimization based on selected voice assistant service
US10586540B1 (en) 2019-06-12 2020-03-10 Sonos, Inc. Network microphone device with command keyword conditioning
US10587430B1 (en) 2018-09-14 2020-03-10 Sonos, Inc. Networked devices, systems, and methods for associating playback devices based on sound codes
US10587928B2 (en) 2013-01-23 2020-03-10 Sonos, Inc. Multiple household management
US10593331B2 (en) 2018-11-15 2020-03-17 Sonos, Inc. Contextualization of voice inputs

Cited By (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012063582A (en) * 2010-09-16 2012-03-29 Denso Corp In-vehicle speech recognition device
US10587928B2 (en) 2013-01-23 2020-03-10 Sonos, Inc. Multiple household management
CN109076285A (en) * 2016-02-22 2018-12-21 搜诺思公司 Acoustic frequency response playback
US9772817B2 (en) 2016-02-22 2017-09-26 Sonos, Inc. Room-corrected voice detection
WO2017147075A1 (en) * 2016-02-22 2017-08-31 Sonos, Inc. Audio response playback
US9811314B2 (en) 2016-02-22 2017-11-07 Sonos, Inc. Metadata exchange involving a networked playback system and a networked microphone system
WO2017147081A1 (en) * 2016-02-22 2017-08-31 Sonos, Inc. Voice control of a media playback system
US10499146B2 (en) 2016-02-22 2019-12-03 Sonos, Inc. Voice control of a media playback system
US9947316B2 (en) 2016-02-22 2018-04-17 Sonos, Inc. Voice control of a media playback system
US9965247B2 (en) 2016-02-22 2018-05-08 Sonos, Inc. Voice controlled media playback system based on user profile
US10555077B2 (en) 2016-02-22 2020-02-04 Sonos, Inc. Music service selection
US10509626B2 (en) 2016-02-22 2019-12-17 Sonos, Inc Handling of loss of pairing between networked devices
US10409549B2 (en) 2016-02-22 2019-09-10 Sonos, Inc. Audio response playback
US10365889B2 (en) 2016-02-22 2019-07-30 Sonos, Inc. Metadata exchange involving a networked playback system and a networked microphone system
US10264030B2 (en) 2016-02-22 2019-04-16 Sonos, Inc. Networked microphone device control
US10097939B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Compensation for speaker nonlinearities
US10097919B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Music service selection
US10095470B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Audio response playback
US10225651B2 (en) 2016-02-22 2019-03-05 Sonos, Inc. Default playback device designation
US10212512B2 (en) 2016-02-22 2019-02-19 Sonos, Inc. Default playback devices
CN109076284A (en) * 2016-02-22 2018-12-21 搜诺思公司 The voice control of media playback system
US10142754B2 (en) 2016-02-22 2018-11-27 Sonos, Inc. Sensor on moving component of transducer
US9820039B2 (en) 2016-02-22 2017-11-14 Sonos, Inc. Default playback devices
US10332537B2 (en) 2016-06-09 2019-06-25 Sonos, Inc. Dynamic player selection for audio signal processing
US9978390B2 (en) 2016-06-09 2018-05-22 Sonos, Inc. Dynamic player selection for audio signal processing
US10297256B2 (en) 2016-07-15 2019-05-21 Sonos, Inc. Voice detection by multiple devices
US10134399B2 (en) 2016-07-15 2018-11-20 Sonos, Inc. Contextualization of voice inputs
US10152969B2 (en) 2016-07-15 2018-12-11 Sonos, Inc. Voice detection by multiple devices
US10021503B2 (en) 2016-08-05 2018-07-10 Sonos, Inc. Determining direction of networked microphone device relative to audio playback device
US10115400B2 (en) 2016-08-05 2018-10-30 Sonos, Inc. Multiple voice services
US10565999B2 (en) 2016-08-05 2020-02-18 Sonos, Inc. Playback device supporting concurrent voice assistant services
US10565998B2 (en) 2016-08-05 2020-02-18 Sonos, Inc. Playback device supporting concurrent voice assistant services
US10354658B2 (en) 2016-08-05 2019-07-16 Sonos, Inc. Voice control of playback device using voice assistant service(s)
US10034116B2 (en) 2016-09-22 2018-07-24 Sonos, Inc. Acoustic position measurement
US9794720B1 (en) 2016-09-22 2017-10-17 Sonos, Inc. Acoustic position measurement
US10582322B2 (en) 2016-09-27 2020-03-03 Sonos, Inc. Audio playback settings for voice interaction
US9942678B1 (en) 2016-09-27 2018-04-10 Sonos, Inc. Audio playback settings for voice interaction
US10075793B2 (en) 2016-09-30 2018-09-11 Sonos, Inc. Multi-orientation playback device microphones
US10117037B2 (en) 2016-09-30 2018-10-30 Sonos, Inc. Orientation-based playback device microphone selection
US10313812B2 (en) 2016-09-30 2019-06-04 Sonos, Inc. Orientation-based playback device microphone selection
US10181323B2 (en) 2016-10-19 2019-01-15 Sonos, Inc. Arbitration-based voice recognition
JP2018194810A (en) * 2017-05-15 2018-12-06 ネイバー コーポレーションNAVER Corporation Device controlling method and electronic apparatus
US10475449B2 (en) 2017-08-07 2019-11-12 Sonos, Inc. Wake-word detection suppression
US10445057B2 (en) 2017-09-08 2019-10-15 Sonos, Inc. Dynamic computation of system response volume
US10446165B2 (en) 2017-09-27 2019-10-15 Sonos, Inc. Robust short-time fourier transform acoustic echo cancellation during audio playback
US10511904B2 (en) 2017-09-28 2019-12-17 Sonos, Inc. Three-dimensional beam forming with a microphone array
US10051366B1 (en) 2017-09-28 2018-08-14 Sonos, Inc. Three-dimensional beam forming with a microphone array
US10482868B2 (en) 2017-09-28 2019-11-19 Sonos, Inc. Multi-channel acoustic echo cancellation
US10466962B2 (en) 2017-09-29 2019-11-05 Sonos, Inc. Media playback system with voice assistance
US10587430B1 (en) 2018-09-14 2020-03-10 Sonos, Inc. Networked devices, systems, and methods for associating playback devices based on sound codes
US10573321B1 (en) 2018-09-25 2020-02-25 Sonos, Inc. Voice detection optimization based on selected voice assistant service
US10593331B2 (en) 2018-11-15 2020-03-17 Sonos, Inc. Contextualization of voice inputs
US10602268B1 (en) 2018-12-20 2020-03-24 Sonos, Inc. Optimization of network microphone devices using noise classification
US10586540B1 (en) 2019-06-12 2020-03-10 Sonos, Inc. Network microphone device with command keyword conditioning

Similar Documents

Publication Publication Date Title
CN105323648B (en) Caption concealment method and electronic device
US9939891B2 (en) Voice dictation systems using earpiece microphone system and method
US9875081B2 (en) Device selection for providing a response
JP6475386B2 (en) Device control method, device, and program
EP3179474B1 (en) User focus activated voice recognition
US20190035399A1 (en) Method and apparatus for executing voice command in electronic device
US20150222977A1 (en) Awareness intelligence headphone
US9600037B2 (en) Housing as an I/O device
CN103310785B (en) Use the electronic installation and method of speech recognition controlled power supply
Wölfel et al. Distant speech recognition
JP3674990B2 (en) Speech recognition dialogue apparatus and speech recognition dialogue processing method
US6219645B1 (en) Enhanced automatic speech recognition using multiple directional microphones
US7438414B2 (en) Gaze discriminating electronic control apparatus, system, method and computer program product
CN1198203C (en) Communication terminal controlled by contact screen and voice recognition and its instruction execution method
US8321219B2 (en) Systems and methods of performing speech recognition using gestures
KR101946364B1 (en) Mobile device for having at least one microphone sensor and method for controlling the same
EP0702351B1 (en) Method and apparatus for analysing audio input events in a speech recognition system
CN102033776B (en) A kind of audio frequency playing method and a kind of calculating equipment
CN100508029C (en) Controlling an apparatus based on speech
KR20160009344A (en) Method and apparatus for recognizing whispered voice
US8421932B2 (en) Apparatus and method for speech recognition, and television equipped with apparatus for speech recognition
JP4788246B2 (en) Input device and input method
US5893063A (en) Data processing system and method for dynamically accessing an application using a voice command
CN104247280A (en) Voice-controlled communication connections
KR100677613B1 (en) Method for controlling operation of multimedia device and apparatus therefore