US20150086034A1 - Audio Routing System for Routing Audio Data to and from a Mobile Device - Google Patents
Audio Routing System for Routing Audio Data to and from a Mobile Device Download PDFInfo
- Publication number
- US20150086034A1 US20150086034A1 US14/065,506 US201314065506A US2015086034A1 US 20150086034 A1 US20150086034 A1 US 20150086034A1 US 201314065506 A US201314065506 A US 201314065506A US 2015086034 A1 US2015086034 A1 US 2015086034A1
- Authority
- US
- United States
- Prior art keywords
- audio
- audio unit
- unit
- routing
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L12/2838—Distribution of signals within a home automation network, e.g. involving splitting/multiplexing signals to/from different paths
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/12—Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/165—Management of the audio stream, e.g. setting of volume, audio stream path
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L12/00—Data switching networks
- H04L12/28—Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
- H04L12/2803—Home automation networks
- H04L2012/2847—Home automation networks characterised by the type of home appliance used
- H04L2012/2849—Audio/video appliances
Definitions
- the present disclosure relates generally to mobile devices and speaker phone extension devices and more particularly to voice recognition systems and routing audio data between devices.
- voice command capability to enable response to a user's voice commands for control of voice operated dialing or other features.
- the voice command system may be capable of waking the mobile device from a low-power sleep mode. This may be accomplished in such existing mobile devices, by having the voice command system active even during the sleep mode and always listening for a specific command word or command phrase. When an appropriate command word or command phrase is recognized, the voice command system wakes the mobile device from sleep mode and appropriately responds to the command. Thus such mobile devices allow for hands-free interaction through voice control.
- One challenge for mobile device voice command systems is in obtaining a clear audio input to the mobile device microphone or microphone array. For example, if the user is too far from the mobile device, or if the mobile device is in a pocket, purse or wedged between sofa cushions such that the microphones are obstructed, the user's voice command will either not be picked up by the microphones or will not be recognizable by the voice command system.
- One way to overcome this challenge is for the user to wear a separate, remote microphone ensemble such as a headset or earpiece that is wirelessly paired to the mobile device.
- the drawbacks to this approach include the limited battery charge time of the headset or earpiece and the encumbrance of wearing the ensemble.
- Another drawback is that the ensemble must be kept within relative close proximity to the mobile device so as to maintain a wireless link. Also, increasing the distance between the mobile device and the ensemble may further drain the batteries of both the ensemble and the mobile device.
- FIG. 1 is a floor plan showing audio units positioned at various locations within various rooms and routing audio data in accordance with the embodiments.
- FIG. 2 is diagram showing various views of an audio unit in accordance with an embodiment.
- FIG. 3A and FIG. 3B are schematic block diagrams that provide examples of various types of audio units that may be used in accordance with various embodiments.
- FIG. 4 is a schematic block diagram of a mobile device in accordance with various embodiments.
- FIG. 5 is a flow chart of a method of operation of an audio routing system in accordance with various embodiments.
- FIG. 6 is a flow chart of a method of operation of a master audio unit in accordance with an embodiment.
- FIG. 7 is a flow chart of a method of operation of an audio routing system in accordance with various embodiments.
- FIG. 8 is a flow chart of a method of operation of an audio routing system in accordance with various embodiments.
- FIG. 9 is a flow chart of a method of operation for handling an incoming phone call by an audio routing system in accordance with various embodiments.
- FIG. 10 is a flow chart providing further details of one operation block of the flow chart of FIG. 8 in accordance with some embodiments.
- FIG. 11 is a flow chart of a method of operation of an audio routing system in accordance with some embodiments.
- FIG. 12 is a flow chart of a method of operation for a “find me feature” (which is a mobile device location feature) as well as other features of the audio routing system in accordance with various embodiments.
- the disclosed embodiments provide, among other things, an audio routing system having features such that a user can transition from room to room in a house, office, etc., and speak voice commands and hear responses without the need for keeping the mobile device in vocal range at all times.
- the audio routing system includes a plurality of audio units and may include a master audio unit in some embodiments.
- Each audio unit includes a speaker and at least one microphone, and plugs directly into a standard wall AC electrical outlet. This allows the audio unit to receive electrical power anytime it is needed and overcomes one drawback of portable electronic devices, namely, discharged batteries.
- Each audio unit also includes wireless data capability or power-line carrier data capability and may communicate with other audio units or with a master audio unit. The audio units may therefore be installed by plugging them into electrical outlets throughout a user's home such as in the kitchen, bathrooms, offices, bedrooms, or living rooms of the home.
- a method of operation includes receiving sound by a first audio unit installed in an electrical outlet, routing audio data corresponding to the received sound from the first audio unit to a second audio unit installed in a second electrical outlet, and sending the audio data to a mobile device using a wireless link between the mobile device and the second audio unit.
- routing the audio data may include receiving the audio data from the first audio unit by a third audio unit, and routing the audio data to the second audio unit by the third audio unit serving as a router. Further, in some embodiments, routing the audio data may include routing the audio data using one of table drive routing or on-demand routing.
- the method of operation may further include performing voice recognition on the audio data and detecting a command word, and routing command word data to the second audio unit.
- routing the audio data to the second audio unit may include sending the command word data from the first audio unit to the third audio unit using a wireless link between the first audio unit and the third audio unit.
- routing the audio data to the second audio unit may be accomplished by sending the audio data from the first audio unit to the third audio unit using a power line carrier connection between the first audio unit and the third audio unit.
- a method of operation includes receiving sound by a group of audio units installed in a corresponding group of electrical outlets. The method further includes selecting a first audio unit from the group of audio units in response to determining that the first audio unit meets a threshold with respect to processing the received sound. Audio data corresponding to the received sound is then routed from the first audio unit installed in a first electrical outlet to a second audio unit installed in a second electrical outlet, and is sent to a mobile device over a wireless communication link between the second audio unit and the mobile device.
- the method of operation further includes performing voice recognition on the audio data and detecting a command word, and routing command word data to the second audio unit.
- performing voice recognition may involve performing voice recognition by one of the first audio unit or a master audio unit serving as a router between the first audio unit and the second audio unit. In that case, the audio data may be routed to the second audio unit only in response to detecting the command word.
- selecting a first audio unit may include a master unit that receives audio data from each audio unit of the group of audio units, and selects the first audio unit based on comparison of one of an audio signal level or a voice confidence metric for a command word detected in the audio data. The metrics are compared for audio data received by the master unit from each of the audio units.
- routing audio data may involve routing the audio data from the first audio unit to the second audio unit using a wireless link between the first audio unit and the second audio unit.
- routing audio data may involve routing the audio data from the first audio unit to the second audio unit using a power line carrier connection between the first audio unit and the second audio unit over an electrical wiring system connected to a first electrical outlet and the second electrical outlet.
- the disclosed embodiments include an audio unit.
- the audio unit includes a housing that has a two-prong alternating current power plug with the two-prongs extending from the housing, and a transformer and a voltage rectifier installed within the housing and coupled to the power plug.
- the voltage rectifier is operative to provide direct current to an internal power bus, which is operatively coupled to the voltage rectifier.
- the audio unit also includes a microphone, a speaker and audio signal conditioning logic, operatively coupled to the microphone and speaker.
- a first transceiver, operative to establish a connection with a second audio unit is operatively coupled to a controller.
- the controller is operative to control the transceiver to send audio data the second audio unit.
- the audio unit also includes a wireless transceiver that is operatively coupled to the controller, and that is operative to establish a wireless connection with a mobile device.
- the audio unit may further include voice recognition logic.
- the voice recognition logic is operatively coupled to the audio signal conditioning logic and to the controller, and is operative to detect a command word in audio data output by the audio signal conditioning logic.
- the controller may be further operative to control the wireless transceiver to send command word data to the mobile device, and may be operative to control the first transceiver to send command word data to the second audio unit.
- the controller is also operative to control the wireless transceiver to receive audio data from the mobile device, and operative to control the first transceiver to receive audio data from the second audio unit and to operate the microphone, speaker and audio signal conditioning logic to provide a speakerphone operation with the mobile device.
- the first transceiver may be a wireless transceiver in some embodiments, or may be a power line carrier transceiver.
- FIG. 1 illustrates a floor plan showing an example audio routing system 100 including various audio units ( 200 A through 200 F) positioned at various locations within various rooms in accordance with the embodiments.
- audio unit 200 A is installed in an electrical outlet within room A which is the same room in which mobile device 102 is located.
- the mobile device 102 is paired to the audio unit 200 A using a wireless link 105 .
- audio unit 200 B 1 and audio unit 200 B 2 are installed in respective electrical outlets that are positioned on walls at a 90° angle with respect to each other.
- the audio routing system 100 also includes audio unit 200 C, audio unit 200 D, audio unit 200 E and audio unit 200 F which are each installed in corresponding respective electrical outlets within various rooms. Additional audio units may also be installed on other floors of a multi-level building. One of the audio units, such as audio unit 200 E, may be installed in an electrical outlet in a relatively centrally located position such as the hallway shown in FIG. 1 and may serve as a “master” audio unit and router in some embodiments. For example, audio data may be routed between audio unit 200 B 1 and audio unit 200 A using the master audio unit 200 E as a router.
- the audio unit 200 B 1 will establish a communication link 107 with the master audio unit 200 E via an appropriate routing path, and the master audio unit 200 E will establish a second communication link 109 with audio unit 200 A via another appropriate routing path.
- These routing paths may be facilitated by peer-to-peer wireless connections between various audio units, or by power-line carrier connections over a building electrical system to which the various AC power outlets are connected.
- the various audio units located in the various rooms serve to extend the audio features and capabilities of the mobile device 102 including voice command features.
- the user 101 may leave the mobile device 102 anywhere within a home or office and still be able to access and control the mobile device 102 using voice commands from other rooms where audio units are installed.
- a user 101 may speak a voice command which may be a query. For example, the user may ask the mobile device 102 “Where are you?” The user's voice creates sound waves 103 and sound waves 104 which may be picked up by microphones within various audio units.
- the user's voice command query will be routed by one of the audio units within a room depending on criteria such as, but not limited to, the “best” received audio signal level, the best voice recognition confidence metric, or some other suitable criteria. For example, assuming the audio unit 200 B 1 is selected as the best audio unit to transmit the user's voice command query, then the audio unit 200 B 1 will establish a routing path 113 to the audio unit 200 A. As mentioned above, the routing may be facilitated by a master audio unit 200 E in some embodiments. The audio unit 200 A will send audio data, or in some embodiments the command query data output from voice recognition logic, to the mobile device 102 using the wireless link 105 .
- the mobile device 102 may send audio report data 106 to the audio unit 200 A using the wireless link 105 .
- the audio unit 200 A will then route the audio report data 106 to the audio unit 200 B 1 using an appropriate routing path 113 .
- the audio unit 200 B 1 will then play an audio report 108 such as, for example, “I′m in another room!!”
- the mobile device 102 may also provide an audio report 111 from its own speaker such as “I′m here!!” such that the user 101 may locate the mobile device 102 as the user 101 moves throughout the various rooms.
- the audio unit 200 A may use the communication link 109 to send the audio report data 106 to the master audio unit 200 E for routing on to the audio unit 200 B 1 using communication link 107 .
- the routing path 113 is facilitated through the master audio unit 200 E.
- a master audio unit may not be present and various ad-hoc like routing protocols may be employed such as, but not limited to, table-driven routing, on-demand routing, or some other suitable routing protocol.
- a routing table may be generated by either one of the audio units or by a distributed function distributed among the various audio units such that the routing table is stored within each of the various audio units.
- a new routing table would be generated based on the new audio routing system 100 topology and the new routing table would be stored in each of the audio units remaining in service.
- one or more audio units receiving sound would broadcast route requests until a route to the audio unit having a wireless link 105 to the mobile device 102 is obtained. In this case, the route may be facilitated directly between two audio units or through any suitable number of audio units.
- communication link 107 and communication link 109 may be wireless links similar to wireless link 105 .
- communication link 107 and communication link 109 may be facilitated by power-line carrier communication links over the electrical system wiring to which each of the electrical outlets is connected.
- power-line carrier case, direct connectivity may be established between two audio units or data may be routed using audio units connected along the pathway.
- audio unit 200 B 1 may be unable to establish a wireless link directly with audio unit 200 A due to impedance cause by various walls or other objects such as furniture that block the wireless path between the two audio units.
- audio unit 200 D is able to communicate wirelessly with audio unit 200 C
- audio unit 200 C is able to communicate wirelessly with audio unit 200 B 2
- audio unit 200 B 2 may be used as an initial routing point for audio data sent wirelessly from audio unit 200 B 1 .
- audio unit 200 B 2 may be used as the receiving point for the user's voice or voice commands based on audio unit 200 B 2 being able to establish a wireless routing path to audio unit 200 A whereas audio unit 200 B 1 is unable to communicate with audio unit 200 A directly.
- FIG. 2 is diagram showing various views of one example of an audio unit 200 in accordance with an embodiment. It is to be understood however that the example audio unit 200 is only one possible physical configuration of an audio unit and is not to be construed as a limitation on the particular form that an audio unit may take in accordance with the various embodiments. Some example features of an audio unit are shown in FIG. 2 that may be present in audio units having different physical configurations different than those shown in the example of FIG. 2 .
- the audio unit 200 includes a housing 201 which has an AC power plug 205 installed within the housing 201 and projecting therefrom such that the audio unit 200 may be installed in a standard AC electrical outlet such as found in a home or office building etc.
- the prongs of the AC power plug 205 may pivot or rotate such that they may be stored within an indentation (not shown) in the housing 201 for storing the AC power plug 205 when the audio unit 200 is not in use.
- the housing 201 includes a speaker port 207 and one or more microphone ports 209 .
- the housing 201 includes a status LED 211 .
- a user interface may also be present in some embodiments such as, for example, control button 202 and control button 203 .
- the control buttons 202 and 203 may be mechanical pushbutton type controls or touch sensitive controls in some embodiments such that there are no mechanical switch mechanisms involved.
- the example audio unit 200 is light weight and portable and may be similar to an AC charger housing such as accessory AC chargers provided with a cord to attach to a mobile device for purposes of charging a battery within the mobile device. However as shown in FIG. 2 , the housing 201 does not include any attached cord with respect to the example audio unit 200 .
- control buttons 202 and 203 may be a volume button (to adjust response volume) and a mute button.
- the mute button may silence audio coming from an audio unit or may turn off the microphone such that the audio unit will no longer receive the user's voice commands.
- the mute button may serve as an “unmute” button if audio is being played from a first audio unit, but the user desires audio to be played from one or more additional audio units. For example, a user may speak a command such as “shuffle music” in proximity of a kitchen audio unit. In that case, the mobile device would start playing music through the kitchen audio unit. If the user then moved to a different room, the user could press the mute/unmute button on a living room audio unit to playback the music through the living room audio unit as well.
- FIG. 3A and FIG. 3B are schematic block diagrams that provide examples of various types of audio units that may be used in accordance with various embodiments.
- FIG. 3A illustrates various embodiments of a basic audio unit 300 A
- FIG. 3B illustrates various embodiments of a master audio unit 300 B.
- Components that may be present in only some embodiments are shown using dotted lines to emphasize that such components may or may not be present depending on the particular embodiment. It is to be understood that with respect to FIG. 3A and FIG. 3B , like reference numerals that are reused between FIG. 3 and FIG. 3B indicate like or similar components.
- the audio unit 300 A or 300 B components may include, but are not limited to, a user interface 307 (which may include control buttons 202 and 203 ), a controller 311 A (or controller/router 311 B in master audio unit 300 B), non-volatile, non-transitory memory 312 , a wireless transceiver 313 , a wireless or power-line carrier transceiver 315 , and an audio signal conditioner 301 .
- One or more microphones 303 (such as a microphone array) and a speaker 305 are operatively coupled to the audio signal conditioner 301 .
- the master audio unit 300 B may or may not include the audio signal conditioner 301 and the corresponding one or more microphones 303 and speaker 305 .
- the audio signal conditioner 301 when present in an embodiment, provides audio signal conditioning logic that may include analog-to-digital converters (ADCs), digital-to-analog converters (DACs), high-pass filters, low-pass filters, band-pass filters, or adjustable band filters that may be applied to filter noise from received audio.
- the audio signal conditioner 301 may include gain including automatic gain control (AGC) in some embodiments.
- AGC automatic gain control
- the audio signal conditioner 301 may also include an echo canceller in some embodiments, to prevent echoing and feedback between the microphones 303 and the speaker 305 during full-duplex communication such as when the audio unit is used as a speakerphone.
- a status LED 323 is operatively coupled to switch logic 321 which is further operatively coupled to the internal communication bus 327 .
- the switch logic 321 is operative to receive control commands from the controller 311 A (or controller/router 311 B) and to light the LED 323 in response to the control commands accordingly.
- the LED 323 may provide one or more indications and may be a multi-color LED in some embodiments. For example, the LED may light to show when the audio unit is paired to another audio unit or to a mobile device. The LED may also provide an indication that the audio unit is in use.
- An AC power plug 319 is connected to a transformer/rectifier 317 which is further connected to an internal power bus (not shown) that provides DC power to the internal components of the audio unit.
- the wireless or power-line carrier transceiver 315 is a power-line transceiver
- the power-line transceiver is operatively coupled to the AC power plug 319 using appropriate safety and filtering circuitry (not shown) so that audio data including voice command data may be sent and received over and electrical system when the AC power plug 319 is inserted into a standard electrical outlet and is thus connected to an electrical wiring system.
- the audio units 300 A or 300 B may include voice recognition logic 309 , and may include a motion sensor 325 .
- the controller 311 A or the controller/router 311 B will be operative to obtain data from the motion sensor 325 and to determine motion in proximity to the respective audio unit.
- the memory 312 may store a library of command words or command phrases (i.e. a vocabulary) that are to be recognized by the voice recognition logic 309 .
- the voice recognition logic 309 may include an integrated memory that stores the command word or command phrase vocabulary.
- the basic audio unit 300 A may include wireless transceiver 313 which is operative to communicate over wireless link 105 with mobile device 102 .
- any basic audio unit may communicate with the mobile device 102 such as audio unit 200 A shown in the example of FIG. 2 .
- the basic audio unit 300 A does not include wireless transceiver 313
- only the master audio unit 300 B will be capable of establishing the wireless link 105 with the mobile device 102 .
- the basic audio units 300 A communicate with each other only, or with the master audio unit 300 B only, by using wireless or power-line carrier transceiver 315 .
- the master audio unit 300 B may also include a docking connector 331 which may be used to connect the mobile device 102 for purposes of charging and also, in some embodiments, for providing a data connection to the mobile device 102 by way of the internal communication bus 327 . That is, the docking connector 331 may be operatively coupled to the internal communication bus 327 and to the power bus (not shown).
- all of the audio units may be include the components of a “master” audio unit 300 B where all audio units include audio signal conditioner 301 , one or more microphones 303 and a speaker 305 . In that case, any audio unit may be designated as a master audio unit while other audio units function only as basic audio units or as slave devices.
- the basic audio unit 300 A may omit some components such voice recognition logic 309 , wireless transceiver 313 , motion sensor 325 or some combination thereof in order to reduce the cost of the basic audio unit 300 A.
- the master audio unit 300 B may omit the audio signal conditioner 301 , microphones 303 and speaker 305 in order to reduce the cost of the master audio unit 300 B.
- the location of the voice recognition logic 309 may depend on whether a central routing function is used or not. That is, if a master audio unit 300 B is used to perform a data routing function, it may also include the voice recognition logic 309 and may perform voice recognition on audio data received from the various basic audio units.
- the basic audio units 300 A may include voice recognition logic 309 , and the master audio unit 300 B may omit this component.
- the basic audio units 300 A need only send a data output from the voice recognition logic 309 upon detection of command words or command phrases that match the vocabulary stored in memory 312 . That is, the basic audio units 300 A need not send all their received audio data when each basic audio unit 300 A includes voice recognition logic 309 which reduces bandwidth requirements.
- the basic audio units 300 A may listen for audio inputs above a certain volume threshold and may discard or never send audio data to the master audio unit if the volume is below a minimum threshold. In such embodiments, the basic audio units 300 A may, or may not, include voice recognition logic 309 . Based on the above examples, one of ordinary skill will understand that various configurations are possible for the audio routing system 100 and for audio unit internal configurations in light of the above description of example basic audio unit 300 A and example master audio unit 300 B.
- the wireless transceiver 313 is operative to establish the wireless link 105 with a mobile device 102 .
- the wireless link 105 may be any suitable type of radio frequency (RF) communication link such as, but not limited to, Bluetooth®, Wi-Fi®, or some other suitable wireless communication link.
- the wireless or power-line carrier transceiver 315 is operative to establish communication link 329 with other audio units so that audio data or other data including command and control signals may be routed and sent to the mobile device 102 .
- RF protocols such as Bluetooth®, Wi-Fi®, or some other suitable wireless communication link similar to that used by wireless transceiver 313 may also be employed.
- the controller 311 A and controller/router 311 B are operative to execute one or more wireless communication protocol stacks and to coordinate communications with other audio units.
- FIG. 4 A schematic block diagram of a mobile device 102 in accordance with various embodiments is shown in FIG. 4 .
- An example mobile device 102 includes an internal communication bus 405 which provides operative coupling between the various components.
- the mobile device 102 components may include, but are not limited to, a wireless transceiver 407 operative to communicate with a network, a wireless transceiver 409 operative to communicate peer-to-peer with the audio units, near field communication logic 411 , location detection logic 413 , a display 415 , a user interface 417 , at least one processor 401 , a non-volatile non-transitory memory 403 , and a sensor hub 419 .
- the sensor hub 419 is operatively coupled to a plurality of sensors 420 which may include thermal sensors, proximity sensors, accelerometers, gyroscopic sensors, light sensors, etc.
- the processor 401 is operative to execute a voice recognition engine 421 .
- the processor 401 is also operative to execute one or more wireless protocol stacks 423 (corresponding to wireless transceiver 407 and wireless transceiver 409 ) and various other applications 425 .
- the voice recognition engine 421 may interact and communicate with the wireless protocol stacks 423 via an application programming interface API 422 and with the various other applications 425 via APIs 424 .
- Each of the various components of the mobile device 102 that are operatively coupled to the communication bus 405 may also accordingly send information to, or receive information from, the processor 401 and are thus operatively coupled to the voice recognition engine 421 , the wireless protocol stacks 423 and the applications 425 via the communication bus 405 and via the various APIs.
- the memory 403 may store voice recognition logic executable code 421 C, wireless protocol stack executable code 423 C, and application executable code 425 C that may be accessed and executed by the processor 401 . That is, the processor 401 is operative to access and execute any executable code stored in memory 403 .
- Example methods of operation of an audio routing system 100 in accordance with various embodiments will now be described with respect to the flowcharts of FIG. 5 through FIG. 12 .
- the flowcharts of FIG. 5 through FIG. 12 will be described with respect to the audio routing system illustrated in FIG. 1 which is useful for understanding the operations.
- the method of operation begins in operation block 501 , and sound is received by an audio unit near the user such as audio unit 200 B 1 .
- the audio unit 200 B 1 then sends audio data to the master audio unit 200 E as shown in operation block 503 .
- audio data is sent to the mobile device 102 over the wireless link from the master audio unit 200 E as shown in operation block 505 . That is, for embodiments in which the method of operation of FIG. 5 applies, wireless communication with the mobile device 102 is facilitated by the master audio unit 200 E.
- command words are detected at the master audio unit 200 E. That is, the master audio unit includes the voice recognition logic 309 .
- the voice recognition logic 309 outputs command word data based on one or more detected command words and sends the command word data to the mobile device 102 via a wireless link as shown in operation block 603 . The method of operation then ends.
- command words may be detected by a basic audio unit such as the audio unit 200 B 1 which thus includes the voice recognition logic 309 .
- the audio unit 200 B 1 may send command word data to the mobile device 102 by a wireless link 105 from a second basic audio unit such as audio unit 200 A which is located near the mobile device 102 .
- audio data may then be routed between the first audio unit 200 B 1 and the second audio unit 200 A and also to and from the mobile device 102 .
- the method of operation then ends. For the method of operation of FIG.
- the routing may be facilitated by the master audio unit 200 E, or may be facilitated by routing through various other basic audio units such as audio unit 200 C, 200 D, or 200 F, or by a direct routing path 113 between audio unit 200 B 1 and audio unit 200 A as was described above previously.
- the method of operation begins and in operation block 801 voice is detected at one or more audio units.
- the user 101 may speak and generate sound waves 103 which are received by audio unit 200 B 1 and sound waves 104 which are received by audio unit 200 B 2 .
- other audio units including those in other rooms such as audio unit 200 C and audio unit 200 D, may also receive sound from the user's voice depending on the direction in which the user 101 is speaking and the volume of the user's voice at that particular time.
- the audio routing system 100 will determine the audio unit in close proximity to the user. For example, audio unit 200 B 1 may be selected.
- This operation may be performed by controller/router 311 B in master audio unit 200 E by comparing the audio signal levels received from the various audio units and selecting the audio unit having the “loudest” received audio.
- One example criteria that may be compared is signal-to-noise ratio however any suitable criteria may be used.
- audio data will then be routed to the mobile device 102 from the audio unit selected as being in close proximity to the user.
- operation block 807 audio data will then be sent from the mobile device to that same audio unit. The method of operation then ends.
- audio unit 200 B 2 is selected as being in close proximity to the user 101 , then audio data from the mobile device 102 will be routed to the audio unit 200 B 2 rather than to the audio unit 200 B 1 such that the audio unit 200 B 2 plays audio over its speaker while the audio unit 200 B 1 speaker may be temporarily muted.
- FIG. 9 is a flowchart of a method of operation for handling an incoming phone call by an audio routing system 100 in accordance with various embodiments.
- a notification of an incoming phone call may be received by audio unit 200 A over the wireless link 105 .
- the notification is then broadcast to all other audio units. This may be facilitated either by using the master audio unit 200 E or by a broadcast notification sent directly from audio unit 200 A to all other audio units.
- all of the audio units will provide an audible alert of the incoming phone call such as a ring tone or some other audible alert.
- the user 101 may respond to the audible alert by speaking a command word.
- command words may be detected at one or more audio units such as audio unit 200 B 1 and audio unit 200 B 2 which happen to be located in the same room as the user 101 .
- the audio unit in close proximity to the user will be determined and, in operation block 911 , voice command word data will be routed to the audio unit in proximity to the mobile device 102 which, in the example of FIG. 1 , is audio unit 200 A.
- the phone call may be answered based on the user command, and assuming that the user commanded the mobile device 102 to answer the phone call, voice data may subsequently be routed to and from the mobile device 102 using the audio units. The method of operation then ends.
- FIG. 10 provides further details of operation block 803 of FIG. 8 which involves determining and selecting an audio unit as being in close proximity to the user. For example, in operation block 1001 , audio levels received at two or more audio units may be compared. In operation block 1003 , the audio unit having an audio level above a threshold value may be selected as being the audio unit in closest proximity to the user and the method of operation then ends as shown. This is only one example of selecting an audio unit in closest proximity to the user and it is to be understood that various other methods may be utilized in the various embodiments. For example, in embodiments where the audio units include voice recognition logic 309 , a voice recognition confidence metric may be used and compared to determine which is the “best” audio unit rather than determining the audio unit in proximity to the user.
- the audio level may not necessarily be an indication of actual proximity of an audio unit to the user but may instead provide an indication of the “best” audio unit to utilize at a particular point in time.
- the audio unit includes a motion sensor 325
- data from the motion sensor 325 may also be used to consider whether or not the particular audio unit receiving audio is the “best” audio unit (or is the audio unit in closest proximity to the user) at that particular time.
- the best or closest proximity audio unit may dynamically change as the user moves about.
- the audio routing system 100 may switch from audio unit 200 B 1 to audio unit 200 E, and then to audio unit 200 D, as the closest proximity or best audio unit based on either audio level comparison received by the respective audio unit microphones as the user walks through the rooms, by using voice confidence metrics, output of a motion sensor 325 , or by using any combination of those information sources in order to make a decision as to when to switch from one audio unit to another.
- one useful feature and advantage of the various embodiments is that the user 101 may engage in a phone call while the mobile device 102 is located in a remote room, even when the user 101 is walking from one room to another, provided that there are a number of audio units such as audio units 200 A through 200 F installed in various rooms and in locations about which the user 101 is traveling and speaking.
- the selected audio unit is also used to route audio data from the mobile device 102 such that the speaker output of the audio units will move along with the user 101 as well.
- audio data is received and sent using a first audio unit determined to be in closest proximity to the user (or the “best” audio unit as discussed above).
- operation block 1103 either the user's voice or user motion is detected by at least a second audio unit.
- decision block 1105 a threshold is tested to see if it is met at a second audio unit.
- the threshold may be tested by the second audio unit itself, such as by controller 311 A, or may be tested by a master audio unit such as by controller/router 311 B.
- the master audio unit may compare results of one or more potential second audio units and select the current closest proximity audio unit.
- a predetermined threshold is used. If the threshold is met, then the method of operation proceeds to decision block 1107 .
- the threshold in decision block 1105 may be a threshold using any of the criteria discussed previously above such as, but not limited to, an audio level threshold, a voice confidence metric threshold, a motion detection threshold or some other appropriate threshold.
- a time duration requirement may also be tested. If either the threshold of decision block 1105 or the time duration of decision block 1107 is not met, then the method of operation may return to operation block 1101 . However, if both the threshold and time duration are met, the method of operation may proceed to operation block 1109 in which the first (previous) audio unit may be muted and placed into a low power mode.
- audio data will then be routed to and from the second audio unit which is now determined to be in closest proximity to the user or the “best” audio unit.
- Hysteresis may also be applied with respect to the time duration of decision block 1107 to prevent the “popcorn effect” of needless unstable switching between audio units.
- the hysteresis may be applied by the controller/router 311 B of a master audio unit in some embodiments. Upon switching to the next audio unit, the method of operation then ends.
- FIG. 11 applying the method of operation of FIG. 11 , the example scenarios described above previously where a user may engage in a telephone call and move from one room to another may be facilitated.
- Other applications may also benefit from the method of operation of FIG. 11 , such as audio playback of music or some other type of audio file playback from the mobile device 102 over the speakers of the audio units.
- the user may listen to music and, as the user moves throughout the building, the music will follow the user by switching from audio unit speaker to audio unit speaker accordingly.
- audio may be received by first audio unit in closest proximity to a user (or the “best” audio unit), for example audio unit 200 B 1 .
- the audio may be examined by a voice recognition system to determine if a voice command is contained in the audio.
- the voice recognition logic 309 may be located in any appropriate location and may therefore be present in either the audio unit 200 B 1 , a master audio unit such as master audio unit 200 E, or the voice recognition may be performed by the mobile device 102 depending on the particular embodiment.
- a decision may be made as to whether or not to route the audio data to a second audio unit, such as audio unit 200 A, which is in proximity to, and paired with, the mobile device 102 .
- audio data may be ignored if it does not contain a voice command as determined by the voice recognition logic 309 . Therefore, in decision block 1203 , if a voice command is not contained in the audio, then the audio unit will continue to receive audio using the first audio unit until a voice command is detected in decision block 1203 .
- the method of operation proceeds to operation block 1205 and command word data is routed to a second audio unit.
- the command word data from the voice recognition logic 309 is then sent to the mobile device 102 using the wireless link 105 .
- the second audio unit 200 A will receive query response audio data, such as audio report data 106 , from the mobile device 102 over the wireless link 105 .
- query response audio data such as audio report data 106
- the query response or audio report data 106 may be modified to provide additional information if appropriate as shown in operation block 1213 .
- the controller/router 311 B of master audio unit 200 E may modify the audio report data 106 accordingly prior to routing it to the next audio unit.
- the audio report data may be modified such that the audio report 108 provided to the user 101 is the message “I′m in another room!!”
- the master audio unit 200 E may be aware that it is communicating with the mobile device 102 via the audio unit 200 A which is located in a different physical position from the audio unit 200 B 1 .
- the mobile device 102 location may be triangulated by the array of home audio units using wireless signals or audio signals. The master audio unit 200 E made thus make an inference that the mobile device 102 is located in another room and change the report data or provide an alternative audio report 108 accordingly.
- the query response audio data such as audio report data 106 , is routed to the audio unit 200 B 1 so that it will be heard by the user 101 .
- the audio report 111 “I′m here” may be played by the speaker of the audio unit 200 B 1 or only by the mobile device 102 speaker. However, in such embodiments, the user 101 will have to travel through the rooms to figure out where exactly “here” is by listening for the audio report 111 emanating from the mobile device 102 speakers. As would be understood, the audio report 111 may repeat for a predetermined period of time so that the user 101 has time to look around and find the mobile device 102 .
- the query response such as audio report 111 or audio report 108
- the audio unit in closest proximity to the user may be played by the audio unit in closest proximity to the user (or the “best” audio unit) using the received audio report data 106 (or modified audio report data).
- the method of operation then ends as shown.
- the voice command does not include a query
- the audio unit in proximity to the user will wait to see if audio data is received from the mobile device as shown in decision block 1219 . If not, the method of operation then ends. If yes, then any audio data is played using the audio unit in closest proximity to the user (or “best” audio unit) as shown in operation block 1221 and the method of operation ends.
- a master audio unit may include capability to turn off audio units or put them into a sleep/standby mode.
- the master audio unit may also turn audio units off or to standby mode when the mobile device 102 is not in wireless data range of the master audio unit or the basic audio units depending on the embodiment.
- the master audio unit may turn the basic audio units on or wake them from standby mode when the mobile device 102 is in wireless data range of the master audio unit.
- Basic audio units may be paired to a master audio unit during an initialization procedure of the audio routing system 100 .
- audio units may be paired by the user manually, simultaneously pressing and holding a control button on the master audio unit and on the basic audio unit to initiate pairing.
- Completed pairing may be indicated by the status LED 323 illuminating (or showing a certain color such as green) or by a sound report from one or both of the master audio unit and the basic audio unit or by both.
- the master audio unit may include a near field communication (NFC) unit and the basic audio units may include NFC tags. In such embodiments, pairing may be accomplished by placing a basic audio unit over or near the master audio unit.
- NFC near field communication
- various Bluetooth® profiles may be used to determine behavior during different audio events. For example, during an incoming call, the Hands-Free Profile (HFP) may be used. This profile supports two-way audio for conversations.
- a master audio unit may as a default setting allow the audio ringer to be played on all audio units connected throughout the home. A user may then answer via voice command or button press on a specific audio unit or may ignore the call via voice command or button press. The default setting may be disabled or changed at the master audio unit if desired by the user either by voice command or control button input.
- Some embodiments may include an “answer all” feature in which a call may be answered and conducted across all available audio units in the home. For example, this feature is useful if a child is calling home from college and the parents are in different rooms of the home and want to participate in the call or for other like scenarios.
- the Advanced Audio Distribution Profile may be used for multimedia audio playback on the audio units.
- the master audio unit may keep basic audio units muted by default. If the master audio unit includes a speaker, it may play the audio. Any other audio unit may be unmuted by voice command (such as “unmute”) or by pressing a control button on an audio unit. Thus, a different command to any audio unit or to the master audio unit (such as “unmute all”) may enable playback to all audio units. Any default settings may be modified by the user at the master audio unit via voice command or by using a control button.
- voice command such as “unmute”
- a different command to any audio unit or to the master audio unit such as “unmute all” may enable playback to all audio units. Any default settings may be modified by the user at the master audio unit via voice command or by using a control button.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Automation & Control Theory (AREA)
- Otolaryngology (AREA)
- Acoustics & Sound (AREA)
- Telephonic Communication Services (AREA)
- Circuit For Audible Band Transducer (AREA)
- Mobile Radio Communication Systems (AREA)
- Telephone Function (AREA)
Abstract
Description
- The present application claims priority to U.S. Provisional Patent Application No. 61/881,965, filed Sep. 25, 2013, entitled “AUDIO ROUTING SYSTEM FOR ROUTING AUDIO DATA TO AND FROM A MOBILE DEVICE,” which is hereby incorporated by reference herein in its entirety, and which is assigned to the same assignee as the present application.
- The present disclosure relates generally to mobile devices and speaker phone extension devices and more particularly to voice recognition systems and routing audio data between devices.
- Many current mobile devices include voice command capability to enable response to a user's voice commands for control of voice operated dialing or other features. In some mobile devices the voice command system may be capable of waking the mobile device from a low-power sleep mode. This may be accomplished in such existing mobile devices, by having the voice command system active even during the sleep mode and always listening for a specific command word or command phrase. When an appropriate command word or command phrase is recognized, the voice command system wakes the mobile device from sleep mode and appropriately responds to the command. Thus such mobile devices allow for hands-free interaction through voice control.
- One challenge for mobile device voice command systems is in obtaining a clear audio input to the mobile device microphone or microphone array. For example, if the user is too far from the mobile device, or if the mobile device is in a pocket, purse or wedged between sofa cushions such that the microphones are obstructed, the user's voice command will either not be picked up by the microphones or will not be recognizable by the voice command system.
- One way to overcome this challenge is for the user to wear a separate, remote microphone ensemble such as a headset or earpiece that is wirelessly paired to the mobile device. The drawbacks to this approach include the limited battery charge time of the headset or earpiece and the encumbrance of wearing the ensemble. Another drawback is that the ensemble must be kept within relative close proximity to the mobile device so as to maintain a wireless link. Also, increasing the distance between the mobile device and the ensemble may further drain the batteries of both the ensemble and the mobile device.
-
FIG. 1 is a floor plan showing audio units positioned at various locations within various rooms and routing audio data in accordance with the embodiments. -
FIG. 2 is diagram showing various views of an audio unit in accordance with an embodiment. -
FIG. 3A andFIG. 3B are schematic block diagrams that provide examples of various types of audio units that may be used in accordance with various embodiments. -
FIG. 4 is a schematic block diagram of a mobile device in accordance with various embodiments. -
FIG. 5 is a flow chart of a method of operation of an audio routing system in accordance with various embodiments. -
FIG. 6 is a flow chart of a method of operation of a master audio unit in accordance with an embodiment. -
FIG. 7 is a flow chart of a method of operation of an audio routing system in accordance with various embodiments. -
FIG. 8 is a flow chart of a method of operation of an audio routing system in accordance with various embodiments. -
FIG. 9 is a flow chart of a method of operation for handling an incoming phone call by an audio routing system in accordance with various embodiments. -
FIG. 10 is a flow chart providing further details of one operation block of the flow chart ofFIG. 8 in accordance with some embodiments. -
FIG. 11 is a flow chart of a method of operation of an audio routing system in accordance with some embodiments. -
FIG. 12 is a flow chart of a method of operation for a “find me feature” (which is a mobile device location feature) as well as other features of the audio routing system in accordance with various embodiments. - The disclosed embodiments provide, among other things, an audio routing system having features such that a user can transition from room to room in a house, office, etc., and speak voice commands and hear responses without the need for keeping the mobile device in vocal range at all times. The audio routing system includes a plurality of audio units and may include a master audio unit in some embodiments. Each audio unit includes a speaker and at least one microphone, and plugs directly into a standard wall AC electrical outlet. This allows the audio unit to receive electrical power anytime it is needed and overcomes one drawback of portable electronic devices, namely, discharged batteries. Each audio unit also includes wireless data capability or power-line carrier data capability and may communicate with other audio units or with a master audio unit. The audio units may therefore be installed by plugging them into electrical outlets throughout a user's home such as in the kitchen, bathrooms, offices, bedrooms, or living rooms of the home.
- In one disclosed embodiment, a method of operation includes receiving sound by a first audio unit installed in an electrical outlet, routing audio data corresponding to the received sound from the first audio unit to a second audio unit installed in a second electrical outlet, and sending the audio data to a mobile device using a wireless link between the mobile device and the second audio unit. In some embodiments, routing the audio data may include receiving the audio data from the first audio unit by a third audio unit, and routing the audio data to the second audio unit by the third audio unit serving as a router. Further, in some embodiments, routing the audio data may include routing the audio data using one of table drive routing or on-demand routing.
- In some embodiments, the method of operation may further include performing voice recognition on the audio data and detecting a command word, and routing command word data to the second audio unit. In some embodiments, routing the audio data to the second audio unit may include sending the command word data from the first audio unit to the third audio unit using a wireless link between the first audio unit and the third audio unit. In other embodiments, routing the audio data to the second audio unit may be accomplished by sending the audio data from the first audio unit to the third audio unit using a power line carrier connection between the first audio unit and the third audio unit.
- In another disclosed embodiment, a method of operation includes receiving sound by a group of audio units installed in a corresponding group of electrical outlets. The method further includes selecting a first audio unit from the group of audio units in response to determining that the first audio unit meets a threshold with respect to processing the received sound. Audio data corresponding to the received sound is then routed from the first audio unit installed in a first electrical outlet to a second audio unit installed in a second electrical outlet, and is sent to a mobile device over a wireless communication link between the second audio unit and the mobile device.
- In some embodiments, the method of operation further includes performing voice recognition on the audio data and detecting a command word, and routing command word data to the second audio unit. Depending on the embodiment, performing voice recognition may involve performing voice recognition by one of the first audio unit or a master audio unit serving as a router between the first audio unit and the second audio unit. In that case, the audio data may be routed to the second audio unit only in response to detecting the command word.
- In some embodiments, selecting a first audio unit may include a master unit that receives audio data from each audio unit of the group of audio units, and selects the first audio unit based on comparison of one of an audio signal level or a voice confidence metric for a command word detected in the audio data. The metrics are compared for audio data received by the master unit from each of the audio units. Depending on the embodiment, routing audio data may involve routing the audio data from the first audio unit to the second audio unit using a wireless link between the first audio unit and the second audio unit. In other embodiments, routing audio data may involve routing the audio data from the first audio unit to the second audio unit using a power line carrier connection between the first audio unit and the second audio unit over an electrical wiring system connected to a first electrical outlet and the second electrical outlet.
- The disclosed embodiments include an audio unit. The audio unit includes a housing that has a two-prong alternating current power plug with the two-prongs extending from the housing, and a transformer and a voltage rectifier installed within the housing and coupled to the power plug. The voltage rectifier is operative to provide direct current to an internal power bus, which is operatively coupled to the voltage rectifier. The audio unit also includes a microphone, a speaker and audio signal conditioning logic, operatively coupled to the microphone and speaker. A first transceiver, operative to establish a connection with a second audio unit is operatively coupled to a controller. The controller is operative to control the transceiver to send audio data the second audio unit. In some embodiments, the audio unit also includes a wireless transceiver that is operatively coupled to the controller, and that is operative to establish a wireless connection with a mobile device.
- In some embodiments, the audio unit may further include voice recognition logic. The voice recognition logic is operatively coupled to the audio signal conditioning logic and to the controller, and is operative to detect a command word in audio data output by the audio signal conditioning logic. The controller may be further operative to control the wireless transceiver to send command word data to the mobile device, and may be operative to control the first transceiver to send command word data to the second audio unit.
- Further, in some embodiments, the controller is also operative to control the wireless transceiver to receive audio data from the mobile device, and operative to control the first transceiver to receive audio data from the second audio unit and to operate the microphone, speaker and audio signal conditioning logic to provide a speakerphone operation with the mobile device. The first transceiver may be a wireless transceiver in some embodiments, or may be a power line carrier transceiver.
- Turning now to the drawings wherein like numerals represent like components,
FIG. 1 illustrates a floor plan showing an exampleaudio routing system 100 including various audio units (200A through 200F) positioned at various locations within various rooms in accordance with the embodiments. Although the exampleaudio routing system 100 is shown on a single level floor plan, the embodiments are not limited to such arrangements and theaudio routing system 100 may cover multiple floors within a building. For example,audio unit 200A is installed in an electrical outlet within room A which is the same room in whichmobile device 102 is located. Themobile device 102 is paired to theaudio unit 200A using awireless link 105. In room B audio unit 200B1 and audio unit 200B2 are installed in respective electrical outlets that are positioned on walls at a 90° angle with respect to each other. Theaudio routing system 100 also includesaudio unit 200C,audio unit 200D,audio unit 200E andaudio unit 200F which are each installed in corresponding respective electrical outlets within various rooms. Additional audio units may also be installed on other floors of a multi-level building. One of the audio units, such asaudio unit 200E, may be installed in an electrical outlet in a relatively centrally located position such as the hallway shown inFIG. 1 and may serve as a “master” audio unit and router in some embodiments. For example, audio data may be routed between audio unit 200B1 andaudio unit 200A using themaster audio unit 200E as a router. - In one example of operation, the audio unit 200B1 will establish a
communication link 107 with themaster audio unit 200E via an appropriate routing path, and themaster audio unit 200E will establish asecond communication link 109 withaudio unit 200A via another appropriate routing path. These routing paths may be facilitated by peer-to-peer wireless connections between various audio units, or by power-line carrier connections over a building electrical system to which the various AC power outlets are connected. - In accordance with the embodiments, the various audio units located in the various rooms (an on various floors in some embodiments) serve to extend the audio features and capabilities of the
mobile device 102 including voice command features. In other words, theuser 101 may leave themobile device 102 anywhere within a home or office and still be able to access and control themobile device 102 using voice commands from other rooms where audio units are installed. In one example application of theaudio routing system 100, auser 101 may speak a voice command which may be a query. For example, the user may ask themobile device 102 “Where are you?” The user's voice createssound waves 103 andsound waves 104 which may be picked up by microphones within various audio units. The user's voice command query will be routed by one of the audio units within a room depending on criteria such as, but not limited to, the “best” received audio signal level, the best voice recognition confidence metric, or some other suitable criteria. For example, assuming the audio unit 200B1 is selected as the best audio unit to transmit the user's voice command query, then the audio unit 200B1 will establish arouting path 113 to theaudio unit 200A. As mentioned above, the routing may be facilitated by amaster audio unit 200E in some embodiments. Theaudio unit 200A will send audio data, or in some embodiments the command query data output from voice recognition logic, to themobile device 102 using thewireless link 105. In response to the query, themobile device 102 may sendaudio report data 106 to theaudio unit 200A using thewireless link 105. Theaudio unit 200A will then route theaudio report data 106 to the audio unit 200B1 using anappropriate routing path 113. The audio unit 200B1 will then play anaudio report 108 such as, for example, “I′m in another room!!” Themobile device 102 may also provide anaudio report 111 from its own speaker such as “I′m here!!” such that theuser 101 may locate themobile device 102 as theuser 101 moves throughout the various rooms. In some embodiments where a master audio unit is employed, such asmaster audio unit 200E, theaudio unit 200A may use thecommunication link 109 to send theaudio report data 106 to themaster audio unit 200E for routing on to the audio unit 200B1 usingcommunication link 107. In other words, in some embodiments, therouting path 113 is facilitated through themaster audio unit 200E. However in other embodiments, a master audio unit may not be present and various ad-hoc like routing protocols may be employed such as, but not limited to, table-driven routing, on-demand routing, or some other suitable routing protocol. - For example, in the case of table driven routing, a routing table may be generated by either one of the audio units or by a distributed function distributed among the various audio units such that the routing table is stored within each of the various audio units. Thus, if the user were to remove an audio unit from an electrical outlet such that the audio unit is taken out of service from the
audio routing system 100, a new routing table would be generated based on the newaudio routing system 100 topology and the new routing table would be stored in each of the audio units remaining in service. In an example of on-demand routing, one or more audio units receiving sound would broadcast route requests until a route to the audio unit having awireless link 105 to themobile device 102 is obtained. In this case, the route may be facilitated directly between two audio units or through any suitable number of audio units. For example, in some embodiments,communication link 107 and communication link 109 may be wireless links similar towireless link 105. However, in other embodiments,communication link 107 and communication link 109 may be facilitated by power-line carrier communication links over the electrical system wiring to which each of the electrical outlets is connected. In the power-line carrier case, direct connectivity may be established between two audio units or data may be routed using audio units connected along the pathway. - In embodiments where the audio units communicate using wireless links, it is possible that the two audio units that need to communicate are outside of the wireless coverage range or are otherwise impeded from communicating directly wirelessly. For example, audio unit 200B1 may be unable to establish a wireless link directly with
audio unit 200A due to impedance cause by various walls or other objects such as furniture that block the wireless path between the two audio units. Thus if, for example,audio unit 200D is able to communicate wirelessly withaudio unit 200C, andaudio unit 200C is able to communicate wirelessly with audio unit 200B2, then audio unit 200B2 may be used as an initial routing point for audio data sent wirelessly from audio unit 200B1. Alternatively, audio unit 200B2 may be used as the receiving point for the user's voice or voice commands based on audio unit 200B2 being able to establish a wireless routing path toaudio unit 200A whereas audio unit 200B1 is unable to communicate withaudio unit 200A directly. -
FIG. 2 is diagram showing various views of one example of anaudio unit 200 in accordance with an embodiment. It is to be understood however that theexample audio unit 200 is only one possible physical configuration of an audio unit and is not to be construed as a limitation on the particular form that an audio unit may take in accordance with the various embodiments. Some example features of an audio unit are shown inFIG. 2 that may be present in audio units having different physical configurations different than those shown in the example ofFIG. 2 . Theaudio unit 200 includes ahousing 201 which has anAC power plug 205 installed within thehousing 201 and projecting therefrom such that theaudio unit 200 may be installed in a standard AC electrical outlet such as found in a home or office building etc. In some embodiments, the prongs of theAC power plug 205 may pivot or rotate such that they may be stored within an indentation (not shown) in thehousing 201 for storing theAC power plug 205 when theaudio unit 200 is not in use. Thehousing 201 includes aspeaker port 207 and one ormore microphone ports 209. In some embodiments thehousing 201 includes astatus LED 211. A user interface may also be present in some embodiments such as, for example,control button 202 andcontrol button 203. Thecontrol buttons example audio unit 200 is light weight and portable and may be similar to an AC charger housing such as accessory AC chargers provided with a cord to attach to a mobile device for purposes of charging a battery within the mobile device. However as shown inFIG. 2 , thehousing 201 does not include any attached cord with respect to theexample audio unit 200. - In some embodiments the
control buttons -
FIG. 3A andFIG. 3B are schematic block diagrams that provide examples of various types of audio units that may be used in accordance with various embodiments.FIG. 3A illustrates various embodiments of abasic audio unit 300A andFIG. 3B illustrates various embodiments of amaster audio unit 300B. Components that may be present in only some embodiments are shown using dotted lines to emphasize that such components may or may not be present depending on the particular embodiment. It is to be understood that with respect toFIG. 3A andFIG. 3B , like reference numerals that are reused betweenFIG. 3 andFIG. 3B indicate like or similar components. - Thus with respect to the example
basic audio unit 300A and examplemaster audio unit 300B, both include aninternal communication bus 327 which provides operative coupling between the various internal components. Theaudio unit control buttons 202 and 203), acontroller 311A (or controller/router 311B inmaster audio unit 300B), non-volatile,non-transitory memory 312, awireless transceiver 313, a wireless or power-line carrier transceiver 315, and anaudio signal conditioner 301. One or more microphones 303 (such as a microphone array) and aspeaker 305 are operatively coupled to theaudio signal conditioner 301. Themaster audio unit 300B may or may not include theaudio signal conditioner 301 and the corresponding one ormore microphones 303 andspeaker 305. Theaudio signal conditioner 301, when present in an embodiment, provides audio signal conditioning logic that may include analog-to-digital converters (ADCs), digital-to-analog converters (DACs), high-pass filters, low-pass filters, band-pass filters, or adjustable band filters that may be applied to filter noise from received audio. Theaudio signal conditioner 301 may include gain including automatic gain control (AGC) in some embodiments. Theaudio signal conditioner 301 may also include an echo canceller in some embodiments, to prevent echoing and feedback between themicrophones 303 and thespeaker 305 during full-duplex communication such as when the audio unit is used as a speakerphone. - A
status LED 323 is operatively coupled to switchlogic 321 which is further operatively coupled to theinternal communication bus 327. Theswitch logic 321 is operative to receive control commands from thecontroller 311A (or controller/router 311B) and to light theLED 323 in response to the control commands accordingly. In some embodiments, theLED 323 may provide one or more indications and may be a multi-color LED in some embodiments. For example, the LED may light to show when the audio unit is paired to another audio unit or to a mobile device. The LED may also provide an indication that the audio unit is in use. AnAC power plug 319 is connected to a transformer/rectifier 317 which is further connected to an internal power bus (not shown) that provides DC power to the internal components of the audio unit. In embodiments where the wireless or power-line carrier transceiver 315 is a power-line transceiver, the power-line transceiver is operatively coupled to theAC power plug 319 using appropriate safety and filtering circuitry (not shown) so that audio data including voice command data may be sent and received over and electrical system when theAC power plug 319 is inserted into a standard electrical outlet and is thus connected to an electrical wiring system. - In some embodiments, the
audio units voice recognition logic 309, and may include amotion sensor 325. In embodiments that include themotion sensor 325, either thecontroller 311A or the controller/router 311B will be operative to obtain data from themotion sensor 325 and to determine motion in proximity to the respective audio unit. In embodiments in whichvoice recognition logic 309 is present, thememory 312 may store a library of command words or command phrases (i.e. a vocabulary) that are to be recognized by thevoice recognition logic 309. In alternative embodiments, thevoice recognition logic 309 may include an integrated memory that stores the command word or command phrase vocabulary. - In some embodiments, the
basic audio unit 300A may includewireless transceiver 313 which is operative to communicate overwireless link 105 withmobile device 102. In such embodiments, any basic audio unit may communicate with themobile device 102 such asaudio unit 200A shown in the example ofFIG. 2 . In embodiments where thebasic audio unit 300A does not includewireless transceiver 313, only themaster audio unit 300B will be capable of establishing thewireless link 105 with themobile device 102. In these embodiments, thebasic audio units 300A communicate with each other only, or with themaster audio unit 300B only, by using wireless or power-line carrier transceiver 315. - The
master audio unit 300B may also include adocking connector 331 which may be used to connect themobile device 102 for purposes of charging and also, in some embodiments, for providing a data connection to themobile device 102 by way of theinternal communication bus 327. That is, thedocking connector 331 may be operatively coupled to theinternal communication bus 327 and to the power bus (not shown). Is to be understood that in some embodiments, all of the audio units may be include the components of a “master”audio unit 300B where all audio units includeaudio signal conditioner 301, one ormore microphones 303 and aspeaker 305. In that case, any audio unit may be designated as a master audio unit while other audio units function only as basic audio units or as slave devices. In other words, thebasic audio unit 300A may omit some components suchvoice recognition logic 309,wireless transceiver 313,motion sensor 325 or some combination thereof in order to reduce the cost of thebasic audio unit 300A. Likewise themaster audio unit 300B may omit theaudio signal conditioner 301,microphones 303 andspeaker 305 in order to reduce the cost of themaster audio unit 300B. The location of thevoice recognition logic 309 may depend on whether a central routing function is used or not. That is, if amaster audio unit 300B is used to perform a data routing function, it may also include thevoice recognition logic 309 and may perform voice recognition on audio data received from the various basic audio units. Alternatively, thebasic audio units 300A may includevoice recognition logic 309, and themaster audio unit 300B may omit this component. In such embodiments, thebasic audio units 300A need only send a data output from thevoice recognition logic 309 upon detection of command words or command phrases that match the vocabulary stored inmemory 312. That is, thebasic audio units 300A need not send all their received audio data when eachbasic audio unit 300A includesvoice recognition logic 309 which reduces bandwidth requirements. In other embodiments, thebasic audio units 300A may listen for audio inputs above a certain volume threshold and may discard or never send audio data to the master audio unit if the volume is below a minimum threshold. In such embodiments, thebasic audio units 300A may, or may not, includevoice recognition logic 309. Based on the above examples, one of ordinary skill will understand that various configurations are possible for theaudio routing system 100 and for audio unit internal configurations in light of the above description of examplebasic audio unit 300A and examplemaster audio unit 300B. - In the various embodiments, the
wireless transceiver 313 is operative to establish thewireless link 105 with amobile device 102. Thewireless link 105 may be any suitable type of radio frequency (RF) communication link such as, but not limited to, Bluetooth®, Wi-Fi®, or some other suitable wireless communication link. The wireless or power-line carrier transceiver 315 is operative to establishcommunication link 329 with other audio units so that audio data or other data including command and control signals may be routed and sent to themobile device 102. In embodiments where wireless or power-line carrier transceiver 315 is a wireless transceiver, RF protocols such as Bluetooth®, Wi-Fi®, or some other suitable wireless communication link similar to that used bywireless transceiver 313 may also be employed. Thecontroller 311A and controller/router 311B are operative to execute one or more wireless communication protocol stacks and to coordinate communications with other audio units. - A schematic block diagram of a
mobile device 102 in accordance with various embodiments is shown inFIG. 4 . An examplemobile device 102 includes aninternal communication bus 405 which provides operative coupling between the various components. Themobile device 102 components may include, but are not limited to, awireless transceiver 407 operative to communicate with a network, awireless transceiver 409 operative to communicate peer-to-peer with the audio units, nearfield communication logic 411,location detection logic 413, adisplay 415, auser interface 417, at least oneprocessor 401, a non-volatilenon-transitory memory 403, and asensor hub 419. Thesensor hub 419 is operatively coupled to a plurality ofsensors 420 which may include thermal sensors, proximity sensors, accelerometers, gyroscopic sensors, light sensors, etc. - In accordance with the embodiments, the
processor 401 is operative to execute avoice recognition engine 421. Theprocessor 401 is also operative to execute one or more wireless protocol stacks 423 (corresponding towireless transceiver 407 and wireless transceiver 409) and variousother applications 425. Thevoice recognition engine 421 may interact and communicate with the wireless protocol stacks 423 via an applicationprogramming interface API 422 and with the variousother applications 425 viaAPIs 424. Each of the various components of themobile device 102 that are operatively coupled to thecommunication bus 405 may also accordingly send information to, or receive information from, theprocessor 401 and are thus operatively coupled to thevoice recognition engine 421, the wireless protocol stacks 423 and theapplications 425 via thecommunication bus 405 and via the various APIs. Thememory 403 may store voice recognition logicexecutable code 421C, wireless protocol stackexecutable code 423C, and applicationexecutable code 425C that may be accessed and executed by theprocessor 401. That is, theprocessor 401 is operative to access and execute any executable code stored inmemory 403. - Example methods of operation of an
audio routing system 100 in accordance with various embodiments will now be described with respect to the flowcharts ofFIG. 5 throughFIG. 12 . The flowcharts ofFIG. 5 throughFIG. 12 will be described with respect to the audio routing system illustrated inFIG. 1 which is useful for understanding the operations. Thus beginning withFIG. 5 , the method of operation begins inoperation block 501, and sound is received by an audio unit near the user such as audio unit 200B1. The audio unit 200B1 then sends audio data to themaster audio unit 200E as shown inoperation block 503. Assuming a wireless link can be established between themaster audio unit 200E and themobile device 102, audio data is sent to themobile device 102 over the wireless link from themaster audio unit 200E as shown inoperation block 505. That is, for embodiments in which the method of operation ofFIG. 5 applies, wireless communication with themobile device 102 is facilitated by themaster audio unit 200E. - Turning to
FIG. 6 , inoperation block 601 command words are detected at themaster audio unit 200E. That is, the master audio unit includes thevoice recognition logic 309. Thevoice recognition logic 309 outputs command word data based on one or more detected command words and sends the command word data to themobile device 102 via a wireless link as shown inoperation block 603. The method of operation then ends. - In the method of operation illustrated in
FIG. 7 , and beginning withoperation block 701, command words may be detected by a basic audio unit such as the audio unit 200B1 which thus includes thevoice recognition logic 309. Inoperation block 703, the audio unit 200B1 may send command word data to themobile device 102 by awireless link 105 from a second basic audio unit such asaudio unit 200A which is located near themobile device 102. Inoperation block 705, audio data may then be routed between the first audio unit 200B1 and thesecond audio unit 200A and also to and from themobile device 102. The method of operation then ends. For the method of operation ofFIG. 7 , the routing may be facilitated by themaster audio unit 200E, or may be facilitated by routing through various other basic audio units such asaudio unit direct routing path 113 between audio unit 200B1 andaudio unit 200A as was described above previously. - In
FIG. 8 , the method of operation begins and inoperation block 801 voice is detected at one or more audio units. For example as shown inFIG. 1 , theuser 101 may speak and generatesound waves 103 which are received by audio unit 200B1 andsound waves 104 which are received by audio unit 200B2. As would be understood, other audio units, including those in other rooms such asaudio unit 200C andaudio unit 200D, may also receive sound from the user's voice depending on the direction in which theuser 101 is speaking and the volume of the user's voice at that particular time. Inoperation block 803, theaudio routing system 100 will determine the audio unit in close proximity to the user. For example, audio unit 200B1 may be selected. This operation may be performed by controller/router 311B inmaster audio unit 200E by comparing the audio signal levels received from the various audio units and selecting the audio unit having the “loudest” received audio. One example criteria that may be compared is signal-to-noise ratio however any suitable criteria may be used. Inoperation block 805, audio data will then be routed to themobile device 102 from the audio unit selected as being in close proximity to the user. Inoperation block 807, audio data will then be sent from the mobile device to that same audio unit. The method of operation then ends. Thus for example, if audio unit 200B2 is selected as being in close proximity to theuser 101, then audio data from themobile device 102 will be routed to the audio unit 200B2 rather than to the audio unit 200B1 such that the audio unit 200B2 plays audio over its speaker while the audio unit 200B1 speaker may be temporarily muted. -
FIG. 9 is a flowchart of a method of operation for handling an incoming phone call by anaudio routing system 100 in accordance with various embodiments. Thus in operation block 901, depending on whether a master audio unit is employed or not, a notification of an incoming phone call may be received byaudio unit 200A over thewireless link 105. Inoperation block 903, the notification is then broadcast to all other audio units. This may be facilitated either by using themaster audio unit 200E or by a broadcast notification sent directly fromaudio unit 200A to all other audio units. Inoperation block 905, all of the audio units will provide an audible alert of the incoming phone call such as a ring tone or some other audible alert. Inoperation block 907, theuser 101 may respond to the audible alert by speaking a command word. - Thus in
operation block 907, command words may be detected at one or more audio units such as audio unit 200B1 and audio unit 200B2 which happen to be located in the same room as theuser 101. Inoperation block 909, the audio unit in close proximity to the user will be determined and, inoperation block 911, voice command word data will be routed to the audio unit in proximity to themobile device 102 which, in the example ofFIG. 1 , isaudio unit 200A. Inoperation block 913, the phone call may be answered based on the user command, and assuming that the user commanded themobile device 102 to answer the phone call, voice data may subsequently be routed to and from themobile device 102 using the audio units. The method of operation then ends. -
FIG. 10 provides further details ofoperation block 803 ofFIG. 8 which involves determining and selecting an audio unit as being in close proximity to the user. For example, inoperation block 1001, audio levels received at two or more audio units may be compared. Inoperation block 1003, the audio unit having an audio level above a threshold value may be selected as being the audio unit in closest proximity to the user and the method of operation then ends as shown. This is only one example of selecting an audio unit in closest proximity to the user and it is to be understood that various other methods may be utilized in the various embodiments. For example, in embodiments where the audio units includevoice recognition logic 309, a voice recognition confidence metric may be used and compared to determine which is the “best” audio unit rather than determining the audio unit in proximity to the user. Also, it is to be understood that the audio level may not necessarily be an indication of actual proximity of an audio unit to the user but may instead provide an indication of the “best” audio unit to utilize at a particular point in time. In embodiments where the audio unit includes amotion sensor 325, data from themotion sensor 325 may also be used to consider whether or not the particular audio unit receiving audio is the “best” audio unit (or is the audio unit in closest proximity to the user) at that particular time. Also, in some embodiments, the best or closest proximity audio unit may dynamically change as the user moves about. - For example, if the user is speaking on a phone call and is walks out of room B, through the hallway, and on to the room in which
audio unit 200D is located, then theaudio routing system 100 may switch from audio unit 200B1 toaudio unit 200E, and then toaudio unit 200D, as the closest proximity or best audio unit based on either audio level comparison received by the respective audio unit microphones as the user walks through the rooms, by using voice confidence metrics, output of amotion sensor 325, or by using any combination of those information sources in order to make a decision as to when to switch from one audio unit to another. In other words, one useful feature and advantage of the various embodiments is that theuser 101 may engage in a phone call while themobile device 102 is located in a remote room, even when theuser 101 is walking from one room to another, provided that there are a number of audio units such asaudio units 200A through 200F installed in various rooms and in locations about which theuser 101 is traveling and speaking. As would be understood, the selected audio unit is also used to route audio data from themobile device 102 such that the speaker output of the audio units will move along with theuser 101 as well. - The method of operation illustrated by the flowchart of
FIG. 11 provides one example of how this “follow me” feature may be accomplished in the various embodiments. Inoperation block 1101, audio data is received and sent using a first audio unit determined to be in closest proximity to the user (or the “best” audio unit as discussed above). Inoperation block 1103, either the user's voice or user motion is detected by at least a second audio unit. Indecision block 1105, a threshold is tested to see if it is met at a second audio unit. The threshold may be tested by the second audio unit itself, such as bycontroller 311A, or may be tested by a master audio unit such as by controller/router 311B. The master audio unit may compare results of one or more potential second audio units and select the current closest proximity audio unit. If the basic audio units perform testing themselves, a predetermined threshold is used. If the threshold is met, then the method of operation proceeds todecision block 1107. The threshold indecision block 1105 may be a threshold using any of the criteria discussed previously above such as, but not limited to, an audio level threshold, a voice confidence metric threshold, a motion detection threshold or some other appropriate threshold. Indecision block 1107, a time duration requirement may also be tested. If either the threshold ofdecision block 1105 or the time duration ofdecision block 1107 is not met, then the method of operation may return tooperation block 1101. However, if both the threshold and time duration are met, the method of operation may proceed tooperation block 1109 in which the first (previous) audio unit may be muted and placed into a low power mode. Inoperation block 1111, audio data will then be routed to and from the second audio unit which is now determined to be in closest proximity to the user or the “best” audio unit. Hysteresis may also be applied with respect to the time duration ofdecision block 1107 to prevent the “popcorn effect” of needless unstable switching between audio units. The hysteresis may be applied by the controller/router 311B of a master audio unit in some embodiments. Upon switching to the next audio unit, the method of operation then ends. - Thus, applying the method of operation of
FIG. 11 , the example scenarios described above previously where a user may engage in a telephone call and move from one room to another may be facilitated. Other applications may also benefit from the method of operation ofFIG. 11 , such as audio playback of music or some other type of audio file playback from themobile device 102 over the speakers of the audio units. For example, the user may listen to music and, as the user moves throughout the building, the music will follow the user by switching from audio unit speaker to audio unit speaker accordingly. - Turning to the flowchart of
FIG. 12 , a “find me feature” (which is a mobile device location feature) as well as other features of theaudio routing system 100 are illustrated. Thus, inoperation block 1201, audio may be received by first audio unit in closest proximity to a user (or the “best” audio unit), for example audio unit 200B1. Indecision block 1203, the audio may be examined by a voice recognition system to determine if a voice command is contained in the audio. Thevoice recognition logic 309 may be located in any appropriate location and may therefore be present in either the audio unit 200B1, a master audio unit such asmaster audio unit 200E, or the voice recognition may be performed by themobile device 102 depending on the particular embodiment. However, in the embodiments where either the closest proximity audio unit 200B1 or themaster audio unit 200E performs voice recognition, a decision may be made as to whether or not to route the audio data to a second audio unit, such asaudio unit 200A, which is in proximity to, and paired with, themobile device 102. In other words, audio data may be ignored if it does not contain a voice command as determined by thevoice recognition logic 309. Therefore, indecision block 1203, if a voice command is not contained in the audio, then the audio unit will continue to receive audio using the first audio unit until a voice command is detected indecision block 1203. When a voice command is detected, the method of operation proceeds tooperation block 1205 and command word data is routed to a second audio unit. Inoperation block 1207, the command word data from thevoice recognition logic 309 is then sent to themobile device 102 using thewireless link 105. - If the voice command includes a query as shown in
decision block 1209, then, inoperation block 1211, thesecond audio unit 200A will receive query response audio data, such asaudio report data 106, from themobile device 102 over thewireless link 105. In some embodiments, such as those where amaster audio unit 200E is utilized, the query response oraudio report data 106 may be modified to provide additional information if appropriate as shown inoperation block 1213. For example, based on a routing table or other information known to themaster audio unit 200E, the location of themobile device 102 with respect to theuser 101 may be known. In that case, the controller/router 311B ofmaster audio unit 200E may modify theaudio report data 106 accordingly prior to routing it to the next audio unit. For example, the audio report data may be modified such that theaudio report 108 provided to theuser 101 is the message “I′m in another room!!” In other words, themaster audio unit 200E may be aware that it is communicating with themobile device 102 via theaudio unit 200A which is located in a different physical position from the audio unit 200B1. In some embodiments, themobile device 102 location may be triangulated by the array of home audio units using wireless signals or audio signals. Themaster audio unit 200E made thus make an inference that themobile device 102 is located in another room and change the report data or provide analternative audio report 108 accordingly. Inoperation block 1215, the query response audio data, such asaudio report data 106, is routed to the audio unit 200B1 so that it will be heard by theuser 101. - Also, in some embodiments, the
audio report 111 “I′m here” may be played by the speaker of the audio unit 200B1 or only by themobile device 102 speaker. However, in such embodiments, theuser 101 will have to travel through the rooms to figure out where exactly “here” is by listening for theaudio report 111 emanating from themobile device 102 speakers. As would be understood, theaudio report 111 may repeat for a predetermined period of time so that theuser 101 has time to look around and find themobile device 102. - As was already described with respect to other embodiments, in
operation block 1217, the query response, such asaudio report 111 oraudio report 108, may be played by the audio unit in closest proximity to the user (or the “best” audio unit) using the received audio report data 106 (or modified audio report data). The method of operation then ends as shown. Returning todecision block 1209, if the voice command does not include a query, then the audio unit in proximity to the user will wait to see if audio data is received from the mobile device as shown indecision block 1219. If not, the method of operation then ends. If yes, then any audio data is played using the audio unit in closest proximity to the user (or “best” audio unit) as shown inoperation block 1221 and the method of operation ends. - In embodiments where a master audio unit is used, other features may include capability to turn off audio units or put them into a sleep/standby mode. The master audio unit may also turn audio units off or to standby mode when the
mobile device 102 is not in wireless data range of the master audio unit or the basic audio units depending on the embodiment. For example, in some embodiments, the master audio unit may turn the basic audio units on or wake them from standby mode when themobile device 102 is in wireless data range of the master audio unit. - Basic audio units may be paired to a master audio unit during an initialization procedure of the
audio routing system 100. For example, audio units may be paired by the user manually, simultaneously pressing and holding a control button on the master audio unit and on the basic audio unit to initiate pairing. Completed pairing may be indicated by thestatus LED 323 illuminating (or showing a certain color such as green) or by a sound report from one or both of the master audio unit and the basic audio unit or by both. In other embodiments, the master audio unit may include a near field communication (NFC) unit and the basic audio units may include NFC tags. In such embodiments, pairing may be accomplished by placing a basic audio unit over or near the master audio unit. - In embodiments where the audio units employ Bluetooth®, various Bluetooth® profiles may be used to determine behavior during different audio events. For example, during an incoming call, the Hands-Free Profile (HFP) may be used. This profile supports two-way audio for conversations. In some embodiments, a master audio unit may as a default setting allow the audio ringer to be played on all audio units connected throughout the home. A user may then answer via voice command or button press on a specific audio unit or may ignore the call via voice command or button press. The default setting may be disabled or changed at the master audio unit if desired by the user either by voice command or control button input.
- Some embodiments may include an “answer all” feature in which a call may be answered and conducted across all available audio units in the home. For example, this feature is useful if a child is calling home from college and the parents are in different rooms of the home and want to participate in the call or for other like scenarios.
- In some embodiments, for multimedia audio playback on the audio units, the Advanced Audio Distribution Profile (A2DP) may be used. For example, in some embodiments, when the master audio unit receives a multimedia audio playback signal, the master audio unit may keep basic audio units muted by default. If the master audio unit includes a speaker, it may play the audio. Any other audio unit may be unmuted by voice command (such as “unmute”) or by pressing a control button on an audio unit. Thus, a different command to any audio unit or to the master audio unit (such as “unmute all”) may enable playback to all audio units. Any default settings may be modified by the user at the master audio unit via voice command or by using a control button. The above profile descriptions are examples only, and other profiles or wireless protocols may be employed in accordance with the various embodiments.
- While various embodiments have been illustrated and described, it is to be understood that the invention is not so limited. Numerous modifications, changes, variations, substitutions and equivalents will occur to those skilled in the art without departing from the scope of the present invention as defined by the appended claims.
Claims (21)
Priority Applications (8)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/065,506 US9668052B2 (en) | 2013-09-25 | 2013-10-29 | Audio routing system for routing audio data to and from a mobile device |
PCT/US2014/050495 WO2015047557A1 (en) | 2013-09-25 | 2014-08-11 | Audio routing system for routing audio data to and from a mobile device |
US15/604,332 US10182293B2 (en) | 2013-09-25 | 2017-05-24 | Audio routing system for routing audio data to and from a mobile device |
US16/246,040 US10735857B2 (en) | 2013-09-25 | 2019-01-11 | Audio routing system for routing audio data to and from a mobile device |
US16/935,651 US10887695B2 (en) | 2013-09-25 | 2020-07-22 | Audio routing system for routing audio data to and from a mobile device |
US17/122,889 US11375313B2 (en) | 2013-09-25 | 2020-12-15 | Audio routing system for routing audio data to and from a mobile device |
US17/850,637 US11689854B2 (en) | 2013-09-25 | 2022-06-27 | Audio routing system for routing audio data to and from a mobile device |
US18/197,507 US12063492B2 (en) | 2013-09-25 | 2023-05-15 | Audio routing system for routing audio data to and from a mobile device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361881965P | 2013-09-25 | 2013-09-25 | |
US14/065,506 US9668052B2 (en) | 2013-09-25 | 2013-10-29 | Audio routing system for routing audio data to and from a mobile device |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/604,332 Continuation US10182293B2 (en) | 2013-09-25 | 2017-05-24 | Audio routing system for routing audio data to and from a mobile device |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150086034A1 true US20150086034A1 (en) | 2015-03-26 |
US9668052B2 US9668052B2 (en) | 2017-05-30 |
Family
ID=52690959
Family Applications (7)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/065,506 Active 2034-08-16 US9668052B2 (en) | 2013-09-25 | 2013-10-29 | Audio routing system for routing audio data to and from a mobile device |
US15/604,332 Active 2033-11-14 US10182293B2 (en) | 2013-09-25 | 2017-05-24 | Audio routing system for routing audio data to and from a mobile device |
US16/246,040 Active US10735857B2 (en) | 2013-09-25 | 2019-01-11 | Audio routing system for routing audio data to and from a mobile device |
US16/935,651 Active US10887695B2 (en) | 2013-09-25 | 2020-07-22 | Audio routing system for routing audio data to and from a mobile device |
US17/122,889 Active US11375313B2 (en) | 2013-09-25 | 2020-12-15 | Audio routing system for routing audio data to and from a mobile device |
US17/850,637 Active US11689854B2 (en) | 2013-09-25 | 2022-06-27 | Audio routing system for routing audio data to and from a mobile device |
US18/197,507 Active US12063492B2 (en) | 2013-09-25 | 2023-05-15 | Audio routing system for routing audio data to and from a mobile device |
Family Applications After (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/604,332 Active 2033-11-14 US10182293B2 (en) | 2013-09-25 | 2017-05-24 | Audio routing system for routing audio data to and from a mobile device |
US16/246,040 Active US10735857B2 (en) | 2013-09-25 | 2019-01-11 | Audio routing system for routing audio data to and from a mobile device |
US16/935,651 Active US10887695B2 (en) | 2013-09-25 | 2020-07-22 | Audio routing system for routing audio data to and from a mobile device |
US17/122,889 Active US11375313B2 (en) | 2013-09-25 | 2020-12-15 | Audio routing system for routing audio data to and from a mobile device |
US17/850,637 Active US11689854B2 (en) | 2013-09-25 | 2022-06-27 | Audio routing system for routing audio data to and from a mobile device |
US18/197,507 Active US12063492B2 (en) | 2013-09-25 | 2023-05-15 | Audio routing system for routing audio data to and from a mobile device |
Country Status (2)
Country | Link |
---|---|
US (7) | US9668052B2 (en) |
WO (1) | WO2015047557A1 (en) |
Cited By (82)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9704489B2 (en) * | 2015-11-20 | 2017-07-11 | At&T Intellectual Property I, L.P. | Portable acoustical unit for voice recognition |
US20170242653A1 (en) * | 2016-02-22 | 2017-08-24 | Sonos, Inc. | Voice Control of a Media Playback System |
WO2018013944A1 (en) * | 2016-07-15 | 2018-01-18 | Sonos, Inc. | Contextualization of voice inputs |
US9947316B2 (en) | 2016-02-22 | 2018-04-17 | Sonos, Inc. | Voice control of a media playback system |
US9965247B2 (en) | 2016-02-22 | 2018-05-08 | Sonos, Inc. | Voice controlled media playback system based on user profile |
US9978390B2 (en) | 2016-06-09 | 2018-05-22 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US10014137B2 (en) | 2015-10-03 | 2018-07-03 | At&T Intellectual Property I, L.P. | Acoustical electrical switch |
US10034116B2 (en) | 2016-09-22 | 2018-07-24 | Sonos, Inc. | Acoustic position measurement |
US10051366B1 (en) | 2017-09-28 | 2018-08-14 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
EP3364410A1 (en) * | 2017-02-21 | 2018-08-22 | Advanced Digital Broadcast S.A. | System and method for voice recognition |
US10075793B2 (en) | 2016-09-30 | 2018-09-11 | Sonos, Inc. | Multi-orientation playback device microphones |
US10095470B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Audio response playback |
US10097939B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Compensation for speaker nonlinearities |
US10115400B2 (en) | 2016-08-05 | 2018-10-30 | Sonos, Inc. | Multiple voice services |
US10152969B2 (en) | 2016-07-15 | 2018-12-11 | Sonos, Inc. | Voice detection by multiple devices |
US20180358009A1 (en) * | 2017-06-09 | 2018-12-13 | International Business Machines Corporation | Cognitive and interactive sensor based smart home solution |
US10181323B2 (en) | 2016-10-19 | 2019-01-15 | Sonos, Inc. | Arbitration-based voice recognition |
US20190020761A1 (en) * | 2017-07-14 | 2019-01-17 | Google Inc. | Computational assistant extension device |
CN109313465A (en) * | 2016-04-05 | 2019-02-05 | 惠普发展公司,有限责任合伙企业 | The audio interface docked for multiple microphones and speaker system with host |
US20190051300A1 (en) * | 2017-08-08 | 2019-02-14 | Premium Loudspeakers (Hui Zhou) Co., Ltd. | Loudspeaker system |
US10264030B2 (en) | 2016-02-22 | 2019-04-16 | Sonos, Inc. | Networked microphone device control |
US10327117B2 (en) * | 2017-02-08 | 2019-06-18 | CliniCloud Inc. | Virtual mesh network for medical voice command devices |
US10365889B2 (en) * | 2016-02-22 | 2019-07-30 | Sonos, Inc. | Metadata exchange involving a networked playback system and a networked microphone system |
US20190273963A1 (en) * | 2016-06-27 | 2019-09-05 | Amazon Technologies, Inc. | Systems and methods for routing content to an associated output device |
US10446165B2 (en) | 2017-09-27 | 2019-10-15 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US10445057B2 (en) | 2017-09-08 | 2019-10-15 | Sonos, Inc. | Dynamic computation of system response volume |
US10466962B2 (en) | 2017-09-29 | 2019-11-05 | Sonos, Inc. | Media playback system with voice assistance |
US10475449B2 (en) | 2017-08-07 | 2019-11-12 | Sonos, Inc. | Wake-word detection suppression |
US10482868B2 (en) | 2017-09-28 | 2019-11-19 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US20200013423A1 (en) * | 2014-04-02 | 2020-01-09 | Plantronics. Inc. | Noise level measurement with mobile devices, location services, and environmental response |
US10573321B1 (en) | 2018-09-25 | 2020-02-25 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US10582322B2 (en) | 2016-09-27 | 2020-03-03 | Sonos, Inc. | Audio playback settings for voice interaction |
US10587430B1 (en) | 2018-09-14 | 2020-03-10 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US10586540B1 (en) | 2019-06-12 | 2020-03-10 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US10602268B1 (en) | 2018-12-20 | 2020-03-24 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US10621981B2 (en) | 2017-09-28 | 2020-04-14 | Sonos, Inc. | Tone interference cancellation |
CN111050265A (en) * | 2019-11-27 | 2020-04-21 | 深圳易科声光科技股份有限公司 | Automatic audio link detection method and device |
WO2020085769A1 (en) * | 2018-10-24 | 2020-04-30 | Samsung Electronics Co., Ltd. | Speech recognition method and apparatus in environment including plurality of apparatuses |
US10681460B2 (en) | 2018-06-28 | 2020-06-09 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
US10692518B2 (en) | 2018-09-29 | 2020-06-23 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US10797667B2 (en) | 2018-08-28 | 2020-10-06 | Sonos, Inc. | Audio notifications |
US10818290B2 (en) | 2017-12-11 | 2020-10-27 | Sonos, Inc. | Home graph |
US10831440B2 (en) * | 2016-09-26 | 2020-11-10 | Lenovo (Singapore) Pte. Ltd. | Coordinating input on multiple local devices |
US10847178B2 (en) | 2018-05-18 | 2020-11-24 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection |
US10867604B2 (en) | 2019-02-08 | 2020-12-15 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
US10871943B1 (en) | 2019-07-31 | 2020-12-22 | Sonos, Inc. | Noise classification for event detection |
US10878811B2 (en) | 2018-09-14 | 2020-12-29 | Sonos, Inc. | Networked devices, systems, and methods for intelligently deactivating wake-word engines |
US10880650B2 (en) | 2017-12-10 | 2020-12-29 | Sonos, Inc. | Network microphone devices with automatic do not disturb actuation capabilities |
US10959029B2 (en) | 2018-05-25 | 2021-03-23 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
US11024331B2 (en) | 2018-09-21 | 2021-06-01 | Sonos, Inc. | Voice detection optimization using sound metadata |
US11076035B2 (en) | 2018-08-28 | 2021-07-27 | Sonos, Inc. | Do not disturb feature for audio notifications |
US11100923B2 (en) | 2018-09-28 | 2021-08-24 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US11120794B2 (en) | 2019-05-03 | 2021-09-14 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
US11132989B2 (en) | 2018-12-13 | 2021-09-28 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US11138975B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US11138969B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US11175880B2 (en) | 2018-05-10 | 2021-11-16 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US11183183B2 (en) * | 2018-12-07 | 2021-11-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11183181B2 (en) | 2017-03-27 | 2021-11-23 | Sonos, Inc. | Systems and methods of multiple voice services |
US11189286B2 (en) | 2019-10-22 | 2021-11-30 | Sonos, Inc. | VAS toggle based on device orientation |
US20210377702A1 (en) * | 2018-02-22 | 2021-12-02 | Amazon Technologies, Inc. | Outputting notifications using device groups |
US11200894B2 (en) | 2019-06-12 | 2021-12-14 | Sonos, Inc. | Network microphone device with command keyword eventing |
US11200900B2 (en) | 2019-12-20 | 2021-12-14 | Sonos, Inc. | Offline voice control |
US11200889B2 (en) | 2018-11-15 | 2021-12-14 | Sonos, Inc. | Dilated convolutions and gating for efficient keyword spotting |
US11212486B1 (en) * | 2016-03-31 | 2021-12-28 | Amazon Technologies, Inc. | Location based device grouping with voice control |
US20220101846A1 (en) * | 2020-09-11 | 2022-03-31 | Samsung Electronics Co., Ltd. | Electronic device for identifying command included in voice and method of opearating the same |
US11308962B2 (en) | 2020-05-20 | 2022-04-19 | Sonos, Inc. | Input detection windowing |
US11308958B2 (en) | 2020-02-07 | 2022-04-19 | Sonos, Inc. | Localized wakeword verification |
US11315556B2 (en) | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
US11343614B2 (en) | 2018-01-31 | 2022-05-24 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US11361756B2 (en) | 2019-06-12 | 2022-06-14 | Sonos, Inc. | Conditional wake word eventing based on environment |
US11393478B2 (en) * | 2018-12-12 | 2022-07-19 | Sonos, Inc. | User specific context switching |
US11482224B2 (en) | 2020-05-20 | 2022-10-25 | Sonos, Inc. | Command keywords with input detection windowing |
US11551700B2 (en) | 2021-01-25 | 2023-01-10 | Sonos, Inc. | Systems and methods for power-efficient keyword detection |
US11556307B2 (en) | 2020-01-31 | 2023-01-17 | Sonos, Inc. | Local voice data processing |
US11562740B2 (en) | 2020-01-07 | 2023-01-24 | Sonos, Inc. | Voice verification for media playback |
CN115766307A (en) * | 2021-09-02 | 2023-03-07 | 华为技术有限公司 | Calling system and calling method |
US20230080895A1 (en) * | 2021-09-15 | 2023-03-16 | International Business Machines Corporation | Dynamic operation of a voice controlled device |
US11698771B2 (en) | 2020-08-25 | 2023-07-11 | Sonos, Inc. | Vocal guidance engines for playback devices |
US11727919B2 (en) | 2020-05-20 | 2023-08-15 | Sonos, Inc. | Memory allocation for keyword spotting engines |
US11899519B2 (en) | 2018-10-23 | 2024-02-13 | Sonos, Inc. | Multiple stage network microphone device with reduced power consumption and processing load |
US11984123B2 (en) | 2020-11-12 | 2024-05-14 | Sonos, Inc. | Network device interaction by range |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9668052B2 (en) | 2013-09-25 | 2017-05-30 | Google Technology Holdings LLC | Audio routing system for routing audio data to and from a mobile device |
US10147441B1 (en) * | 2013-12-19 | 2018-12-04 | Amazon Technologies, Inc. | Voice controlled system |
US10467509B2 (en) * | 2017-02-14 | 2019-11-05 | Microsoft Technology Licensing, Llc | Computationally-efficient human-identifying smart assistant computer |
US10121494B1 (en) * | 2017-03-30 | 2018-11-06 | Amazon Technologies, Inc. | User presence detection |
US10051600B1 (en) * | 2017-12-12 | 2018-08-14 | Amazon Technologies, Inc. | Selective notification delivery based on user presence detections |
EP4343757A3 (en) * | 2018-04-23 | 2024-05-29 | Google LLC | Transferring an automated assistant routine between client devices during execution of the routine |
CN108900950A (en) * | 2018-06-11 | 2018-11-27 | 杨俊杰 | A kind of integrated communicaton and driving are in the power carrier speaker driving chip of one |
US11528575B2 (en) * | 2020-07-28 | 2022-12-13 | Arris Enterprises Llc | System and method for dynamic control of wireless speaker systems |
US11575992B2 (en) * | 2020-10-02 | 2023-02-07 | Arris Enierprises Llc | System and method for dynamic line-of-sight multi-source audio control |
SE546166C2 (en) * | 2021-07-07 | 2024-06-18 | Pink Nectarine Health Ab | A system including a backend system, a self-organizing network (SON) including at least three network nodes, and a wearable device, and methods for improving the audio experience of an individual carrying a low power wearable device connected to a network |
SE546167C2 (en) * | 2021-07-07 | 2024-06-18 | Pink Nectarine Health Ab | A system comprising a backend system, a self-organizing network (SON) and a wearable device, and methods for establishment of a communication connection for a call involving a wearable device connected to a network |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030210770A1 (en) * | 2002-05-10 | 2003-11-13 | Brian Krejcarek | Method and apparatus for peer-to-peer voice communication using voice recognition and proper noun identification |
US20050026560A1 (en) * | 2003-07-28 | 2005-02-03 | Fellowes Inc. | Audio communications system including wireless microphone and wireless speaker |
US8054987B2 (en) * | 2005-12-29 | 2011-11-08 | Henri Seydoux | System for wireless audio signal distribution between a plurality of active loudspeakers |
US20130057215A1 (en) * | 2011-09-07 | 2013-03-07 | Lakshman Rajeswaran | Attachable charger |
US20130103765A1 (en) * | 2011-10-19 | 2013-04-25 | Matthew Nicholas Papakipos | Social Ad Hoc Networking Protocol and Presentation Layer |
US20130343568A1 (en) * | 2006-08-31 | 2013-12-26 | Avrum G. Mayman | System With Speaker, Transceiver and Related Devices |
US8983383B1 (en) * | 2012-09-25 | 2015-03-17 | Rawles Llc | Providing hands-free service to multiple devices |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6181739B1 (en) * | 1995-11-22 | 2001-01-30 | Telefonaktiebolaget Lm Ericsson (Publ) | Signal-to-noise ratio determination using digital signal processing |
US6104923A (en) * | 1997-10-03 | 2000-08-15 | Karen Kite | Remote operational screener |
US6658091B1 (en) | 2002-02-01 | 2003-12-02 | @Security Broadband Corp. | LIfestyle multimedia security system |
US7391315B2 (en) | 2004-11-16 | 2008-06-24 | Sonitrol Corporation | System and method for monitoring security at a plurality of premises |
US7764775B2 (en) | 2005-03-30 | 2010-07-27 | Onq/Legrand, Inc. | Distributed intercom system |
US20100087182A1 (en) * | 2006-11-02 | 2010-04-08 | Stewart Robert J | System and method for calling party to specify a ring tone used by a called party's mobile phone |
US20120188067A1 (en) | 2011-01-24 | 2012-07-26 | Weihao Xiao | Alarm Sound Activated Module for Remote Notification |
US20130089026A1 (en) | 2011-07-18 | 2013-04-11 | geoffrey Chilton Piper | Wireless Audio Transmission |
US8885623B2 (en) | 2011-09-22 | 2014-11-11 | American Megatrends, Inc. | Audio communications system and methods using personal wireless communication devices |
US20130150114A1 (en) | 2011-09-23 | 2013-06-13 | Revolabs, Inc. | Wireless multi-user audio system |
US9251792B2 (en) * | 2012-06-15 | 2016-02-02 | Sri International | Multi-sample conversational voice verification |
US9668052B2 (en) | 2013-09-25 | 2017-05-30 | Google Technology Holdings LLC | Audio routing system for routing audio data to and from a mobile device |
-
2013
- 2013-10-29 US US14/065,506 patent/US9668052B2/en active Active
-
2014
- 2014-08-11 WO PCT/US2014/050495 patent/WO2015047557A1/en active Application Filing
-
2017
- 2017-05-24 US US15/604,332 patent/US10182293B2/en active Active
-
2019
- 2019-01-11 US US16/246,040 patent/US10735857B2/en active Active
-
2020
- 2020-07-22 US US16/935,651 patent/US10887695B2/en active Active
- 2020-12-15 US US17/122,889 patent/US11375313B2/en active Active
-
2022
- 2022-06-27 US US17/850,637 patent/US11689854B2/en active Active
-
2023
- 2023-05-15 US US18/197,507 patent/US12063492B2/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030210770A1 (en) * | 2002-05-10 | 2003-11-13 | Brian Krejcarek | Method and apparatus for peer-to-peer voice communication using voice recognition and proper noun identification |
US20050026560A1 (en) * | 2003-07-28 | 2005-02-03 | Fellowes Inc. | Audio communications system including wireless microphone and wireless speaker |
US8054987B2 (en) * | 2005-12-29 | 2011-11-08 | Henri Seydoux | System for wireless audio signal distribution between a plurality of active loudspeakers |
US20130343568A1 (en) * | 2006-08-31 | 2013-12-26 | Avrum G. Mayman | System With Speaker, Transceiver and Related Devices |
US20130057215A1 (en) * | 2011-09-07 | 2013-03-07 | Lakshman Rajeswaran | Attachable charger |
US20130103765A1 (en) * | 2011-10-19 | 2013-04-25 | Matthew Nicholas Papakipos | Social Ad Hoc Networking Protocol and Presentation Layer |
US8983383B1 (en) * | 2012-09-25 | 2015-03-17 | Rawles Llc | Providing hands-free service to multiple devices |
Cited By (243)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200013423A1 (en) * | 2014-04-02 | 2020-01-09 | Plantronics. Inc. | Noise level measurement with mobile devices, location services, and environmental response |
US11404228B2 (en) | 2015-10-03 | 2022-08-02 | At&T Intellectual Property I, L.P. | Smart acoustical electrical switch |
US10672572B2 (en) | 2015-10-03 | 2020-06-02 | At&T Intellectual Property I, L.P. | Smart acoustical electrical switch |
US10014137B2 (en) | 2015-10-03 | 2018-07-03 | At&T Intellectual Property I, L.P. | Acoustical electrical switch |
US10091021B2 (en) | 2015-11-20 | 2018-10-02 | At&T Intellectual Property I, L.P. | Portable acoustical unit |
US9704489B2 (en) * | 2015-11-20 | 2017-07-11 | At&T Intellectual Property I, L.P. | Portable acoustical unit for voice recognition |
US10958468B2 (en) * | 2015-11-20 | 2021-03-23 | At&T Intellectual Property I, L. P. | Portable acoustical unit |
US20180375682A1 (en) * | 2015-11-20 | 2018-12-27 | At&T Intellectual Property I, L.P. | Portable Acoustical Unit |
AU2022291481B2 (en) * | 2016-02-22 | 2023-02-23 | Sonos, Inc. | Audio response playback |
US11832068B2 (en) | 2016-02-22 | 2023-11-28 | Sonos, Inc. | Music service selection |
US11184704B2 (en) | 2016-02-22 | 2021-11-23 | Sonos, Inc. | Music service selection |
US11137979B2 (en) | 2016-02-22 | 2021-10-05 | Sonos, Inc. | Metadata exchange involving a networked playback system and a networked microphone system |
US11947870B2 (en) | 2016-02-22 | 2024-04-02 | Sonos, Inc. | Audio response playback |
KR102343902B1 (en) | 2016-02-22 | 2021-12-27 | 소노스 인코포레이티드 | Audio response playback |
US10095470B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Audio response playback |
US10097939B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Compensation for speaker nonlinearities |
US10097919B2 (en) | 2016-02-22 | 2018-10-09 | Sonos, Inc. | Music service selection |
US20170242653A1 (en) * | 2016-02-22 | 2017-08-24 | Sonos, Inc. | Voice Control of a Media Playback System |
US11212612B2 (en) * | 2016-02-22 | 2021-12-28 | Sonos, Inc. | Voice control of a media playback system |
KR20210157487A (en) * | 2016-02-22 | 2021-12-28 | 소노스 인코포레이티드 | Audio response playback |
US10142754B2 (en) | 2016-02-22 | 2018-11-27 | Sonos, Inc. | Sensor on moving component of transducer |
US11042355B2 (en) | 2016-02-22 | 2021-06-22 | Sonos, Inc. | Handling of loss of pairing between networked devices |
EP3974957A1 (en) * | 2016-02-22 | 2022-03-30 | Sonos, Inc. | Audio response playback |
CN109076285A (en) * | 2016-02-22 | 2018-12-21 | 搜诺思公司 | Acoustic frequency response playback |
US11863593B2 (en) | 2016-02-22 | 2024-01-02 | Sonos, Inc. | Networked microphone device control |
AU2019236722B2 (en) * | 2016-02-22 | 2021-05-27 | Sonos, Inc. | Audio response playback |
US11006214B2 (en) | 2016-02-22 | 2021-05-11 | Sonos, Inc. | Default playback device designation |
EP3420736A4 (en) * | 2016-02-22 | 2019-01-23 | Sonos, Inc. | Audio response playback |
AU2017222436B2 (en) * | 2016-02-22 | 2019-01-31 | Sonos, Inc. | Voice control of a media playback system |
US10970035B2 (en) | 2016-02-22 | 2021-04-06 | Sonos, Inc. | Audio response playback |
KR20190014494A (en) * | 2016-02-22 | 2019-02-12 | 소노스 인코포레이티드 | How to play an audio response |
KR20190014495A (en) * | 2016-02-22 | 2019-02-12 | 소노스 인코포레이티드 | Voice control method of media playback system |
US10971139B2 (en) | 2016-02-22 | 2021-04-06 | Sonos, Inc. | Voice control of a media playback system |
US10212512B2 (en) | 2016-02-22 | 2019-02-19 | Sonos, Inc. | Default playback devices |
KR102234804B1 (en) | 2016-02-22 | 2021-04-01 | 소노스 인코포레이티드 | Voice control of a media playback system |
US10225651B2 (en) | 2016-02-22 | 2019-03-05 | Sonos, Inc. | Default playback device designation |
US10264030B2 (en) | 2016-02-22 | 2019-04-16 | Sonos, Inc. | Networked microphone device control |
WO2017147075A1 (en) | 2016-02-22 | 2017-08-31 | Sonos, Inc. | Audio response playback |
AU2019202257B2 (en) * | 2016-02-22 | 2021-02-04 | Sonos, Inc. | Voice control of a media playback system |
US20180226074A1 (en) * | 2016-02-22 | 2018-08-09 | Sonos, Inc. | Voice control of a media playback system |
KR102422270B1 (en) | 2016-02-22 | 2022-07-18 | 소노스 인코포레이티드 | Audio response playback |
US11983463B2 (en) | 2016-02-22 | 2024-05-14 | Sonos, Inc. | Metadata exchange involving a networked playback system and a networked microphone system |
AU2017223395B2 (en) * | 2016-02-22 | 2019-07-04 | Sonos, Inc. | Audio response playback |
US20240244368A1 (en) * | 2016-02-22 | 2024-07-18 | Sonos, Inc. | Voice Control of a Media Playback System |
US10365889B2 (en) * | 2016-02-22 | 2019-07-30 | Sonos, Inc. | Metadata exchange involving a networked playback system and a networked microphone system |
KR20200138421A (en) * | 2016-02-22 | 2020-12-09 | 소노스 인코포레이티드 | Audio response playback |
US10409549B2 (en) | 2016-02-22 | 2019-09-10 | Sonos, Inc. | Audio response playback |
KR102187147B1 (en) | 2016-02-22 | 2020-12-04 | 소노스 인코포레이티드 | Audio response playback |
US11726742B2 (en) | 2016-02-22 | 2023-08-15 | Sonos, Inc. | Handling of loss of pairing between networked devices |
US10847143B2 (en) * | 2016-02-22 | 2020-11-24 | Sonos, Inc. | Voice control of a media playback system |
KR102095250B1 (en) | 2016-02-22 | 2020-04-01 | 소노스 인코포레이티드 | Voice control method of media playback system |
US11736860B2 (en) * | 2016-02-22 | 2023-08-22 | Sonos, Inc. | Voice control of a media playback system |
AU2021212112B2 (en) * | 2016-02-22 | 2022-09-29 | Sonos, Inc. | Audio response playback |
US10499146B2 (en) * | 2016-02-22 | 2019-12-03 | Sonos, Inc. | Voice control of a media playback system |
US10509626B2 (en) | 2016-02-22 | 2019-12-17 | Sonos, Inc | Handling of loss of pairing between networked devices |
AU2023200179B2 (en) * | 2016-02-22 | 2023-03-02 | Sonos, Inc. | Voice control of a media playback system |
US9965247B2 (en) | 2016-02-22 | 2018-05-08 | Sonos, Inc. | Voice controlled media playback system based on user profile |
KR20200034839A (en) * | 2016-02-22 | 2020-03-31 | 소노스 인코포레이티드 | Voice control of a media playback system |
US11514898B2 (en) | 2016-02-22 | 2022-11-29 | Sonos, Inc. | Voice control of a media playback system |
US10764679B2 (en) | 2016-02-22 | 2020-09-01 | Sonos, Inc. | Voice control of a media playback system |
KR102080002B1 (en) * | 2016-02-22 | 2020-02-24 | 소노스 인코포레이티드 | How to Play Audio Response |
US20230054164A1 (en) * | 2016-02-22 | 2023-02-23 | Sonos, Inc. | Voice Control of a Media Playback System |
US10740065B2 (en) | 2016-02-22 | 2020-08-11 | Sonos, Inc. | Voice controlled media playback system |
KR20200022513A (en) * | 2016-02-22 | 2020-03-03 | 소노스 인코포레이티드 | Audio response playback |
US10743101B2 (en) | 2016-02-22 | 2020-08-11 | Sonos, Inc. | Content mixing |
AU2021202786B2 (en) * | 2016-02-22 | 2023-02-23 | Sonos, Inc. | Voice control of a media playback system |
CN111510821A (en) * | 2016-02-22 | 2020-08-07 | 搜诺思公司 | Audio response playback |
US12047752B2 (en) | 2016-02-22 | 2024-07-23 | Sonos, Inc. | Content mixing |
US11513763B2 (en) | 2016-02-22 | 2022-11-29 | Sonos, Inc. | Audio response playback |
US10555077B2 (en) | 2016-02-22 | 2020-02-04 | Sonos, Inc. | Music service selection |
US11405430B2 (en) | 2016-02-22 | 2022-08-02 | Sonos, Inc. | Networked microphone device control |
US11750969B2 (en) | 2016-02-22 | 2023-09-05 | Sonos, Inc. | Default playback device designation |
US11556306B2 (en) | 2016-02-22 | 2023-01-17 | Sonos, Inc. | Voice controlled media playback system |
US9947316B2 (en) | 2016-02-22 | 2018-04-17 | Sonos, Inc. | Voice control of a media playback system |
US11902707B1 (en) | 2016-03-31 | 2024-02-13 | Amazon Technologies, Inc. | Location based device grouping with voice control |
US11212486B1 (en) * | 2016-03-31 | 2021-12-28 | Amazon Technologies, Inc. | Location based device grouping with voice control |
EP3440527A4 (en) * | 2016-04-05 | 2019-11-27 | Hewlett-Packard Development Company, L.P. | Audio interface for multiple microphones and speaker systems to interface with a host |
CN109313465A (en) * | 2016-04-05 | 2019-02-05 | 惠普发展公司,有限责任合伙企业 | The audio interface docked for multiple microphones and speaker system with host |
US10848866B2 (en) | 2016-04-05 | 2020-11-24 | Hewett-Packard Development Company, L.P. | Audio interface for multiple microphones and speaker systems to interface with a host |
US11133018B2 (en) | 2016-06-09 | 2021-09-28 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US9978390B2 (en) | 2016-06-09 | 2018-05-22 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US10714115B2 (en) | 2016-06-09 | 2020-07-14 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US11545169B2 (en) | 2016-06-09 | 2023-01-03 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US10332537B2 (en) | 2016-06-09 | 2019-06-25 | Sonos, Inc. | Dynamic player selection for audio signal processing |
US11064248B2 (en) * | 2016-06-27 | 2021-07-13 | Amazon Technologies, Inc. | Systems and methods for routing content to an associated output device |
US20190273963A1 (en) * | 2016-06-27 | 2019-09-05 | Amazon Technologies, Inc. | Systems and methods for routing content to an associated output device |
US10297256B2 (en) | 2016-07-15 | 2019-05-21 | Sonos, Inc. | Voice detection by multiple devices |
US10134399B2 (en) | 2016-07-15 | 2018-11-20 | Sonos, Inc. | Contextualization of voice inputs |
US11184969B2 (en) | 2016-07-15 | 2021-11-23 | Sonos, Inc. | Contextualization of voice inputs |
US11979960B2 (en) | 2016-07-15 | 2024-05-07 | Sonos, Inc. | Contextualization of voice inputs |
US10699711B2 (en) | 2016-07-15 | 2020-06-30 | Sonos, Inc. | Voice detection by multiple devices |
US10152969B2 (en) | 2016-07-15 | 2018-12-11 | Sonos, Inc. | Voice detection by multiple devices |
CN109690672A (en) * | 2016-07-15 | 2019-04-26 | 搜诺思公司 | Voice is inputted and carries out contextualization |
US10593331B2 (en) | 2016-07-15 | 2020-03-17 | Sonos, Inc. | Contextualization of voice inputs |
US11664023B2 (en) | 2016-07-15 | 2023-05-30 | Sonos, Inc. | Voice detection by multiple devices |
WO2018013944A1 (en) * | 2016-07-15 | 2018-01-18 | Sonos, Inc. | Contextualization of voice inputs |
US10565999B2 (en) | 2016-08-05 | 2020-02-18 | Sonos, Inc. | Playback device supporting concurrent voice assistant services |
US10847164B2 (en) | 2016-08-05 | 2020-11-24 | Sonos, Inc. | Playback device supporting concurrent voice assistants |
US11531520B2 (en) | 2016-08-05 | 2022-12-20 | Sonos, Inc. | Playback device supporting concurrent voice assistants |
US10565998B2 (en) | 2016-08-05 | 2020-02-18 | Sonos, Inc. | Playback device supporting concurrent voice assistant services |
US10115400B2 (en) | 2016-08-05 | 2018-10-30 | Sonos, Inc. | Multiple voice services |
US10354658B2 (en) | 2016-08-05 | 2019-07-16 | Sonos, Inc. | Voice control of playback device using voice assistant service(s) |
US11934742B2 (en) | 2016-08-05 | 2024-03-19 | Sonos, Inc. | Playback device supporting concurrent voice assistants |
US10034116B2 (en) | 2016-09-22 | 2018-07-24 | Sonos, Inc. | Acoustic position measurement |
US10831440B2 (en) * | 2016-09-26 | 2020-11-10 | Lenovo (Singapore) Pte. Ltd. | Coordinating input on multiple local devices |
US10582322B2 (en) | 2016-09-27 | 2020-03-03 | Sonos, Inc. | Audio playback settings for voice interaction |
US11641559B2 (en) | 2016-09-27 | 2023-05-02 | Sonos, Inc. | Audio playback settings for voice interaction |
US10313812B2 (en) | 2016-09-30 | 2019-06-04 | Sonos, Inc. | Orientation-based playback device microphone selection |
US11516610B2 (en) | 2016-09-30 | 2022-11-29 | Sonos, Inc. | Orientation-based playback device microphone selection |
US10075793B2 (en) | 2016-09-30 | 2018-09-11 | Sonos, Inc. | Multi-orientation playback device microphones |
US10873819B2 (en) | 2016-09-30 | 2020-12-22 | Sonos, Inc. | Orientation-based playback device microphone selection |
US10117037B2 (en) | 2016-09-30 | 2018-10-30 | Sonos, Inc. | Orientation-based playback device microphone selection |
US11308961B2 (en) | 2016-10-19 | 2022-04-19 | Sonos, Inc. | Arbitration-based voice recognition |
US10181323B2 (en) | 2016-10-19 | 2019-01-15 | Sonos, Inc. | Arbitration-based voice recognition |
US11727933B2 (en) | 2016-10-19 | 2023-08-15 | Sonos, Inc. | Arbitration-based voice recognition |
US10614807B2 (en) | 2016-10-19 | 2020-04-07 | Sonos, Inc. | Arbitration-based voice recognition |
US10327117B2 (en) * | 2017-02-08 | 2019-06-18 | CliniCloud Inc. | Virtual mesh network for medical voice command devices |
EP3364410A1 (en) * | 2017-02-21 | 2018-08-22 | Advanced Digital Broadcast S.A. | System and method for voice recognition |
US11183181B2 (en) | 2017-03-27 | 2021-11-23 | Sonos, Inc. | Systems and methods of multiple voice services |
US11853648B2 (en) | 2017-06-09 | 2023-12-26 | International Business Machines Corporation | Cognitive and interactive sensor based smart home solution |
US20180358009A1 (en) * | 2017-06-09 | 2018-12-13 | International Business Machines Corporation | Cognitive and interactive sensor based smart home solution |
US10983753B2 (en) * | 2017-06-09 | 2021-04-20 | International Business Machines Corporation | Cognitive and interactive sensor based smart home solution |
US20190020761A1 (en) * | 2017-07-14 | 2019-01-17 | Google Inc. | Computational assistant extension device |
US11005993B2 (en) * | 2017-07-14 | 2021-05-11 | Google Llc | Computational assistant extension device |
US11558505B2 (en) | 2017-07-14 | 2023-01-17 | Google Llc | Computational assistant extension device |
US11900937B2 (en) | 2017-08-07 | 2024-02-13 | Sonos, Inc. | Wake-word detection suppression |
US11380322B2 (en) | 2017-08-07 | 2022-07-05 | Sonos, Inc. | Wake-word detection suppression |
US10475449B2 (en) | 2017-08-07 | 2019-11-12 | Sonos, Inc. | Wake-word detection suppression |
US20190051300A1 (en) * | 2017-08-08 | 2019-02-14 | Premium Loudspeakers (Hui Zhou) Co., Ltd. | Loudspeaker system |
CN109391884A (en) * | 2017-08-08 | 2019-02-26 | 惠州超声音响有限公司 | Speaker system and the method for manipulating loudspeaker |
US10445057B2 (en) | 2017-09-08 | 2019-10-15 | Sonos, Inc. | Dynamic computation of system response volume |
US11080005B2 (en) | 2017-09-08 | 2021-08-03 | Sonos, Inc. | Dynamic computation of system response volume |
US11500611B2 (en) | 2017-09-08 | 2022-11-15 | Sonos, Inc. | Dynamic computation of system response volume |
US11646045B2 (en) | 2017-09-27 | 2023-05-09 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US11017789B2 (en) | 2017-09-27 | 2021-05-25 | Sonos, Inc. | Robust Short-Time Fourier Transform acoustic echo cancellation during audio playback |
US10446165B2 (en) | 2017-09-27 | 2019-10-15 | Sonos, Inc. | Robust short-time fourier transform acoustic echo cancellation during audio playback |
US10511904B2 (en) | 2017-09-28 | 2019-12-17 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US10621981B2 (en) | 2017-09-28 | 2020-04-14 | Sonos, Inc. | Tone interference cancellation |
US10880644B1 (en) | 2017-09-28 | 2020-12-29 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US12047753B1 (en) | 2017-09-28 | 2024-07-23 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US10051366B1 (en) | 2017-09-28 | 2018-08-14 | Sonos, Inc. | Three-dimensional beam forming with a microphone array |
US11769505B2 (en) | 2017-09-28 | 2023-09-26 | Sonos, Inc. | Echo of tone interferance cancellation using two acoustic echo cancellers |
US10482868B2 (en) | 2017-09-28 | 2019-11-19 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US10891932B2 (en) | 2017-09-28 | 2021-01-12 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US11538451B2 (en) | 2017-09-28 | 2022-12-27 | Sonos, Inc. | Multi-channel acoustic echo cancellation |
US11302326B2 (en) | 2017-09-28 | 2022-04-12 | Sonos, Inc. | Tone interference cancellation |
US10606555B1 (en) | 2017-09-29 | 2020-03-31 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US11893308B2 (en) | 2017-09-29 | 2024-02-06 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US11175888B2 (en) | 2017-09-29 | 2021-11-16 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US11288039B2 (en) | 2017-09-29 | 2022-03-29 | Sonos, Inc. | Media playback system with concurrent voice assistance |
US10466962B2 (en) | 2017-09-29 | 2019-11-05 | Sonos, Inc. | Media playback system with voice assistance |
US11451908B2 (en) | 2017-12-10 | 2022-09-20 | Sonos, Inc. | Network microphone devices with automatic do not disturb actuation capabilities |
US10880650B2 (en) | 2017-12-10 | 2020-12-29 | Sonos, Inc. | Network microphone devices with automatic do not disturb actuation capabilities |
US11676590B2 (en) | 2017-12-11 | 2023-06-13 | Sonos, Inc. | Home graph |
US10818290B2 (en) | 2017-12-11 | 2020-10-27 | Sonos, Inc. | Home graph |
US11343614B2 (en) | 2018-01-31 | 2022-05-24 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US11689858B2 (en) | 2018-01-31 | 2023-06-27 | Sonos, Inc. | Device designation of playback and network microphone device arrangements |
US20210377702A1 (en) * | 2018-02-22 | 2021-12-02 | Amazon Technologies, Inc. | Outputting notifications using device groups |
US11812344B2 (en) * | 2018-02-22 | 2023-11-07 | Amazon Technologies, Inc. | Outputting notifications using device groups |
US11797263B2 (en) | 2018-05-10 | 2023-10-24 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US11175880B2 (en) | 2018-05-10 | 2021-11-16 | Sonos, Inc. | Systems and methods for voice-assisted media content selection |
US10847178B2 (en) | 2018-05-18 | 2020-11-24 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection |
US11715489B2 (en) | 2018-05-18 | 2023-08-01 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection |
US11792590B2 (en) | 2018-05-25 | 2023-10-17 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
US10959029B2 (en) | 2018-05-25 | 2021-03-23 | Sonos, Inc. | Determining and adapting to changes in microphone performance of playback devices |
US10681460B2 (en) | 2018-06-28 | 2020-06-09 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
US11696074B2 (en) | 2018-06-28 | 2023-07-04 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
US11197096B2 (en) | 2018-06-28 | 2021-12-07 | Sonos, Inc. | Systems and methods for associating playback devices with voice assistant services |
US11563842B2 (en) | 2018-08-28 | 2023-01-24 | Sonos, Inc. | Do not disturb feature for audio notifications |
US11973893B2 (en) | 2018-08-28 | 2024-04-30 | Sonos, Inc. | Do not disturb feature for audio notifications |
US11482978B2 (en) | 2018-08-28 | 2022-10-25 | Sonos, Inc. | Audio notifications |
US10797667B2 (en) | 2018-08-28 | 2020-10-06 | Sonos, Inc. | Audio notifications |
US11076035B2 (en) | 2018-08-28 | 2021-07-27 | Sonos, Inc. | Do not disturb feature for audio notifications |
US11432030B2 (en) | 2018-09-14 | 2022-08-30 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US11551690B2 (en) | 2018-09-14 | 2023-01-10 | Sonos, Inc. | Networked devices, systems, and methods for intelligently deactivating wake-word engines |
US11778259B2 (en) | 2018-09-14 | 2023-10-03 | Sonos, Inc. | Networked devices, systems and methods for associating playback devices based on sound codes |
US10878811B2 (en) | 2018-09-14 | 2020-12-29 | Sonos, Inc. | Networked devices, systems, and methods for intelligently deactivating wake-word engines |
US10587430B1 (en) | 2018-09-14 | 2020-03-10 | Sonos, Inc. | Networked devices, systems, and methods for associating playback devices based on sound codes |
US11790937B2 (en) | 2018-09-21 | 2023-10-17 | Sonos, Inc. | Voice detection optimization using sound metadata |
US11024331B2 (en) | 2018-09-21 | 2021-06-01 | Sonos, Inc. | Voice detection optimization using sound metadata |
US10811015B2 (en) | 2018-09-25 | 2020-10-20 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11727936B2 (en) | 2018-09-25 | 2023-08-15 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11031014B2 (en) | 2018-09-25 | 2021-06-08 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US10573321B1 (en) | 2018-09-25 | 2020-02-25 | Sonos, Inc. | Voice detection optimization based on selected voice assistant service |
US11100923B2 (en) | 2018-09-28 | 2021-08-24 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US11790911B2 (en) | 2018-09-28 | 2023-10-17 | Sonos, Inc. | Systems and methods for selective wake word detection using neural network models |
US12062383B2 (en) | 2018-09-29 | 2024-08-13 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US10692518B2 (en) | 2018-09-29 | 2020-06-23 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US11501795B2 (en) | 2018-09-29 | 2022-11-15 | Sonos, Inc. | Linear filtering for noise-suppressed speech detection via multiple network microphone devices |
US11899519B2 (en) | 2018-10-23 | 2024-02-13 | Sonos, Inc. | Multiple stage network microphone device with reduced power consumption and processing load |
WO2020085769A1 (en) * | 2018-10-24 | 2020-04-30 | Samsung Electronics Co., Ltd. | Speech recognition method and apparatus in environment including plurality of apparatuses |
US11741948B2 (en) | 2018-11-15 | 2023-08-29 | Sonos Vox France Sas | Dilated convolutions and gating for efficient keyword spotting |
US11200889B2 (en) | 2018-11-15 | 2021-12-14 | Sonos, Inc. | Dilated convolutions and gating for efficient keyword spotting |
US11881223B2 (en) * | 2018-12-07 | 2024-01-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11183183B2 (en) * | 2018-12-07 | 2021-11-23 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11557294B2 (en) | 2018-12-07 | 2023-01-17 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US20230215433A1 (en) * | 2018-12-07 | 2023-07-06 | Sonos, Inc. | Systems and methods of operating media playback systems having multiple voice assistant services |
US11393478B2 (en) * | 2018-12-12 | 2022-07-19 | Sonos, Inc. | User specific context switching |
US11790920B2 (en) | 2018-12-12 | 2023-10-17 | Sonos, Inc. | Guest access for voice control of playback devices |
US11132989B2 (en) | 2018-12-13 | 2021-09-28 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US11538460B2 (en) | 2018-12-13 | 2022-12-27 | Sonos, Inc. | Networked microphone devices, systems, and methods of localized arbitration |
US11540047B2 (en) | 2018-12-20 | 2022-12-27 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US10602268B1 (en) | 2018-12-20 | 2020-03-24 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US11159880B2 (en) | 2018-12-20 | 2021-10-26 | Sonos, Inc. | Optimization of network microphone devices using noise classification |
US11315556B2 (en) | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
US11646023B2 (en) | 2019-02-08 | 2023-05-09 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
US10867604B2 (en) | 2019-02-08 | 2020-12-15 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing |
US11120794B2 (en) | 2019-05-03 | 2021-09-14 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
US11798553B2 (en) | 2019-05-03 | 2023-10-24 | Sonos, Inc. | Voice assistant persistence across multiple network microphone devices |
US11501773B2 (en) | 2019-06-12 | 2022-11-15 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US11854547B2 (en) | 2019-06-12 | 2023-12-26 | Sonos, Inc. | Network microphone device with command keyword eventing |
US11200894B2 (en) | 2019-06-12 | 2021-12-14 | Sonos, Inc. | Network microphone device with command keyword eventing |
US10586540B1 (en) | 2019-06-12 | 2020-03-10 | Sonos, Inc. | Network microphone device with command keyword conditioning |
US11361756B2 (en) | 2019-06-12 | 2022-06-14 | Sonos, Inc. | Conditional wake word eventing based on environment |
US11138975B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US11551669B2 (en) | 2019-07-31 | 2023-01-10 | Sonos, Inc. | Locally distributed keyword detection |
US11354092B2 (en) | 2019-07-31 | 2022-06-07 | Sonos, Inc. | Noise classification for event detection |
US11714600B2 (en) | 2019-07-31 | 2023-08-01 | Sonos, Inc. | Noise classification for event detection |
US10871943B1 (en) | 2019-07-31 | 2020-12-22 | Sonos, Inc. | Noise classification for event detection |
US11710487B2 (en) | 2019-07-31 | 2023-07-25 | Sonos, Inc. | Locally distributed keyword detection |
US11138969B2 (en) | 2019-07-31 | 2021-10-05 | Sonos, Inc. | Locally distributed keyword detection |
US11862161B2 (en) | 2019-10-22 | 2024-01-02 | Sonos, Inc. | VAS toggle based on device orientation |
US11189286B2 (en) | 2019-10-22 | 2021-11-30 | Sonos, Inc. | VAS toggle based on device orientation |
CN111050265A (en) * | 2019-11-27 | 2020-04-21 | 深圳易科声光科技股份有限公司 | Automatic audio link detection method and device |
US11200900B2 (en) | 2019-12-20 | 2021-12-14 | Sonos, Inc. | Offline voice control |
US11869503B2 (en) | 2019-12-20 | 2024-01-09 | Sonos, Inc. | Offline voice control |
US11562740B2 (en) | 2020-01-07 | 2023-01-24 | Sonos, Inc. | Voice verification for media playback |
US11556307B2 (en) | 2020-01-31 | 2023-01-17 | Sonos, Inc. | Local voice data processing |
US12118273B2 (en) | 2020-01-31 | 2024-10-15 | Sonos, Inc. | Local voice data processing |
US11308958B2 (en) | 2020-02-07 | 2022-04-19 | Sonos, Inc. | Localized wakeword verification |
US11961519B2 (en) | 2020-02-07 | 2024-04-16 | Sonos, Inc. | Localized wakeword verification |
US11482224B2 (en) | 2020-05-20 | 2022-10-25 | Sonos, Inc. | Command keywords with input detection windowing |
US11308962B2 (en) | 2020-05-20 | 2022-04-19 | Sonos, Inc. | Input detection windowing |
US12119000B2 (en) | 2020-05-20 | 2024-10-15 | Sonos, Inc. | Input detection windowing |
US11727919B2 (en) | 2020-05-20 | 2023-08-15 | Sonos, Inc. | Memory allocation for keyword spotting engines |
US11694689B2 (en) | 2020-05-20 | 2023-07-04 | Sonos, Inc. | Input detection windowing |
US11698771B2 (en) | 2020-08-25 | 2023-07-11 | Sonos, Inc. | Vocal guidance engines for playback devices |
US20220101846A1 (en) * | 2020-09-11 | 2022-03-31 | Samsung Electronics Co., Ltd. | Electronic device for identifying command included in voice and method of opearating the same |
US11810561B2 (en) * | 2020-09-11 | 2023-11-07 | Samsung Electronics Co., Ltd. | Electronic device for identifying command included in voice and method of operating the same |
US11984123B2 (en) | 2020-11-12 | 2024-05-14 | Sonos, Inc. | Network device interaction by range |
US11551700B2 (en) | 2021-01-25 | 2023-01-10 | Sonos, Inc. | Systems and methods for power-efficient keyword detection |
CN115766307A (en) * | 2021-09-02 | 2023-03-07 | 华为技术有限公司 | Calling system and calling method |
US20230080895A1 (en) * | 2021-09-15 | 2023-03-16 | International Business Machines Corporation | Dynamic operation of a voice controlled device |
Also Published As
Publication number | Publication date |
---|---|
US12063492B2 (en) | 2024-08-13 |
US20230283954A1 (en) | 2023-09-07 |
US11689854B2 (en) | 2023-06-27 |
US20210105562A1 (en) | 2021-04-08 |
US20170265001A1 (en) | 2017-09-14 |
US9668052B2 (en) | 2017-05-30 |
US10735857B2 (en) | 2020-08-04 |
US20200351588A1 (en) | 2020-11-05 |
WO2015047557A1 (en) | 2015-04-02 |
US11375313B2 (en) | 2022-06-28 |
US10887695B2 (en) | 2021-01-05 |
US20220329944A1 (en) | 2022-10-13 |
US10182293B2 (en) | 2019-01-15 |
US20190149922A1 (en) | 2019-05-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US12063492B2 (en) | Audio routing system for routing audio data to and from a mobile device | |
TWI763642B (en) | Lighting and sound system and method of controlling the same | |
CN108496374B (en) | Earphone working mode switching method and device, audio circuit, earphone and earphone system | |
US9521360B2 (en) | Communication system and method | |
CN108399917B (en) | Speech processing method, apparatus and computer readable storage medium | |
US8761841B2 (en) | Personal conferencing device | |
US20110316664A1 (en) | Remote control for sound system | |
KR101071540B1 (en) | Wireless mic system for automatically pairing | |
KR101381289B1 (en) | Wire and wireless earset using in ear-type microphone | |
KR101122767B1 (en) | Hearing aid combination Bluetooth-Headset and Control Method | |
US9900715B2 (en) | Method and system for estimating acoustic noise levels | |
KR101330476B1 (en) | Wire and wireless earphone apparatus using in ear-type microphone | |
JP2018019205A (en) | Remote control system for water heater | |
US10038948B2 (en) | Battery charging adaptor for a wireless microphone | |
US11606852B2 (en) | Voice communication between lamp and remote device plus lighting control via remote device | |
WO2011083143A1 (en) | A personal conferencing device | |
US20160112574A1 (en) | Audio conferencing system for office furniture | |
US20110053509A1 (en) | Modular Dual Radio Headset | |
US20170064594A1 (en) | Apparatus and method for receiving an audio signal | |
KR101418929B1 (en) | Pairing Sensitivity Control Methode for Wireless MIC system | |
GB2549401A (en) | Audio systems | |
KR101946900B1 (en) | Media peripherals Wireless connection system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MOTOROLA MOBILITY LLC, ILLINOIS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LOMBARDI, MICHAEL J;PATEL, MITUL R;PIERCE, AMBER M;REEL/FRAME:031496/0083 Effective date: 20131024 |
|
AS | Assignment |
Owner name: GOOGLE TECHNOLOGY HOLDINGS LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MOTOROLA MOBILITY LLC;REEL/FRAME:034227/0095 Effective date: 20141028 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |