US11356789B2 - Signal processing device, channel setting method, and speaker system - Google Patents
Signal processing device, channel setting method, and speaker system Download PDFInfo
- Publication number
- US11356789B2 US11356789B2 US16/980,973 US201916980973A US11356789B2 US 11356789 B2 US11356789 B2 US 11356789B2 US 201916980973 A US201916980973 A US 201916980973A US 11356789 B2 US11356789 B2 US 11356789B2
- Authority
- US
- United States
- Prior art keywords
- speaker
- speakers
- signal processing
- processing device
- notification
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R5/00—Stereophonic arrangements
- H04R5/04—Circuit arrangements, e.g. for selective connection of amplifier inputs/outputs to loudspeakers, for loudspeaker detection, or for adaptation of settings to personal preferences or hearing impairments
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/12—Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S3/00—Systems employing more than two channels, e.g. quadraphonic
- H04S3/002—Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2205/00—Details of stereophonic arrangements covered by H04R5/00 but not provided for in any of its subgroups
- H04R2205/024—Positioning of loudspeaker enclosures for spatial sound reproduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2420/00—Details of connection covered by H04R, not provided for in its groups
- H04R2420/07—Applications of wireless loudspeakers or wireless microphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/01—Multi-channel, i.e. more than two input channels, sound reproduction with two speakers wherein the multi-channel information is substantially preserved
Definitions
- the present technology relates to a signal processing device, a channel setting method, a program, and a speaker system, and particularly relates to a technology of channel setting to each speaker.
- a user For use of a surround audio system that connects a plurality of speakers, a user needs to set a channel to each speaker (output channel), correctly.
- a user who is usually unfamiliar with such a system has difficulty in understanding its setting method and thus often performs wrong channel setting.
- a channel for each speaker is determined in advance and each speaker has a mark representing its determined channel in some cases.
- the marks include “FL” for the front left speaker, “FR” for the front right speaker, “SUR L” for the rear left speaker, and “SUR R” for the rear right speaker. Comparing such marks with the ideal layout of the speakers illustrated, for example, on the instruction manual, a user needs to arrange each speaker at the correct position in advance.
- Patent Document 1 Regarding channel setting to speakers, Patent Document 1 below has been also known.
- Patent Document 1 Japanese Patent Application Laid-Open No. 2002-345100
- Patent Document 1 discloses a technique in which, once a master unit and speakers are connected, a test tone is reproduced from each speaker at initial setup and an output channel is set to each speaker in order of test-tone reproduction.
- channel setting needs to be performed manually to all the speakers in connection.
- a user who is unfamiliar with the concept of left and right of speaker layout in such a system performs channel setting wrongly because of a reason similar to the above.
- an object of the present technology is to enable, in a case where a plurality of speakers is arranged like a surround system, facilitation of speaker channel setting and additionally even a user who is unfamiliar with such a system, to perform speaker channel setting accurately.
- a signal processing device includes: a relative-position recognition unit configured to perform processing of recognizing two arrangement reference speakers with reception of notification that a designation operation has been received from a user, from two speakers among N number of speakers, the N being three or more, and processing of acquiring distance information between each speaker, the relative-position recognition unit being configured to recognize a relative-position relationship between the N number of speakers, with the two arrangement reference speakers and the distance information between each speaker; and
- a channel setting unit configured to automatically set a channel to each speaker, on the basis of the relative-position relationship recognized by the relative-position recognition unit.
- a multichannel speaker system having three or more channels is assumed.
- Examples of the multichannel speaker system include a 5-channel surround audio system, a 7-channel surround audio system, and the like.
- the two speakers among the N number of speakers are detected by the notification of the designation operation of the user.
- the relative-position relationship between each speaker is recognized with the distance information between the N number of speakers grasped.
- the signal processing device further includes a channel signal processing unit configured to perform signal processing to an input sound signal and generate N channels of sound signals to be supplied one to one to the N number of speakers, in which the channel signal processing unit generates the N channels of sound signals as transmission signals one to one to the speakers, on the basis of the channels set by the channel setting unit.
- a channel signal processing unit configured to perform signal processing to an input sound signal and generate N channels of sound signals to be supplied one to one to the N number of speakers, in which the channel signal processing unit generates the N channels of sound signals as transmission signals one to one to the speakers, on the basis of the channels set by the channel setting unit.
- the channel signal processing unit generates a sound signal to each channel.
- the respective sound signals generated to the channels are allocated and transmitted one to one to the speakers, in accordance with channel assignment of each speaker by the channel setting unit.
- the N number of speakers each include an operation detection unit that detects a designation operation from the user, and the relative-position recognition unit issues an instruction for activation of the operation detection unit to each speaker, and additionally recognizes, as an arrangement reference speaker, a speaker having issued a notification of the operation detection unit having had detection during a period of activation.
- Each speaker is provided with the operation detection unit including a sensing device of some kind, such as a touch sensor, a button, a microphone, or an optical sensor.
- the relative-position recognition unit of the signal processing device performs activation control such that the operation detection unit of each speaker is activated. In a case where an operation has been detected during the period of the activation, it is recognized that a designation operation has been received from the user.
- the relative-position recognition unit recognizes, as a front left speaker and a front right speaker, the two arrangement reference speakers that have received the notification that the designation operation has been received from the user.
- the arrangement reference speakers determined are the front left speaker and the front right speaker arranged ahead of the face of the user at the time of listening.
- the relative-position recognition unit distinguishes the two arrangement reference speakers as the front left speaker and the front right speaker in order of the designation operations from the user.
- the user is prompted to perform sequential operations. Then, the front left speaker and the front right speaker are determined in the order.
- the relative-position recognition unit issues, in a case where a first designation operation is performed by the user, the instruction for activation of the operation detection unit to each speaker different from a speaker having transmitted a notification of the first designation operation, and waits for a second designation operation.
- the speaker is controlled so as not to issue a notification of a designation operation.
- the relative-position recognition unit causes, for acquisition of the distance information between each speaker, each speaker sequentially to output a test sound.
- One speaker is caused to output the test sound and the other speakers are caused to collect the test sound through the respective microphones. All the speakers are sequentially caused to output the test sound as described above.
- each speaker includes a sound detection unit and is capable of transmitting detection time information regarding test sound from another speaker
- the relative-position recognition unit calculates, from output start time information regarding test sound from a speaker and detection time information from another speaker, a distance between the speaker and the another speaker.
- the another speaker Because all the speakers are synchronized in time, for example, the another speaker generates a file in which the test sound is recorded together with the time information, and transmits the file to the relative-position recognition unit.
- the signal processing device further includes a virtual speaker setting unit configure to set a virtual speaker arrangement, on the basis of the relative-position relationship recognized by the relative-position recognition unit and the channel setting performed by the channel setting unit.
- a virtual speaker is a speaker virtually arranged in position differently from the actual speaker arrangement.
- the signal processing device further includes a channel signal processing unit configured to perform signal processing to an input sound signal and generate N channels of sound signals to be supplied one to one to the N number of speakers, in which the channel signal processing unit generates, in a case where the virtual speaker arrangement is set by the virtual speaker setting unit, the N channels of sound signals with which the virtual speaker arrangement is achieved, as transmission signals one to one to the speakers.
- a channel signal processing unit configured to perform signal processing to an input sound signal and generate N channels of sound signals to be supplied one to one to the N number of speakers, in which the channel signal processing unit generates, in a case where the virtual speaker arrangement is set by the virtual speaker setting unit, the N channels of sound signals with which the virtual speaker arrangement is achieved, as transmission signals one to one to the speakers.
- the channel signal processing unit performs processing to the respective channel sound signals to be transmitted to the actual speakers such that the position of audio output and the localized state of each virtual speaker are achieved in accordance with the virtual speaker setting.
- the virtual speaker setting unit displaces position of the virtual speaker arrangement in a direction of rotation, in accordance with an operation signal.
- the position of the virtual speaker arrangement is displaced in the direction of left-handed rotation or in the direction of right-handed rotation, in accordance with a rotational operation in the direction of left-handed/right-handed rotation of the user.
- the signal processing device further includes a to-be-used speaker setting unit configured to control switching between audio output with the N number of speakers and audio output with part of the N number of speakers, in accordance with a user operation.
- designation of one speaker by the user with all the speakers in use causes only the one speaker to perform audio output.
- a channel setting method to be performed by a signal processing device, includes: recognizing two arrangement reference speakers with reception of notification that a designation operation has been received from a user, from two speakers among N number of speakers, the N being three or more; acquiring distance information between each speaker; recognizing a relative-position relationship between the N number of speakers, with the two arrangement reference speakers and the distance information between each speaker; and automatically setting a channel to each speaker, on the basis of the relative-position relationship recognized.
- the signal processing device includes an information processing device, and the information processing device performs processing following the above steps.
- a program according to the present technology causes the information processing device to perform such processing. This arrangement causes achievement of the channel setting method according to the present technology, in the signal processing device including the information processing device.
- a speaker system includes the signal processing device and N number of speakers. This arrangement causes achievement of the speaker system easy and accurate for speaker arrangement and channel setting.
- the present technology enables easy and accurate channel setting to speakers. In particular, even a user who is unfamiliar with speaker systems can perform correct channel setting easily, so that proper audio output can be acquired.
- FIG. 1 is an explanatory view of an exemplary arrangement of a speaker system according to an embodiment of the present technology.
- FIG. 2 is an explanatory diagram of the equipment configuration of the speaker system according to the embodiment.
- FIG. 3 is an explanatory view of remote controllers to be used in the speaker system according to the embodiment.
- FIG. 4 is a block diagram of the respective internal configurations of a signal processing device and a speaker according to the embodiment.
- FIG. 5 is an explanatory diagram of the functional configuration of the signal processing device according to the embodiment.
- FIGS. 6A and 6B are explanatory views of channel setting steps according to the embodiment.
- FIGS. 7A and 7B are explanatory views of channel setting steps according to the embodiment.
- FIGS. 8A and 8B are explanatory views of channel setting steps according to the embodiment.
- FIGS. 9A and 9B are explanatory views of a channel setting step and virtual speaker setting according to the embodiment.
- FIG. 10 is a flowchart of channel setting processing according to the embodiment.
- FIG. 11 is a flowchart of channel setting processing according to the embodiment.
- FIGS. 12A, 12B and 12C are explanatory views of speaker arrangement direction changing according to the embodiment.
- FIGS. 13A, 13B, 13C, 13D, 13E, 13F, 13G and 13H are explanatory views of speaker arrangement direction changing according to the embodiment.
- FIG. 14 is a flowchart of arrangement change processing according to the embodiment.
- FIGS. 15A and 15B are explanatory views of to-be-used speaker selection according to the embodiment.
- FIG. 16 is a flowchart of to-be-used speaker selection processing according to the embodiment.
- FIG. 17 is an explanatory sequence diagram of an operation of to-be-used speaker selection according to the embodiment.
- FIG. 18 is a flowchart of to-be-used speaker selection processing according to the embodiment.
- FIG. 19 is an explanatory sequence diagram of an operation of to-be-used speaker selection according to the embodiment.
- channels are allowed to be set to the speakers easily.
- FIG. 1 An exemplary surround audio system with four speakers 3 ( 3 A, 3 B, 3 C, and 3 D) as in FIG. 1 will be described below.
- channels for the speakers 3 , 4 channels are assumed and are defined as a front L channel, a front R channel, a surround L channel, and a surround R channel.
- the front L channel, the front R channel, the surround L channel, and the surround R channel are referred to as “FL channel”, “FR channel”, “SL channel”, and “SR channel”, respectively.
- the 4 channels are exemplary for description.
- 5 channels, 5.1 channels, 7 channels, or 7.1 channels can be thought.
- the front left speaker having the front L channel, the front right speaker having the front R channel, the rear left speaker having the surround L channel, and the rear right speaker having the surround R channel will be given the term “FL speaker”, the term “FR speaker”, the term “SL speaker”, and the term “SR speaker”, respectively.
- the term “FL speaker 3 A” will be given, in some cases.
- FIG. 1 illustrates an exemplary arrangement of the surround audio system, for example, in a living room.
- the surround audio system according to the embodiment is provided as a speaker system including a signal processing device 1 and the speakers 3 A, 3 B, 3 C, and 3 D. Moreover, in some cases, the speaker system includes a remote controller 5 .
- the speaker system is used in audio reproduction of video content that is displayed on a monitor device 9 , for example, as a television receiver or the like.
- the speaker system is used in audio reproduction, such as music or environmental music, even in a case where the monitor device 9 does not perform video display.
- the monitor device 9 is arranged in position on the front side of a user, for example, in front of a sofa 8 . Then, in the example, the signal processing device 1 is arranged in proximity to the monitor device 9 .
- the FL speaker 3 A and the FR speaker 3 B are arranged on the left and the right of the monitor device 9 , respectively.
- the SL speaker 3 C and the SR speaker 3 D are arranged on the rear left and the rear right of the sofa 8 , respectively.
- the arrangement above is a typical exemplary arrangement of the monitor device 9 and the 4-channel speaker system. Needless to say, the actual arrangement varies depending on, for example, the taste of the user, the arrangement of furniture, the size of the room, or the shape of the room.
- the speakers 3 A, 3 B, 3 C, and 3 D are preferably arranged in suitable position as the FL channel, the FR channel, the SL channel, and the SR channel.
- FIG. 2 illustrates an exemplary configuration of the speaker system according to the embodiment.
- the speaker system enables communication between the signal processing device 1 as a master unit and the speakers 3 A, 3 B, 3 C, and 3 D as slave units.
- wireless communication in a communication scheme such as Wi-Fi (registered trademark) or Bluetooth (registered trademark) may be performed between the signal processing device 1 and each speaker 3 .
- wireless communication in a communication scheme, such as Wi-Fi (registered trademark) or Bluetooth (registered trademark) may be performed between the signal processing device 1 and each speaker 3 .
- LAN local area network
- USB universal serial bus
- connections may be made with dedicated cables including audio cables and control cables.
- a sound signal (digital sound signal or analog sound signal), control data, or notification data is transmitted between the signal processing device 1 and the speakers 3 by such wireless communication or wired communication.
- the speakers 3 A, 3 B, 3 C, and 3 D each are synchronized in time, for example, through the signal processing device 1 .
- the speakers 3 A, 3 B, 3 C, and 3 D may be capable of communicating with each other. Alternatively, it can be thought that the speakers 3 A, 3 B, 3 C, and 3 D do not particularly communicate with each other.
- the speakers 3 A, 3 B, 3 C, and 3 D are subjected to channel setting (channel assignment) by the signal processing device 1 .
- the speakers 3 A, 3 B, 3 C, and 3 D each have an individual speaker ID, for example, as an identifier.
- the speakers 3 A, 3 B, 3 C, and 3 D each are identical in configuration and are not necessarily a dedicated device to a certain channel.
- the speaker 3 A can be used as any of the FL speaker, the FR speaker, the SL speaker, and the SR speaker.
- the other speakers 3 B, 3 C, and 3 D can be used, similarly.
- the user is only required to arrange the speakers 3 A, 3 B, 3 C, and 3 D in position around the user, for example, as in FIG. 1 .
- Each speaker 3 is subjected to channel assignment by the signal processing device 1 as described later, so that a channel is determined to each speaker 3 with the signal processing device 1 as the base.
- the signal processing device 1 inputs a sound signal from a sound source device 2 thereto and performs necessary signal processing to the sound signal. Then, the signal processing device 1 transmits the respective sound signals allocated to the channels to the corresponding assigned speakers 3 . Each speaker 3 receives the corresponding channel sound signal from the signal processing device 1 and performs sound output. This arrangement causes performance of 4-channel surround audio output.
- the sound source device 2 illustrated in FIG. 2 is, for example, the monitor device 9 , a reproduction device (audio player) not illustrated, or the like.
- the sound source device 2 supplies a sound signal having an L-and-R stereo channel (digital sound signal or analog sound signal) or a multichannel-surround-enabled sound signal to the signal processing device 1 .
- the signal processing device 1 allocates or generates sound signals to the channels corresponding to the installed speakers 3 .
- sound signals are generated to the FL channel, the FR channel, the SL channel, and the SR channel, and then are transmitted to the corresponding speakers 3 A, 3 B, 3 C, and 3 D.
- Each speaker 3 includes a speaker unit 32 , and performs sound output with the speaker unit 32 driven by the transmitted sound signal.
- each speaker 3 includes a microphone 33 that can be used in channel setting to be described later.
- FIG. 3 illustrates remote controllers 5 A and 5 B as examples of the remote controller 5 .
- the remote controllers 5 A and 5 B each transmit operation information by the user to the signal processing device 1 .
- the remote controllers 5 A and 5 B include respective operators 50 ( 50 A and 50 B) for rotational operation.
- the operator 50 A is, for example, a rotary encoder capable of transmitting information regarding the amount of rotational operation.
- the operator 50 B is a button capable of issuing an instruction for a predetermined angle of rotation, for example, by a single press operation.
- each speaker 3 that is a slave unit is capable of identifying communication to itself from a slave address given to itself.
- each speaker 3 causes transmission information to include its identifier (speaker ID), so that the signal processing device 1 can identify from which speaker 3 communication has come from.
- the signal processing device 1 includes a central processing unit (CPU) 11 , an output signal formation unit 12 , a radio frequency (RF) module 13 , and a reception unit 14 .
- CPU central processing unit
- RF radio frequency
- the output signal formation unit 12 performs processing regarding a sound signal to be output to each speaker 3 .
- the output signal formation unit 12 in cooperation with the CPU 11 performs allocation of a sound signal to each channel or generation processing of a channel sound signal, generation processing of a sound signal to each speaker for virtual speaker output to be described later, such as signal processing including, for example, channel mixing, adjustment in localization, and delay.
- the output signal formation unit 12 performs, for example, amplification processing, tone processing, equalizing, or band-pass filtering processing to the sound signal of each channel.
- the output signal formation unit 12 performs processing of generating a sound signal as a test tone to be used in channel setting.
- An RF module 13 transmits a sound signal and a control signal to each speaker 3 or receives a signal from each speaker 3 .
- the RF module 13 performs encoding processing for wireless transmission and transmission processing to a sound signal and a control signal to be transmitted, supplied from the CPU 11 . Moreover, the RF module 13 performs, for example, reception processing to signals transmitted from the speakers 3 , decoding processing to reception data, and transfer to the CPU 11 .
- the reception unit 14 receives an operation signal from the remote controller 5 , demodulates/decodes the received operation signal, and transmits the operation information to the CPU 11 .
- the CPU 11 performs, for example, arithmetic processing to a sound signal supplied from the sound source device 2 , channel setting processing, or processing regarding virtual speakers.
- the CPU 11 is provided with functions illustrated in FIG. 5 by an implemented program (software), and performs arithmetic processing as the functions. That is, the CPU 11 has functions as a relative-position recognition unit 11 a , a channel setting unit 11 b , a virtual speaker setting unit 11 c , a channel signal processing unit 11 d , and a to-be-used speaker setting unit 11 e.
- the relative-position recognition unit 11 a and the channel setting unit 11 b perform processing for channel setting to each speaker 3 to be described later.
- the relative-position recognition unit 11 a performs processing of recognizing two arrangement reference speakers with reception of notification that a designation operation has been received from the user, from two speakers 3 among N number of speakers 3 installed (N is four in the present example). Moreover, the relative-position recognition unit 11 a performs processing of acquiring distance information between each speaker 3 . Furthermore, the relative-position recognition unit 11 a performs processing of recognizing the relative-position relationship between the N number of (four) speakers 3 , with the two arrangement reference speakers and the distance information between each speaker.
- the channel setting unit 11 b performs processing of automatically setting a channel to each speaker 3 , on the basis of the relative-position relationship recognized by the relative-position recognition unit.
- the virtual speaker setting unit 11 c performs processing of setting a virtual speaker arrangement, on the basis of the relative-position relationship recognized by the relative-position recognition unit 11 a and the channel setting performed by the channel setting unit 11 b .
- a virtual speaker is a speaker virtually arranged in position differently from the arrangement of an actual speaker 3 .
- Setting virtual speakers by the virtual speaker setting unit 11 c includes applying predetermined processing to a sound signal to each speaker 3 and performing audio output localized in position differently from the arrangement of the actual speakers 3 .
- the virtual speaker setting unit 11 c performs processing of changing the virtual speaker arrangement, in accordance with the operation.
- the channel signal processing unit 11 d performs processing of generating N channels of sound signals to be supplied one to one to the N number of speakers 3 , on the basis of an input sound signal, and transferring the N channels of sound signals to the RF module 13 .
- the channel signal processing unit 11 d in cooperation with the output signal formation unit 12 performs processing of generating, as respective transmission signals to the speakers 3 , the N channels of sound signals processed so as to be localized for achievement of virtual speakers.
- the to-be-used speaker setting unit 11 e performs processing of controlling switching between audio output with the N number of speakers 3 and audio output with part of the N number of speakers.
- the to-be-used speaker setting unit 11 e delivers information regarding a to-be-used speaker, to the channel signal processing unit 11 d . Then, the channel signal processing unit 11 d performs, for example, generation of a to-be-used channel sound signal and mute control to a not-to-be-used channel such that audio output is performed by only the to-be-used speaker.
- the information regarding a to-be-used speaker as a user operation is detected, for example, by reception of information from the corresponding speaker 3 through the RF module 13 .
- the to-be-used speaker setting unit 11 e may perform processing in accordance therewith.
- the speaker 3 includes a CPU 31 , a speaker unit 32 , a microphone 33 , a touch sensor 34 , an RF module 35 , an amplifier 36 , and a microphone input unit 37 .
- the CPU 31 performs communication processing or speaker internal control.
- the RF module 35 performs wireless communication with the RF module 13 of the signal processing device 1 .
- the RF module 35 receives a sound signal and a control signal transmitted from the signal processing device 1 and performs decoding processing to the sound signal and the control signal. Then, the RF module 35 transfers the decoded signals to the CPU 31 .
- the RF module 35 also performs processing of encoding, for wireless transmission, to a control signal and a notification signal transferred from the CPU 31 and transmitting the encoded signals to the signal processing device 1 .
- the CPU 31 supplies a sound signal transmitted from the signal processing device 1 to the amplifier 36 .
- the amplifier 36 converts the sound signal, for example, as digital data, transferred from the CPU 31 , into an analog signal and amplifies the analog signal.
- the amplifier 36 outputs the amplified signal to the speaker unit 32 . This arrangement causes audio output to be performed from the speaker unit 32 .
- the amplifier 36 is only required to output a digital sound signal.
- the microphone 33 collects external sound.
- the microphone input unit 37 amplifies a sound signal acquired by the microphone 33 , and converts the amplified signal into, for example, digital sound data. Then, the microphone input unit 37 supplies the digital sound data to the CPU 31 .
- the CPU 31 is capable of storing, as a microphone input sound signal, a sound signal together with time information (time stamp), for example, in an internal random access memory (RAM). Alternatively, in a case where a specific sound signal is detected as a test sound to be described later, the CPU 31 may store only time information without storing the sound signal.
- the CPU 31 transfers the stored information to the RF module 35 with predetermined timing and causes the RF module 35 to transmit the transferred information to the signal processing device 1 .
- the touch sensor 34 is a contact detection sensor, for example, as a touch pad or the like formed at a position that the user can touch easily, such as the upper face or front face of the casing of the speaker 3 .
- the touch sensor 34 detects a touch operation of the user, so that detection information is transmitted to the CPU 31 .
- the CPU 31 causes the RF module 35 to transmit detection information regarding the touch operation, to the signal processing device 1 .
- the touch sensor 34 is an exemplary device that detects an operation of the user to the speaker 3 .
- a device capable of detecting an operation or action of the user may be provided, such as an image pickup device (camera), an operation button, or a capacitive sensor.
- the microphone 33 detects a sound due to a touch operation (sound of contact) without providing the touch sensor 34 or the like.
- each speaker 3 is defined as being arranged on the same plane.
- the user only touches some speakers 3 , so that output channels can be set to all the speakers 3 , correctly.
- FIG. 6A illustrates the state where the signal processing device 1 and the four speakers 3 A, 3 B, 3 C, and 3 D are installed, for example, as described with FIG. 1 .
- each speaker 3 because channel setting to each speaker 3 is not determined in advance, the user installs the speakers 3 A, 3 B, 3 C, and 3 D at any positions without caring about channel setting. Rightfully, each speaker 3 has not been subjected to channel setting.
- supply of power to the signal processing device 1 that is the master unit and each speaker 3 causes wireless communication connection between the signal processing device 1 and each speaker 3 , for example, by WiFi or the like as illustrated, resulting in a start of initial setup.
- the user touches the speaker 3 A placed on the left of the monitor device 9 as indicated with solid line H 1 in FIG. 6B and subsequently touches the speaker 3 B placed on the right of the monitor device 9 as indicated with broken line H 2 .
- the speaker system may provide a guide sound, such as “Please touch the speaker on the left facing the front” or may display the message on the monitor device 9 .
- the user performs an operation of touching the touch sensor 34 of the speaker 3 A on the left facing ahead (arrow DRU).
- the direction in which the user faces the monitor device 9 is the front.
- the speaker system After it is detected that the user has performed, for example, an operation of touching the touch sensor 34 of the speaker 3 A, subsequently, the speaker system provides a guidance of the content “Please touch the speaker on the right facing the front”.
- the user subsequently performs an operation of touching the touch sensor 34 of the speaker 3 B.
- FIG. 7A illustrates the state where the speakers 3 A and 3 B are set as the FL speaker and the FR speaker.
- the speaker system can specify the FL speaker 3 A and the FR speaker 3 B and additionally can estimate the orientation of the user at the time of listening as the relative-position relationship to the set FL speaker 3 A and FR speaker 3 B.
- the speaker system automatically measures the distance between each speaker 3 .
- PTP precision time protocol
- a test sound reproduced by one speaker 3 is detected by the other speakers 3 , and arrival times of the sound are measured.
- a test sound reproduced by the speaker unit 32 of the FL speaker 3 A is picked up by the respective microphones 33 with which the FR speaker 3 B and the speakers 3 C and 3 D are equipped, so that each picked-up test sound is stored together with a time stamp (time information).
- the distance between the speakers 3 A and 3 B, the distance between the speakers 3 A and 3 C, and the distance between the speakers 3 A and 3 D indicated with broken lines can be measured.
- test sound for example, an electronic sound having a predetermined frequency needs at least outputting for a moment. Needless to say, a sound continuing, for example, for one second or for a few seconds, may be provided. In any case, a sound needs at least measurement of arrival time.
- Such an operation is performed with a speaker 3 for reproduction changed.
- the speaker 3 A reproduces a test sound
- the speakers 3 B, 3 C, and 3 D each store the test sound and time information.
- the speaker unit 32 of the speaker 3 B reproduces a test sound.
- the respective microphones 33 of the speakers 3 A, 3 C, and 3 D pick up the test sound, and then the speakers 3 A, 3 C, and 3 D each store the test sound and time information.
- This arrangement causes measurement of the distance between the speakers 3 B and 3 A, the distance between the speakers 3 B and 3 C, and the distance between the speakers 3 B and 3 D indicated with broken lines.
- the speaker 3 C reproduces a test sound
- the speakers 3 A, 3 B, and 3 D each store the test sound and time information. This arrangement causes measurement of the distance between the speakers 3 C and 3 A, the distance between the speakers 3 C and 3 B, and the distance between the speakers 3 C and 3 D.
- the speaker 3 D reproduces a test sound
- the speakers 3 A, 3 B, and 3 C each store the test sound and time information. This arrangement causes measurement of the distance between the speakers 3 D and 3 A, the distance between the speakers 3 D and 3 B, and the distance between the speakers 3 D and 3 C.
- reproduction/storage of test sounds as above enables two times of measurement of time difference (distance) in one combination.
- the average of two times of measurement is acquired to reduce measurement error.
- the processing of reproduction/storage of test sounds may finish at the point in time of completion of measurement in all the combinations.
- reproduction of a test sound from the speaker 3 D may be omitted.
- any speaker 3 that has already performed reproduction does not necessarily perform the processing of storage.
- the speaker 3 A enables measurement of the respective distances to the speakers 3 B, 3 C, and 3 D from the speaker 3 A after the finish of reproduction of the speaker 3 A.
- the speaker 3 A does not necessarily perform storage at the time of reproduction of a test sound from each of the speakers 3 B and 3 C.
- the speaker 3 B does not necessarily perform storage at the time of reproduction of a test sound from the speaker 3 C.
- the signal processing device 1 can grasp either the arrangement state in FIG. 8A or the arrangement state in FIG. 8B .
- the arrangement in FIG. 8A and the arrangement in FIG. 8B are in the relationship between mirror images identical in the distance between each speaker 3 .
- the signal processing device 1 can specify that the arrangement state in FIG. 8A is actual one.
- the signal processing device 1 automatically sets channels (SL and SR) to all the remaining speakers.
- the SR channel and the SL channel are automatically set to the speaker 3 C and the speaker 3 D, respectively.
- the signal processing device 1 has set the FL speaker 3 A, the FR speaker 3 B, the SR speaker 3 C, and the SL speaker 3 D. That is, the FL channel, the FR channel, the SL channel, and the SR channel have been assigned to the four speakers 3 arranged arbitrarily, in accordance with the respective arrangement positions.
- localization control with the mixing ratio of each channel sound signal or delay time setting corresponding to the difference in position between the set virtual speakers 4 and the real speakers 3 enables creation of an audio space in which sound is audible from the positions of the virtual speakers 4 A, 4 B, 4 C, and 4 D although the sound is actually output from the speakers 3 A, 3 B, 3 C, and 3 D.
- Such virtual speaker setting enables achievement of a further surround audio environment even in a case where a speaker arrangement that is not necessarily proper is made as a surround audio system (or even in a case where no proper arrangement can be made because of the condition of the room).
- virtual speaker setting may be performed sequentially.
- FIG. 10 illustrates the processing of the signal processing device 1 on the left and the processing of each speaker 3 on the right.
- the processing of the signal processing device 1 is performed mainly by the functions of the relative-position recognition unit 11 a and the channel setting unit 11 b in the CPU 11 .
- the processing of each speaker 3 is indicated as the processing of the corresponding CPU 31 .
- FIG. 10 illustrates the processing from the point in time initial setup starts after establishment of communication between the signal processing device 1 and each speaker 3 .
- step S 100 the CPU 11 of the signal processing device 1 issues an instruction for touch sensor on to all the speakers 3 that are slave units.
- the CPU 31 of each of the speakers 3 A, 3 B, 3 C, and 3 D turns on the touch sensor 34 in step S 201 to start the processing for a monitoring loop in steps S 202 and S 203 .
- the CPUs 31 each check the presence or absence of a user operation to the touch sensor 34 .
- the CPUs 31 each check the presence or absence of an instruction for touch sensor off from the signal processing device 1 .
- step S 102 the CPU 11 of the signal processing device 1 performs guidance control. That is, the CPU 11 performs control such that a guide output is performed to the user, such as “Please touch the speaker on the left facing the front”. For example, a sound signal of such a message sound may be transmitted to part or all of the speakers 3 such that sound output is performed. Alternatively, in a case where the speakers 3 store guidance sound source data, the CPU 11 may instruct each CPU 31 to output a guidance sound based on the sound source data. Furthermore, the CPU 11 may instruct the monitor device 9 to perform guide display.
- step S 103 the CPU 11 waits for a notification from a slave unit (speaker 3 ).
- the user touches the front left speaker 3 , so that the CPU 31 of the speaker 3 arranged on the front left detects an operation to the touch sensor in step S 202 .
- step S 204 the CPU 31 of the speaker 3 notifies the signal processing device 1 that is the master unit that a touch operation has been detected. Then, in step S 205 , the CPU 13 turns off the touch sensor.
- step S 103 After detecting, from a speaker 3 , that the touch sensor 34 has detected an operation, the CPU 11 of the signal processing device 1 proceeds from step S 103 to step S 104 and issues an instruction for touch sensor off to each speaker.
- This arrangement causes the CPU 31 of each speaker 3 to which a touch operation has not been performed to recognize the instruction for touch sensor off in step S 203 . Then, the CPU 31 of each speaker 3 proceeds to step S 205 to turn off the touch sensor.
- step S 105 the CPU 11 of the signal processing device 1 sets the speaker 3 having transmitted the effect that a touch operation has been detected, as the FL-channel speaker.
- the speaker 3 A has been set as the FL channel speaker.
- step S 106 the CPU 11 instructs each of the speakers 3 B, 3 C, and 3 D different from the speaker 3 A set, for example, to the FL channel, to turn on the touch sensor 34 .
- the CPU 11 of the FL speaker 3 A does not particularly perform corresponding processing because the control is not performed to itself, but the other speakers 3 B, 3 C, and 3 D each recognize the instruction for touch sensor on in step S 201 again and then start the processing for a monitoring loop in steps S 202 and S 203 .
- step S 107 the CPU 11 of the signal processing device 1 performs second guidance control. That is, the CPU 11 performs control such that a guide output is performed to the user, such as “Please touch the speaker on the right facing the front”. Then, in step S 108 , the CPU 11 waits for a notification from a slave unit (speaker 3 ).
- the user touches the front right speaker 3 , so that the CPU 31 of the speaker 3 arranged on the front right detects an operation to the touch sensor in step S 202 .
- step S 204 the CPU 31 of the speaker 3 notifies the signal processing device 1 that is the master unit that a touch operation has been detected. Then, in step S 205 , the CPU 31 turns off the touch sensor.
- step S 108 After detecting, from a speaker 3 , that the touch sensor 34 has detected an operation, the CPU 11 of the signal processing device 1 proceeds from step S 108 to step S 109 and issues an instruction for touch sensor off to each speaker.
- This arrangement causes the CPU 31 of each speaker 3 to which a touch operation has not been performed to recognize the instruction for touch sensor off in step S 203 . Then, the CPU 31 of each speaker 3 proceeds to step S 205 to turn off the touch sensor.
- step S 110 the CPU 11 of the signal processing device 1 sets the speaker 3 having transmitted the effect that a touch operation has been detected, as the FR-channel speaker.
- the speaker 3 B has been set as the FR channel speaker.
- FIG. 11 illustrates the processing of the CPU 11 on the left and the processing of the CPU 31 of a storage-side speaker 3 and the processing of the CPU 31 of a reproduction-side speaker 3 on the right.
- every single speaker 3 in sequence outputs a test sound as a production-side speaker and the other three speakers 3 each store the sound and time information as a storage-side speaker.
- the CPU 11 of the signal processing device 1 repeats the processing in steps S 150 , S 151 , and S 152 as loop processing LP 1 N number of times, the N corresponding to the number of speakers 3 .
- step S 150 the CPU 11 issues an instruction for storage start to a plurality of speakers 3 different from the i-th speaker 3 .
- step S 151 the CPU 11 performs control such that the i-th speaker 3 reproduces a test sound, for example, at a designated time.
- step S 152 the CPU 11 performs reception processing to information regarding a stored file from each of the plurality of speakers 3 different from the i-th speaker 3 .
- the CPU 11 repeats the above processing while incrementing the variable i sequentially.
- the CPU 31 of the i-th speaker 3 performs processing as a production-side speaker.
- step S 260 with reception of an instruction for test-sound reproduction based on step S 151 on the signal processing device 1 side, the CPU 31 reproduces a test sound at the designated time.
- the CPU 31 of each of the plurality of speakers 3 different from the i-th speaker 3 performs processing as a storage-side speaker. That is, in step S 250 , with reception of an instruction for storage start based on step S 150 on the signal processing device 1 side, each CPU 31 starts storage of a sound that the microphone 33 inputs and time information.
- each CPU 31 transmits the storage file to the signal processing device 1 that is the master unit.
- the reproduction time of a test sound that the reproduction-side speaker outputs is defined as a time length of 0.5 seconds.
- the storage-side speakers 3 each perform a predetermined period of audio sound storage, such as one second.
- Each frame included in the sound signal at this time includes a time stamp as the current time information. Then, for example, one second of sound signal storage file is generated, and then, in step S 251 , the generated storage file is transmitted to the signal processing device 1 .
- each storage file to be transmitted to the signal processing device 1 includes a first period in which no test sound is present (period including silence or ambient noise), a middle period in which a test sound is present, and a last period in which no test sound is present.
- the CPU 11 of the signal processing device 1 that has received such a storage file specifies the first frame storing the test sound, so that the arrival time of the test sound at the speaker 3 can be detected from the time stamp of the frame.
- the CPUs 31 each may analyze the storage start time of the test sound detected by the microphone 33 , and may transmit only the time information to the information processing device 1 .
- the CPU 11 of the signal processing device 1 performs the loop processing LP 1 four times, resulting in achievement of the operation described with FIGS. 7A and 7B .
- this arrangement enables the CPU 11 to detect the sound arrival time at each speaker 3 with reception of the storage files from the storage-side speakers 3 at each point in time.
- step S 153 the CPU 11 calculates the distance between each speaker.
- the CPU 11 can calculate the sound arrival time between the speakers 3 A and 3 B, the sound arrival time between the speakers 3 A and 3 C, and the sound arrival time between the speakers 3 A and 3 D from the test-sound output start time of the speaker 3 A and the storage files received from the speakers 3 B, 3 C, and 3 D. Therefore, the distance between the speakers 3 A and 3 B, the distance between the speakers 3 A and 3 C, and the distance between the speakers 3 A and 3 D can be calculated.
- Such calculation acquires the distance between each speaker 3 .
- step S 154 the CPU 11 performs coordinate calculation. Since the distance between each speaker 3 has been specified, the position of each speaker 3 is mapped on the coordinates such that the specified inter-speaker distances are expressed. Furthermore, since the FL speaker and the FR speaker have been specified, the speakers are defined to be arranged ahead.
- This arrangement causes expression of the speaker arrangement relationship as in FIG. 8A on the coordinates.
- step S 155 the CPU 11 performs channel assignment in accordance with the positions of all the speakers 3 .
- step S 156 virtual speaker setting may be performed.
- the user can arrange the speakers 3 without caring about output channel setting to the speakers 3 . Moreover, improper channel assignment can be prevented from being performed to the arrangement state of the speakers 3 .
- a surround speaker system assumes that a user always listens facing in the same direction once setting is performed.
- use of virtual speakers 4 enables the user with a simple action to change the direction of listening arbitrarily.
- FIG. 12A illustrates a model assumed as the interior of a room like a living dining kitchen.
- the actual speakers 3 A, 3 B, 3 C, and 3 D are arranged at the corners as in the figure.
- virtual speakers 4 are set suitably to the listening position of the user 100 of FIG. 12A , namely, suitably to a case where the user 100 faces the monitor device 9 like an arrow DRU.
- virtual speakers 4 virtual speakers 4 FL, 4 FR, 4 SL, and 4 SR will be given for indication of each channel.
- the arrangement of the virtual speakers 4 FL, 4 FR, 4 SL, and 4 SR is moved in the direction of rotation, for example, as illustrated in the figure.
- the user can change (rotate) the arrangement of the virtual speakers 4 arbitrarily.
- an operation with the remote controller 5 enables the user to rotate the arrangement of the virtual speakers 4 .
- the remote controllers 5 A and 5 B are provided with the respective operators 50 ( 50 A and 50 B) for rotational operation.
- the user performs a rotational operation with either operator 50 .
- the user operates the operator 50 by an amount of rotation suitable to the orientation of the user.
- the speaker system changes the arrangement positions of the virtual speakers 4 by a designated amount of rotation.
- FIGS. 13A, 13B, 13C, 13D, 13E, 13F, 13G, and 13H illustrate situations in which the arrangement of the virtual speakers 4 is rotated.
- FIG. 13A illustrates an arrangement state suitable to the user facing in the direction of the arrow DRU.
- the signal processing device 1 rotates the arrangement of the virtual speakers 4 , for example, from FIG. 13B to FIG. 13C and from FIG. 13C to FIG. 13D .
- FIG. 13D illustrates the arrangement positions in FIG. 13A rotated by 90 degrees clockwise.
- FIG. 13E illustrates the arrangement positions in FIG. 13D further rotated by 90 degrees clockwise. Moreover, in a case where the user performs a right-handed rotational operation from FIG. 13E , the signal processing device 1 rotates the arrangement of the virtual speakers 4 , for example, from FIG. 13F to FIG. 13G and from FIG. 13G to FIG. 13H .
- the signal processing device 1 rotates the arrangement of the virtual speakers 4 counterclockwise.
- an operation of the user to the remote controller 5 causes the positions of the virtual speakers 4 to rotate every previously determined angle changing step.
- the direction of listening can be switched suitably to the direction indicated with the arrow DRU.
- the angle changing step can be set freely in rate, and thus the direction of listening adaptable to the arrangement situation the speakers 3 is not limited.
- Use of the mechanism enables the direction of listening to be easily changed meeting use cases different in the direction of listening, such as the time when the user listens sitting on a sofa and the time when the user listens in the kitchen, in the surround speaker system arranged in the living dining kitchen, for example, as illustrated in FIGS. 12A and 12B .
- the virtual speakers 4 can be not only rotated but also shifted forward and backward.
- a forward-and-backward operation or crosswise operation of the user may be allowed such that the state of FIG. 12B can be changed to the arrangement state of FIG. 12C .
- FIG. 14 illustrates exemplary processing of the CPU 11 of the signal processing device 1 for rotation of the arrangement state as in FIGS. 13A, 13B, 13C, 13D, 13E, 13F, 13G, and 13H .
- the processing of FIG. 14 is performed mainly by the virtual speaker setting unit 11 c in the CPU 11 .
- step S 170 for example, the CPU 11 monitors a rotational operation of the user through the remote controller 5 .
- step S 170 the CPU 11 proceeds from step S 170 to step S 171 , and determines the amount of rotational operation and the direction of rotation in the operation per unit of time.
- step S 172 from the amount of rotational operation, the CPU 11 calculates the amount of movement of the virtual speakers 4 , namely, in this case, the angle of rotational movement of the virtual speakers 4 .
- the CPU 11 calculates by how many steps of angle rotation should be performed to the minimum step angle as the resolution of arrangement.
- step S 173 the CPU 11 determines positions to which the virtual speakers 4 are to be changed. For example, on the coordinates, the positions (coordinate values) rotationally moved on the basis of the angle and the direction corresponding to the rotational operation are determined as new positions to the virtual speakers 4 FL, 4 FR, 4 SL, and 4 SR.
- step S 174 the CPU 11 controls signal processing such that a sound field is formed on the basis of the new positions of the virtual speakers 4 FL, 4 FR, 4 SL, and 4 SR.
- coefficient changing for example, for the localized state due to the mixing rate between the respective channel sound signals to be output to the speakers 3 A, 3 B, 3 C, and 3 D or for delay time setting is performed such that an audio space is created on the basis of the new positions of the virtual speakers 4 FL, 4 FR, 4 SL, and 4 SR with audio outputs of the actual speakers 3 A, 3 B, 3 C, and 3 D.
- This processing causes rotational movement of the virtual speakers 4 .
- step S 175 the CPU 11 verifies whether or not the rotational operation of the user has been performed continuously. In a case where the rotational operation has been performed continuously (e.g., a case where the operator 50 A as a rotary encoder has been rotated continuously, or other cases), the CPU 11 goes back to step S 171 and then performs similar processing.
- the rotational operation has been performed continuously (e.g., a case where the operator 50 A as a rotary encoder has been rotated continuously, or other cases)
- the CPU 11 goes back to step S 171 and then performs similar processing.
- the CPU 11 finishes the processing of FIG. 14 from step S 175 .
- the CPU 11 starts the processing of FIG. 14 , again.
- the above processing causes the arrangement of the virtual speakers 4 to be rotated in accordance with an operation of the user.
- This arrangement enables the user to change the arrangement of the virtual speakers 4 with a high degree of freedom, meeting the direction of listening of the user.
- the user selects an arbitrary speaker 3 , so that switching can be performed between the mode in which sound is audible only from the selected speaker 3 and the mode in which sound is audible from all the speakers 3 .
- FIG. 15A illustrates the state where all the speakers 3 A, 3 B, 3 C, and 3 D are used
- FIG. 15B illustrates the state where only the speaker 3 C designated by the user is used.
- the state of FIG. 15A with an indoor wide area defined as a reproduction area AS 1 is suitable for listening in an ordinary surround audio environment.
- a reproduction area AS 1 there is a case where the user wants to listen to music or the like at a relatively small volume when the user is alone in the kitchen early in the morning and the like.
- designation of the speaker 3 C arranged near the kitchen enables sound output suitable for listening in a reproduction area AS 2 on the periphery of the kitchen as in FIG. 15B .
- to-be-used speaker selection can be performed easily with an intuitive operation.
- the speakers 3 each include the touch sensor 34 .
- use of the touch sensor 34 enables designation of a to-be-used speaker and additionally switching between the use of all the speakers and the use of a single speaker.
- a long-press touch of the user to the touch sensor 34 enables a necessary operation.
- FIGS. 16 and 17 Exemplary processing of the signal processing device 1 for the above, will be described with FIGS. 16 and 17 .
- FIG. 16 illustrates exemplary processing of the CPU 11 of the signal processing device 1 .
- the CPU 11 performs the processing with the function as the to-be-used speaker setting unit 11 e.
- step S 180 of FIG. 16 the CPU 11 instructs all the speakers 3 (slave units) to turn on the respective touch sensors 34 .
- Each speaker 3 turns on the touch sensor 34 , in accordance with the instruction.
- step S 181 the CPU 11 verifies whether or not a long-press notification has been received from any of the speakers 3 .
- the CPU 11 proceeds to step S 185 and verifies an instruction for finish.
- the instruction for finish is, for example, an instruction for finishing sound output in the speaker system.
- step S 181 If no instruction for finish has been detected, verification of a long-press notification in step S 181 is continued.
- the user In a case where the user wants to switch from the ordinary state where the four speakers 3 are used to the state where only one close speaker 3 is used, the user performs a long-press operation to the touch sensor 34 of a close speaker 3 . For example, the user keeps touching for a predetermined time or more, such as approximately one to two seconds.
- the CPU 31 of the speaker 3 transmits a long-press notification to the signal processing device 1
- step S 181 the CPU 11 proceeds from step S 181 to step S 182 and performs the processing, on the basis of whether reproduction-area limiting control is currently on or off.
- the reproduction-area limiting control means limiting reproduction area narrowly with use of only part of the speakers 3 as in FIG. 15B .
- the CPU 11 proceeds to step S 183 and turns on the reproduction-area limiting control.
- mute control is performed to each speaker 3 different from the speaker 3 having transmitted the long-press notification.
- control of changing channel signal processing is performed for switching to the state where only the speaker 3 having transmitted the long-press notification is used. For example, channel signal processing is changed such that a monaural sound signal is transmitted to the speaker 3 having transmitted the long-press notification.
- This arrangement causes the CPU 31 of each speaker 3 having received the mute control (speakers 3 different from the speaker 3 subjected to a long press) to perform the mute control to its audio output. Sound output is thus stopped. Meanwhile, the signal processing device 1 transmits a monaural sound signal to the speaker 3 having transmitted the long-press notification. Therefore, acquired is the state where the monaural sound signal is output only from the speaker 3 .
- the supply of the monaural sound signal is exemplary.
- a sound signal having two channels of L and R may be generated so as to be supplied to the speaker 3 .
- acquired is the state where sound output is performed only from the speaker 3 subjected to a long-press touch of the user.
- the user in a case where the user wants to return from the state where the reproduction-area limiting control is being performed to the original surround audio environment, the user is only required to perform a long-press operation again.
- step S 184 the CPU 31 turns off the reproduction-area limiting control.
- mute release control is performed to all the speakers 3 .
- control of changing channel signal processing is performed such that the present state is returned to the surround audio environment.
- This arrangement causes the CPU 31 of each speaker 3 having received the mute-release control to release the mute control to its audio output.
- This arrangement causes the present state to return the state where all the speakers 3 perform sound output. Then, the signal processing device 1 transmits respective assigned channel sound signals to the speakers 3 . Therefore, the present state is returned to the original surround audio environment.
- step S 186 the CPU 31 instructs each speaker 3 to turn off the touch sensor 34 , and then finishes the processing. In accordance with the instruction, each speaker 3 turns off the touch sensor 34 .
- the above processing is expressed on a time-series basis in FIG. 17 .
- FIG. 17 illustrates operations of the user and the operations of the signal processing device 1 (CPU 11 ) and the speakers 3 A, 3 B, 3 C, and 3 D.
- the user performs a long-press touch to the speaker 3 A.
- the CPU 31 of the speaker 3 A detects the long-press touch (S 300 ), and issues a long-press notification to the signal processing device 1 (S 301 ).
- the CPU 11 of the signal processing device 1 detects the long-press notification in the processing in step S 181 , and performs the processing in step S 183 . That is, the CPU 11 of the signal processing device 1 transmits an instruction for mute to the speakers 3 B, 3 C, and 3 D. In accordance with the instruction for mute, the CPU 31 of each of the speakers 3 B, 3 C, and 3 D mutes its sound output (S 350 ).
- This arrangement causes the reproduction-area limiting control to be turned on for use of the speaker 3 A.
- the user After that, for example, at a point in time, the user performs a long-press touch to the speaker 3 A again.
- This arrangement causes the CPU 31 of the speaker 3 A to detect the long-press touch (S 310 ), and then the CPU 31 of the speaker 3 A issues a long-press notification to the signal processing device 1 (S 311 ).
- the CPU 11 of the signal processing device 1 detects the long-press notification in the processing in step S 181 , and next performs the processing in step S 184 . That is, the CPU 11 of the signal processing device 1 transmits an instruction for mute release to all the speakers 3 A, 3 B, 3 C, and 3 D. In accordance with the instruction for mute release, the CPU 31 of each of the speakers 3 B, 3 C, and 3 D finishes the mute to its sound output (S 351 ). Because the speaker 3 A has not muted its sound output, the CPU 31 of the speaker 3 A does not particularly need to follow the instruction for mute release.
- the signal processing device 1 transmits respective channel sound signals to the speakers 3 .
- This arrangement causes resumption of reproduction of the surround audio system with the speakers 3 A, 3 B, 3 C, and 3 D, so that the reproduction-area limiting control is turned off.
- an operation at the time of release of the reproduction-area limiting control is not limited to a long-press touch to the speaker 3 in use, and may be a long-press touch to the other speakers 3 .
- the user wants to release the reproduction-area limiting control, the user is only required to perform a long-press touch to a close speaker 3 . In that case, the CPU 31 proceeds to the processing in step S 184 , so that all the speakers 3 are released from the mute.
- the user performs a long-press touch to the touch sensor 34 provided on, for example, the top face of a speaker 3 , so that toggle switching can be performed between the mode of reproduction only from the touched speaker 3 and the mode of reproduction from all the speakers.
- the user when the user wants reproduction only from the speaker 3 right in front of the user, the user performs an intuitive operation of simply touching the speaker 3 right in front of the user, resulting in switching to the state where sound is audible only from the speaker operated for selection.
- FIG. 18 illustrates other exemplary processing of the CPU 11 . Note that pieces of processing same as those in FIG. 16 are denoted with the same step numbers, and thus the descriptions thereof will be omitted. Steps S 180 to S 186 are the same as those in FIG. 16 .
- the CPU 11 monitors a long-press notification in step S 181 as well as a short-press notification in step S 190 .
- a short press is a touch operation for a short time, such as 100 ms or less, for example.
- step S 190 the CPU 11 proceeds from step S 190 to step S 191 and performs the processing, on the basis of whether the reproduction-area limiting control is currently on or off. If the reproduction-area limiting control is off, the CPU 11 goes back to step S 181 through step S 185 with doing nothing particularly.
- step S 191 the CPU 11 proceeds to step S 192 and verifies whether or not the speaker 3 having transmitted the short-press notification is a speaker in mute control.
- the CPU 11 proceeds to step S 193 and transmits an instruction for mute release to the speaker 3 .
- the CPU 31 of the speaker 3 having transmitted the short-press notification releases the mute, resulting in resumption of sound output.
- the CPU 11 of the signal processing device 1 changes channel signal processing such that sound output is performed with part of the speakers, the part being currently released from the mute.
- This arrangement causes addition of the speaker 3 subjected to a short press of the user in reproduction-area limiting control, to sound reproduction.
- the CPU 11 proceeds to step S 194 and transmits an instruction for mute to the speaker 3 .
- the CPU 31 of the speaker 3 having transmitted the short-press notification resumes mute to stop the sound output.
- the CPU 11 of the signal processing device 1 changes channel signal processing so as to be suitable for sound output without the speaker 3 having received the instruction for mute.
- This arrangement causes the speaker 3 performing reproduction in reproduction-area limiting control to be removed from the sound reproduction by a short press of the user.
- FIG. 19 illustrates operations of the user and the operations of the signal processing device 1 (CPU 11 ) and the speakers 3 A, 3 B, 3 C, and 3 D.
- the user performs a long-press touch to the speaker 3 A.
- the CPU 31 of the speaker 3 A detects the long-press touch (S 300 ), and issues a long-press notification to the signal processing device 1 (S 301 ).
- the CPU 11 of the signal processing device 1 detects the long-press notification (S 181 ) and transmits an instruction for mute to the speakers 3 B, 3 C, and 3 D as the processing in step S 183 .
- the CPU 31 of each of the speakers 3 B, 3 C, and 3 D mutes its sound output (S 350 ). That is, the reproduction-area limiting control is turned on for use of the speaker 3 A.
- the user After that, for example, at a point in time, the user performs a short-press touch to the speaker 3 B.
- the CPU 31 of the speaker 3 B When detecting the short-press touch (S 370 ), the CPU 31 of the speaker 3 B issues a short-press notification to the signal processing device 1 (S 371 ).
- the CPU 11 of the signal processing device 1 detects the short-press notification in step S 190 , and then performs the processing in step S 193 . That is, the CPU 11 of the signal processing device 1 issues an instruction for mute release to the speaker 3 B. In accordance with the instruction, the CPU 31 of the speaker 3 B releases the mute (S 372 ).
- the CPU 31 of the speaker 3 B When detecting the short-press touch (S 380 ), the CPU 31 of the speaker 3 B issues a short-press notification to the signal processing device 1 (S 381 ).
- the CPU 11 of the signal processing device 1 detects the short-press notification in step S 190 , and performs, in this case, the processing in step S 194 . That is, the CPU 11 of the signal processing device 1 issues an instruction for mute to the speaker 3 B. In accordance with the instruction, the CPU 31 of the speaker 3 B performs mute control (S 382 ).
- This arrangement causes the state where reproduction is performed only with the speaker 3 A, again.
- the user After that, for example, the user performs a long-press touch to the speaker 3 A.
- This arrangement causes the CPU 31 of the speaker 3 A to issue a long-press notification to the signal processing device 1 (S 310 ).
- the CPU 11 of the signal processing device 1 detects the long-press notification in the processing in step S 181 . In this time, in step S 184 , the CPU 11 of the signal processing device 1 transmits an instruction for mute release to all the speakers 3 A, 3 B, 3 C, and 3 D.
- the signal processing device 1 transmits respective channel sound signals to the speakers 3 .
- This arrangement causes resumption of reproduction of the surround audio system with the speakers 3 A, 3 B, 3 C, and 3 D, so that the reproduction-area limiting control is turned off.
- the reproduction-area limiting control of FIGS. 18 and 19 enables not only use of a single speaker 3 but also use of an arbitrary number of speakers that the user designates with a short-press touch.
- This arrangement enables the user to intuitively and more freely select the state where part of the speakers 3 is used.
- a speaker selection operation is preferably an operation of some kind performed to a speaker itself.
- the signal processing device 1 performs: processing of recognizing two arrangement reference speakers (FL speaker and FR speaker) with reception of notification that a designation operation has been received from the user, from two speakers among N number of speakers 3 , the N being three or more (S 102 to S 110 of FIG. 10 ); processing of acquiring distance information between each speaker 3 (S 150 to S 153 of FIG. 11 ); and processing of recognizing the relative-position relationship between the N number of speakers, with the two arrangement reference speakers and the distance information between each speaker (S 154 ).
- the signal processing device 1 performs channel setting such that a channel is automatically set to each speaker 3 , on the basis of the recognized relative-position relationship (S 155 ).
- the signal processing device 1 first recognizes the FL speaker and the FR speaker as the arrangement reference speakers, so that, for example, the front direction of the user (listener) can be determined.
- the actual speaker arrangement can be specified. Therefore, in accordance with the speaker arrangement, channel assignment can be automatically performed.
- each speaker 3 arranged arbitrarily by the user is automatically assigned to a proper channel only by the operation. This arrangement achieves proper channel setting with no trouble to the user. Furthermore, even if the user has no knowledge about channels, intuitive operations of simply touching two speakers 3 enables proper channel setting. This arrangement enables formation of an environment enabling optimum audio reproduction with no burden to the user.
- the present technology is applicable to a speaker system that connects three or more speakers 3 .
- the same operability enables simple and correct setting of output channels to all the speakers 3 .
- the signal processing device 1 includes the channel signal processing unit ( 11 d and 12 ) that performs signal processing to an input sound signal and generates N channels of sound signals to be supplied one to one to the N number of speakers 3 .
- the channel signal processing unit ( 11 d and 12 ) On the basis of the channels set by the channel setting unit 11 b , the channel signal processing unit ( 11 d and 12 ) generates N channels of sound signals as transmission signals one to one to the speakers 3 .
- This arrangement causes channel signal processing to be performed based on automatic channel assignment based on automatic recognition of the speaker positional relationship to be performed, resulting in achievement of proper surround audio output.
- the N number of speakers 3 each include an operation detection unit (touch sensor 34 ) that detects a designation operation from the user. Then, the relative-position recognition unit 11 a of the signal processing device 1 issues an instruction for activation of the operation detection unit to each speaker 3 (S 100 ) and additionally recognizes, as an arrangement reference speaker, a speaker 3 having issued a notification of the operation detection unit having had detection during a period of activation (S 102 to S 110 ). That is, for specification of an arrangement reference speaker, the CPU 11 (relative-position recognition unit 11 b ) performs control such that the touch sensor 34 of each speaker 3 is temporarily activated.
- the CPU 11 for specification of an arrangement reference speaker
- This arrangement causes the touch sensors 34 of the speakers 3 to function when the CPU 11 needs to specify the FL speaker and the FR speaker as arrangement reference speakers. Therefore, the CPU 11 can recognize, as designation operations for the FL speaker and the FR speaker, notifications of the touch sensors 34 each having received an operation, at the time of need, such as initial setup.
- the speakers 3 each do not necessarily supply any power to the operation detection unit and perform operation detection processing.
- temporal activation is useful to power saving and reduction of load in processing.
- the user may designate a speaker 3 with a gesture operation.
- each speaker 3 is provided with an operation detection unit including a sensing device of some kind, such as a button, a microphone, or an optical sensor.
- an operation can be thought in which successive switching is performed between the speakers 3 each output a test sound by, for example, operations to the remote controller 5 and a determination button is pressed when a desired speaker 3 outputs a test sound, to designate the speaker 3 .
- the signal processing device 1 recognizes, as the front left speaker (FL speaker) and the front right speaker (FR speaker), the two arrangement reference speakers that have received the notification that a designation operation has been received from the user.
- Determination of the front left speaker and the front right speaker enables determination of the orientation of the user at the time of listening.
- This arrangement enables determination of which arrangement state of two arrangement states in mirror symmetry is actual one, as speaker arrangements assumed, with the relative-position relationship between all the speakers grasped, so that a channel can be properly set to each speaker, on the basis of the front left and front right speakers 3 .
- the FL speaker and the FR speaker are not necessarily provided as arrangement reference speakers.
- the user may touch a surround speaker.
- the FL speaker and the FR speaker are present and the respective positions are easy for the user to find, it can be thought that the FL speaker and the FR speaker each are favorable to an instruction for touch operation to the user and hardly any wrong touch operation is performed.
- the FL speaker and the FR speaker are preferable as arrangement reference speakers.
- the two arrangement reference speakers are distinguished as the front left speaker and the front right speaker in order of designation operations from the user.
- the front left speaker and the front right speaker can be discriminated clearly. This arrangement enables accurate specification of the FL speaker and the FR speaker even in a case where each speaker 3 is identical in configuration and transmits an identical touch sensor detection signal.
- an instruction for activation of the operation detection unit is issued to each speaker 3 different from the speaker 3 having transmitted a notification of the designation operation, and a second designation operation is waited for (S 106 to S 108 ).
- This arrangement enables, in a case where the user touches the same speaker 3 two times, prevention of the speaker 3 from transmitting a needless notification to the signal processing device 1 .
- each speaker in order to acquire the distance information between each speaker 3 , each speaker is caused to output a test sound in sequence (LP 1 : S 150 to S 152 of FIG. 11 ).
- This arrangement enables measurement of the arrival time of a sound from one speaker to each speaker, so that the inter-speaker distances can be calculated.
- each speaker 3 includes the microphone 33 and is capable of transmitting detection time information regarding the test sound from each of the other speakers 3 .
- the signal processing device 1 calculates the distances between the one speaker and the other speakers (S 153 ). This arrangement enables accurate calculation of the inter-speaker distances.
- the signal processing device 1 includes the virtual speaker setting unit 11 c that sets the arrangement of virtual speakers 4 , on the basis of the relative-position relationship recognized by the relative-position recognition unit 11 a and the channel setting performed by the channel setting unit 11 b.
- the system that generates the virtual speakers 4 enables installation of the actual speakers 3 at various positions.
- the user may be unsure which channels should be selected, resulting in extreme difficulty in channel setting.
- channel setting in the present embodiment even in a case where the virtual speakers 4 are used, for example, it is sufficient that two speakers 3 placed on both sides of the monitor device 9 are selected. Thus, difficulty in channel setting due to speaker installed positions vanishes.
- the channel signal processing unit ( 11 d and 12 ) generates, as respective transmission signals to the speakers 3 , N channels of sound signals with which the virtual speaker arrangement is achieved.
- This arrangement causes sound output from the actual speakers to form an audio space in the virtual speaker arrangement. Therefore, even in a case where speaker arrangement in which optimum surround effect is not acquired is made due to, for example, the speaker arrangement positions corresponding to, for example, the indoor shape, the taste of the user, or the arrangement of furniture, a proper audio space can be provided to the user.
- the example in which the virtual speaker setting unit 11 c displaces the arrangement positions of the virtual speakers 4 in the direction of rotation in accordance with an operation signal, has been given (refer to FIG. 14 ).
- the arrangement positions of the virtual speakers are displaced in the direction of left-handed rotation or in the direction of right-handed rotation, in accordance with a rotational operation in the direction of left-handed/right-handed rotation of the user.
- This arrangement enables provision of an audio space in which the virtual speakers are arranged with the directivity that the user desires.
- an optimum surround audio space can be provided.
- an operation of the user is not limited to a rotational operation.
- a button operation or a directional operation is assumed.
- the signal processing device 1 includes the to-be-used speaker setting unit 11 e that controls, in accordance with a user operation, switching between audio output with the N number of speakers and audio output with part of the N number of speakers, has been given (refer to FIGS. 16 to 19 ).
- This arrangement enables provision of an audio space in a state that the user desires.
- optimum audio output can be provided with a simple operation.
- the reproduction area meeting user's usage scene can be controlled simply by direct selection of the speaker right in front of the user.
- a program causes, for example, a CPU, a digital signal processor (DSP), or the like to perform the functions as the relative-position recognition unit 11 a , the channel setting unit 11 b , the virtual speaker setting unit 11 c , the channel signal processing unit 11 d , and the to-be-used speaker setting unit 11 e , or causes an information processing device as a device including the units to perform the functions.
- DSP digital signal processor
- the program according to the embodiment causes the information processing device to perform: processing of recognizing two arrangement reference speakers with reception of notification that a designation operation has been received from the user, from two speakers among N number of speakers, the N being three or more; processing of acquiring distance information between each speaker; processing of recognizing the relative-position relationship between the N number of speakers, with the two arrangement reference speakers and the distance information between each speaker; and processing of automatically setting a channel to each speaker, on the basis of the recognized relative-position relationship.
- Such a program enables achievement of the signal processing device 1 according to the present disclosure.
- Such a program can be in advance recorded on a hard disk drive (HDD) as a recording medium built in equipment, such as a computer device, a ROM in a microcomputer including a CPU, or the like.
- HDD hard disk drive
- Such a program can be temporarily or permanently stored (recorded) in a removable recording medium, such as a flexible disk, a compact disc read only memory (CD-ROM), a magnet optical (MO) disc, a digital versatile disc (DVD), a Blu-ray Disc (registered trademark), a magnetic disk, a semiconductor memory, or a memory card.
- a removable recording medium such as a flexible disk, a compact disc read only memory (CD-ROM), a magnet optical (MO) disc, a digital versatile disc (DVD), a Blu-ray Disc (registered trademark), a magnetic disk, a semiconductor memory, or a memory card.
- a removable recording medium can be provided as so-called packaged software.
- Such a program can be downloaded from a download site through a network, such as a local area network (LAN) or the Internet, in addition to being installed from a removable recording medium to, for example, a personal computer.
- a network such as a local area network (LAN) or the Internet
- downloading such a program to various types of equipment including an arithmetic processing device, such as audio equipment, a personal computer, a mobile information processing device, a mobile phone, game equipment, video equipment, and a personal digital assistant (PDA), enables the various types of equipment to be provided as the signal processing device 1 according to the present disclosure.
- an arithmetic processing device such as audio equipment, a personal computer, a mobile information processing device, a mobile phone, game equipment, video equipment, and a personal digital assistant (PDA)
- PDA personal digital assistant
- a signal processing device including:
- a relative-position recognition unit configured to perform processing of recognizing two arrangement reference speakers with reception of notification that a designation operation has been received from a user, from two speakers among N number of speakers, the N being three or more, and processing of acquiring distance information between each speaker, the relative-position recognition unit being configured to recognize a relative-position relationship between the N number of speakers, with the two arrangement reference speakers and the distance information between each speaker;
- a channel setting unit configured to automatically set a channel to each speaker, on the basis of the relative-position relationship recognized by the relative-position recognition unit.
- the signal processing device further including:
- a channel signal processing unit configured to perform signal processing to an input sound signal and generate N channels of sound signals to be supplied one to one to the N number of speakers, in which
- the channel signal processing unit generates the N channels of sound signals as transmission signals one to one to the speakers, on the basis of the channels set by the channel setting unit.
- the N number of speakers each include an operation detection unit that detects a designation operation from the user, and
- the relative-position recognition unit issues an instruction for activation of the operation detection unit to each speaker, and additionally recognizes, as an arrangement reference speaker, a speaker having issued a notification of the operation detection unit having had detection during a period of activation.
- the relative-position recognition unit recognizes, as a front left speaker and a front right speaker, the two arrangement reference speakers that have received the notification that the designation operation has been received from the user.
- the relative-position recognition unit distinguishes the two arrangement reference speakers as the front left speaker and the front right speaker in order of the designation operations from the user.
- the relative-position recognition unit issues, in a case where a first designation operation is performed by the user, the instruction for activation of the operation detection unit to each speaker different from a speaker having transmitted a notification of the first designation operation, and waits for a second designation operation.
- the relative-position recognition unit causes, for acquisition of the distance information between each speaker, each speaker sequentially to output a test sound.
- each speaker includes a sound detection unit and is capable of transmitting detection time information regarding the test sound from another speaker, and
- the relative-position recognition unit calculates, from output start time information regarding the test sound from a speaker and detection time information from another speaker, a distance between the speaker and the another speaker.
- the signal processing device according to any of (1) to (8) above, further including:
- a virtual speaker setting unit configure to set a virtual speaker arrangement, on the basis of the relative-position relationship recognized by the relative-position recognition unit and the channel setting performed by the channel setting unit.
- the signal processing device according to (9) above, further including,
- a channel signal processing unit configured to perform signal processing to an input sound signal and generate N channels of sound signals to be supplied one to one to the N number of speakers, in which
- the channel signal processing unit generates, in a case where the virtual speaker arrangement is set by the virtual speaker setting unit, the N channels of sound signals with which the virtual speaker arrangement is achieved, as transmission signals one to one to the speakers.
- the virtual speaker setting unit displaces position of the virtual speaker arrangement in a direction of rotation, in accordance with an operation signal.
- the signal processing device according to any of (1) to (11), further including:
- a to-be-used speaker setting unit configured to control switching between audio output with the N number of speakers and audio output with part of the N number of speakers, in accordance with a user operation.
- a channel setting method to be performed by a signal processing device including:
- a speaker system including:
- N number of speakers the N being three or more;
- the signal processing device includes:
- a relative-position recognition unit configured to perform processing of recognizing two arrangement reference speakers with reception of notification that a designation operation has been received from a user, from two speakers among the N number of speakers, and processing of acquiring distance information between each speaker, the relative-position recognition unit being configured to recognize a relative-position relationship between the N number of speakers, with the two arrangement reference speakers and the distance information between each speaker;
- a channel setting unit configured to automatically set a channel to each speaker, on the basis of the relative-position relationship recognized by the relative-position recognition unit.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- Stereophonic System (AREA)
Abstract
Description
- 1 Signal processing device
- 3 Speaker
- 4 Virtual speaker
- 5 Remote controller
- 11, 31 CPU
- 11 a Relative-position recognition unit
- 11 b Channel setting unit
- 11 c Virtual speaker setting unit
- 11 d Channel signal processing unit
- 11 e To-be-used speaker setting unit
- 12 Output signal formation unit
- 13, 35 RF module
- 14 Reception unit
- 32 Speaker unit
- 33 Microphone
- 34 Touch sensor
Claims (12)
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2018-082949 | 2018-04-24 | ||
JP2018082949 | 2018-04-24 | ||
JPJP2018-082949 | 2018-04-24 | ||
PCT/JP2019/010683 WO2019208012A1 (en) | 2018-04-24 | 2019-03-14 | Signal processing device, channel setting method, program and speaker system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20210037331A1 US20210037331A1 (en) | 2021-02-04 |
US11356789B2 true US11356789B2 (en) | 2022-06-07 |
Family
ID=68294480
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/980,973 Active US11356789B2 (en) | 2018-04-24 | 2019-03-14 | Signal processing device, channel setting method, and speaker system |
Country Status (5)
Country | Link |
---|---|
US (1) | US11356789B2 (en) |
EP (1) | EP3787318A4 (en) |
JP (1) | JP7298599B2 (en) |
CN (1) | CN111886881B (en) |
WO (1) | WO2019208012A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US12058509B1 (en) * | 2021-12-09 | 2024-08-06 | Amazon Technologies, Inc. | Multi-device localization |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3787318A4 (en) | 2018-04-24 | 2021-06-30 | Sony Corporation | Signal processing device, channel setting method, program and speaker system |
FR3111040B1 (en) * | 2020-06-02 | 2022-07-01 | Sagemcom Broadband Sas | Centralized processing of an input audio stream |
CN114205716B (en) * | 2020-09-18 | 2023-03-24 | 华为技术有限公司 | Method, system, electronic device and storage medium for determining sound channel role of sound box |
Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05122800A (en) | 1991-10-28 | 1993-05-18 | Sony Corp | Acoustic device |
JP2002345100A (en) | 2001-05-22 | 2002-11-29 | Funai Electric Co Ltd | Av amplifier and channel setting method of output terminal of the av amplifier |
EP1443804A2 (en) | 2003-02-03 | 2004-08-04 | Denon, Ltd. | A multichannel reproducing apparatus |
WO2006131894A2 (en) | 2005-06-09 | 2006-12-14 | Koninklijke Philips Electronics N.V. | A method of and system for automatically identifying the functional positions of the loudspeakers of an audio-visual system |
WO2007028094A1 (en) | 2005-09-02 | 2007-03-08 | Harman International Industries, Incorporated | Self-calibrating loudspeaker |
WO2012164444A1 (en) | 2011-06-01 | 2012-12-06 | Koninklijke Philips Electronics N.V. | An audio system and method of operating therefor |
US20140169569A1 (en) * | 2012-12-17 | 2014-06-19 | Nokia Corporation | Device Discovery And Constellation Selection |
US20140362725A1 (en) * | 2013-06-06 | 2014-12-11 | The Boeing Company | Device identification via peer exchange |
US20150381296A1 (en) | 2014-06-25 | 2015-12-31 | Cheng Yih Jenq | Wireless transmission of sound signals to multiple discrete speakers |
US20160080886A1 (en) | 2013-05-16 | 2016-03-17 | Koninklijke Philips N.V. | An audio processing apparatus and method therefor |
US20160080884A1 (en) * | 2013-04-27 | 2016-03-17 | Intellectual Discovery Co., Ltd. | Audio signal processing method |
DE102015106114A1 (en) | 2015-04-21 | 2016-10-27 | D & B Audiotechnik Gmbh | METHOD AND DEVICE FOR POSITION DETECTION OF SPEAKER BOXES OF A SPEAKER BOX ARRANGEMENT |
JP2017184174A (en) | 2016-03-31 | 2017-10-05 | 株式会社バンダイナムコエンターテインメント | Simulation system and program |
CN107396250A (en) | 2017-09-15 | 2017-11-24 | 深圳市禾音视频科技有限公司 | The integrated-type intelligent cinema system realized based on Doby |
US9854362B1 (en) | 2016-10-20 | 2017-12-26 | Sony Corporation | Networked speaker system with LED-based wireless communication and object detection |
WO2019208012A1 (en) | 2018-04-24 | 2019-10-31 | ソニー株式会社 | Signal processing device, channel setting method, program and speaker system |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3909749B2 (en) * | 2001-12-25 | 2007-04-25 | アルパイン株式会社 | Audio field adjustment device for audio |
US9749769B2 (en) | 2014-07-30 | 2017-08-29 | Sony Corporation | Method, device and system |
-
2019
- 2019-03-14 EP EP19793355.9A patent/EP3787318A4/en active Pending
- 2019-03-14 WO PCT/JP2019/010683 patent/WO2019208012A1/en unknown
- 2019-03-14 US US16/980,973 patent/US11356789B2/en active Active
- 2019-03-14 JP JP2020516096A patent/JP7298599B2/en active Active
- 2019-03-14 CN CN201980020047.5A patent/CN111886881B/en active Active
Patent Citations (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05122800A (en) | 1991-10-28 | 1993-05-18 | Sony Corp | Acoustic device |
JP2002345100A (en) | 2001-05-22 | 2002-11-29 | Funai Electric Co Ltd | Av amplifier and channel setting method of output terminal of the av amplifier |
EP1443804A2 (en) | 2003-02-03 | 2004-08-04 | Denon, Ltd. | A multichannel reproducing apparatus |
US20040151476A1 (en) | 2003-02-03 | 2004-08-05 | Denon, Ltd. | Multichannel reproducing apparatus |
JP2004241820A (en) | 2003-02-03 | 2004-08-26 | Denon Ltd | Multichannel reproducing apparatus |
WO2006131894A2 (en) | 2005-06-09 | 2006-12-14 | Koninklijke Philips Electronics N.V. | A method of and system for automatically identifying the functional positions of the loudspeakers of an audio-visual system |
WO2007028094A1 (en) | 2005-09-02 | 2007-03-08 | Harman International Industries, Incorporated | Self-calibrating loudspeaker |
US20140161265A1 (en) | 2005-09-02 | 2014-06-12 | Harman International Industries, Incorporated | Self-calibration loudspeaker system |
WO2012164444A1 (en) | 2011-06-01 | 2012-12-06 | Koninklijke Philips Electronics N.V. | An audio system and method of operating therefor |
US20140169569A1 (en) * | 2012-12-17 | 2014-06-19 | Nokia Corporation | Device Discovery And Constellation Selection |
US20160080884A1 (en) * | 2013-04-27 | 2016-03-17 | Intellectual Discovery Co., Ltd. | Audio signal processing method |
US20160080886A1 (en) | 2013-05-16 | 2016-03-17 | Koninklijke Philips N.V. | An audio processing apparatus and method therefor |
US20200186956A1 (en) | 2013-05-16 | 2020-06-11 | Koninklijke Philips N.V. | Audio Processing Apparatus and Method Therefor |
US20210136512A1 (en) | 2013-05-16 | 2021-05-06 | Koninklijke Philips N.V. | Audio Processing Apparatus and Method Therefor |
US20210144507A1 (en) | 2013-05-16 | 2021-05-13 | Koninklijke Philips N.V. | Audio Processing Apparatus and Method Therefor |
US20140362725A1 (en) * | 2013-06-06 | 2014-12-11 | The Boeing Company | Device identification via peer exchange |
US20150381296A1 (en) | 2014-06-25 | 2015-12-31 | Cheng Yih Jenq | Wireless transmission of sound signals to multiple discrete speakers |
DE102015106114A1 (en) | 2015-04-21 | 2016-10-27 | D & B Audiotechnik Gmbh | METHOD AND DEVICE FOR POSITION DETECTION OF SPEAKER BOXES OF A SPEAKER BOX ARRANGEMENT |
JP2017184174A (en) | 2016-03-31 | 2017-10-05 | 株式会社バンダイナムコエンターテインメント | Simulation system and program |
CN107277736A (en) | 2016-03-31 | 2017-10-20 | 株式会社万代南梦宫娱乐 | Simulation System, Sound Processing Method And Information Storage Medium |
US9854362B1 (en) | 2016-10-20 | 2017-12-26 | Sony Corporation | Networked speaker system with LED-based wireless communication and object detection |
CN107396250A (en) | 2017-09-15 | 2017-11-24 | 深圳市禾音视频科技有限公司 | The integrated-type intelligent cinema system realized based on Doby |
WO2019208012A1 (en) | 2018-04-24 | 2019-10-31 | ソニー株式会社 | Signal processing device, channel setting method, program and speaker system |
US20210037331A1 (en) | 2018-04-24 | 2021-02-04 | Sony Corporation | Signal processing device, channel setting method, program, and speaker system |
Non-Patent Citations (3)
Title |
---|
Extended European Search Report of EP Application No. 19793355.9, dated May 28, 2021, 08 pages. |
International Search Report & Written Opinion of PCT Application No. PCT/JP2019/010683, dated Jun. 4, 2019, 09 pages of ISRWO. |
Office Action for CN Patent Application No. 201980020047.5, dated Nov. 12, 2021, 09 pages of English Translation and 14 pages of Office Action. |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US12058509B1 (en) * | 2021-12-09 | 2024-08-06 | Amazon Technologies, Inc. | Multi-device localization |
Also Published As
Publication number | Publication date |
---|---|
EP3787318A1 (en) | 2021-03-03 |
WO2019208012A1 (en) | 2019-10-31 |
EP3787318A4 (en) | 2021-06-30 |
JP7298599B2 (en) | 2023-06-27 |
US20210037331A1 (en) | 2021-02-04 |
CN111886881A (en) | 2020-11-03 |
JPWO2019208012A1 (en) | 2021-04-30 |
CN111886881B (en) | 2022-11-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11356789B2 (en) | Signal processing device, channel setting method, and speaker system | |
CN112205004B (en) | Method, device and system for interacting with earphone of media playback system | |
KR102187147B1 (en) | Audio response playback | |
JP6668661B2 (en) | Parameter control device and parameter control program | |
US10798488B2 (en) | Device for reproducing sound | |
WO2015199927A1 (en) | Audio systems and related methods and devices | |
JP2005524253A (en) | Method for enabling interaction using a portable device | |
US11758326B2 (en) | Wearable audio device within a distributed audio playback system | |
JP6481341B2 (en) | Content playback device | |
JP6597248B2 (en) | System and control method | |
US11463836B2 (en) | Information processing apparatus and information processing method | |
US20160180880A1 (en) | Multitrack recording system with wireless lan function | |
JP6587281B2 (en) | Acoustic system, acoustic device, and terminal device | |
JP6747563B2 (en) | Recording/playback device and co-listening system with wireless LAN function | |
US12032872B2 (en) | Intelligent user interfaces for playback devices | |
JP6690751B2 (en) | Multi-track recording system with wireless LAN function | |
US20230099144A1 (en) | Enabling and Disabling Microphones and Voice Assistants | |
JP2016021640A (en) | Audio device and audio system | |
WO2014193482A2 (en) | System and constituent media device components and media device-based ecosystem |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TAMAKI, TATSUYA;OHURA, YOSHIKAZU;YORIMOTO, KENJI;AND OTHERS;SIGNING DATES FROM 20200820 TO 20200830;REEL/FRAME:053773/0241 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |