WO2016140058A1 - 音声信号再生装置、音声信号再生方法、プログラム、および記録媒体 - Google Patents
音声信号再生装置、音声信号再生方法、プログラム、および記録媒体 Download PDFInfo
- Publication number
- WO2016140058A1 WO2016140058A1 PCT/JP2016/054480 JP2016054480W WO2016140058A1 WO 2016140058 A1 WO2016140058 A1 WO 2016140058A1 JP 2016054480 W JP2016054480 W JP 2016054480W WO 2016140058 A1 WO2016140058 A1 WO 2016140058A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- sound source
- virtual sound
- user
- audio signal
- detected
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
- H04S7/302—Electronic adaptation of stereophonic sound system to listener position or orientation
- H04S7/303—Tracking of listener position or orientation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1626—Constructional details or arrangements for portable computers with a single-body enclosure integrating a flat display, e.g. Personal Digital Assistants [PDAs]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1686—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated camera
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1688—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being integrated loudspeakers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/012—Head tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/0304—Detection arrangements using opto-electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S3/00—Systems employing more than two channels, e.g. quadraphonic
- H04S3/008—Systems employing more than two channels, e.g. quadraphonic in which the audio signals are in digital form, i.e. employing more than two discrete digital channels
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2460/00—Details of hearing devices, i.e. of ear- or headphones covered by H04R1/10 or H04R5/033 but not provided for in any of their subgroups, or of hearing aids covered by H04R25/00 but not provided for in any of its subgroups
- H04R2460/07—Use of position data from wide-area or local-area positioning systems in hearing devices, e.g. program or information selection
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R2499/00—Aspects covered by H04R or H04S not otherwise provided for in their subgroups
- H04R2499/10—General applications
- H04R2499/11—Transducers incorporated or for use in hand-held devices, e.g. mobile phones, PDA's, camera's
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/01—Multi-channel, i.e. more than two input channels, sound reproduction with two speakers wherein the multi-channel information is substantially preserved
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/11—Positioning of individual sound objects, e.g. moving airplane, within a sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/01—Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/03—Application of parametric coding in stereophonic audio systems
Definitions
- the present invention relates to an audio signal reproduction device and an audio signal reproduction method for reproducing an audio signal. Furthermore, the present invention relates to a program for operating a computer as the audio signal reproducing apparatus and a recording medium on which such a program is recorded.
- Information transmitted by the information terminal itself or an application executed in the information terminal includes visual information such as image information or text information.
- the image information or the character information is provided to the user by being displayed as an image or a character on a display provided in the main body of the information terminal or an external display terminal connected to the information terminal.
- Patent Document 1 discloses a spectacle-type image information display terminal as an example of such an information terminal.
- Patent Document 1 displays notification information to be transmitted as an image
- the user has performed until just before the user to watch the image displayed on the image display unit such as a display.
- the image display unit such as a display.
- a certain amount of time is required for the user, such as the necessity to interrupt the work.
- information transmitted from the information terminal or application includes an auditory situation such as voice information.
- the sound information is provided to the user by being output as sound in a stereo format or a stereophonic sound format through a speaker provided in the information terminal or headphones or earphones connected to the information terminal. Therefore, in order to avoid interruption of the user's work, it is conceivable to notify the information by outputting sound instead of notifying the user by displaying images or characters.
- Patent Document 2 discloses a stereophonic sound control device that gives directions to a listener by voice.
- the device operated in the cited document 2 does not take into account the user's behavior when presenting information to the user by means of an audio signal. Therefore, the problem which inhibits a user's action may arise because the voice which does not match a user's action is shown to a user.
- An object of the present invention is to provide an audio signal reproducing apparatus, an audio signal reproducing method, a program, and a recording medium for reproducing an audio signal so that audio suitable for the user's operation is output.
- an audio signal reproduction device includes an acquisition unit that acquires an audio signal indicating audio, a detection unit that detects a user's operation, and the detected operation. Based on the acquired sound, the determination unit that determines the position of the virtual sound source from which the sound represented by the acquired sound signal is output, and the stereophonic sound processing according to the determined position of the virtual sound source. A processing unit applied to the signal and a playback unit that plays back the audio signal subjected to the stereophonic sound processing are provided.
- an audio signal reproduction method includes an acquisition step of acquiring an audio signal indicating audio, a detection step of detecting a user action, and the detected user's operation. Based on the operation, the determination step of determining the position of the virtual sound source from which the sound represented by the acquired sound signal is output, and the stereophonic sound processing corresponding to the determined position of the virtual sound source are acquired. It has a processing step applied to the audio signal and a reproduction step of reproducing the audio signal subjected to the stereophonic sound processing.
- FIG. 1 It is a block diagram which shows the principal part structure of the audio
- (A) is a figure which shows an example of the position of the virtual sound source when the user is stationary in Embodiment 1 of the present invention
- (b) is a diagram where the user is stationary in Embodiment 1 of the present invention. It is a figure which shows an example of the position of the virtual sound source at the time of being.
- (A) is a figure which shows an example of the position of the virtual sound source when the user is stationary in the first embodiment of the present invention
- (b) is a diagram showing the movement of the user in the first embodiment of the present invention. It is a figure which shows an example of the position of the virtual sound source at the time of being.
- (A) is a figure which shows an example of the position of the virtual sound source when the user is stationary in the first embodiment of the present invention
- (b) is a diagram showing the movement of the user in the first embodiment of the present invention. It is a figure which shows an example of the position of the virtual sound source at the time of being.
- (A) is a figure which shows an example of the position of the some virtual sound source when the user is stationary in Embodiment 1 of this invention
- (b) is a user in Embodiment 1 of this invention. It is a figure which shows an example of the position of the some virtual sound source at the time of moving.
- (A) is a figure which shows an example of the position of the some virtual sound source when the user is stationary in Embodiment 1 of this invention
- (b) is a user in Embodiment 1 of this invention. It is a figure which shows an example of the position of the some virtual sound source at the time of moving.
- (A) is a figure which shows an example of the position of several virtual sound sources when the user is not gazing at a fixed gaze area in Embodiment 1 of the present invention
- (b) is an embodiment of the present invention.
- 2 is a diagram illustrating an example of positions of a plurality of virtual sound sources when the user is gazing at the gaze area in FIG. 1, and (c) is a gaze at the gaze area in Embodiment 1 of the present invention.
- (c) is further the position of the several virtual sound sources when the user is gazing at the gaze area in Embodiment 1 of this invention.
- It is a figure which shows another example.
- FIG. 2 It is a block diagram which shows the principal part structure of the audio
- (A)-(c) is a figure which shows each example of the metadata in Embodiment 2 of this invention.
- (A) is a figure which shows an example of the position of the some virtual sound source when the user is still in Embodiment 2 of this invention,
- (b) is a user in Embodiment 2 of this invention. It is a figure which shows an example of the position of the some virtual sound source at the time of moving.
- Embodiment 1 of the present invention will be described below with reference to FIGS.
- the audio signal reproduction device 1 acquires at least one audio signal and detects the operation of the user of the audio signal reproduction device 1. Then, based on the detected user action, the acquired audio signal is virtually arranged at a position based on a preset arbitrary rule. Thereafter, each audio signal is converted into a stereophonic audio signal based on the arranged virtual audio signal position, and the converted audio signal is reproduced.
- FIG. 1 is a block diagram showing a main configuration of an audio signal reproduction device 1 according to Embodiment 1 of the present invention.
- the audio signal reproduction device 1 includes an audio signal acquisition unit 11 (acquisition unit), a control unit 12, an audio signal reproduction unit 13 (reproduction unit), and a storage unit 14.
- the control unit 12 includes a user action detection unit 121 (detection unit), a virtual sound source position determination unit 122 (determination unit), and an audio signal processing unit 123 (processing unit).
- the audio signal acquisition unit 11 acquires at least one audio signal from the outside of the audio signal reproduction device 1.
- the audio signal may be either a stereo format or a monaural format.
- the audio signal acquisition unit 11 can also acquire an audio signal in a format in which a plurality of audio signals are interleaved. In this case, the audio signal acquisition unit 11 deinterleaves the acquired audio signal into a plurality of monaural audio signals.
- the control unit 12 controls the audio signal acquisition unit 11, the audio signal reproduction unit 13, and the storage unit 14, and inputs / outputs data to / from these members.
- the control unit 12 is realized, for example, by a CPU (Central Processing Unit) executing a program stored in a predetermined memory.
- CPU Central Processing Unit
- the audio signal reproduction unit 13 outputs audio through the earphone 24 by reproducing each audio signal that has been subjected to the stereophonic sound processing (acoustic effect processing) by the control unit 12.
- the storage unit 14 is configured by a secondary storage device for storing predetermined data used by the control unit 12.
- the storage unit 14 is realized as, for example, a magnetic disk, an optical disk, or a flash memory, specifically, an HDD (Hard Disk Drive), an SSD (Solid State Drive), a BD (Blu-Ray Disc, registered trademark), or the like.
- the control unit 12 can read data from the storage unit 14 or record data in the storage unit 14 as necessary.
- the user operation detection unit 121 detects a user's action or operation based on the output of various connected sensors.
- the virtual sound source position determination unit 122 determines the position of the virtual sound source from which each audio signal is virtually output based on the detected user behavior or action.
- the audio signal processing unit 123 performs stereophonic sound processing on each audio signal acquired by the audio signal acquisition unit 11 based on the determined position of each virtual output.
- FIG. 2 is a block diagram showing a main configuration of the audio signal reproduction system 2 according to Embodiment 1 of the present invention. As shown in this figure, in addition to the audio signal reproduction device 1, the audio signal reproduction system 2 further includes a signal reception device 21, a digital-analog conversion device (DAC) 22, an amplification device 23, and an earphone 24.
- DAC digital-analog conversion device
- the signal receiving device 21 receives an audio signal from the outside of the audio signal reproducing device 1 by wired communication or wireless communication.
- a wireless transmission technology such as Bluetooth (registered trademark) or WiFi (Wireless Fidelity, registered trademark) can be used, but is not limited thereto.
- the audio signal acquisition unit 11 receives an audio signal from the portable terminal 25 that is a music player or a smartphone via the signal reception device 21.
- the audio signal acquisition unit 11 acquires an audio signal that is a digital signal by wireless communication using Bluetooth.
- the DAC 22 converts the input digital audio signal into an analog audio signal and outputs the analog audio signal to the amplifier 23.
- the amplifying device 23 amplifies the input audio signal and outputs it to the earphone 24.
- the earphone 24 outputs sound based on the input sound signal.
- the user operation detection unit 121 acquires the current operation of the user who uses the audio signal reproduction device 1 and outputs the acquired operation to the virtual sound source position determination unit 122.
- the user action notified at this time is used as an index when the virtual sound source position determination unit 122 determines the position of each audio signal.
- the user operation detection unit 121 detects the operation of the user 31 based on outputs from various sensors or systems for detecting the operation of the user 31.
- the sensor is, for example, an acceleration sensor or a gyro sensor, and the system is, for example, GPS (Global Positioning System), but is not limited thereto. These sensors are desirably installed on a part of the user's body (preferably the head) in order to detect the user's movement.
- the user motion detection unit 121 detects the motion of the user 31 based on the output of an acceleration sensor (not shown). Specifically, the user motion detection unit 121 detects (determines) the motion of the user 31 as “movement” if the output of the acceleration sensor is equal to or greater than a predetermined threshold Th. On the other hand, if the output of the acceleration sensor is smaller than the threshold Th, the operation of the user 31 is detected (determined) as “still”.
- the threshold value Th is stored in the storage unit 14 in advance.
- the operation of the user 31 changes from moment to moment, and based on the operation of the user 31 from time to time, the operation of the user 31 notified from the user operation detection unit 121 to the virtual sound source position determination unit 122 is “moving” or “still”. Change.
- the virtual sound source position determination unit 122 is a virtual sound source that virtually outputs a sound represented by each sound signal acquired by the sound signal acquisition unit 11 based on a user action detected by the user action detection unit 121. And the audio signal processing unit 123 is notified. In addition to the detected user action, the virtual sound source position determination unit 122 is based on the preconditions stored in advance in the storage unit 14 (for example, the position and number of virtual sound sources arranged around the user 31). The position of the virtual sound source in each audio signal may be determined.
- FIG. 3 is a diagram illustrating the relationship between the user 31 and the position of the virtual sound source 33 according to the first embodiment of the present invention.
- the user (listener) 31 does not perceive the sound represented by each sound signal as the sound output from the position of the actual sound source (that is, the earphone 24). Instead, the user 31 perceives the sound represented by each sound signal as sound output from a virtual sound source set to each sound signal by the stereophonic sound processing of the sound signal processing unit 123.
- a sound source in which sound represented by each sound signal is virtually output is referred to as “virtual sound source”.
- the position of the virtual sound source is a relative position of the virtual sound source arranged around the user 31 with respect to the user 31.
- the position of the virtual sound source is represented by a predetermined coordinate system having the origin 32 as an intermediate position between the right ear and the left ear of the user 31.
- FIG. 3 shows a virtual sound source 33 as an example of the virtual sound source.
- this coordinate system is a two-dimensional polar coordinate system composed of a distance (radial radius) r from the origin 32 to the virtual sound source 33 and an angle (deflection angle) ⁇ of the virtual sound source 33 with respect to the origin 32. It is. That is, the position of the virtual sound source 33 is expressed as a combination of the distance r and the angle ⁇ .
- the angle ⁇ of the virtual sound source is an angle formed by a straight line L1 passing through the origin 32 and a straight line L2 connecting the origin 32 and the virtual sound source 33.
- the distance r of the virtual sound source 33 is assumed to take one of three values (r 1 , r 2 , or r 3 ). Note that the relationship r 1 ⁇ r 2 ⁇ r 3 is satisfied.
- the virtual sound source 33 when the distance of the virtual sound source 33 is r 1, the virtual sound source 33 is located at any position on the circumference C1. Further, when the distance of the virtual sound source 33 is r 2, the virtual sound source 33 is located at any position on the circumference C2. If the distance of the virtual sound source 33 is r 3, the virtual sound source 33 is located at any position on the circumference C3.
- the audio signal processing unit 123 performs predetermined stereophonic sound processing on each audio signal input from the audio signal acquisition unit 11 according to the position of the virtual sound source of each audio signal notified from the virtual sound source position determination unit 122. Thus, a virtual sound source that virtually outputs sound from the notified position is set for each sound signal.
- the audio signal processing unit 123 outputs each processed audio signal to the audio signal reproduction unit 13.
- the audio signal processing unit 123 converts each audio signal input from the audio signal acquisition unit 11 into each audio signal of a stereophonic sound system by applying a head related transfer function (HRTF).
- HRTF head related transfer function
- HL n (z) is an HRTF for the left ear at the position (deflection angle) of the virtual sound source set in the input signal I n (z).
- HR n (z) is an HRTF for the right ear at the position (deflection angle) of the virtual sound source set in the input signal I n (z).
- these HRTFs are stored in advance in the storage unit 14 as discrete table information.
- the coefficient d indicates an attenuation amount based on the distance r from the origin 32 of each virtual sound source, and is represented by the following equation (2) in the present embodiment.
- Equation (2) r represents the distance of the virtual sound source from the origin 32, and ⁇ is a preset coefficient.
- the audio signal reproduction unit 13 converts the left ear signal L OUT and the right ear signal R OUT generated by the audio signal processing unit 123 into a digital audio signal in an arbitrary audio file format.
- the audio signal reproducing unit 13 reproduces the converted digital audio signal by outputting it to the outside of the audio signal reproducing apparatus 1.
- the audio signal reproduction unit 13 when the audio signal reproduction apparatus 1 is provided in the audio signal reproduction system 2, the audio signal reproduction unit 13 generates a digital audio signal in, for example, an Inter-IC Sound (I2S) format to generate a DAC 22. Output to.
- I2S Inter-IC Sound
- the audio signal acquisition unit 11 acquires at least one audio signal from the outside of the audio signal reproduction device 1 (for example, the mobile terminal 25) through the signal reception device 21 and outputs the acquired audio signal to the virtual sound source position determination unit 122.
- the user operation detection unit 121 detects the current operation of the user 31 of the audio signal reproduction device 1 and notifies the virtual sound source position determination unit 122 of the current operation.
- the virtual sound source position determination unit 122 determines a position where each sound represented by each input sound signal is virtually output based on the notified operation of the user 31.
- the virtual sound source position determination unit 122 outputs the determined positions to the audio signal processing unit 123 in association with the corresponding audio signals.
- the audio signal processing unit 123 performs stereophonic sound processing on each input audio signal based on the position of each virtual sound source determined by the virtual sound source position determination unit 122, so that the left ear signal L OUT and the right ear signal L OUT are processed.
- a signal R OUT is generated and output to the audio signal reproduction unit 13.
- the audio signal reproduction unit 13 converts the input left ear signal L OUT and right ear signal R OUT into a digital audio signal in an arbitrary audio file format, and outputs the digital audio signal to the DAC 22.
- the DAC 22 converts the input digital audio signal into an analog audio signal and outputs the analog audio signal to the amplifying device 23.
- the amplifying device 23 amplifies the input analog audio signal and outputs it to the earphone 24.
- the earphone 24 outputs sound represented by the input analog sound signal.
- the user 31 perceives each sound represented by each sound signal acquired by the sound signal acquisition unit 11 as sound output from each virtual sound source set in each sound signal.
- FIG. 4A is a diagram illustrating an example of the position of the virtual sound source 41 when the user 31 is stationary in Embodiment 1 of the present invention
- FIG. 4B is a diagram illustrating the implementation of the present invention. It is a figure which shows an example of the position of the virtual sound source 41 in the form 1 when the user 31 is stationary.
- the audio signal acquisition unit 11 acquires one audio signal from the outside of the audio signal reproduction device 1.
- the user motion detection unit 121 detects the motion of the user 31 as “still”. Further, the virtual sound source position determination unit 122 determines the position of one virtual sound source 41 corresponding to one acquired sound signal as a position (r 2 , ⁇ 1 ) corresponding to the detected “stillness”. Thereby, as illustrated in FIG. 4A, the virtual sound source 41 is disposed at a position (r 2 , ⁇ 1 ) in front of the user 31.
- the user action detection unit 121 detects that the action of the user 31 is “movement”. Further, the virtual sound source position determination unit 122 determines the position of one virtual sound source 41 corresponding to one acquired sound signal as a position (r 3, ⁇ 1 ) corresponding to the detected “movement”. As a result, as shown in FIG. 4A, the virtual sound source 41 is arranged at a position (r 3 , ⁇ 1 ) in front of the user 31.
- the virtual sound source position determination unit 122 makes the distance of the virtual sound source 41 when “moving” is detected longer than the distance of the virtual sound source 41 when “still” is detected. Then, the position of the virtual sound source 41 is determined. As a result, when the operation of the user 31 changes from “still” to “move”, the position of the virtual sound source 41 moves away from the position (r 2 , ⁇ 1 ) to the position (r 3 , ⁇ 1 ). At this time, the angle of the virtual sound source 41 remains ⁇ 1 and the distance of the virtual sound source 41 changes from r 2 to r 3 . As a result, the user 31 can hear the sound heard from the closer virtual sound source 41 when stationary, from the farther virtual sound source 41 when moving.
- the position where the virtual sound source 41 is arranged is determined in advance in the audio signal reproduction device 1.
- Position information indicating each position is stored in advance in the storage unit 14, and the virtual sound source position determination unit 122 reads out and uses the information from the storage unit 14, so that the virtual sound source corresponding to the operation of the user 31 is performed. 41 is determined.
- the information indicating the position of the virtual sound source 41 may specify the position of the virtual sound source 41 as an absolute value, or may indicate information as a relative value.
- the virtual sound source position determination unit 122 uses the coefficient ⁇ satisfying the following expression (3) with reference to the position (r 2 , ⁇ 1 ) of the virtual sound source 41 at the time of “still” shown in FIG. Is used to determine the position of the virtual sound source 41 at the time of “operation” as the position (r 2 + ⁇ , ⁇ 1 ).
- FIG. 5A is a diagram illustrating an example of the position of the virtual sound source 51 when the user 31 is stationary in Embodiment 1 of the present invention
- FIG. 5B is a diagram illustrating the implementation of the present invention. It is a figure which shows an example of the position of the virtual sound source 51 when the user 31 is moving in the form 1.
- the virtual sound source 51 corresponding to one acquired audio signal is arranged at a position (r 2 , ⁇ 1 ) in front of the user 31.
- the virtual sound source 51 is arranged at a position (r 2 , ⁇ 2 ) behind the user 31. At this time, the distance of the virtual sound source 51 remains r 2 , and the angle of the virtual sound source 51 changes from ⁇ 1 to ⁇ 2 .
- FIG. 6A is a diagram illustrating an example of the position of the virtual sound source 61 when the user 31 is stationary in Embodiment 1 of the present invention
- FIG. 5B is a diagram illustrating the implementation of the present invention. It is a figure which shows an example of the position of the virtual sound source 61 when the user 31 is moving in the form 1.
- the virtual sound source 61 corresponding to the acquired one audio signal is arranged at a position (r 2 , ⁇ 1 ) in front of the user 31.
- the virtual sound source 51 is arranged at a position (r 3 , ⁇ 2 ) behind the user 31. At this time, the distance of the virtual sound source 51 changes from r 2 to r 3 , and the angle of the virtual sound source 51 changes from ⁇ 1 to ⁇ 2 .
- the virtual sound source position determination unit 122 determines at least one of the distance and the angle from the origin 32 to the virtual sound source with reference to the user 31 based on the detected operation of the user 31.
- the position of the virtual sound source can be determined to be different.
- the audio signal acquisition unit 11 can also acquire a plurality of audio signals from the outside of the audio signal reproduction device 1. In this case, the position of each virtual sound source corresponding to each audio signal is detected based on the detected operation of the user 31.
- FIG. 7A is a diagram showing an example of the positions of the plurality of virtual sound sources 71 to 73 when the user 31 is stationary in Embodiment 1 of the present invention
- FIG. FIG. 7 is a diagram illustrating an example of positions of a plurality of virtual sound sources 71 to 73 when a user 31 is moving in the first embodiment of the present invention.
- a case where only the distance between the virtual sound sources 71 to 73 changes when the operation of the user 31 changes will be described.
- FIG. 7A since the user 31 is currently stationary, the virtual sound sources 71 to 73 corresponding to the three acquired audio signals are positioned around the user 31 (r 2 , ⁇ 1 ), position (R 2 , ⁇ 2 ) and position (r 2 , ⁇ 3 ) are arranged.
- FIG. 7B since the user 31 is currently moving forward (direction X), the virtual sound sources 71 to 73 are positioned around the user 31 (r 3 , ⁇ 1 ), position (r 3 , ⁇ 2 ), and position (r 3 , ⁇ 3 ). At this time, none of the angles of the virtual sound sources 71 to 73 change, while all the distances of the virtual sound sources 71 to 73 change from r 2 to r 3 .
- FIG. 8A is a diagram illustrating an example of the positions of the plurality of virtual sound sources 81 to 83 when the user 31 is stationary in Embodiment 1 of the present invention
- FIG. 7 is a diagram illustrating an example of positions of a plurality of virtual sound sources 81 to 83 when a user 31 is moving in the first embodiment of the present invention.
- both the distance and the angle of each of the virtual sound sources 81 to 83 change when the operation of the user 31 changes.
- FIG. 8A since the user 31 is currently stationary, the virtual sound sources 81 to 83 corresponding to the three acquired audio signals are positioned around the user 31 (r 2 , ⁇ 1 ), position (R 2 , ⁇ 2 ) and position (r 2 , ⁇ 3 ) are arranged.
- FIG. 8B since the user 31 is currently moving forward (direction X), the virtual sound sources 81 to 83 are positioned around the user 31 (r 3 , ⁇ 4 ), position (r 3 , ⁇ 5 ), and position (r 3 , ⁇ 6 ).
- the distances of the virtual sound sources 71 to 73 all change from r 2 to r 3 , and the angles of the virtual sound sources 71 to 73 are changed from ⁇ 1 , ⁇ 2 , and ⁇ 3 to ⁇ 4 , ⁇ 5 , and ⁇ . Change to 6 respectively.
- the user operation detection unit 121 detects “still” or “movement” as the operation of the user 31, and the virtual sound source position determination unit 122 responds to these detected “still” or “movement”.
- An example of determining the position of the virtual sound source has been described.
- the processes of the user motion detection unit 121 and the virtual sound source position determination unit 122 are not limited to this.
- the user action detection unit 121 represents “gazing” indicating that the user 31 is gazing at a certain area as the action of the user 31 or “noting that the user 31 is not gazing at the certain area”. It is also possible to detect “non-gaze”. In this case, the virtual sound source position determination unit 122 determines the position of the virtual sound source to a position based on the detected “gaze” or “non-gaze”.
- the user operation detection unit 121 detects whether the operation of the user 31 is “gazing” or “non-gazing” based on outputs from various sensors attached to the user 31. Specifically, when a gyro sensor is attached to the user 31, the user motion detection unit 121 sets a change in output (that is, an angle) of the gyro sensor for each preset unit time t 1 . If it is within the threshold Th 1 , “gazing” is detected as the operation of the user 31. On the other hand, when the change in the angle is greater than the threshold value Th 1, for detecting the "inattentive” as the operation of the user 31.
- the user operation detection unit 121 calculates a gaze area 94 (a constant area) in which the user 31 is actually gazing based on an output from the gyro sensor.
- the virtual sound source position determination unit 122 is notified of the detection of “gaze” and the calculated gaze area 94.
- the virtual sound source position determination unit 122 determines the position of the virtual sound source 91 so that the position of the virtual sound source 91 located in the gaze area 94 varies based on the detected “gaze” or “non-gaze”. More specifically, the virtual sound source position determination unit 122 uses the distance of the virtual sound source 91 located in the gaze area 94 when “gaze” is detected as the gaze area 94 when “non-gaze” is detected.
- the position of the virtual sound source 91 is determined so as to be shorter than the distance of the virtual sound source 91 located inside.
- FIG. 9 is a diagram illustrating an example of the positions of the plurality of virtual sound sources 91 to 93 when the user 31 is not gazing at the fixed gaze area 94 in the first embodiment of the present invention.
- FIG. 9B is a diagram illustrating an example of the positions of the plurality of virtual sound sources 91 to 93 when the user 31 is gazing at the gaze area 94 in the first embodiment of the present invention.
- FIG. 9C is a diagram illustrating another example of positions of a plurality of virtual sound sources when the user is gazing at the gaze area in the first embodiment of the present invention.
- FIG. 9D is a diagram illustrating still another example of the positions of the plurality of virtual sound sources when the user is gazing at the gaze area in the first embodiment of the present invention.
- the three virtual sound sources 91 to 93 are moved to the position (r 2 , ⁇ 1 ) and the position (r 2 , ⁇ , respectively). 2 ) and the position (r 2 , ⁇ 3 ).
- the user operation detection unit 121 detects that the user 31 is gazing at the gaze area 94.
- the position of the virtual sound source 91 included in the gaze area 94 is closer to the position (r 1 , ⁇ 1 ) closer to the user 31 as shown in FIG. 9B. Be changed.
- the positions of the remaining virtual sound sources 92 and 93 are not changed.
- the sound emitted from the virtual sound source 91 in the gaze area 94 can be heard closer to the user 31 than before gaze. That is, among the three virtual sound sources 91 to 93, the sound from the virtual sound source 91 that the user 31 seems to be interested in can be heard more loudly, so that the user 31 can be more satisfied.
- the position change of the three virtual sound sources 91 to 93 at the time of gaze is not limited to the example shown in FIG. 9B.
- the positions of the three virtual sound sources 91 to 93 may be changed to the positions shown in (c) of FIG.
- the position of the virtual sound source 91 included in the gaze area 94 is not changed, while the positions of the virtual sound sources 92 and 93 are farther from the user 31 (r 3 , ⁇ 2 ). And the position (r 3 , ⁇ 3 ).
- each sound emitted from the virtual sound sources 92 and 93 outside the gaze area 94 can be heard farther from the user 31 than before gaze.
- the sound from the virtual sound source 91 that the user 31 seems to be interested in can be heard relatively louder, so that the user 31 can be more satisfied.
- the position change of the three virtual sound sources 91 to 93 at the time of gaze is not limited to the example shown in FIG. 9B and the example shown in FIG.
- the positions of the three virtual sound sources 91 to 93 may be changed to the positions shown in (d) of FIG.
- the position of the virtual sound source 91 included in the gaze area 94 is not changed, while the positions of the virtual sound sources 92 and 93 are the areas 95 (other areas on the opposite side of the gaze area 94).
- the region 95 is not limited to the opposite side of the gaze region 94 and may be any region in a different location from the gaze region.
- the operation of the user 31 detected by the user operation detection unit 121 is not limited to “gaze” and “non-gaze”.
- the user motion detection unit 121 can detect the state in which the user lies based on both the output from the gyro sensor and the output from the acceleration sensor.
- the user operation detection unit 121 detects “sleep” indicating that the user 31 is asleep as the operation of the user 31 and notifies the virtual sound source position determination unit 122 of it.
- the virtual sound source position determination unit 122 determines to lower the volume of sound output from each virtual sound source or to stop output of sound from each virtual sound source. May be. As a result, it is possible to prevent the sleep of the user 31 from being inhibited by the sound from each virtual sound source.
- the present invention is not limited to this.
- the operation of the user 31 detected by the virtual sound source position determination unit 122 may be any of three or more different operations.
- the audio signal reproduction system 2 including the audio signal reproduction apparatus 1 that reproduces the audio signal and the earphone 24 that outputs the audio has been described.
- the embodiment of the present invention has such a configuration. It is not specifically limited to.
- the audio signal reproducing device 1 is configured to transmit an audio signal to a speaker in an audio system (a component may be a separate component or a single component) that includes a tuner, a player, and a speaker. It can be realized as an output player.
- the audio signal reproduction device 1 is also realized as a single device including an audio signal reproduction device (player) capable of performing audio signal processing and an audio output device (for example, a speaker or headphones) capable of outputting audio.
- Embodiment 2 A second embodiment according to the present invention will be described below with reference to FIGS.
- Each member common to Embodiment 1 described above is denoted by the same reference numeral, and detailed description thereof is omitted.
- the virtual sound source position determination unit 122 has been described based on an example in which the position of each virtual sound source is determined based only on the operation of the user 31 detected by the user operation detection unit 121.
- the present invention is not limited to this.
- the virtual sound source position determination unit 122 determines the position of each virtual sound source based on both the detected operation of the user 31 and predetermined metadata (additional data) added to each audio signal. An example will be described.
- FIG. 10 is a block diagram showing a main configuration of an audio signal reproduction device 1a according to Embodiment 2 of the present invention.
- the audio signal reproduction device 1 further includes a metadata acquisition unit 15 (additional data acquisition unit) in addition to each member provided in the audio signal reproduction device 1 according to the first embodiment.
- the content of the process by the virtual sound source position determining unit 122 according to the present embodiment is different from the content of the process by the virtual sound source position determining unit 122 according to the first embodiment. Since the contents of processing by other members such as the audio signal acquisition unit 11 are the same as those in the first embodiment, detailed description thereof is omitted.
- the metadata acquisition unit 15 acquires predetermined metadata added to each audio signal acquired by the audio signal acquisition unit 11 and outputs it to the virtual sound source position determination unit 122.
- the metadata acquisition unit 15 acquires the metadata added to each audio signal from the outside of the audio signal reproduction device 1 a separately from the audio signal acquired by the audio signal acquisition unit 11.
- the metadata acquisition unit 15 can acquire metadata by either wired communication or wireless communication.
- the metadata acquisition unit 15 may acquire metadata from the storage unit 14.
- the audio signal and the metadata corresponding to the audio signal may be managed as audio data in an arbitrary audio file format in which these are paired outside the audio signal reproduction device 1a.
- a decoder (not shown) provided in either the audio signal reproducing device 1a or the audio signal reproducing system 2 acquires the audio data and appropriately decodes the audio data, thereby converting the audio data into the audio signal, the metadata, To separate.
- the audio signal acquisition unit 11 acquires the audio signal separated by the decoder
- the metadata acquisition unit 15 acquires the metadata separated from the audio data by the decoder.
- the audio signal acquisition unit 11 acquires at least one audio signal from the outside of the audio signal reproduction device 1 (for example, the mobile terminal 25) through the signal reception device 21, and outputs the acquired audio signal to the virtual sound source position determination unit 122.
- the user operation detection unit 121 detects the current operation of the user 31 of the audio signal reproduction device 1 and notifies the virtual sound source position determination unit 122 of the current operation.
- the virtual sound source position determination unit 122 determines the position of the virtual sound source from which each sound represented by each input sound signal is virtually output based on the notified operation of the user 31 and the input metadata. decide. Further, the virtual sound source position determination unit 122 may determine the position of the virtual sound source based on the notified operation of the user 31 and position information indicating the position of each virtual sound source stored in advance in the storage unit 14. it can. The virtual sound source position determination unit 122 outputs the determined positions to the audio signal processing unit 123 in association with the corresponding audio signals. Since the processing by the audio signal processing unit 123 is the same as that of the first embodiment, description thereof is omitted.
- (A) to (c) of FIG. 11 are diagrams showing examples of metadata in Embodiment 2 of the present invention.
- the metadata shown in (a) of FIG. 11 defines whether or not the change of the position of each of the three virtual sound sources is permitted.
- the metadata shown in (b) of FIG. 11 is important for the output of sound from the virtual sound source to the user 31 according to each detected motion (stationary or moving) of the user 31 for each of the three virtual sound sources. The degree is specified.
- the metadata shown in (c) of FIG. 11 shows the absolute position on the two-dimensional polar coordinates of each of the three virtual sound sources, indicating the position of the virtual sound source according to each detected motion (stationary or moving) of the user 31. It is specified as a value.
- FIG. 12 (Virtual sound source position) (A) of FIG. 12 is a diagram illustrating an example of the positions of the plurality of virtual sound sources 121 to 123 when the user 31 is stationary in Embodiment 2 of the present invention
- FIG. 10 is a diagram illustrating an example of positions of a plurality of virtual sound sources 121 to 123 when a user 31 is moving in Embodiment 2 of the present invention.
- the audio signal acquisition unit 11 acquires three different audio signals from the outside of the audio signal reproduction device 1a.
- the metadata acquisition part 15 acquires the metadata shown to (a) of FIG.
- the user motion detection unit 121 detects the motion of the user 31 as “still”.
- the storage unit 14 stores positions (r 2 , ⁇ 1 ), positions (r 2 , ⁇ 2 ), and (r 2 , ⁇ ) as the positions of the three virtual sound sources when the operation of the user 31 is “still”.
- Position information indicating each of 3 ) is stored. Therefore, the virtual sound source position determination unit 122 determines the positions of the three virtual sound sources 121 to 123 corresponding to the acquired three audio signals as the position (r 2 , ⁇ 1 ), the position (r 2 , ⁇ 2 ), and (R 2 , ⁇ 3 ) As a result, as shown in FIG. 12A, the virtual sound sources 121 to 123 are moved to positions (r 2 , ⁇ 1 ), positions (r 2 , ⁇ 2 ), and (r 2 , ⁇ ) around the user 31. 3 ), respectively.
- the user operation detection unit 121 detects that the operation of the user 31 is “movement”.
- the storage unit 14 stores positions (r 3 , ⁇ 1 ), positions (r 3 , ⁇ 2 ), and (r 3 , ⁇ ) as the positions of the three virtual sound sources when the operation of the user 31 is “move”.
- Position information indicating each of 3 ) is stored. Therefore, the virtual sound source position determination unit 122 determines the positions of the three virtual sound sources 121 to 123 corresponding to the acquired three audio signals as “movement” detected and the metadata shown in FIG. The position is determined according to both.
- the virtual sound source position determination unit 122 sets the positions of the virtual sound source 121 and the virtual sound source 123 to positions (r 3 , ⁇ 1 ) and positions (r 3 , ⁇ 3 ) defined in the position information corresponding to “movement”. Change to On the other hand, since the metadata shown in FIG. 11A stipulates that the movement of the position of the virtual sound source 121 corresponding to the sound source number 1 and the virtual sound source 123 corresponding to the sound source number 3 is permitted.
- the virtual sound source position determination unit 122 sets the positions of the virtual sound source 121 and the virtual sound source 123 to positions (r 3 , ⁇ 1 ) and positions (r 3 , ⁇ 3 ) defined in the position information corresponding to “movement”. Change to On the other hand, since the metadata shown in FIG.
- the virtual sound source position determination unit 122 stipulates that the virtual sound source 122 corresponding to the sound source number 2 is not allowed to move, the virtual sound source position determination unit 122 The position is not changed to the position (r 3 , ⁇ 2 ) defined in the position information corresponding to “movement”, but is maintained at the position (r 2 , ⁇ 2 ) corresponding to “still”. Thereby, as shown in FIG. 12B, the virtual sound sources 121 to 123 are positioned around the user 31 at the position (r 3 , ⁇ 1 ), the position (r 2 , ⁇ 2 ), and the position (r 3 , ⁇ 3 ), respectively.
- each virtual sound source is arranged at a position corresponding to the operation of the user 31 and the metadata. Therefore, the position of each virtual sound source according to the operation of the user 31 can be flexibly controlled by appropriately changing the content of the metadata.
- Embodiment 3 Embodiment 3 according to the present invention will be described below with reference to FIG.
- Each member common to the above-described first or second embodiment is denoted by the same reference numeral, and detailed description thereof is omitted.
- Embodiment 1 has described an example in which the audio signal acquisition unit 11 acquires an audio signal unrelated to the operation of the user 31 detected by the user operation detection unit 121, but the present invention is not limited to this. In the present embodiment, an example in which the audio signal acquisition unit 11 acquires an appropriate audio signal according to the detected operation of the user 31 will be described below.
- FIG. 13 is a block diagram showing a main configuration of an audio signal reproduction device 1b according to Embodiment 3 of the present invention.
- each member provided in the audio signal reproduction device 1b is the same as each member provided in the audio signal reproduction device 1 according to the first embodiment.
- the content of the processing by the audio signal acquisition unit 11 and the user motion detection unit 121 is different from the content of the processing by the virtual sound source position determination unit 122 according to the first embodiment. Since the contents of processing by other members such as the virtual sound source position determining unit 122 are the same as those in the first embodiment, detailed description thereof is omitted.
- the user operation detection unit 121 notifies the detected operation of the user 31 not only to the virtual sound source position determination unit 122 but also to the audio signal acquisition unit 11.
- voice signal acquisition part 11 acquires the audio
- the audio signal acquisition unit 11 changes the audio signal acquired from the portable terminal 25 based on the notified operation of the user 31.
- the mobile terminal 25 is provided with different audio signals corresponding to each operation of the user 31 in advance.
- the portable terminal 25 is prepared in advance with an audio signal A corresponding to “still” and an audio signal corresponding to “move”.
- the audio signal A is a signal that is preferably reproduced when the user 31 is stationary, while the audio signal B is a signal that is preferably reproduced when the user 31 is moving.
- the voice signal acquisition unit 11 requests the portable terminal 25 for the voice signal A corresponding to “still”.
- the portable terminal 25 provides the audio signal A to the audio signal acquisition unit 11 in response to the request.
- the audio signal A that is preferably reproduced when the user 31 is stationary is reproduced when the actual user 31 is stationary. Therefore, the user 31 perceives a sound suitable for the state as a sound output from a virtual sound source at an appropriate position corresponding to the stationary state when the user 31 is stationary.
- the audio signal acquisition unit 11 requests the mobile terminal 25 for an audio signal B corresponding to “move”.
- the portable terminal 25 provides the audio signal B to the audio signal acquisition unit 11 in response to the request.
- the audio signal B that is preferably reproduced when the user 31 moves is reproduced when the user 31 is actually moving. Accordingly, when the user 31 moves, the user 31 perceives a sound appropriate for the situation as a sound output from a virtual sound source at an appropriate position corresponding to the movement.
- the user operation detection unit 121 detects the current operation of the user 31 of the audio signal reproduction device 1 and notifies the audio signal acquisition unit 11 and the virtual sound source position determination unit 122.
- the audio signal acquisition unit 11 acquires at least one audio signal corresponding to the notified operation of the user 31 from the outside of the audio signal reproduction device 1 (for example, the mobile terminal 25) through the signal reception device 21, and the virtual sound source position
- the data is output to the determination unit 122.
- the virtual sound source position determination unit 122 determines the position of the virtual sound source from which each sound represented by each input sound signal is virtually output based on the notified operation of the user 31.
- the virtual sound source position determination unit 122 outputs the determined positions to the audio signal processing unit 123 in association with the corresponding audio signals.
- the audio signal processing unit 123 performs stereophonic sound processing according to the determined position on the audio signal.
- An audio signal reproduction device provides: An acquisition unit (audio signal acquisition unit 11) that acquires an audio signal indicating audio; A detection unit (user operation detection unit 121) for detecting a user operation; A determination unit (virtual sound source position determination unit 122) that determines a position of a virtual sound source from which the sound represented by the acquired sound signal is output based on the detected operation; A processing unit (audio signal processing unit 123) that performs stereophonic sound processing according to the determined position of the virtual sound source on the acquired audio signal; And a reproduction unit (audio signal reproduction unit 13) that reproduces the audio signal subjected to the stereophonic processing.
- the virtual sound source is arranged at an appropriate position according to the detected user action.
- the audio signal reproduction device can reproduce the audio signal so that the audio is output from a position suitable for the user's operation.
- An audio signal reproduction device is the above-described aspect 1,
- the determining unit may vary at least one of a distance from a predetermined origin to the virtual sound source with respect to the user and an angle of the virtual sound source with respect to the origin based on the detected user's operation. The position of the virtual sound source is determined.
- the position of the virtual sound source is determined so that at least one of the distance and the angle of the virtual sound source has an appropriate value according to the user's operation.
- the audio signal reproduction device is the above-described aspect 2,
- the detection unit detects movement indicating that the user is moving or stillness indicating that the user is stationary as the user's operation,
- the determining unit determines the position of the virtual sound source so that the position of the virtual sound source varies based on the detected movement or the stillness detected.
- the virtual sound source is arranged at an appropriate position according to the user's operation when the user is stationary and moving.
- the audio signal reproduction device is the above-described aspect 3,
- the determining unit determines the position of the virtual sound source to be behind the user when the movement is detected, and determines the position of the virtual sound source to be in front of the user when the stillness is detected. It is a feature.
- the user can hear the sound heard from the virtual sound source in front of himself / herself when standing still from the virtual sound source behind him / her when moving.
- no voice can be emitted from the front of the user when moving, so that it is possible to prevent the user from hesitating to move forward in consideration of the voice emitted from the front.
- the user is not obstructed by his / her voice output from the virtual sound source.
- the audio signal reproducing device is the above aspect 3 or 4,
- the determination unit determines the position of the virtual sound source so that the distance of the virtual sound source when the movement is detected is longer than the distance of the virtual sound source when the stillness is detected. It is characterized by that.
- An audio signal reproduction device is the above-described aspect 2,
- the detection unit detects, as the operation, gaze indicating that the user is gazing at a certain area or non-gaze indicating that the user is not gazing at the certain area,
- the determining unit determines the position of the virtual sound source so that the position of the virtual sound source located in the certain region is different based on the detected gaze or the non-gaze.
- the virtual sound source is arranged at an appropriate position according to the user's operation at the time of user's gaze and non-gaze.
- An audio signal reproduction device is the above aspect 6,
- the determination unit determines the distance of the virtual sound source located in the certain area when the gaze is detected, and the distance of the virtual sound source located in the certain area when the non-gaze is detected.
- the position of the virtual sound source is determined so as to be shorter than the distance.
- the sound emitted from the virtual sound source in the area can be heard closer to the user than before gaze. That is, since the sound from the virtual sound source that the user is currently interested in can be heard louder, the user can be more satisfied.
- the determination unit determines the distance of the virtual sound source located outside the certain area when the gaze is detected, The position of the virtual sound source is determined so as to be longer than the distance of the virtual sound source located outside the certain area when non-gaze is detected.
- the sound emitted from the virtual sound source outside the area can be heard farther from the user than before gaze.
- the sound from the virtual sound source that the user is currently interested in can be heard relatively louder, so that the user can be more satisfied.
- the audio signal reproduction device is the audio signal reproduction device according to any one of the above aspects 6 to 8, wherein the determination unit detects the position of the virtual sound source located outside the certain region when the gaze is detected. The position of the virtual sound source is determined so as to be moved to another area different from the certain area.
- the sound emitted from the virtual sound source outside the area can be heard from another area (for example, behind the user) different from the area.
- another area for example, behind the user
- An audio signal reproduction device is any one of the above aspects 1 to 9, An additional data acquisition unit that acquires predetermined additional data added to the audio signal; The determining unit determines the position of the virtual sound source based on both the detected motion and the acquired additional data.
- the position of the virtual sound source according to the user's operation can be flexibly controlled.
- An audio signal reproduction device is any of the above aspects 1 to 10,
- the acquisition unit acquires the audio signal corresponding to the detected user action.
- the user can perceive a sound appropriate for his / her action as a sound output from a virtual sound source at an appropriate position according to the action.
- An audio signal reproduction method includes: An acquisition step of acquiring an audio signal indicating audio; A detection process for detecting user actions; A determination step of determining a position of a virtual sound source from which the sound represented by the acquired sound signal is output based on the detected user action; A processing step of performing stereophonic sound processing according to the determined position of the virtual sound source on the acquired audio signal; And a reproduction step of reproducing the audio signal subjected to the stereophonic processing.
- the audio signal reproduction device may be realized by a computer.
- a control program for the audio signal reproduction device that causes the audio signal reproduction device to be realized by the computer by operating the computer as each unit included in the audio signal reproduction device, and a computer-readable recording medium that records the control program are also provided. It falls within the scope of the present invention.
- Each functional block of the audio signal reproduction device 1 shown in FIG. 1 may be realized by a logic circuit (hardware) formed in an integrated circuit (IC chip) or the like, or software using a CPU (Central Processing Unit). It may be realized by.
- the audio signal reproduction device 1 includes a CPU that executes instructions of a program that is software that realizes each function, and a ROM (Read Only Memory) in which the program and various data are recorded so as to be readable by a computer (or CPU) ) Or a storage device (these are referred to as “recording media”), a RAM (Random Access Memory) for expanding the program, and the like.
- a computer or CPU
- a “non-temporary tangible medium”, for example, a tape, a disk, a card, a semiconductor memory, a programmable logic circuit, or the like can be used as the recording medium.
- the program may be supplied to the computer via an arbitrary transmission medium (such as a communication network or a broadcast wave) that can transmit the program.
- the present invention can also be realized in the form of a data signal embedded in a carrier wave in which the program is embodied by electronic transmission.
- the audio signal reproducing device 1a shown in FIG. 10 and the audio signal reproducing device 1b shown in FIG. 13 can also be realized by a shift in hardware or software.
- the present invention can be suitably used as various portable music players or audio systems.
- Audio signal reproduction device 1, 1a, 1b Audio signal reproduction device 2 Audio signal reproduction system 11 Audio signal acquisition unit (acquisition unit) 12 control unit 14 storage unit 15 metadata acquisition unit (additional data acquisition unit) 121 User motion detection unit (detection unit) 122 Virtual sound source position determination unit (determination unit) 123 Audio signal processing unit (processing unit) 22 DAC 23 Amplifier 24 Earphone
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Signal Processing (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- General Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Health & Medical Sciences (AREA)
- Stereophonic System (AREA)
Abstract
Description
本発明の実施形態1について、図1~図9を参照して以下に説明する。
図2は、本発明の実施形態1に係る音声信号再生システム2の要部構成を示すブロック図である。この図に示すように、音声信号再生システム2は、音声信号再生装置1に加え、信号受信装置21、デジタル-アナログ変換装置(DAC)22、増幅装置23、およびイヤホン24をさらに備えている。
図3を参照して、仮想音源の位置についてより詳細に説明する。図3は、本発明の実施形態1におけるユーザ31と仮想音源33の位置との関係を説明する図である。ユーザ(聴取者)31は、各音声信号によって表される音声を、実際の音源(すなわちイヤホン24)の位置から出力された音声としては、知覚しない。その代わりにユーザ31は、各音声信号によって表される音声を、音声信号処理部123の立体音響処理によって各音声信号に設定される仮想的な音源から出力される音声して、知覚する。以下では、各音声信号によって表される音声が仮想的に出力される音源を、「仮想音源」と表記する。仮想音源の位置は、ユーザ31の周囲に配置される仮想音源の、ユーザ31に対する相対的な位置である。
音声信号再生システム2における音声信号再生処理の流れの一例について、以下に説明する。音声信号取得部11は、信号受信装置21を通じて、音声信号再生装置1の外部(たとえば携帯端末25)から少なくとも1つの音声信号を取得し、仮想音源位置決定部122に出力する。一方、ユーザ動作検知部121は、音声信号再生装置1のユーザ31の現在動作を検知し、仮想音源位置決定部122に通知する。
上述したように、各仮想音源の位置は、ユーザ31の動作に基づき決定される。その一例について、図4を参照して以下に説明する。
図4の例では、ユーザ31の動作が変化した場合、仮想音源41の距離のみが変化する。しかし本発明はこれに限らず、ユーザ31の動作が変化した場合、仮想音源の距離ではなく角度のみが変化してもよい。この例について、図5を参照して以下に説明する。
ユーザ31の動作が変化した場合、仮想音源の距離および角度の双方が変化してもよい。この例について、図6を参照して以下に説明する。
音声信号取得部11は、複数の音声信号を音声信号再生装置1の外部から取得することもできる。この場合、各音声信号に対応する各仮想音源の位置が、検知されたユーザ31の動作に基づきそれぞれ検知される。
図8の(a)は、本発明の実施形態1における、ユーザ31が静止している時の複数の仮想音源81~83の位置の一例を示す図であり、図8の(b)は、本発明の実施形態1における、ユーザ31が移動している時の複数の仮想音源81~83の位置の一例を示す図である。以下では、ユーザ31の動作が変化した場合に、各仮想音源81~83の距離および角度の双方が変化する場合を説明する。
本実施形態では、ユーザ動作検知部121が、ユーザ31の動作として「静止」または「移動」を検知し、仮想音源位置決定部122は、検知されたこれらの「静止」または「移動」に応じた仮想音源の位置を決定する例を説明した。しかし、ユーザ動作検知部121および仮想音源位置決定部122の処理は、これに限定されない。たとえば、ユーザ動作検知部121は、ユーザ31の動作として、ユーザ31が一定の領域を注視していることを表す「注視」、または、ユーザ31が一定の領域を注視していないことを表す「非注視」を検知することもできる。この場合、仮想音源位置決定部122は、仮想音源の位置を、検知された「注視」または「非注視」に基づく位置に決定する。
本発明に係る実施形態2について、図10~図12に基づき以下に説明する。上述した実施形態1と共通する各部材には同じ符号を付し、詳細な説明を省略する。
まず、図10を参照して、本実施形態に係る音声信号再生装置1aの構成の概要について説明する。図10は、本発明の実施形態2に係る音声信号再生装置1aの要部構成を示すブロック図である。この図に示すように、音声信号再生装置1は、実施形態1に係る音声信号再生装置1に備えられる各部材に加えて、さらに、メタデータ取得部15(付加データ取得部)を備えている。これに伴い、本実施形態に係る仮想音源位置決定部122による処理の内容が、実施形態1に係る仮想音源位置決定部122による処理の内容と異なっている。音声信号取得部11等のその他の部材による処理の内容は実施形態1におけるそれと同一であるため、詳細な説明を省略する。
本実施形態に係る音声信号再生装置1aによって実行される処理の詳細について、図10の各矢印によって示されるデータの流れに基づき、以下に説明する。
図11の(a)~(c)は、本発明の実施形態2におけるメタデータの各例を示す図である。図11の(a)に示すメタデータは、3つの仮想音源のそれぞれについて、その位置の変更を許可するか否かを規定している。図11の(b)に示すメタデータは、3つの仮想音源のそれぞれについて、検知されたユーザ31の各動作(静止または移動)に応じた、仮想音源からのユーザ31への音声の出力の重要度を規定している。図11の(c)に示すメタデータは、3つの仮想音源のそれぞれについて、検知されたユーザ31の各動作(静止または移動)に応じた仮想音源の位置を、2次元極座標上における絶対的な値として規定している。
図12の(a)は、本発明の実施形態2における、ユーザ31が静止している時の複数の仮想音源121~123の位置の一例を示す図であり、図12の(b)は、本発明の実施形態2における、ユーザ31が移動している時の複数の仮想音源121~123の位置の一例を示す図である。この図の例では、音声信号取得部11は、音声信号再生装置1aの外部から3つの異なる音声信号を取得する。また、メタデータ取得部15は、図11の(a)に示すメタデータを取得する。
本発明に係る実施形態3について、図13に基づき以下に説明する。上述した実施形態1または2と共通する各部材には同じ符号を付し、詳細な説明を省略する。
まず、図13を参照して、本実施形態に係る音声信号再生装置1bの構成の概要について説明する。図13は、本発明の実施形態3に係る音声信号再生装置1bの要部構成を示すブロック図である。この図に示すように、音声信号再生装置1bが備える各部材は、実施形態1に係る音声信号再生装置1に備えられる各部材と同一である。ただし、本実施形態では、音声信号取得部11およびユーザ動作検知部121による処理の内容が、実施形態1に係る仮想音源位置決定部122による処理の内容と異なっている。仮想音源位置決定部122等のその他の部材による処理の内容は実施形態1におけるそれと同一であるため、詳細な説明を省略する。
本実施形態に係る音声信号再生装置1bによって実行される処理の詳細について、図13の各矢印によって示されるデータの流れに基づき、以下に説明する。
本発明の態様1に係る音声信号再生装置は、
音声を示す音声信号を取得する取得部(音声信号取得部11)と、
ユーザの動作を検知する検知部(ユーザ動作検知部121)と、
検知された上記動作に基づき、取得された上記音声信号によって表される上記音声が出力される仮想音源の位置を決定する決定部(仮想音源位置決定部122)と、
決定された上記仮想音源の位置に応じた立体音響処理を、取得された上記音声信号に施す処理部(音声信号処理部123)と、
上記立体音響処理が施された上記音声信号を再生する再生部(音声信号再生部13)とを備えていることを特徴としている。
上記決定部は、上記ユーザを基準にした所定の原点から上記仮想音源までの距離および上記原点に対する上記仮想音源の角度のうち少なくともいずれかを、検知された上記ユーザの動作に基づき異ならせるように、上記仮想音源の位置を決定することを特徴としている。
上記検知部は、上記ユーザの動作として、上記ユーザが移動していることを示す移動または上記ユーザが静止していることを示す静止を検知し、
上記決定部は、上記仮想音源の位置を、検知された上記移動または上記静止に基づき異ならせるように、上記仮想音源の位置を決定することを特徴としている。
上記決定部は、上記移動が検知された場合、上記仮想音源の位置を上記ユーザの後方に決定し、上記静止が検知された場合、上記仮想音源の位置を上記ユーザの前方に決定することを特徴としている。
上記決定部は、上記移動が検知された場合の上記仮想音源の上記距離を、上記静止が検知された場合の上記仮想音源の上記距離よりも長くするように、上記仮想音源の位置を決定することを特徴としている。
上記検知部は、上記動作として、上記ユーザが一定の領域を注視していることを表す注視または上記ユーザが上記一定の領域を注視していないことを表す非注視を検知し、
上記決定部は、上記一定の領域内に位置する上記仮想音源の位置を、検知された上記注視または上記非注視に基づき異ならせるように、上記仮想音源の位置を決定することを特徴としている。
上記決定部は、上記注視が検知された場合の上記一定の領域内に位置する上記仮想音源の上記距離を、上記非注視が検知された場合の上記一定の領域内に位置する上記仮想音源の上記距離よりも短くするように、上記仮想音源の位置を決定することを特徴としている。
上記音声信号に付加される所定の付加データを取得する付加データ取得部をさらに備えており、
上記決定部は、検知された上記動作と取得された上記付加データとの双方に基づき、上記仮想音源の位置を決定することを特徴としている。
上記取得部は、検知された上記ユーザの動作に応じた上記音声信号を取得することを特徴としている。
音声を示す音声信号を取得する取得工程と、
ユーザの動作を検知する検知工程と、
検知された上記ユーザの動作に基づき、取得された上記音声信号によって表される上記音声が出力される仮想音源の位置を決定する決定工程と、
決定された上記仮想音源の位置に応じた立体音響処理を、取得された上記音声信号に施す処理工程と、
上記立体音響処理が施された上記音声信号を再生する再生工程とを有することを特徴としている。
図1に示す音声信号再生装置1の各機能ブロックは、集積回路(ICチップ)等に形成された論理回路(ハードウェア)によって実現してもよいし、CPU(Central Processing Unit)を用いてソフトウェアによって実現してもよい。
2 音声信号再生システム
11 音声信号取得部(取得部)
12 制御部
14 記憶部
15 メタデータ取得部(付加データ取得部)
121 ユーザ動作検知部(検知部)
122 仮想音源位置決定部(決定部)
123 音声信号処理部(処理部)
22 DAC
23 増幅装置
24 イヤホン
Claims (11)
- 音声を示す音声信号を取得する取得部と、
ユーザの動作を検知する検知部と、
検知された上記動作に基づき、取得された上記音声信号によって表される上記音声が出力される仮想音源の位置を決定する決定部と、
決定された上記仮想音源の位置に応じた立体音響処理を、取得された上記音声信号に施す処理部と、
上記立体音響処理が施された上記音声信号を再生する再生部とを備えていることを特徴とする音声信号再生装置。 - 上記決定部は、上記ユーザを基準にした所定の原点から上記仮想音源までの距離および上記原点に対する上記仮想音源の角度のうち少なくともいずれかを、検知された上記ユーザの動作に基づき異ならせるように、上記仮想音源の位置を決定することを特徴とする請求項1に記載の音声信号再生装置。
- 上記検知部は、上記ユーザの動作として、上記ユーザが移動していることを示す移動または上記ユーザが静止していることを示す静止を検知し、
上記決定部は、上記仮想音源の位置を、検知された上記移動または上記静止に基づき異ならせるように、上記仮想音源の位置を決定することを特徴とする請求項2に記載の音声信号再生装置。 - 上記決定部は、上記移動が検知された場合、上記仮想音源の位置を上記ユーザの後方に決定し、上記静止が検知された場合、上記仮想音源の位置を上記ユーザの前方に決定することを特徴とする請求項3に記載の音声信号再生装置。
- 上記決定部は、上記移動が検知された場合の上記仮想音源の上記距離を、上記静止が検知された場合の上記仮想音源の上記距離よりも長くするように、上記仮想音源の位置を決定することを特徴とする請求項3または4に記載の音声信号再生装置。
- 上記検知部は、上記動作として、上記ユーザが一定の領域を注視していることを表す注視または上記ユーザが上記一定の領域を注視していないことを表す非注視を検知し、
上記決定部は、上記一定の領域内に位置する上記仮想音源の位置を、検知された上記注視または上記非注視に基づき異ならせるように、上記仮想音源の位置を決定することを特徴とする請求項2に記載の音声信号再生装置。 - 上記決定部は、上記注視が検知された場合の上記一定の領域内に位置する上記仮想音源の上記距離を、上記非注視が検知された場合の上記一定の領域内に位置する上記仮想音源の上記距離よりも短くするように、上記仮想音源の位置を決定することを特徴とする請求項6に記載の音声信号再生装置。
- 上記決定部は、上記注視が検知された場合の上記一定の領域外に位置する上記仮想音源の上記距離を、上記非注視が検知された場合の上記一定の領域外に位置する上記仮想音源の上記距離よりも長くするように、上記仮想音源の位置を決定することを特徴とする請求項6または7に記載の音声信号再生装置。
- 上記決定部は、上記注視が検知された場合、上記一定の領域外に位置する上記仮想音源の上記位置を上記一定の領域とは異なる他の領域に移動させるように、上記仮想音源の位置を決定することを特徴とする請求項6~7のいずれか1項に記載の音声信号再生装置。
- 上記音声信号に付加される所定の付加データを取得する付加データ取得部をさらに備えており、
上記決定部は、検知された上記動作と取得された上記付加データとの双方に基づき、上記仮想音源の位置を決定することを特徴とする請求項1~9のいずれか1項に記載の音声信号再生装置。 - 音声を示す音声信号を取得する取得工程と、
ユーザの動作を検知する検知工程と、
検知された上記ユーザの動作に基づき、取得された上記音声信号によって表される上記音声が出力される仮想音源の位置を決定する決定工程と、
決定された上記仮想音源の位置に応じた立体音響処理を、取得された上記音声信号に施す処理工程と、
上記立体音響処理が施された上記音声信号を再生する再生工程とを有することを特徴とする音声信号再生方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2017503403A JP6522105B2 (ja) | 2015-03-04 | 2016-02-16 | 音声信号再生装置、音声信号再生方法、プログラム、および記録媒体 |
US15/555,068 US10085107B2 (en) | 2015-03-04 | 2016-02-16 | Sound signal reproduction device, sound signal reproduction method, program, and recording medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2015043003 | 2015-03-04 | ||
JP2015-043003 | 2015-03-04 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2016140058A1 true WO2016140058A1 (ja) | 2016-09-09 |
Family
ID=56849296
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2016/054480 WO2016140058A1 (ja) | 2015-03-04 | 2016-02-16 | 音声信号再生装置、音声信号再生方法、プログラム、および記録媒体 |
Country Status (3)
Country | Link |
---|---|
US (1) | US10085107B2 (ja) |
JP (1) | JP6522105B2 (ja) |
WO (1) | WO2016140058A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3747205A1 (en) * | 2018-02-01 | 2020-12-09 | Qualcomm Incorporated | Scalable unified audio renderer |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10990169B2 (en) * | 2016-06-28 | 2021-04-27 | Rec Room Inc. | Systems and methods for assisting virtual gestures based on viewing frustum |
CN109983786B (zh) * | 2016-11-25 | 2022-03-01 | 索尼公司 | 再现方法、装置及介质、信息处理方法及装置 |
US11096004B2 (en) * | 2017-01-23 | 2021-08-17 | Nokia Technologies Oy | Spatial audio rendering point extension |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2010122369A (ja) * | 2008-11-18 | 2010-06-03 | Advanced Telecommunication Research Institute International | 音声再生装置 |
JP2011521511A (ja) * | 2008-04-18 | 2011-07-21 | ソニー エリクソン モバイル コミュニケーションズ, エービー | 拡張現実を強化した音声 |
WO2012120810A1 (ja) * | 2011-03-08 | 2012-09-13 | パナソニック株式会社 | 音声制御装置および音声制御方法 |
Family Cites Families (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1995022235A1 (fr) * | 1994-02-14 | 1995-08-17 | Sony Corporation | Systeme de reproduction visuelle et sonore |
JP2002005675A (ja) * | 2000-06-16 | 2002-01-09 | Matsushita Electric Ind Co Ltd | 音響ナビゲーション装置 |
US7415123B2 (en) * | 2001-09-26 | 2008-08-19 | The United States Of America As Represented By The Secretary Of The Navy | Method and apparatus for producing spatialized audio signals |
KR101118214B1 (ko) * | 2004-09-21 | 2012-03-16 | 삼성전자주식회사 | 청취 위치를 고려한 2채널 가상 음향 재생 방법 및 장치 |
US8023659B2 (en) * | 2005-06-21 | 2011-09-20 | Japan Science And Technology Agency | Mixing system, method and program |
JP4411547B2 (ja) | 2006-03-20 | 2010-02-10 | ソニー株式会社 | 画像表示装置 |
JP2008151766A (ja) | 2006-11-22 | 2008-07-03 | Matsushita Electric Ind Co Ltd | 立体音響制御装置及び立体音響制御方法 |
US20080118074A1 (en) | 2006-11-22 | 2008-05-22 | Shuichi Takada | Stereophonic sound control apparatus and stereophonic sound control method |
JP5114981B2 (ja) * | 2007-03-15 | 2013-01-09 | 沖電気工業株式会社 | 音像定位処理装置、方法及びプログラム |
US8620009B2 (en) * | 2008-06-17 | 2013-12-31 | Microsoft Corporation | Virtual sound source positioning |
KR101673232B1 (ko) * | 2010-03-11 | 2016-11-07 | 삼성전자주식회사 | 수직 방향 가상 채널을 생성하는 장치 및 그 방법 |
US9491560B2 (en) * | 2010-07-20 | 2016-11-08 | Analog Devices, Inc. | System and method for improving headphone spatial impression |
US8767968B2 (en) * | 2010-10-13 | 2014-07-01 | Microsoft Corporation | System and method for high-precision 3-dimensional audio for augmented reality |
JP5716451B2 (ja) * | 2011-02-25 | 2015-05-13 | ソニー株式会社 | ヘッドホン装置およびヘッドホン装置の音声再生方法 |
NL2006997C2 (en) * | 2011-06-24 | 2013-01-02 | Bright Minds Holding B V | Method and device for processing sound data. |
EP2804402B1 (en) * | 2012-01-11 | 2021-05-19 | Sony Corporation | Sound field control device, sound field control method and program |
JP6065370B2 (ja) * | 2012-02-03 | 2017-01-25 | ソニー株式会社 | 情報処理装置、情報処理方法、及びプログラム |
GB201211512D0 (en) * | 2012-06-28 | 2012-08-08 | Provost Fellows Foundation Scholars And The Other Members Of Board Of The | Method and apparatus for generating an audio output comprising spartial information |
JP5949311B2 (ja) * | 2012-08-15 | 2016-07-06 | 富士通株式会社 | 推定プログラム、推定装置、及び推定方法 |
JP5983313B2 (ja) * | 2012-10-30 | 2016-08-31 | 富士通株式会社 | 情報処理装置、音像定位強調方法、及び音像定位強調プログラム |
JP6275557B2 (ja) * | 2014-06-10 | 2018-02-07 | 株式会社Nttドコモ | 音声出力装置 |
EP3103269B1 (en) * | 2014-11-13 | 2018-08-29 | Huawei Technologies Co., Ltd. | Audio signal processing device and method for reproducing a binaural signal |
-
2016
- 2016-02-16 WO PCT/JP2016/054480 patent/WO2016140058A1/ja active Application Filing
- 2016-02-16 US US15/555,068 patent/US10085107B2/en not_active Expired - Fee Related
- 2016-02-16 JP JP2017503403A patent/JP6522105B2/ja not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011521511A (ja) * | 2008-04-18 | 2011-07-21 | ソニー エリクソン モバイル コミュニケーションズ, エービー | 拡張現実を強化した音声 |
JP2010122369A (ja) * | 2008-11-18 | 2010-06-03 | Advanced Telecommunication Research Institute International | 音声再生装置 |
WO2012120810A1 (ja) * | 2011-03-08 | 2012-09-13 | パナソニック株式会社 | 音声制御装置および音声制御方法 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3747205A1 (en) * | 2018-02-01 | 2020-12-09 | Qualcomm Incorporated | Scalable unified audio renderer |
Also Published As
Publication number | Publication date |
---|---|
JP6522105B2 (ja) | 2019-05-29 |
US20180048978A1 (en) | 2018-02-15 |
JPWO2016140058A1 (ja) | 2018-01-11 |
US10085107B2 (en) | 2018-09-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
KR101844388B1 (ko) | 개인용 오디오의 전달을 위한 시스템들 및 방법들 | |
JP4612728B2 (ja) | 音声出力装置、及び音声処理システム | |
US20140079238A1 (en) | Automated left-right headphone earpiece identifier | |
US11006202B2 (en) | Automatic user interface switching | |
EP3001422A1 (en) | Media player automated control based on detected physiological parameters of a user | |
EP2827614A1 (en) | Audio playing method and device | |
WO2015163031A1 (ja) | 情報処理装置、情報処理方法及びプログラム | |
WO2016140058A1 (ja) | 音声信号再生装置、音声信号再生方法、プログラム、および記録媒体 | |
US10354651B1 (en) | Head-mounted device control based on wearer information and user inputs | |
US11683654B2 (en) | Audio content format selection | |
US20160360316A1 (en) | Electronic device and vibration information generation device | |
KR20170026786A (ko) | 능동 소음 제어 기능을 가지는 스마트 보청기 시스템 및 그 보청기의 제어 장치 | |
WO2014057921A1 (ja) | 電子機器、及び音響再生方法 | |
US20200301653A1 (en) | System and method for processing audio between multiple audio spaces | |
TW201928945A (zh) | 音訊場景處理技術 | |
EP1771039A2 (en) | Audio control system | |
US11877133B2 (en) | Audio output using multiple different transducers | |
EP2887698B1 (en) | Hearing aid for playing audible advertisement | |
TW201345277A (zh) | 音頻播放裝置及其控制方法 | |
CN112385143B (zh) | 取决于背景水平的动态音量水平 | |
WO2016009850A1 (ja) | 音声信号再生装置、音声信号再生方法、プログラム、および、記録媒体 | |
CN113196800A (zh) | 无线头戴式耳机的混合麦克风 | |
JP4922462B1 (ja) | コンテンツ再生装置およびコンテンツ再生方法 | |
US20240323627A1 (en) | System for outputting audio for a user, and a method thereof | |
WO2024090309A1 (ja) | 音出力装置、音出力方法及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 16758752 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 15555068 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2017503403 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 16758752 Country of ref document: EP Kind code of ref document: A1 |