EP1816895B1 - Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs - Google Patents

Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs Download PDF

Info

Publication number
EP1816895B1
EP1816895B1 EP07010496A EP07010496A EP1816895B1 EP 1816895 B1 EP1816895 B1 EP 1816895B1 EP 07010496 A EP07010496 A EP 07010496A EP 07010496 A EP07010496 A EP 07010496A EP 1816895 B1 EP1816895 B1 EP 1816895B1
Authority
EP
European Patent Office
Prior art keywords
acoustic
filter
position information
acoustic characteristics
section
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP07010496A
Other languages
German (de)
English (en)
Other versions
EP1816895A3 (fr
EP1816895A2 (fr
Inventor
Naoshi Matsuo
Kaori Suzuki
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujitsu Ltd
Original Assignee
Fujitsu Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP23170595A external-priority patent/JP3810110B2/ja
Priority claimed from JP04610596A external-priority patent/JP4306815B2/ja
Application filed by Fujitsu Ltd filed Critical Fujitsu Ltd
Publication of EP1816895A2 publication Critical patent/EP1816895A2/fr
Publication of EP1816895A3 publication Critical patent/EP1816895A3/fr
Application granted granted Critical
Publication of EP1816895B1 publication Critical patent/EP1816895B1/fr
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S1/00Two-channel systems
    • H04S1/002Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
    • H04S1/005For headphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S1/00Two-channel systems
    • H04S1/002Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S1/00Two-channel systems
    • H04S1/007Two-channel systems in which the audio signals are in digital form
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/01Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation

Definitions

  • the present invention relates to a three-dimensional acoustic apparatus for positioning a sound image using a virtual sound source.
  • the acoustic output device such as a speaker or a headphone
  • the former acoustic characteristics are added to the sound source and the latter characteristics are removed from the sound source, so that even using a speaker or a headphone it is possible to reproduce to the listener the sound image of the original sound image of the original sound field, or so that it is possible to accurately localize the position of the original sound image.
  • a FIR (finite impulse response, non-recursive) filter having coefficients that are the impulse responses of each of the acoustic spatial paths was used as a filter to emulate the transfer characteristics of the acoustic spatial path and the reverse of the acoustic characteristics of the reproducing sound field up to the listener.
  • the number of taps of the FIR which represent those characteristics when using an audio-signal sampling frequency of 44.1 kHz is several thousand or even greater. Even in the case of the inverse of the transfer characteristics of a headphone, the number of taps required is several hundred or even greater.
  • DE 32 38 933 A1 discloses a method for creating sound signals with respect to video games, which are displayed using the combination of a video gaming device with a screen device, which is typically a TV set.
  • the aim of the invention is to improve the previously simple acoustic accompaniment and to produce and replay the acoustic information in a way that the user of a video gaming device is attracted not only optically but also acoustically by the programmed action of the video game.
  • the acoustic information which accompany the respective game are available in the video gaming device and stored with the own frequency access and phase relationships of the artificial stereophonic head. Play back thereof is carried out via at least one set of stereo headphones.
  • US 3,920,904 discloses that sound signals are created at the eardrums of a listener to correspond to sound signals which would be created at the eardrums of such listener in a predetermined acoustical environment in response to first electrical signals applied to a loudspeaker having known sound-reproducing characteristics.
  • the sound waves at the eardrums of the listener are measured as a function of frequency to determine the pressure signals at the eardrums of the listener when the listener is listening to acoustical information transmitted by the loudspeaker.
  • An electro acoustical transfer function relating the electrical input signal of the loudspeaker to the pressure actually impinging upon the eardrums is determined.
  • a compensating network is connected to the earphones.
  • the compensating network has a transfer function corresponding to the quotient of the first transfer function divided by the second transfer function.
  • EP-A-0 438 281 discloses an acoustic signal reproducing apparatus for reproducing acoustic signals by headphone devices.
  • the left channel and right channel acoustic signals are provided by a device for processing the acoustic signals with constant transmission characteristics from an imaginary sound source to both of the listener's ears.
  • the left channel and right channel acoustic signals, processed in this manner by the device are provided by an acoustic signal processing device with a level difference and time difference consistent with changes in orientation of the listener's head.
  • a three-dimensional acoustic apparatus for positioning a sound image using a virtual sound source as set out in Claim 1.
  • a three-dimensional acoustic processor which localizes a sound image using a virtual sound source has a first acoustic characteristics adding filter which is formed by a linear synthesis filter which has filter coefficients that are the linear predictive coefficients obtained by linear predictive analysis of the impulse response which represents each of the acoustic characteristics of one or each of a plurality of propagation paths to the left ear to be added to the original signal, a first acoustic characteristics elimination filter which is connected in series with the above-noted first acoustic characteristics adding filter, and which is formed by a linear predictive filter having filter coefficients which represent the inverse of acoustic characteristics for the purpose of eliminating the acoustic characteristics of an acoustic output device to the left ear, these filter coefficients being obtained by a linear predictive analysis of the impulse response representing the acoustic characteristics of the above-noted acoustic output device, a second acoustic characteristics adding filter which is
  • the above-noted first and second acoustic characteristics adding filters are configured from a common section which adds characteristics which are common to each of the acoustic characteristics of the acoustic path, and an individual characteristic section which adds characteristics individual to each of the acoustic characteristics of each acoustic path.
  • a storage medium into which is stored the calculation results for the above-noted common section of the desired sound source, and a readout/indication section which reads out the above-noted stored calculation results, the readout/indication section directly to the above-noted individual characteristic section the read out calculation results, by means of the readout it performs.
  • the storage medium can also store the calculation results of the corresponding first or second acoustic characteristics elimination filter.
  • the above-noted first acoustic characteristics adding filter and second acoustic characteristics adding filter further have a delay section which imparts a delay time between the two ears, so that by making the delay time of the delay section of either the first or the second acoustic characteristics adding filter the reference (zero delay time), it is possible to eliminate the delay section which has this delay of zero.
  • the above-noted first acoustic characteristics adding filter and second acoustic characteristics adding filter each further have an amplification section which enables variable setting of the output signal level thereof, the above-noted selection setting section relatively varying the output signal levels of the first and the second acoustic characteristics adding filters by setting the gain of these amplification sections in response to position information of the sound image, thereby enabling movement of the localized position of the sound image.
  • the above-noted first and second acoustic characteristics adding filters can be left-to-right symmetrical about the center of the front of the listener, in which case, the parameters for the above-noted delay sections and amplification sections are shared in common between positions which correspond in this left-to-right symmetry.
  • the above-noted three-dimensional acoustic processor has a position information interpolation section which interpolates intermediate position information from past and future sound image position information, interpolated position information from this position information interpolation section being given to the selection setting section as position information.
  • a position information prediction section which performs predictive interpolation of future position information from past and current sound image position information, the future position information from this position information prediction section being given to the selection setting section as position information.
  • the above-noted position information prediction section further includes a regularity judgment section which performs a judgment with regard to the existence of regularity with regard to the movement direction, based on past and current sound image position information, and in the case in which the regularity judgment section judges that regularity exists, the above-noted position information prediction section provides the above-noted future position information. It is possible to use the visual image position information from image display information for a visual image which generates a sound image in place of the above-noted sound image position information. So that the above-noted selection setting section can further provide and maintain a good audible environment for the listener, it can move the above-noted environment in response to position information given with regard to the listener.
  • filters are configured by dividing the acoustic characteristics to be added to the input signal into characteristics which are common to each position at which the sound image is to be localized and individual characteristics. In the case of adding acoustic characteristics, these filters are connected in series. By doing this, it is possible to reduce the overall amount of calculations performed. In this case, the larger the number of individual characteristics, the larger will be the effect of the above-noted reduction in the amount of calculations.
  • a sound image it is possible to move a sound image continuously by moving the sound image in accordance with the interpolated positions of a visual image which is moving discontinuously. Also, by inputting the user's auditory and visual environment into an image controller and a sound image controller it is possible to achieve apparent agreement between the movement of the visual image and the movement of the sound image, by using this information to control the movement of the visual image and sound image.
  • Fig. 1 shows the case of listening to a sound image from a two-channel stereo apparatus in the past.
  • Fig. 2 shows the basic block diagram circuit configuration which achieves an acoustic space that is equivalent to that created by the headphone in Fig. 1 .
  • the transfer characteristics for each of the acoustic space paths from the left and right speakers (L, R) 1 and 2 to the left and right ears (1, r) of the listener 3 are expressed as Ll, Lr, Rr, and Rl.
  • the inverse characteristic (Hl -1 and Hr -1 ) 15 and 16 of each of the characteristics from the left and right earphones of headphone (HL and HR) 5 and 6 to the left and right ears are added.
  • Fig. 3 shows an example of configuration of a circuit of an FIR filter (non-recursive filter) of the past for the purpose of achieving the above-noted transfer characteristics.
  • an FIR filter non-recursive filter having coefficients that represent the impulse response of each of the acoustic space paths, this being expressed by Equation (1).
  • the filter coefficients obtained from the impulse response obtained from, for example, an acoustic measurement or an acoustic simulation for each path are used as the filter coefficients (a0, a1, a2, ..., an) which represent the transfer characteristics 11 to 14 of each of the acoustic space paths.
  • the impulse response which represents the characteristics of each of the paths are convoluted via these filters.
  • the filter coefficients (a0, a1, a2, ..., an) of the inverse characteristics (Hl -1 and Hr -1 ) 15 and 16 of the headphone, shown in Fig. 2 are determined in the frequency domain. First, the frequency characteristics of the headphone are measured and the inverse characteristics thereof determined, after which these results are restored to the time domain to obtain the impulse response which is used as the filter coefficients.
  • Fig. 4 shows an example of the basic system configuration for the case of moving a sound image to match a visual image on a computer graphics (CG) display.
  • CG computer graphics
  • the controller 26 of the CG display apparatus 24 drives a CG accelerator 25, which performs image display, and also provides to a controller 29 of the three-dimensional acoustic apparatus 27 position information of the sound image which is synchronized with the image.
  • an acoustic characteristics adder 28 controls the audio output signal level from each of the channel speakers 22 and 23 (or headphone) by means of control from the controller 29, so that the sound image is localized at a visual image position within the display screen of the display 21 or so that it is localized at a virtual position outside the display screen of the display 21.
  • Fig. 5 shows the basic configuration of the acoustic characteristics adder 28 which is shown in Fig. 4 .
  • the acoustic characteristics adder 28 comprises acoustic characteristics adding filters 35 and 37 which use the FIR filter of Fig. 3 and which give the transfer characteristics Sl and Sr of each of the acoustic space path from the sound source to the ears, acoustic characteristics elimination filters 36 and 38 for headphone channels L and R, and a filter coefficients selection section 39, which selectively gives the filter coefficients of each of the acoustic characteristics adding filters 35 and 37, based on the above-noted position information.
  • Figs. 6 through 8B illustrate the sound image localization technology of the past, which used the acoustic characteristics adder 28.
  • Fig. 6 shows the general relationship between a sound source and a listener.
  • the transfer characteristics Sl and Sr between the sound source 30 and the listener 31 are similar to those described above in relation to Fig. 1 .
  • Fig. 7A shows an example of acoustic characteristics adding filters (S ⁇ l) 35 and (S ⁇ r) 37 between the sound source (S) 30 and the listener 31 and the inverse transfer characteristics (h -1 ) 36 and 38 of the earphones of headphone 33 and 34 for the case of localizing one sound source.
  • Fig. 7B shows the configuration of the acoustic characteristics adding filters 35 and 37 for the case in which the sound source 30 is further localized at a plurality of sound image positions P through Q.
  • Fig. 8A and Fig. 8B show a specific circuit block diagram of the acoustic characteristics adding filters 35 and 37 of Fig. 7B .
  • Fig. 8B is the same as Fig. 8A .
  • the gains of each of the acoustic characteristics adding filters 35 and 37 are controlled in response to the position information provided by one for one of the sound image positions P through Q, thereby localizing the sound image 30 at one of the sound image positions P through Q.
  • Fig. 9A and Fig. 9B show an example of moving a sound image by means of output interpolation between a plurality of virtual sound sources.
  • Fig. 9A shows an example of a circuit configuration for the purpose of localization a sound image among three virtual sound sources (A through C) 30-1 through 30-3.
  • Fig. 9B three types of acoustic characteristics adding filters, 35-1 and 37-1, 35-2 and 37-2, and 35-3 and 37-3 are provided in accordance with the transfer characteristics of each of the acoustic space paths leading to the left and right ears of the listener 31, these corresponding to each of the virtual sound sources 30-1, 30-2, and 30-3.
  • Each of these acoustic characteristics adding filters have filter coefficients and a filter memory which holds past input signals, the above-noted filter calculation output results being input to the subsequent stages of variable amplifiers (gA through gC).
  • Fig. 10 shows an example of a surround-type sound image localization.
  • Fig. 10 the example shown is that of a surround system in which five speakers (L, C, R, SR, and SL) surround the listener 31.
  • the output levels from the five sound sources are controlled in relation to one another, enabling the localization of a sound image in the region surrounding the listener 31.
  • L and SL shown in Fig. 10 .
  • Fig. 11 shows the conceptual configuration for the purpose of determining a linear synthesis filter for the purpose of adding acoustic characteristics.
  • an anechoic chamber which is free of reflected sound and residual sound, is used to measure the impulse responses of each of the acoustic space paths which represent the above-noted acoustic characteristics, these being used as the basis for performing linear predictive analysis processing 41 to determine the linear predictive coefficients of the impulse responses.
  • the above-noted linear predictive coefficients are further subjected to compensation processing 42, the resulting coefficients being set as the filter coefficients of a linear synthesis filter 40 which is configured as an IIR filter.
  • an original signal which is passed through the above-noted linear synthesis filter 40 has added to it the frequency characteristics of the acoustic characteristics of the above-noted acoustic space path.
  • Fig. 12 shows an example of the configuration of a linear synthesis filter for the purpose of adding acoustic characteristics.
  • the linear synthesis filter 40 comprises a short-term synthesis filter 44 and a pitch synthesis filter 43, these being represented, respectively, by the following Equation (2) and Equation (3).
  • Y Z X Z 1 1 - b ⁇ 1 ⁇ Z - 1 + b ⁇ 2 ⁇ Z - 2 + ... + bm ⁇ Z - m
  • Y Z X Z 1 1 - bL ⁇ Z - L
  • the short-term synthesis filter 44 (Equation (2)) is configured as an IIR filter having linear predictive coefficients which are obtained from a linear predictive analysis of the impulse response which represents each of the transfer characteristics, this providing a sense of directivity to the listener.
  • the pitch synthesis filter 43 (Equation (3)) further provides the sound source with initial reflected sound and reverberation.
  • Fig. 13 shows the method of determining the linear predictive coefficients (b1, b2, ..., bm) of the short-term synthesis filter 44 and the pitch coefficients L and bL of the pitch synthesis filter 43.
  • an IIR filter By configuring an IIR filter using linear predictive coefficients, it is possible to add the frequency characteristics, which are transfer characteristics, using a number of filter taps which is much reduced from the number of samples of the impulse response. For example, in the case of 256 taps, it is possible to reduce the number of taps to approximately 10.
  • Equation (2) and Equation (4) by passing through the above-noted short-term predictive filter 47, it is possible to eliminate the frequency characteristics component that is equivalent to that added by the short-term synthesis filter 44. As a result, it is possible, by the pitch extraction processing 48 performed at the next stage, to determine the above-noted delay (Z -L ) and gain (bL) from the remaining time component.
  • Fig. 14 shows the block diagram configuration of the pitch synthesis filter 43, in which separate pitch synthesis filters are used for so-called direct sound and reflected sound.
  • the impulse response which is obtained by measuring a sound field generally starts with a part that has a large attenuation factor (direct sound), this being followed by a part that has a small attenuation factor (reflected sound).
  • the pitch synthesis filter 43 can be configured, as shown in Fig. 14 , by a pitch synthesis filter 49 related to the direct sound, a pitch synthesis filter 51 related to the reflected sound, and a delay section 50 which provides the delay time therebetween. It is also possible to configure the direct sound part using an FIR filter and to make the configuration so that there is overlap between the direct sound and reflected sound parts.
  • Fig. 15 shows an example of compensation processing on the linear predictive coefficients obtained as described above.
  • the evaluation processing 52 of time-domain envelope and spectrum of Fig. 15 a comparison is performed between the series linking of the first obtained short-term synthesis filter 44 and the pitch synthesis filter 43 and the impulse response having the desired acoustic characteristics, the filter coefficients being compensated based on this, so that the time-domain envelope and spectrum of the linear synthesis filter impulse response are the same as or close to the original impulse response.
  • Fig. 16 shows an example of the configuration of a filter which represents the inverse characteristics Hl -1 and Hr -1 of the transfer characteristics of the headphone.
  • the filter 53 in Fig. 16 has the same configuration as the short-term prediction filter 47 which is shown in Fig. 13 , this performing linear predictive analysis in determining the auto-correlation coefficients of the impulse response of the headphone, the thus-obtained linear predictive coefficients (cl, c2, ..., cm) being used to configure an FIR-type linear predictive filter.
  • the filter 53 in Fig. 16 has the same configuration as the short-term prediction filter 47 which is shown in Fig. 13 , this performing linear predictive analysis in determining the auto-correlation coefficients of the impulse response of the headphone, the thus-obtained linear predictive coefficients (cl, c2, ..., cm) being used to configure an FIR-type linear predictive filter.
  • Fig. 17 shows an example of the frequency characteristics of acoustic characteristics adding filter according to the present background example, in comparison with the prior art.
  • the solid line represents the frequency characteristics of a prior art acoustic characteristics adding filter made up of 256 taps as shown in Fig. 3
  • the broken line represents the frequency characteristics of an acoustic characteristics adding filter (using only a short-term synthesis filter) having 10 taps. It can be seen that it is possible to obtain a spectral approximation with a number of taps greatly reduced from the number in the past.
  • Figs. 18A through 18C show the conceptual configuration for determining the linear predictive coefficients in a further improvement of the above-noted background example.
  • Fig. 18A shows the most basic processing block diagram.
  • the impulse response is first input to a critical bandwidth pre-processor which considers the critical bandwidth according to the present background example.
  • the auto-correlation calculation section 45 and linear predictive analysis section 46 of this example are the same as, for example, that shown in Fig. 13 .
  • the "critical bandwidth" as defined by Fletcher is the bandwidth of a bandpass filter having a center frequency that varies continuously, such that when frequency analysis is performed using a bandpass filter having a center frequency closest to a signal sound, the influence of noise components in masking the signal sound is limited to frequency components within the passband of the filter.
  • the above-noted bandpass filter is also known as an “auditory” filter, and a variety of measurements have verified that, between the center frequency and the bandwidth, the critical bandwidth is narrow when the center frequency of the filter is low and wide when the center frequency is high. For example, at a center frequency of below 500 kHz, the critical bandwidth is virtually constant at 100 Hz.
  • Bark 13 arctan 0.76 ⁇ f + 3.5 arctan f / 5.5 2
  • Fig. 18B and Fig. 18C show examples of the internal block diagram configuration of the critical bandwidth pre-processor 110 of Fig. 18A .
  • the impulse response signal has a fast Fourier transform applied to it by the FFT processor 111, thereby converting it from the time domain to the frequency domain.
  • Fig. 19 shows an example of the power spectrum of an impulse response of an acoustic space path, as measured in an anechoic chamber, from a sound source localized at an angle of 45 degrees to the left-front of a listener to the left ear of the listener.
  • the above-noted band-limited signal is divided into a plurality of bands having a Bark scale value of 1.0, by the following stages, the critical bandwidth processing sections 112 and 114.
  • the power spectra within each critical bandwidth are summed, this summed value being used to represent the signal sound of the band-limited signal.
  • the average value of the power spectra is used to represent the signal sound of the band-limited signal.
  • Fig. 20 shows the example of dividing the power spectrum of Fig. 19 into critical bandwidths and determining the maximum value of the power spectrum of each band shown in Fig. 18C .
  • output interpolation processing is performed, which applies smoothing between the summed power spectrum values and maximum or averaged values determined for each of the above-noted critical bandwidths.
  • This interpolation is performed by means of either linear interpolation or a high-order Taylor series.
  • Fig. 21 shows an example of output interpolation of the power spectrum, whereby the power spectrum is smoothed.
  • a power spectrum which is smooth as described above is subjected to an inverse Fourier transform by the Inverse FFT processor 113, thereby restoring the frequency-domain signal to the time domain.
  • the phase spectrum used is the original impulse response phase spectrum without any change.
  • the above-noted reproduced impulse response signal is further processed as described previously.
  • the characteristic part of a signal sound is extracted using critical bandwidths, without causing a change in the auditory perception, these being smoothed by means of interpolation, after which the result is reproduced as an approximation of the impulse response.
  • Fig. 22 shows an example of the circuit configuration of a synthesis filter (IIR) 121 which uses the linear predictive coefficients (an, ..., a2, al) which are obtained from the processing shown in Fig. 18A .
  • Fig. 23 shows an example of a power spectrum determined from the impulse response after approximation using a 10th order synthesis filter which uses the linear predictive coefficients of Fig. 22 . From this, it can be seen that there is an improvement in the accuracy of approximation in the peak part of the power spectrum.
  • IIR synthesis filter
  • Fig. 24 shows an example of the processing configuration for compensation of the synthesis filter 121 which uses the linear predictive coefficients shown in Fig. 22 .
  • a compensation filter 122 is connected in series therewith to form the acoustic characteristics adding filter 120.
  • Fig. 25 and Fig. 26 show, respectively, examples of each of these filters.
  • Fig. 25 shows the example of a compensation filter (FIR) for the purpose of approximating the valley part of the frequency band
  • Fig. 26 shows the example of a delay/amplification circuit for the purpose of compensating for the difference in delay times and level between the two ears.
  • FIR compensation filter
  • an impulse response signal representing actual acoustic characteristics is applied to one input of the error calculator 130, the impulse signal being applied to the input of the above-noted acoustic characteristics adding filter 120. Because of the input of the above-noted impulse signal, the time-domain acoustic characteristics adding characteristic signal is output from the acoustic characteristics adding filter 120. This output signal is applied to the other input of the error calculator 130, and a comparison is made with this input and the above-noted impulse response signal which represents actual acoustic characteristics. The compensation filter 122 is then adjusted so as to minimize the error component.
  • An example of using an n-th order FIR filter 122 is shown in Fig.
  • Fig. 27 shows an example of using the above-noted compensation filter 122 to change the frequency characteristics of the synthesis filter 121 which uses the linear predictive coefficients.
  • the broken line in Fig. 27 represents an example of the frequency characteristics of the synthesis filter 121 before compensation, and the solid line in Fig. 27 represents an example of changing these frequency characteristics by using the compensation filter 122. It can be seen from this example that the compensation has the effect of making the valley parts of the frequency characteristics prominent.
  • Fig. 28 shows an example of the application of the above-described present background example.
  • the acoustic characteristics adding filters 35 and 37 and the inverse characteristics filters 36 and 38 for the headphone were each determined separately and then connected in series.
  • the previous stage filter 35 (or 37) has 128 taps and the following stage filter 36 (or 38) has 128 taps, to guarantee signal convergence when these are connected in series, approximately double this number, 255 taps, were required.
  • a single filter 141 (or 142) is used, this being the combination of the acoustic characteristics adding filter and the headphone inverse characteristics filter.
  • preprocessing which considers the critical bandwidth is performed before performing linear predictive analysis of the acoustic characteristics.
  • extraction of characteristics of the signal sound are extracted and interpolation processing is performed, so that there is no auditorilly perceived change.
  • the filter circuit can be simplified in comparison to the prior art approach, in which two series connected stages were used.
  • Fig. 29 shows an example of the inverse characteristics (h -1 ) of the power spectrum of a headphone.
  • Fig. 30 shows an example of the power spectrum of a combined filter comprising actual acoustic characteristics and the headphone inverse characteristics (S ⁇ l * h -1 ).
  • Fig. 31 shows the results of using the maximum value of each band to represent each band when division is done of the power spectrum of Fig. 30 into critical bandwidths.
  • Fig. 32 shows an example of the base of performing interpolation processing on the representative values of the power spectrum shown in Fig. 31 . It can be seen from a comparison of the power spectra of Fig. 30 and Fig. 32 that the latter is a more accurate approximation using linear predictive analysis with a lower order.
  • Fig. 33 shows the basic block diagram configuration for the purpose of localizing a sound image using an acoustic filter that employs linear predictive analysis according to a background example.
  • Fig. 33 corresponds to the acoustic characteristics adder 28 of Fig. 4 and Fig. 5 , the acoustic characteristics adding filters 35 and 37 thereof comprising the IIR filters 54 and 55, respectively, which add frequency characteristics using linear predictive coefficients, the delay sections 56 and 57, which serve as the input stages for the filters 35 and 37, respectively, and which provide, for example, pitch and time difference to reach the left and right ears, and amplifiers 58 and 59 which control the individual gains and serve as the output stages for the filters 35 and 37, respectively.
  • the filters 36 and 38, which eliminate the acoustic characteristics of the headphone on the left and right channels are FIR filters using linear predictive coefficients.
  • the IIR filters 54 and 55 are the short-term synthesis filter 44 which was described in relationship to Fig. 12
  • the delay sections 55 and 56 are the delay circuit (Z -d ) of Fig. 12
  • the filters 36 and 38 which eliminate the acoustic characteristics of the headphone are the FIR-type linear predictive filters 53 of Fig. 16 . Therefore, the above-noted filters will not be explained again at this point.
  • the filter coefficient selection means 39 performs selective setting of the filter coefficients, pitch/delay time, and gain as parameters of the above-noted filters.
  • Fig. 34 shows an example of an implementation of sound image localization as illustrated in Fig. 10 , using the acoustic characteristics adder 28.
  • Five virtual sound sources made of 10 filters (C1 to SR1 and Cr to SRr) 54 to 57, corresponding to the five speakers shown in Fig. 10 (L, C, R, SR, and SL) are in the same kind of placement, and the acoustic characteristics of the earphones of headphone 33 and 34 are eliminated by the acoustic characteristics eliminating filters 36 and 38. Because, as seen from the listener, this environment is the same as in Fig. 10 , as described with regard to Fig.
  • Fig. 35 shows an example of the configuration of an acoustic characteristics adder according to a background example, this having the same type of configuration as described above with regard to Fig. 33 , except for the addition of a position information interpolation/prediction section 60 and a regularity judgment section 61.
  • Figs. 36 through 40 illustrate the functioning of the position information interpolation/prediction section 60 and the regularity judgment section 61 shown in Fig. 35 .
  • Figs. 36 through 38 are related to the interpolation of position information.
  • the future position information is pre-read to the sound image controller 63 (corresponding to the three-dimensional acoustic apparatus 27 in Fig. 4 ) from the visual image controller 62 (corresponding to the CG display apparatus 24 in Fig. 4 ) before performing visual image processing, which requires a long processing time.
  • the above-noted position information interpolation/prediction section 60 which is included in the sound image controller 63 of Fig. 36 , performs interpolation of the sound image position information on the display 21 (refer to Fig. 4 ) using the future, current, and past positions of the visual image.
  • the method of performing x-axis value interpolation for a system of (x, y, z) orthogonal axes for the visual image is as follows. It is also possible to perform interpolation in the same way for y-axis and z-axis values.
  • t0 is the current time
  • t-1, ..., t-m are past times
  • t+1 is a future time.
  • the predictive coefficients b1, ..., bn in the above equation are determined by performing linear predictive analysis by means of an auto-correlation of the current and past values x(t), ..., x(t-1). It is also possible to determine this by trial-and-error, by using a method such as the steepest descent method.
  • Fig. 39 and Fig. 40 show a method of predicting a future position by making a judgment as to whether or not regularity exists in the movement of a visual image.
  • Equation (5.4) when the above-noted Equation (5.4) is used to determine the predictive coefficients b1, ..., bn using linear predictive analysis, the regularity judgment section 64 of Fig. 39 which corresponds to the regularity judgment section 61 of Fig. 35 judges that regularity in the movement of the visual image if a set of stable predictive coefficients is obtained.
  • the movement of the visual image when using a prescribed adaptive algorithm to determine the predictive coefficients b1, ..., bn, by trial-and-error, the movement of the visual image is judged have regularity if the coefficients converge to within a certain value. Only when such a judgment result occurs are the coefficients determined from Equation (5.4) used as the future position.
  • Fig. 41 and Fig. 42 show examples of optimal localization of a sound image in accordance with listener position information.
  • Fig. 41 show an example in which in the system of Fig. 4 , the listener 31 moves away from the proper listening/viewing environment, which is marked by hatching lines, so that as seen from the listener 31 the sound image position and visual image position do not match.
  • the method described above can be applied as is. That is, the right and left channel signals are controlled so as to move the range of the listening/viewing environment toward the user.
  • Fig. 43A and Fig. 43B show an embodiment of improved efficiency calculation according to the present invention.
  • Fig. 43A and Fig. 43B by extracting the common acoustic characteristics in each of the acoustic characteristics adding filters 35 and 37 of Fig. 33 or Fig. 35 , these are divided between the common calculation sections (C ⁇ l) 64 and (C ⁇ r) 65 and the individual calculation sections (P ⁇ l) through (Q ⁇ r) 66 through 69, thereby avoiding calculations that are duplications, the result being that it is possible to achieve an even greater improvement in calculation efficiency and speed in comparison with the prior art as described with regard to Fig. 8A and Fig. 8B .
  • the common calculation sections 64 and 65 are connected in series with the individual calculation sections 66 through 69, respectively.
  • Each of the individual calculation sections 66 through 69 has connected to it an amplifier g pl through g Qr , for the purpose of controlling the difference in level between the two ears and the position of the sound image.
  • the common acoustic characteristics are the acoustic characteristics from a virtual sound source (C), which is positioned between two or more real sound sources (P through Q), to a listener.
  • Fig. 44A shows the processing system for determining the common characteristics linear predictive coefficients using an impulse response which represents the acoustic characteristics from the above-noted virtual sound source C to the listener.
  • this example happens to show the acoustic characteristics of C ⁇ l, the same would apply to the acoustic characteristics for C ⁇ r.
  • the virtual sound source positioned directly in front of the listener, it is possible to assume that the C ⁇ l and C ⁇ r acoustic characteristics are equal.
  • a Hamming window or the like is used for the windowing processing 70, with linear predictive analysis 71 being performed by the Levinson-Durbin recursion method.
  • Fig. 44B and Fig. 44C show the processing system for determining the linear predictive coefficient which represent the individual acoustic characteristics from the real sound sources P through Q to the listener.
  • Each of the acoustic characteristics is input to the filter (C ⁇ l) -1 72 or (C ⁇ r) -1 73 which eliminates the common acoustic characteristics of the impulse response, the corresponding outputs being subjected to linear predictive analysis, thereby determining the linear predictive coefficients which represent the individual parts of each of the acoustic characteristics.
  • the above filters 72 and 73 have set into them linear predictive coefficients for the common characteristics, using a method similar to that described with regard to Fig. 13 .
  • the common characteristics parts are removed from each of the individual impulse responses beforehand, the linear predictive coefficients for the filter characteristics of each individual filter (P ⁇ l) through (Q ⁇ l) and (P ⁇ r) through (Q ⁇ r) being determined.
  • Fig. 45 and Fig. 46 show an embodiment in which the common and individual parts of the characteristics are separated, acoustic characteristics adding filters 35 and 37 being connected is series therebetween.
  • the common parts 64 and 65 of Fig. 45 are formed by the linear synthesis filter, described with relation to Fig. 12 , which comprises a short-term synthesis filter and a pitch synthesis filter.
  • Individual parts 66 through 69 are formed by, in addition to short-term synthesis filter which represent each of the individual frequency characteristics, delay devices Z -DP and Z -DQ which control the time difference between the two ears, and amplifiers g pl through g Ql for the purpose of controlling the level difference and position of the sound image.
  • Fig. 46 shows and example of an acoustic characteristics adding filter between two sound sources L and R and a listener.
  • Figs. 47A through 49 there is no pitch synthesis filter used in the common parts 64 and 65.
  • Figs. 47A through 49 show an example of the frequency characteristics of the acoustic characteristics adding filter shown in Fig. 46 .
  • the two sound sources L and R in Fig. 46 correspond, respectively, to the sound sources S1 and S2 shown in Figs. 47A and 47B , these being disposed with an angle of 30 degrees between them, as seen from the listener.
  • Fig. 47B is a block diagram representation of the acoustic characteristics adding filter of Fig. 46
  • Figs. 48 and 49 show the measurement system.
  • the broken line of Fig. 48 indicates the frequency characteristics of the common part (C ⁇ l) in Fig. 47B
  • the broken line in Fig. 49 indicates the frequency characteristics when the common part and individual part are connected in series.
  • the solid lines here indicate the case of 256 taps for a prior art filter, the broken lines indicating the number taps for a short-term synthesis filter, this being 6 taps for C ⁇ l and 4 taps for sl ⁇ l, for a total of 10 taps.
  • a pitch synthesis filter is not used, the more individual parts there are, the greater is the effect of reducing the amount of calculation.
  • Fig. 50 shows the example of using a hard disk or the like as a storage medium 74 for use with sound signal data to which the common characteristics of common parts 64 and 65 have already been added.
  • Fig. 51 shows the example of reading a signal from the storage medium 74, to which the common characteristics have already been added, rather than performing calculations of the common characteristics, and providing this to the individual parts 66 through 69.
  • the listener performs the required operation of the acoustic control apparatus 75, thereby enabling readout of the signal from the storage medium which has already be subjected to common characteristics calculations.
  • the thus readout signal is then subjected to calculations which add to it the individual characteristics and adjust the output gain thereof, to achieve the desired position for the sound image.
  • the signal stored in the storage medium 74 can include, in addition to the above-noted common characteristics, the processing for the inverse of the acoustic characteristics of the headphone, this processing being fixed.
  • Fig. 52 two virtual sound sources A and B are used, the levels g Al , g Ar , g Bl , and g Br , between them being used to localize the sound image S.
  • the processing is performed with left-to-right symmetry with respect to the center line of the listener. That is, the virtual sound sources A and B to the left of the listener and the virtual sound sources A and B to the right of the listener are said to form the same type of acoustic environment with respect to the listener.
  • Fig. 52 two virtual sound sources A and B are used, the levels g Al , g Ar , g Bl , and g Br , between them being used to localize the sound image S.
  • the processing is performed with left-to-right symmetry with respect to the center line of the listener. That is, the virtual sound sources A and B to the left of the listener and the virtual sound sources A and B to the right of the listener are said to form the same type of acoustic environment with respect to
  • the area surrounding the listener is divided into n equal parts, with virtual sound sources A and B placed on each of the borders therebetween, the acoustic characteristics of the propagation path from each of the virtual sound sources to the ears of the listener being left-to-right symmetrical as shown in Fig. 54 .
  • it is sufficient to have only 0, ..., n/2-1 coefficients in reality.
  • the position of the sound image with respect to the listener is expressed as the angle ⁇ as measured in, for example, the counterclockwise direction from the direct front direction.
  • the Equation (6) given below is used to determine in what region of the n equal-sized regions the sound image is localized, from the angle ⁇ .
  • Region number Integer part of ( ⁇ /(2 ⁇ /n)) ... (6)
  • the angle ⁇ is converted as shown by Equation (7).
  • 0 ⁇ ⁇ ⁇ ⁇ . or 2 ⁇ - ⁇ ( ⁇ 2 ⁇ )
  • Fig. 56 shows an example of an acoustic characteristics adding filter for the purpose of processing a system such as described above, in which there is left-to-right symmetry.
  • a feature of this acoustic characteristics adding filter is that, by performing the delay processing for the propagation paths A ⁇ r and B ⁇ r with reference to the delays of A ⁇ l and B ⁇ l, it is possible to eliminate the delay processing for A ⁇ l and B ⁇ l. Therefore, it is possible to halve the delay processing to represent the time difference between the two ears.
  • Fig. 57A and Fig. 57B show the conceptual configuration for the processing of a sound image, using output interpolation between a plurality of virtual sound sources.
  • Fig. 57A in order to add the transfer characteristics of each of the acoustic space paths from the virtual sound sources at two locations (A, B) 30-1 and 30-2 to the left and right ears of the listener 31, four acoustic characteristics calculation filters 151 through 154 are provided. These are followed by amplifiers for the adjustment of the gain of each, so that it is possible to either localize a sound image between the above-noted virtual sound sources 30-1 and 30-2 or move the sound image thereamong.
  • Fig. 57B when localizing a sound image between the virtual sound sources (B, C) 30-2 and 30-3 or moving the sound image thereamong, of the four acoustic characteristics calculation filters 151 through 154, the two acoustic characteristics calculation filters 151 and 152 are allocated to the virtual sound source 30-1. In this case, the acoustic characteristics calculation filters 153 and 154 of the virtual sound source 30-2 remain unchanged and are used as is. Similar to the case of Fig. 57A , amplifiers after these filters are provided to adjust each of the gains, enabling positioning of a sound image between virtual sound sources 30-2 and 30-3 or smooth movement of the sound image thereamong.
  • the amount of hardware, in terms of, for example, memory capacity, that is required for movement of a sound image is minimized, thereby providing not only a simplification of the processing control, but also an increase in speed.
  • the virtual sound source (B) of (3) when switching between sound-generation areas, only the virtual sound source (B) of (3) generates sound, the other virtual sound sources (A and C) having amplifier gains of zero. Therefore, no click noise is generated from the above-noted switch of sound-generation areas.
  • Fig. 58 and Fig. 59 each show a specific embodiment of Fig. 57A and Fig. 57B .
  • new position information is given, from which a filter controller 155 performs setting of filter coefficients and selection of memory, a gain controller 156 being provided to perform calculation of the gain with respect to the amplifier for each sound image position.
  • a sound image is localized by using a plurality of virtual sound sources, even when the number or position of the sound images change, it is not necessary to change the acoustic characteristics from each virtual sound source to the listener, thereby eliminating the need to use a linear synthesis filter. Additionally, it is possible to add the desired acoustic characteristics to the original signal with a filter having a small number of taps. It is further possible, by considering the critical bandwidth, to smooth the original impulse response so that there is no audible change, thereby enabling an even further improvement in the accuracy of approximation when approximating frequency characteristics using linear predictive coefficients of low order. In doing this, by compensating for the waveform of the impulse response in the time domain, it is possible to facilitate control of the time and level difference and the like between the two ears of the listener.
  • the number of filter taps can be reduced without changing the overall acoustic characteristics, making it easy to implement control of a three-dimension sound image using digital signal processor or the like.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Stereophonic System (AREA)

Claims (16)

  1. Appareil acoustique tridimensionnel pour positionner une image sonore en utilisant une source sonore virtuelle, comprenant :
    un premier filtre d'ajout de caractéristiques acoustiques (35) configuré comme un filtre de synthèse linéaire ayant des coefficients de filtrage qui sont des coefficients prédictifs linéaires obtenus par une analyse prédictive linéaire d'une réponse impulsionnelle qui représente des caractéristiques acoustiques d'un chemin ou de chacun d'une pluralité de chemins acoustiques vers l'oreille gauche à ajouter à un signal original ; et
    un second filtre d'ajout de caractéristiques acoustiques (37) configuré comme un filtre de synthèse linéaire ayant des coefficients de filtrage qui sont des coefficients prédictifs linéaires obtenus par une analyse prédictive linéaire d'une réponse impulsionnelle qui représente des caractéristiques acoustiques d'un chemin ou de chacun d'une pluralité de chemins acoustiques vers l'oreille droite à ajouter à un signal original ;
    caractérisé en ce que ledit appareil comprend en outre :
    un premier filtre d'élimination de caractéristiques acoustique (36), qui est relié en série audit premier filtre d'ajout de caractéristiques acoustiques (35), et qui est configuré comme un filtre de synthèse linéaire ayant des coefficients qui sont obtenus par une analyse prédictive linéaire d'une réponse impulsionnelle qui représente les caractéristiques acoustiques d'un dispositif de sortie acoustique vers l'oreille gauche, ces coefficients de filtrage communiquant des caractéristiques acoustiques audit premier filtre d'élimination de caractéristiques acoustiques représentant des caractéristiques inverses des caractéristiques acoustiques du dispositif de sortie acoustique, de façon à éliminer lesdites caractéristiques ;
    un second filtre d'élimination de caractéristiques acoustique (38), qui est relié en série avec ledit second filtre d'ajout de caractéristiques acoustiques (37), et qui est configuré comme un filtre de synthèse linéaire ayant des coefficients qui sont obtenus par une analyse prédictive linéaire d'une réponse impulsionnelle qui représente les caractéristiques acoustiques d'un dispositif de sortie acoustique vers l'oreille droite, ces coefficients de filtrage communiquant des caractéristiques acoustiques audit second filtre d'élimination de caractéristiques acoustique représentant des caractéristiques inverses des caractéristiques acoustiques du dispositif de sortie acoustique, de façon à éliminer lesdites caractéristiques ; et
    une section de paramétrage sélectif (39) pour paramétrer de manière sélective des paramètres prédéterminés dudit premier filtre d'ajout de caractéristiques acoustiques et dudit second filtre d'ajout de caractéristiques acoustiques, en réponse à des informations de position d'image sonore,
    dans lequel lesdits premier et second filtres d'ajout de caractéristiques acoustiques (35, 37) sont formés chacun par une section commune distincte (64, 65) pour ajouter des caractéristiques qui sont communes à chaque chemin acoustique, et des sections individuelles (66-67, 68-69) pour ajouter des caractéristiques qui sont individuelles à chaque chemin acoustique, ladite section commune et lesdites sections individuelles étant reliées en série pour ajouter les caractéristiques acoustiques globales.
  2. Appareil acoustique tridimensionnel selon la revendication 1, comprenant en outre un support de stockage (74) pour stocker des résultats desdites sections communes (64, 65) en ce qui concerne un son original prédéterminé, et une section de commande de lecture (75) pour commander la lecture de résultats de calcul qui sont stockés dans ledit support de stockage (74), ladite section de commande de lecture (75) étant agencée pour fournir directement lesdits résultats de calcul lus auxdites sections individuelles (66-69).
  3. Appareil acoustique tridimensionnel selon la revendication 2, dans lequel ledit support de stockage, en plus d'être agencé pour stocker des résultats de calcul provenant desdites sections communes (64, 65) en ce qui concerne un son original prédéterminé, est également agencé pour stocker des résultats de calcul du filtre correspondant desdits premier ou second filtres d'ajout de caractéristiques acoustiques.
  4. Appareil acoustique tridimensionnel selon la revendication 1, dans lequel ledit premier filtre d'ajout de caractéristiques acoustiques (35) et ledit second filtre d'ajout de caractéristiques acoustiques (37) comprennent en outre respectivement des sections de retard (56, 57) pour appliquer un temps de retard entre les deux oreilles.
  5. Appareil acoustique tridimensionnel selon la revendication 4, dans lequel le temps de retard d'une des sections de retard (56, 57) des premier et second filtres d'ajout de caractéristiques acoustiques est fait pour être un temps de retard de référence de zéro.
  6. Appareil acoustique tridimensionnel selon la revendication 1, dans lequel le premier filtre d'ajout de caractéristiques acoustiques (35) et le second filtre d'ajout de caractéristiques acoustiques (37) comprennent en outre respectivement des sections d'amplification (58, 59) qui valident des paramétrages variables des niveaux de sortie provenant desdits premier filtre d'ajout de caractéristiques acoustiques (35) et second filtre d'ajout de caractéristiques acoustiques (37).
  7. Appareil acoustique tridimensionnel selon la revendication 6, dans lequel ladite section de paramétrage sélectif (39) est agencée pour déplacer la position d'une image sonore en faisant varier les niveaux de signal de sortie relatifs provenant du premier filtre d'ajout de caractéristiques acoustiques (35) et du second filtre d'ajout de caractéristiques acoustiques (37) en paramétrant le gain de ladite section d'amplification (58) en réponse auxdites informations de position d'image sonore.
  8. Appareil acoustique tridimensionnel selon la revendication 5, dans lequel ledit premier filtre d'ajout de caractéristiques acoustiques (35) et ledit second filtre d'ajout de caractéristiques acoustiques (37) sont configurés de façon à être symétriques de gauche à droite par rapport à une ligne centrale à l'avant de l'auditeur, des paramètres desdites sections de retard (56, 57) étant partagés entre des positions qui correspondent dans ladite symétrie.
  9. Appareil acoustique tridimensionnel selon la revendication 7, dans lequel ledit premier filtre d'ajout de caractéristiques acoustiques (35) et ledit second filtre d'ajout de caractéristiques acoustiques (37) sont configurés de façon à être symétriques de gauche à droite par rapport à une ligne centrale à l'avant de l'auditeur, des paramètres desdites sections d'amplification (58, 59) étant partagées entre des positions qui correspondent dans ladite symétrie.
  10. Appareil acoustique tridimensionnel selon la revendication 1, comprenant en outre une section d'interpolation d'informations de position (60) pour interpoler des information de position entre des informations de position d'image sonore passées et futures, ladite section d'interpolation d'informations de position (60) étant agencée pour donner des informations de position interpolées à ladite section de paramétrage sélectif (39) en tant qu'informations de position.
  11. Appareil acoustique tridimensionnel selon la revendication 1, comprenant en outre une section de prédiction d'informations de position (60) pour effectuer une prédiction par interpolation d'informations de position futures à partir d'informations de position d'image sonore passées et actuelles, ladite section de prédiction d'informations de position étant agencée pour donner des informations de position futures à ladite section de paramétrage sélectif en tant qu'informations de position.
  12. Appareil acoustique tridimensionnel selon la revendication 11, dans lequel ladite section de prédiction d'informations de position (60) comprend en outre une section de détermination de régularité (61) pour effectuer une détermination quant à l'existence de régularité par rapport au mouvement, en se basant sur des informations de position d'image sonore passées et actuelles, et dans lequel lorsque ladite section de détermination de régularité (61) détermine qu'une régularité existe, ladite section de prédiction d'informations de position est agencée pour donner lesdites informations de position futures.
  13. Appareil acoustique tridimensionnel selon la revendication 1, comprenant en outre une section d'interpolation d'informations de position (60) pour interpoler des informations de position entre des informations d'image visuelle passées et futures provenant d'un appareil d'affichage d'image sur lequel une image visuelle qui produit un son est affichée, ladite section d'interpolation d'informations de position (60) étant agencée pour donner des informations de position interpolées à ladite section de paramétrage sélectif (39) en tant qu'informations de position.
  14. Appareil acoustique tridimensionnel selon la revendication 1, comprenant en outre une section de prédiction d'informations de position (60) pour effectuer une prédiction par interpolation d'informations de position futures à partir d'informations d'image visuelle passées et actuelles provenant d'un appareil d'affichage d'image sur lequel une image visuelle qui produit un son est affichée, ladite section de prédiction d'informations de position étant agencée pour donner des informations de position futures à ladite section de paramétrage sélectif en tant qu'informations de position.
  15. Appareil acoustique tridimensionnel selon la revendication 11, dans lequel ladite section de prédiction d'informations de position (60) comprend en outre une section de détermination de régularité (61) pour effectuer une détermination quant à l'existence d'une régularité en ce qui concerne un mouvement, en se basant sur des informations d'image visuelle passées et actuelles provenant d'un appareil d'affichage d'image sur lequel une image visuelle qui produit un son est affichée, et dans lequel lorsque ladite section de détermination de régularité (61) détermine qu'une régularité existe, ladite section de prédiction d'informations de position est agencée pour donner lesdites informations de position futures.
  16. Appareil acoustique tridimensionnel selon la revendication 1, dans lequel, afin de fournir à un auditeur un bon environnement d'écoute, ladite section de paramétrage sélectif (39) est agencée pour déplacer ledit environnement d'écoute en réponse à des informations de position d'auditeur données.
EP07010496A 1995-09-08 1996-08-20 Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs Expired - Lifetime EP1816895B1 (fr)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP23170595A JP3810110B2 (ja) 1995-09-08 1995-09-08 線型予測係数を用いた立体音響処理装置
JP04610596A JP4306815B2 (ja) 1996-03-04 1996-03-04 線形予測係数を用いた立体音響処理装置
EP96113318A EP0762804B1 (fr) 1995-09-08 1996-08-20 Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
EP96113318.8 Division 1996-08-20
EP96113318A Division EP0762804B1 (fr) 1995-09-08 1996-08-20 Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs

Publications (3)

Publication Number Publication Date
EP1816895A2 EP1816895A2 (fr) 2007-08-08
EP1816895A3 EP1816895A3 (fr) 2007-09-05
EP1816895B1 true EP1816895B1 (fr) 2011-10-12

Family

ID=26386227

Family Applications (2)

Application Number Title Priority Date Filing Date
EP96113318A Expired - Lifetime EP0762804B1 (fr) 1995-09-08 1996-08-20 Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs
EP07010496A Expired - Lifetime EP1816895B1 (fr) 1995-09-08 1996-08-20 Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs

Family Applications Before (1)

Application Number Title Priority Date Filing Date
EP96113318A Expired - Lifetime EP0762804B1 (fr) 1995-09-08 1996-08-20 Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs

Country Status (3)

Country Link
US (3) US6023512A (fr)
EP (2) EP0762804B1 (fr)
DE (1) DE69637736D1 (fr)

Families Citing this family (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3976360B2 (ja) * 1996-08-29 2007-09-19 富士通株式会社 立体音響処理装置
JP4088725B2 (ja) * 1998-03-30 2008-05-21 ソニー株式会社 オーディオ再生装置
JP2000041294A (ja) * 1998-07-23 2000-02-08 Sony Corp ヘッドホンおよびその送信回路
FR2782228B1 (fr) * 1998-08-05 2001-05-25 Ct Scient Tech Batiment Cstb Dispositif de simulation sonore et procede pour realiser un tel dispositif
JP3918315B2 (ja) * 1998-08-20 2007-05-23 ヤマハ株式会社 インパルス応答測定方法
JP4240683B2 (ja) * 1999-09-29 2009-03-18 ソニー株式会社 オーディオ処理装置
US6980592B1 (en) * 1999-12-23 2005-12-27 Agere Systems Inc. Digital adaptive equalizer for T1/E1 long haul transceiver
US6366862B1 (en) * 2000-04-19 2002-04-02 National Instruments Corporation System and method for analyzing signals generated by rotating machines
JP3624805B2 (ja) * 2000-07-21 2005-03-02 ヤマハ株式会社 音像定位装置
GB2374507B (en) * 2001-01-29 2004-12-29 Hewlett Packard Co Audio user interface with audio cursor
GB2372923B (en) * 2001-01-29 2005-05-25 Hewlett Packard Co Audio user interface with selective audio field expansion
GB2374502B (en) * 2001-01-29 2004-12-29 Hewlett Packard Co Distinguishing real-world sounds from audio user interface sounds
US20030227476A1 (en) * 2001-01-29 2003-12-11 Lawrence Wilcock Distinguishing real-world sounds from audio user interface sounds
GB2374506B (en) * 2001-01-29 2004-11-17 Hewlett Packard Co Audio user interface with cylindrical audio field organisation
US20030123676A1 (en) * 2001-03-22 2003-07-03 Schobben Daniel Willem Elisabeth Method of deriving a head-related transfer function
KR20030021250A (ko) * 2001-05-11 2003-03-12 코닌클리케 필립스 일렉트로닉스 엔.브이. 디바이스들의 세트의 동작
US6757622B2 (en) 2002-01-17 2004-06-29 Seagate Technology Llc Predicting disc drive acoustic sound power from mechanical vibration
JP3682032B2 (ja) * 2002-05-13 2005-08-10 株式会社ダイマジック オーディオ装置並びにその再生用プログラム
US7324598B2 (en) * 2002-07-15 2008-01-29 Intel Corporation Apparatus and method to reduce quantization error
US7113610B1 (en) * 2002-09-10 2006-09-26 Microsoft Corporation Virtual sound source positioning
US7243064B2 (en) * 2002-11-14 2007-07-10 Verizon Business Global Llc Signal processing of multi-channel data
AU2003255673A1 (en) * 2003-06-13 2005-01-28 France Telecom Binaural signal processing with improved efficiency
JP2005051660A (ja) * 2003-07-31 2005-02-24 Onkyo Corp 映像信号および音声信号の再生システム
JP4127156B2 (ja) * 2003-08-08 2008-07-30 ヤマハ株式会社 オーディオ再生装置、ラインアレイスピーカユニットおよびオーディオ再生方法
CN1886780A (zh) * 2003-12-15 2006-12-27 法国电信 声音合成和空间化方法
JP2006030443A (ja) * 2004-07-14 2006-02-02 Sony Corp 記録媒体、記録装置及び方法、データ処理装置及び方法、データ出力装置及び方法
KR101118214B1 (ko) * 2004-09-21 2012-03-16 삼성전자주식회사 청취 위치를 고려한 2채널 가상 음향 재생 방법 및 장치
JP2006203850A (ja) * 2004-12-24 2006-08-03 Matsushita Electric Ind Co Ltd 音像定位装置
DE102005033238A1 (de) * 2005-07-15 2007-01-25 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Vorrichtung und Verfahren zum Ansteuern einer Mehrzahl von Lautsprechern mittels eines DSP
DE102005033239A1 (de) * 2005-07-15 2007-01-25 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Vorrichtung und Verfahren zum Steuern einer Mehrzahl von Lautsprechern mittels einer graphischen Benutzerschnittstelle
JP4602204B2 (ja) * 2005-08-31 2010-12-22 ソニー株式会社 音声信号処理装置および音声信号処理方法
PL1938661T3 (pl) * 2005-09-13 2014-10-31 Dts Llc System i sposób przetwarzania dźwięku
JP4479644B2 (ja) * 2005-11-02 2010-06-09 ソニー株式会社 信号処理装置および信号処理方法
JP4637725B2 (ja) * 2005-11-11 2011-02-23 ソニー株式会社 音声信号処理装置、音声信号処理方法、プログラム
WO2007123788A2 (fr) * 2006-04-03 2007-11-01 Srs Labs, Inc. Traitement de signal audio
JP2007280485A (ja) * 2006-04-05 2007-10-25 Sony Corp 記録装置、再生装置、記録再生装置、記録方法、再生方法および記録再生方法並びに記録媒体
JP4835298B2 (ja) * 2006-07-21 2011-12-14 ソニー株式会社 オーディオ信号処理装置、オーディオ信号処理方法およびプログラム
JP4894386B2 (ja) * 2006-07-21 2012-03-14 ソニー株式会社 音声信号処理装置、音声信号処理方法および音声信号処理プログラム
JP5082327B2 (ja) * 2006-08-09 2012-11-28 ソニー株式会社 音声信号処理装置、音声信号処理方法および音声信号処理プログラム
KR101238361B1 (ko) * 2007-10-15 2013-02-28 삼성전자주식회사 어레이 스피커 시스템에서 근접장 효과를 보상하는 방법 및장치
KR101438389B1 (ko) * 2007-11-15 2014-09-05 삼성전자주식회사 오디오 매트릭스 디코딩 방법 및 장치
US8126172B2 (en) * 2007-12-06 2012-02-28 Harman International Industries, Incorporated Spatial processing stereo system
CN103152499B (zh) * 2008-06-11 2014-12-10 三菱电机株式会社 回声消除器
US20090312849A1 (en) * 2008-06-16 2009-12-17 Sony Ericsson Mobile Communications Ab Automated audio visual system configuration
JP2010140235A (ja) * 2008-12-11 2010-06-24 Sony Corp 画像処理装置、画像処理方法およびプログラム
US8848909B2 (en) * 2009-07-22 2014-09-30 Harris Corporation Permission-based TDMA chaotic communication systems
TWM423331U (en) * 2011-06-24 2012-02-21 Zinwell Corp Multimedia player device
JP6006627B2 (ja) * 2012-12-05 2016-10-12 日本放送協会 インパルス応答長変換装置、インパルス応答長変換方法、インパルス方法変換プログラム
JP6147603B2 (ja) * 2013-07-31 2017-06-14 Kddi株式会社 音声伝達装置、音声伝達方法
JP6561718B2 (ja) * 2015-09-17 2019-08-21 株式会社Jvcケンウッド 頭外定位処理装置、及び頭外定位処理方法
GB2544458B (en) * 2015-10-08 2019-10-02 Facebook Inc Binaural synthesis
US11256768B2 (en) 2016-08-01 2022-02-22 Facebook, Inc. Systems and methods to manage media content items
WO2019106848A1 (fr) 2017-12-01 2019-06-06 株式会社ソシオネクスト Dispositif de traitement du signal, et procédé de traitement du signal

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE2244162C3 (de) 1972-09-08 1981-02-26 Eugen Beyer Elektrotechnische Fabrik, 7100 Heilbronn «system
JPS5419242B2 (fr) * 1973-06-22 1979-07-13
DE3238933A1 (de) * 1982-10-21 1984-04-26 Sennheiser Electronic Kg, 3002 Wedemark Verfahren zur akustischen ausgestaltung von video-spielen
JPH0736866B2 (ja) * 1989-11-28 1995-04-26 ヤマハ株式会社 ホール音場支援装置
US5495534A (en) * 1990-01-19 1996-02-27 Sony Corporation Audio signal reproducing apparatus
JP2964514B2 (ja) 1990-01-19 1999-10-18 ソニー株式会社 音響信号再生装置
JP3258195B2 (ja) * 1995-03-27 2002-02-18 シャープ株式会社 音像定位制御装置

Also Published As

Publication number Publication date
EP1816895A3 (fr) 2007-09-05
US6553121B1 (en) 2003-04-22
DE69637736D1 (de) 2008-12-18
EP0762804A2 (fr) 1997-03-12
US6023512A (en) 2000-02-08
EP1816895A2 (fr) 2007-08-08
EP0762804B1 (fr) 2008-11-05
US6269166B1 (en) 2001-07-31
EP0762804A3 (fr) 2006-08-02

Similar Documents

Publication Publication Date Title
EP1816895B1 (fr) Processeur acoustique tridimensionnel utilisant des coefficients linéaires prédictifs
US9918179B2 (en) Methods and devices for reproducing surround audio signals
JP7139409B2 (ja) 少なくとも一つのフィードバック遅延ネットワークを使ったマルチチャネル・オーディオに応答したバイノーラル・オーディオの生成
JP7183467B2 (ja) 少なくとも一つのフィードバック遅延ネットワークを使ったマルチチャネル・オーディオに応答したバイノーラル・オーディオの生成
JP5285626B2 (ja) 音声空間化及び環境シミュレーション
US6078669A (en) Audio spatial localization apparatus and methods
US5982903A (en) Method for construction of transfer function table for virtual sound localization, memory with the transfer function table recorded therein, and acoustic signal editing scheme using the transfer function table
US6421446B1 (en) Apparatus for creating 3D audio imaging over headphones using binaural synthesis including elevation
US6990205B1 (en) Apparatus and method for producing virtual acoustic sound
US7912225B2 (en) Generating 3D audio using a regularized HRTF/HRIR filter
JP2013524562A (ja) マルチチャンネル音響再生方法及び装置
US20090041254A1 (en) Spatial audio simulation
KR100647338B1 (ko) 최적 청취 영역 확장 방법 및 그 장치
CA2314374A1 (fr) Systeme d'annulation de la diaphonie acoustique
US6178245B1 (en) Audio signal generator to emulate three-dimensional audio signals
Otani et al. Binaural Ambisonics: Its optimization and applications for auralization
CN117376784A (zh) 单声道立体声场扩展方法、电子设备及存储介质
Kahana et al. A multiple microphone recording technique for the generation of virtual acoustic images
JPH09191500A (ja) 仮想音像定位用伝達関数表作成方法、その伝達関数表を記録した記憶媒体及びそれを用いた音響信号編集方法
JP4306815B2 (ja) 線形予測係数を用いた立体音響処理装置
JP3810110B2 (ja) 線型予測係数を用いた立体音響処理装置
JP2755081B2 (ja) 音像定位制御方法
Filipanits Design and implementation of an auralization system with a spectrum-based temporal processing optimization
JP2924539B2 (ja) 音像定位制御方法
JP2000059892A (ja) 複数チャンネルオ―ディオ信号処理方法

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

PUAL Search report despatched

Free format text: ORIGINAL CODE: 0009013

AC Divisional application: reference to earlier application

Ref document number: 0762804

Country of ref document: EP

Kind code of ref document: P

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): DE GB NL

AK Designated contracting states

Kind code of ref document: A3

Designated state(s): DE GB NL

17P Request for examination filed

Effective date: 20080304

AKX Designation fees paid

Designated state(s): DE GB NL

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AC Divisional application: reference to earlier application

Ref document number: 0762804

Country of ref document: EP

Kind code of ref document: P

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): DE GB NL

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

RIN2 Information on inventor provided after grant (corrected)

Inventor name: SUZUKI, KAORI

Inventor name: MATSUO, NAOSHI

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 69638435

Country of ref document: DE

Effective date: 20111201

REG Reference to a national code

Ref country code: NL

Ref legal event code: VDEP

Effective date: 20111012

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20111012

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed

Effective date: 20120713

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 69638435

Country of ref document: DE

Effective date: 20120713

GBPC Gb: european patent ceased through non-payment of renewal fee

Effective date: 20120820

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20120820

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20130528

Year of fee payment: 18

REG Reference to a national code

Ref country code: DE

Ref legal event code: R119

Ref document number: 69638435

Country of ref document: DE

REG Reference to a national code

Ref country code: DE

Ref legal event code: R119

Ref document number: 69638435

Country of ref document: DE

Effective date: 20150303

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20150303