EP0790753A1 - System für Raumklangeffekt und Verfahren dafür - Google Patents
System für Raumklangeffekt und Verfahren dafür Download PDFInfo
- Publication number
- EP0790753A1 EP0790753A1 EP97400248A EP97400248A EP0790753A1 EP 0790753 A1 EP0790753 A1 EP 0790753A1 EP 97400248 A EP97400248 A EP 97400248A EP 97400248 A EP97400248 A EP 97400248A EP 0790753 A1 EP0790753 A1 EP 0790753A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- sound
- channels
- transfer functions
- user
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S3/00—Systems employing more than two channels, e.g. quadraphonic
- H04S3/002—Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
- H04S3/004—For headphones
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/01—Multi-channel, i.e. more than two input channels, sound reproduction with two speakers wherein the multi-channel information is substantially preserved
Definitions
- the present invention relates to a sound spatialization system, as well as to a personalization method making it possible to implement sound spatialization.
- An airplane pilot in particular a fighter airplane, wears a stereophonic helmet which restores not only radio communications, but also various alarms and on-board communications. Radiocommunications can be satisfied with a stereophonic, or even monophonic reproduction, while the alarms and on-board communications cannot be located in relation to the pilot (or the co-pilot ).
- the subject of the present invention is an audiophonic communication system, which makes it possible to easily discriminate the location of a determined sound source, in particular in the case of the existence of several sound sources close to the user.
- the sound spatialization system comprises, for each monophonic channel to be spatialized, a binaural processor with two channels of convolution filters combined linearly in each channel, this processor (s) being connected to one orienting device for calculating the spatial location of sound sources, itself connected to localization devices, characterized in that it includes, for at least part of the channels, a complementary sound illustration device connected to the corresponding binaural processor, this device additional sound illustration comprising at least one of the following circuits: bandwidth widening circuit, background noise production circuit, circuit for simulating the acoustic behavior of a room, Doppler effect simulation circuit, circuit producing different sound symbols each corresponding to a specific source or alarm.
- the personalization method according to the invention consists in estimating the head transfer functions of the user by measuring these functions at a finite number of points in the surrounding space, then by interpolating the values thus measured to be calculated the head transfer functions, for each of the user's two ears, at the point in the space where the sound source is located, and to create the "spatialized” signal from the monophonic signal to be processed by convolving it with each of the two transfer functions thus estimated. It is thus possible to "personalize" the convolution filters for each user of the system implementing this method. Each user will then be able to locate in the best possible way the virtual sound source reproduced by his audio equipment.
- the invention is described below with reference to an aircraft audio system, in particular a combat aircraft, but it is understood that it is not limited to such an application, and that it can be implementation as well in other types of vehicles (land or sea) as in fixed installations.
- the user of this system is, in this case, the pilot of a combat aircraft, but it is understood that there can be several users simultaneously, in particular if it is a combat aircraft. civil transport, specific devices for each user being provided in corresponding number.
- the role of the spatialization module 1 shown in FIG. 1 is to make sound signals (tones, speech, alarms, etc.) heard using a stereo headset so that they are perceived by the listener as if they came from a particular point in space, this point can be the actual position of the sound source or an arbitrary position.
- the pilot of a combat aircraft hears the voice of his co-pilot as if it actually came from behind him, or else an audible missile attack alert is positioned spatially at the point of arrival of the threat.
- the position of the sound source changes in function of the pilot's head movements and of the airplane's movements: for example, an alarm generated at the azimuth "3 hours” must be found at "noon” if the pilot turns his head 90 degrees to the right.
- the module 1 is for example connected to a digital bus 2 from which it receives information supplied by: a head position detector 3, an inertial unit 4 and / or a location device such as a goniometer, a radar, etc. ., countermeasures devices 5 (detection of external threats such as missiles) and an alarm management device 6 (signaling in particular breakdowns of aircraft instruments or equipment).
- a head position detector 3 an inertial unit 4 and / or a location device such as a goniometer, a radar, etc. .
- countermeasures devices 5 detection of external threats such as missiles
- an alarm management device 6 signalaling in particular breakdowns of aircraft instruments or equipment.
- the module 1 includes an interpolator 7, the input of which is connected to the bus 2 to which various sound sources are connected (microphones, alarms, etc.). In general, these sources are sampled at relatively low frequencies (6, 12 or 24 kHz for example).
- the interpolator 7 makes it possible to raise these frequencies to a multiple common, for example 48 kHz in the present case, a frequency necessary for the processors located downstream.
- This interpolator 7 is connected to n binaural processors, referenced 8 as a whole, n being the maximum number of channels to be spatialized simultaneously.
- the outputs of the processors 8 are connected to an adder 9 whose output constitutes the output of the module 1.
- the module 1 also comprises in the connection between at least one output of the interpolator 7 and the input of the corresponding processor of the assembly 8 an adder 10, the other input of which is connected to the output of a device 11 for additional sound illustration.
- This device 11 produces a sound signal covering in particular the high frequencies (for example from 5 to 16 kHz) of the audio spectrum. It thus completes the useful bandwidth of the transmission channel to which its output signal is added.
- This transmission channel can advantageously be a radio channel, but it is understood that any other channel can be completed in this way, and that several channels can be completed in the same system, by providing a corresponding number of adders such as 10.
- radio communications use reduced bandwidths (3 to 4 kHz in general). Such a bandwidth is insufficient for correct spatialization of the sound signal. Tests have shown that high frequencies (above about 14 kHz), located beyond the limit of the vocal spectrum, allow a better localization of the source of the sound. The device 11 is then a bandwidth widening device.
- the additional sound signal can for example be background noise characteristic of a radio link.
- the device 11 can also be, for example, a device simulating the acoustic behavior of a room, a building, or a device simulating a Doppler effect, or even a device producing different sound symbols each corresponding to a source or a specific alarm.
- the processors 8 each generate a signal of stereophonic type from the monophonic signal coming from the interpolator 7 to which is added if necessary, the signal from the device 11, taking into account the data supplied by the head position detector 3 of the pilot.
- the module 1 also includes a device 12 for managing the sources to be spatialized followed by an orienter 13 with n inputs (n being defined above) controlling the n different processors of the set 8.
- the device 13 is a calculator calculating, from the data provided by the pilot's head position detector, the orientation of the aircraft relative to the terrestrial reference (provided by the aircraft's inertial unit) and the location of the source, the coordinates in the space of the point where the sounds emitted by this source must appear to come from.
- n2 is advantageously equal to four at most.
- the device 12 for managing the n sources to be spatialized is a computer which receives, via bus 2, information concerning the characteristics of the sources to be spatialized (site, deposit and distance by report to the pilot), personalization criteria at the user's choice and priority information (threats, alarms, important radio communications, etc.).
- the device 12 receives from the device 4 information concerning the evolution of the location of certain sources (or of all the sources, if applicable). From this information, the device 12 selects the source (or at most the n2 sources) to be spatialized.
- a memory card reader 15 is used 16 for the device in order to personalize the management of sound sources by the device 12.
- the reader 15 is connected to the bus 2.
- the card 16 then contains the characteristics of the filtering carried out by the flags of each user's ears. In the preferred embodiment, it is a set of pairs of digital filters (that is to say coefficients representing their impulse responses) corresponding to the acoustic filtering "left ear" and "right ear” produced for various points in the space surrounding the user.
- the database thus formed is loaded, via bus 2, into the memory associated with the various processors 8.
- the processors 8 each essentially comprise two channels (called “left ear” and “right ear”) for convolution filtering. More precisely, the role of each of the processors 8 is on the one hand to calculate by interpolation the head transfer functions (right and left) at the point at which the source will be placed, on the other hand to create the spatial signal on two channels from the original monophonic signal.
- head transfer functions require spatial sampling: these transfer functions are only measured in a finite number of points (of the order of 100). However to correctly "spatialize” a sound, it would be necessary to know the transfer functions at the source point of the source, determined by the orienter 13. It is therefore necessary to be satisfied with an estimation of these functions: this operation is carried out by an interpolation "barycentric" of the four pairs of functions associated with the four measurement points closest to the point of the calculated space.
- FIG. 2 shows a part of the "grid" G thus obtained for the points Pm, Pm + 1, Pm + 2, ... Pp, Pp + 1 .... Or a point P of said sphere, determined by the orienter 13 as being located in the direction of the sound source to be "spatialized".
- This point P is inside the curvilinear quadrilateral delimited by the points Pm + 1, Pm + 2, Pn + 1, Pn + 2.
- the barycentric interpolation is therefore carried out for the position of P with respect to these four points.
- the different equipment determining the orientation of the sound source and the orientation and location of the user's head provide their respective data every 20 or 40 ms ( ⁇ T), that is to say that every ⁇ T a couple of transfer functions are available.
- ⁇ T 20 or 40 ms
- the signal to be spatialized is in fact convoluted by a pair of filters obtained by "temporal" interpolation performed between the convolution filters spatially interpolated at times T and T + ⁇ T. It then remains only to convert the digital signals thus obtained into analog before their restitution in the user's earphones.
- FIG. 3 which relates to a channel to be spatialized, the various attitude (position) sensors used have been shown. These are: a head attitude sensor 17, a sound source attitude sensor 18, and an attitude sensor 19 of the carrier mobile (airplane for example).
- the information from these sensors is supplied to the orienter 13, which determines from this information the spatial position of the source relative to the head of the user (in line of sight and in distance).
- the orienter 13 is connected to a database 20 (included in the card 16) of which it controls the loading to the processors 8 of the "left" and "right” transfer functions of the four points closest to the position of the source (see Figure 2), or possibly the measurement point (if the position of the source coincides with that of one of the measurement points in grid G).
- transfer functions are subjected to a spatial interpolation at 21, then to a temporal interpolation at 22, and the resulting values are convolved at 23 with the signal 24 to be spatialized.
- functions 21 and 23 are performed by the same interpolator (interpolator 7 in FIG. 1), and the convolutions are performed by the binaural processor 8 corresponding to the spatialized channel.
- a digital-analog conversion is carried out, in 25, and the sound reproduction (amplification and sending to a stereo headset) at 26.
- operations 20 to 23 and 25, 26 are done separately for the left channel and for the right channel.
- the “personalized” convolution filters constituting the previously mentioned database are established from measurements using a method described below with reference to FIG. 4.
- an automated mechanical tool 27 is installed, consisting of a semi-circular rail 28 mounted on a motorized pivot 29 fixed to the floor of this chamber.
- the rail 28 is arranged vertically, so that its ends are on the same perpendicular.
- a support 30 moves on which a broadband speaker 31 is mounted. This device makes it possible to place the speaker at any point on the sphere defined by the rail when the latter performs a rotation of 360 degrees around a vertical axis passing through pivot 29.
- the precision of the positioning of the loudspeaker is one degree in elevation and in bearing, for example.
- a first series of readings is taken: the loudspeaker 31 is placed successively at X points of the sphere, that is to say that the space is "discretized”: it is a spatial sampling. At each measurement point, a pseudo-random code is generated and reproduced by the loudspeaker 31.
- the sound signal emitted is picked up by a pair of reference microphones placed at the center 32 of the sphere (the distance separating the microphones is the order of the width of the head of the subject whose transfer functions are to be collected), in order to measure the resulting sound pressure as a function of frequency.
- the method is the same but this time, the subject is placed so that his ears are located at the location of the microphones (the subject checks the position of his head by video feedback).
- the subject is provided with individualized shutter earplugs in which miniature microphones are placed.
- Complete obturation of the duct has the following advantages: the ear is acoustically protected, and the stapedial reflex (nonexistent in this case) does not modify the acoustic impedance of the assembly.
- the database of transfer functions can be made up either of pairs of frequency responses (convolution by multiplication in the frequency domain) or of pairs of impulse responses (classical temporal convolution), inverse Fourier transforms of the previous ones.
- acoustic sources emitting pseudo-random binary signals tends to be generalized in the impulse response measurement technique, especially with regard to the characterization of an acoustic room by the correlation method.
- these signals make the assumption of linearity of the acoustic collection system acceptable. They also make it possible to overcome variations in the acoustic impedance of the ossicular chain by stapedial reflex, by limiting the emission level (70 dBa). Preferably, pseudo-random binary signals of maximum length sequence are produced. The advantage of maximum length sequences lies in their spectral characteristics (white noise) and their generation mode which allows optimization of the processing processor.
- the impulse response is obtained over time (2n-1) / fe where N is the order of the sequence and where fe is the sampling frequency. It is up to the experimenter to choose a couple of values (sequence order, fe) sufficient to have all the useful decrease of the response.
- the sound spatialization device described above makes it possible to increase the intelligibility of the sound sources which it processes, to decrease the reaction time of the operator with respect to alarm, alert or warning signals. 'other sound indicators, the sources of which seem to be located respectively at different points in space, therefore easier to discriminate between them and easier to classify in order of importance or urgency.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FR9601740 | 1996-02-13 | ||
FR9601740A FR2744871B1 (fr) | 1996-02-13 | 1996-02-13 | Systeme de spatialisation sonore, et procede de personnalisation pour sa mise en oeuvre |
Publications (2)
Publication Number | Publication Date |
---|---|
EP0790753A1 true EP0790753A1 (de) | 1997-08-20 |
EP0790753B1 EP0790753B1 (de) | 2004-01-28 |
Family
ID=9489132
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP97400248A Expired - Lifetime EP0790753B1 (de) | 1996-02-13 | 1997-02-05 | System für Raumklangeffekt und Verfahren dafür |
Country Status (6)
Country | Link |
---|---|
US (1) | US5987142A (de) |
EP (1) | EP0790753B1 (de) |
JP (1) | JPH1042399A (de) |
CA (1) | CA2197166C (de) |
DE (1) | DE69727328T2 (de) |
FR (1) | FR2744871B1 (de) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1758336B (zh) * | 2001-07-10 | 2010-08-18 | 编码技术股份公司 | 用于低比特率音频编码应用的高效可标度参数立体声编码 |
FR2977335A1 (fr) * | 2011-06-29 | 2013-01-04 | France Telecom | Procede et dispositif de restitution de contenus audios |
FR3002205A1 (fr) * | 2013-08-14 | 2014-08-22 | Airbus Operations Sas | Systeme indicateur d'attitude d'un aeronef par spatialisation sonore tridimensionnelle |
Families Citing this family (44)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2744277B1 (fr) * | 1996-01-26 | 1998-03-06 | Sextant Avionique | Procede de reconnaissance vocale en ambiance bruitee, et dispositif de mise en oeuvre |
FR2765715B1 (fr) | 1997-07-04 | 1999-09-17 | Sextant Avionique | Procede de recherche d'un modele de bruit dans des signaux sonores bruites |
AUPP272898A0 (en) * | 1998-03-31 | 1998-04-23 | Lake Dsp Pty Limited | Time processed head related transfer functions in a headphone spatialization system |
FR2786107B1 (fr) | 1998-11-25 | 2001-02-16 | Sextant Avionique | Masque inhalateur d'oxygene avec dispositif de prise de son |
WO2001055833A1 (en) * | 2000-01-28 | 2001-08-02 | Lake Technology Limited | Spatialized audio system for use in a geographical environment |
JP4304845B2 (ja) * | 2000-08-03 | 2009-07-29 | ソニー株式会社 | 音声信号処理方法及び音声信号処理装置 |
WO2002052895A1 (de) * | 2000-12-22 | 2002-07-04 | Harman Audio Electronic Systems Gmbh | Anordnung zur auralisation eines lautsprechers in einem abhörraum bei beliebigen eingangssignalen |
GB2374502B (en) * | 2001-01-29 | 2004-12-29 | Hewlett Packard Co | Distinguishing real-world sounds from audio user interface sounds |
US20030227476A1 (en) * | 2001-01-29 | 2003-12-11 | Lawrence Wilcock | Distinguishing real-world sounds from audio user interface sounds |
GB2374507B (en) * | 2001-01-29 | 2004-12-29 | Hewlett Packard Co | Audio user interface with audio cursor |
GB2374506B (en) * | 2001-01-29 | 2004-11-17 | Hewlett Packard Co | Audio user interface with cylindrical audio field organisation |
GB0127776D0 (en) * | 2001-11-20 | 2002-01-09 | Hewlett Packard Co | Audio user interface with multiple audio sub-fields |
GB2372923B (en) * | 2001-01-29 | 2005-05-25 | Hewlett Packard Co | Audio user interface with selective audio field expansion |
US7346172B1 (en) * | 2001-03-28 | 2008-03-18 | The United States Of America As Represented By The United States National Aeronautics And Space Administration | Auditory alert systems with enhanced detectability |
US7079658B2 (en) * | 2001-06-14 | 2006-07-18 | Ati Technologies, Inc. | System and method for localization of sounds in three-dimensional space |
US6956955B1 (en) * | 2001-08-06 | 2005-10-18 | The United States Of America As Represented By The Secretary Of The Air Force | Speech-based auditory distance display |
FR2842064B1 (fr) * | 2002-07-02 | 2004-12-03 | Thales Sa | Systeme de spatialisation de sources sonores a performances ameliorees |
WO2004047489A1 (en) | 2002-11-20 | 2004-06-03 | Koninklijke Philips Electronics N.V. | Audio based data representation apparatus and method |
GB0419346D0 (en) * | 2004-09-01 | 2004-09-29 | Smyth Stephen M F | Method and apparatus for improved headphone virtualisation |
US7756281B2 (en) * | 2006-05-20 | 2010-07-13 | Personics Holdings Inc. | Method of modifying audio content |
JP4780119B2 (ja) | 2008-02-15 | 2011-09-28 | ソニー株式会社 | 頭部伝達関数測定方法、頭部伝達関数畳み込み方法および頭部伝達関数畳み込み装置 |
JP2009206691A (ja) | 2008-02-27 | 2009-09-10 | Sony Corp | 頭部伝達関数畳み込み方法および頭部伝達関数畳み込み装置 |
WO2009115299A1 (de) * | 2008-03-20 | 2009-09-24 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e. V. | Vorrichtung und verfahren zur akustischen anzeige |
JP5540581B2 (ja) | 2009-06-23 | 2014-07-02 | ソニー株式会社 | 音声信号処理装置および音声信号処理方法 |
JP5163685B2 (ja) * | 2010-04-08 | 2013-03-13 | ソニー株式会社 | 頭部伝達関数測定方法、頭部伝達関数畳み込み方法および頭部伝達関数畳み込み装置 |
JP5024418B2 (ja) * | 2010-04-26 | 2012-09-12 | ソニー株式会社 | 頭部伝達関数畳み込み方法および頭部伝達関数畳み込み装置 |
JP5533248B2 (ja) | 2010-05-20 | 2014-06-25 | ソニー株式会社 | 音声信号処理装置および音声信号処理方法 |
JP2012004668A (ja) | 2010-06-14 | 2012-01-05 | Sony Corp | 頭部伝達関数生成装置、頭部伝達関数生成方法及び音声信号処理装置 |
US9031256B2 (en) | 2010-10-25 | 2015-05-12 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for orientation-sensitive recording control |
US9552840B2 (en) | 2010-10-25 | 2017-01-24 | Qualcomm Incorporated | Three-dimensional sound capturing and reproducing with multi-microphones |
US8855341B2 (en) | 2010-10-25 | 2014-10-07 | Qualcomm Incorporated | Systems, methods, apparatus, and computer-readable media for head tracking based on recorded sound signals |
JP6065370B2 (ja) * | 2012-02-03 | 2017-01-25 | ソニー株式会社 | 情報処理装置、情報処理方法、及びプログラム |
US8929573B2 (en) * | 2012-09-14 | 2015-01-06 | Bose Corporation | Powered headset accessory devices |
WO2014073968A2 (en) * | 2012-11-09 | 2014-05-15 | Nederlandse Organisatie Voor Toegepast-Natuurwetenschappelijk Onderzoek Tno | Vehicle spacing control |
US10075795B2 (en) | 2013-04-19 | 2018-09-11 | Electronics And Telecommunications Research Institute | Apparatus and method for processing multi-channel audio signal |
WO2014171791A1 (ko) | 2013-04-19 | 2014-10-23 | 한국전자통신연구원 | 다채널 오디오 신호 처리 장치 및 방법 |
US9319819B2 (en) | 2013-07-25 | 2016-04-19 | Etri | Binaural rendering method and apparatus for decoding multi channel audio |
EP3090576B1 (de) | 2014-01-03 | 2017-10-18 | Dolby Laboratories Licensing Corporation | Verfahren und vorrichtung für die erstellung und die anwendung numerisch optimierter binauraler raumimpulsantworten |
CN105120419B (zh) * | 2015-08-27 | 2017-04-12 | 武汉大学 | 一种多声道系统效果增强方法及系统 |
WO2017135063A1 (ja) * | 2016-02-04 | 2017-08-10 | ソニー株式会社 | 音声処理装置、および音声処理方法、並びにプログラム |
US9832587B1 (en) | 2016-09-08 | 2017-11-28 | Qualcomm Incorporated | Assisted near-distance communication using binaural cues |
KR102283964B1 (ko) * | 2019-12-17 | 2021-07-30 | 주식회사 라온에이엔씨 | 인터콤시스템 통신명료도 향상을 위한 다채널다객체 음원 처리 장치 |
US11363402B2 (en) | 2019-12-30 | 2022-06-14 | Comhear Inc. | Method for providing a spatialized soundfield |
FR3110762B1 (fr) | 2020-05-20 | 2022-06-24 | Thales Sa | Dispositif de personnalisation d'un signal audio généré automatiquement par au moins un équipement matériel avionique d'un aéronef |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4700389A (en) * | 1985-02-15 | 1987-10-13 | Pioneer Electronic Corporation | Stereo sound field enlarging circuit |
FR2633125A1 (fr) * | 1988-06-17 | 1989-12-22 | Sgs Thomson Microelectronics | Appareil acoustique avec carte de filtrage vocal |
WO1990007172A1 (en) * | 1988-12-19 | 1990-06-28 | Honeywell Inc. | System and simulator for in-flight threat and countermeasures training |
WO1994001933A1 (en) * | 1992-07-07 | 1994-01-20 | Lake Dsp Pty. Limited | Digital filter having high accuracy and efficiency |
EP0664660A2 (de) * | 1990-01-19 | 1995-07-26 | Sony Corporation | Einrichtung zur Wiedergabe von Audiosignalen |
US5438623A (en) * | 1993-10-04 | 1995-08-01 | The United States Of America As Represented By The Administrator Of National Aeronautics And Space Administration | Multi-channel spatialization system for audio signals |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2652164A1 (fr) * | 1989-09-15 | 1991-03-22 | Thomson Csf | Procede de formation de voies pour sonar, notamment pour sonar remorque. |
EP0661906A1 (de) * | 1990-01-19 | 1995-07-05 | Sony Corporation | Kopfhörervorrichtung |
FR2700055B1 (fr) * | 1992-12-30 | 1995-01-27 | Sextant Avionique | Procédé de débruitage vectoriel de la parole et dispositif de mise en Óoeuvre. |
US5371799A (en) * | 1993-06-01 | 1994-12-06 | Qsound Labs, Inc. | Stereo headphone sound source localization system |
US5659619A (en) * | 1994-05-11 | 1997-08-19 | Aureal Semiconductor, Inc. | Three-dimensional virtual audio display employing reduced complexity imaging filters |
-
1996
- 1996-02-13 FR FR9601740A patent/FR2744871B1/fr not_active Expired - Fee Related
-
1997
- 1997-02-05 DE DE69727328T patent/DE69727328T2/de not_active Expired - Fee Related
- 1997-02-05 EP EP97400248A patent/EP0790753B1/de not_active Expired - Lifetime
- 1997-02-10 CA CA002197166A patent/CA2197166C/fr not_active Expired - Fee Related
- 1997-02-11 US US08/797,212 patent/US5987142A/en not_active Expired - Fee Related
- 1997-02-13 JP JP9029372A patent/JPH1042399A/ja active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4700389A (en) * | 1985-02-15 | 1987-10-13 | Pioneer Electronic Corporation | Stereo sound field enlarging circuit |
FR2633125A1 (fr) * | 1988-06-17 | 1989-12-22 | Sgs Thomson Microelectronics | Appareil acoustique avec carte de filtrage vocal |
WO1990007172A1 (en) * | 1988-12-19 | 1990-06-28 | Honeywell Inc. | System and simulator for in-flight threat and countermeasures training |
EP0664660A2 (de) * | 1990-01-19 | 1995-07-26 | Sony Corporation | Einrichtung zur Wiedergabe von Audiosignalen |
WO1994001933A1 (en) * | 1992-07-07 | 1994-01-20 | Lake Dsp Pty. Limited | Digital filter having high accuracy and efficiency |
US5438623A (en) * | 1993-10-04 | 1995-08-01 | The United States Of America As Represented By The Administrator Of National Aeronautics And Space Administration | Multi-channel spatialization system for audio signals |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1758336B (zh) * | 2001-07-10 | 2010-08-18 | 编码技术股份公司 | 用于低比特率音频编码应用的高效可标度参数立体声编码 |
FR2977335A1 (fr) * | 2011-06-29 | 2013-01-04 | France Telecom | Procede et dispositif de restitution de contenus audios |
FR3002205A1 (fr) * | 2013-08-14 | 2014-08-22 | Airbus Operations Sas | Systeme indicateur d'attitude d'un aeronef par spatialisation sonore tridimensionnelle |
Also Published As
Publication number | Publication date |
---|---|
DE69727328T2 (de) | 2004-10-21 |
FR2744871A1 (fr) | 1997-08-14 |
EP0790753B1 (de) | 2004-01-28 |
US5987142A (en) | 1999-11-16 |
CA2197166A1 (fr) | 1997-08-14 |
FR2744871B1 (fr) | 1998-03-06 |
JPH1042399A (ja) | 1998-02-13 |
DE69727328D1 (de) | 2004-03-04 |
CA2197166C (fr) | 2005-08-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP0790753B1 (de) | System für Raumklangeffekt und Verfahren dafür | |
US10433098B2 (en) | Apparatus and method for generating a filtered audio signal realizing elevation rendering | |
EP3320692B1 (de) | Räumliche audioverarbeitungsvorrichtung | |
EP2898707B1 (de) | Optimierte kalibrierung eines klangwiedergabesystems mit mehreren lautsprechern | |
US10334357B2 (en) | Machine learning based sound field analysis | |
EP3441966A1 (de) | System und verfahren zur ermittlung von audiocontext in augmented reality anwendungen | |
EP1992198B1 (de) | Optimierung des binauralen raumklangeffektes durch mehrkanalkodierung | |
US9237398B1 (en) | Motion tracked binaural sound conversion of legacy recordings | |
CN104756526A (zh) | 信号处理装置、信号处理方法、测量方法及测量装置 | |
BR112015014380B1 (pt) | Filtro e método para filtragem espacial informada utilizando múltiplas estimativas da direção de chegada instantânea | |
EP1658755B1 (de) | Tonquelle-raumklangssystem | |
JP2007158731A (ja) | 収音・再生方法および装置 | |
EP3260875A1 (de) | Automobilprüfsystem, verfahren und computerprogrammprodukt | |
EP2042000B1 (de) | Fehlerdiagnosevorrichtung und -methode für ein beschallungssystem | |
EP1502475B1 (de) | Verfahren und system zum repräsentieren eines schallfeldes | |
FR2858403A1 (fr) | Systeme et procede de determination d'une representation d'un champ acoustique | |
FR2805433A1 (fr) | Procede et dispositif de comparaison de signaux pour le controle de transducteurs et systeme de controle de transducteurs | |
JP5867799B2 (ja) | 収音再生装置、プログラム及び収音再生方法 | |
FR3112017A1 (fr) | Equipement électronique comprenant un simulateur de distorsion | |
US20240137720A1 (en) | Generating restored spatial audio signals for occluded microphones | |
EP0356327B1 (de) | Einrichtung zur Aufnahme von Schallsignalen mit Störgeräuschunterdrückung | |
Iida et al. | Acoustic VR System | |
WO2024086449A1 (en) | Generating restored spatial audio signals for occluded microphones | |
WO2023043963A1 (en) | Systems and methods for efficient and accurate virtual accoustic rendering | |
JP2023122230A (ja) | 音響信号処理装置、および、プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): DE GB NL |
|
17P | Request for examination filed |
Effective date: 19980115 |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: THOMSON-CSF SEXTANT |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: THALES AVIONICS S.A. |
|
17Q | First examination report despatched |
Effective date: 20021203 |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): DE GB NL |
|
REG | Reference to a national code |
Ref country code: GB Ref legal event code: FG4D Free format text: NOT ENGLISH |
|
REF | Corresponds to: |
Ref document number: 69727328 Country of ref document: DE Date of ref document: 20040304 Kind code of ref document: P |
|
GBT | Gb: translation of ep patent filed (gb section 77(6)(a)/1977) | ||
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed |
Effective date: 20041029 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: NL Payment date: 20090203 Year of fee payment: 13 Ref country code: DE Payment date: 20090129 Year of fee payment: 13 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20090204 Year of fee payment: 13 |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: V1 Effective date: 20100901 |
|
GBPC | Gb: european patent ceased through non-payment of renewal fee |
Effective date: 20100205 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: NL Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20100901 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20100901 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20100205 |