EP1226454A1 - Method to determine whether an acoustic source is near or far from a pair of microphones - Google Patents
Method to determine whether an acoustic source is near or far from a pair of microphonesInfo
- Publication number
- EP1226454A1 EP1226454A1 EP00976703A EP00976703A EP1226454A1 EP 1226454 A1 EP1226454 A1 EP 1226454A1 EP 00976703 A EP00976703 A EP 00976703A EP 00976703 A EP00976703 A EP 00976703A EP 1226454 A1 EP1226454 A1 EP 1226454A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- microphones
- acoustic signal
- direct
- acoustic
- reverberant
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000000034 method Methods 0.000 title claims abstract description 31
- 238000004590 computer program Methods 0.000 claims abstract description 10
- 230000006870 function Effects 0.000 description 10
- 238000010586 diagram Methods 0.000 description 8
- 238000012545 processing Methods 0.000 description 6
- 230000000875 corresponding effect Effects 0.000 description 4
- 230000004807 localization Effects 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 239000000872 buffer Substances 0.000 description 2
- 238000003491 array Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 230000001052 transient effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S11/00—Systems for determining distance or velocity not using reflection or reradiation
- G01S11/14—Systems for determining distance or velocity not using reflection or reradiation using ultrasonic, sonic, or infrasonic waves
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
- H04R3/005—Circuits for transducers, loudspeakers or microphones for combining the signals of two or more microphones
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L2021/02082—Noise filtering the noise being echo, reverberation of the speech
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L21/0216—Noise filtering characterised by the method used for estimating noise
- G10L2021/02161—Number of inputs available containing the signal or the noise to be suppressed
- G10L2021/02166—Microphone arrays; Beamforming
Definitions
- This invention relates to systems and methods for estimating the distance of the source of an acoustic signal within a reverberant space.
- the inventor has determined that it would be desirable to be able to estimate the range of a sound source independently of its inherent loudness using only two microphones.
- the present invention provides a system and method for determining the range of an acoustic signal within a reverberant space that avoids the limitations of prior techniques.
- the invention includes a method, apparatus, and computer program to determine whether a sound source is situated near or far from a pair of microphones situated in a reverberant space.
- the input signal may represent human speech, but it should be recognized that the invention could be used to localize any type of acoustic data, such as musical instruments and transient noise.
- the preferred embodiment of the invention localizes input signals as follows. Sound input signals from a pair of microphones are digitized into binary data. A signal detector is applied to the data; only the data that passes the signal detector is processed any further. The signal at the two microphones is compared to obtain the angular distribution of acoustic power. The concentration of acoustic power in the direct path versus reverberant paths is determined and used to compute a direct-to-reverberant ratio. If this ratio is greater than a threshold, the source is determined to be near the microphones. Otherwise, the source is determined to be far from the microphones, and ignored by the system.
- the invention includes a method for estimating the distance of an acoustic signal within a reverberant space utilizing two microphones, including: optionally passing the acoustic signal through a signal detector to discriminate against noises that are not of the same class as the acoustic signal; determining the angular distribution of acoustic power from the acoustic signal with respect to the two microphones; estimating the direct-to-reverberant ratio from the angular distribution of acoustic power; optionally passing the direct-to-reverberant ratio through a threshold detector; and outputting an indication as to whether the distance of the acoustic signal to the pair of microphones is near or far based on the output of the threshold detector.
- the invention has the following advantages: since the direct-to-reverberant ratio is independent of source volume or loudness, the range estimate will also be independent of volume; and, the invention needs only two microphones to work.
- FIG. 1 is block diagram of a prior art programmable computer system suitable for implementing the signal enhancement technique of the invention.
- FIG. 2 is a diagram that depicts direct and reflected sound rays for a source near the a pair of microphones.
- FIG. 3 is a diagram that depicts direct and reflected sound rays for a source far from a pair of microphones.
- FIG. 4 is a flow diagram showing the basic method of the preferred embodiment of the invention. Like reference numbers and designations in the various drawings indicate like elements.
- FIG. 1 shows a block diagram of a typical prior art programmable processing system that may be used for implementing the signal enhancement system of the invention.
- An acoustic signal is received at a pair of transducer microphones 10, which each generate a corresponding electrical signal representation of the acoustic signal.
- the signal from the transducer microphones 10 is then preferably amplified by corresponding amplifiers 12 before being digitized by corresponding analog-to-digital converters 14.
- the output of each analog-to-digital converter 14 is applied to a processing system that applies the enhancement techniques of the invention.
- the processing system preferably includes a CPU 16, RAM 20, ROM 18 (which may be writable, such as a flash ROM), coupled by a CPU bus 22 as shown.
- the output of the localization process can be applied to other processing systems, such as an automatic speech recognition system, or transmitted for the benefit of a remote listener, or captured by a recording system.
- An optional component of the invention is a signal detector function. This step allows the system to discriminate against noises that are not of the same class as the signal. For speaker localization, a voice detector is applied at this step.
- the next functional component of the system is the determination of the distribution of acoustic power as a function of angle relative to the two microphones 10. This is done by comparing the time difference (time lag) between the input from the two microphones 10, because the time lag corresponds to an angle of incidence of acoustic power.
- time lag corresponds to an angle of incidence of acoustic power.
- the acoustic power as a function of angle is estimated by measuring acoustic power as a function of time lag.
- a direct-to-reverberant ratio is then determined from the angular acoustic power distribution. This is done by assuming that the power at or near the peak of the angular distribution is from the direct path between the sound source and the microphones 10, and that the rest of the angular power distribution comes from the reverberant paths originating at the source and around a space (e.g., a room) confining or enclosing the microphones 10. Most of those paths will arrive at the microphones 10 from angles different from that of the direct path.
- the ratio of the power between the direct path and the rest of the power distribution is a measure of the DTR.
- the DTR can be used to estimate source range.
- FIG. 2 is a diagram that depicts direct and reflected sound rays for a source near the a pair of microphones.
- FIG. 3 is a diagram that depicts direct and reflected sound rays for a source far from a pair of microphones.
- a sound source 202 emits sound along a direct path 204 toward a pair of microphones 210, and along multiple indirect paths 206 such that the sound first reflects from the walls 208 of a reverberant space before being directed toward the microphone pair 210.
- the sound along the direct path 204 is relatively loud compared to the sound from the reflected paths 206.
- FIG. 3 represents the analogous situation when a sound source 302 is situated far from a microphone pair 310.
- the sound source 202 emits sound along a direct path 304 toward the microphone pair 310, and along multiple indirect paths 306 such that the sound first reflects from the walls 308 of a reverberant space before being directed toward the microphone pair 310.
- the sound along the direct path 304 is relatively weak compared to the sound from the reflected paths 306.
- the DTR value may be compared against a preset but adjustable threshold. If the DTR value is greater than the threshold, the sound source is determined to be in the distant field. On the other hand, if the DTR value is smaller than the threshold, the sound source is determined to be in the near field. An indication is output as to whether the distance of the sound source to the pair of microphones is near or far.
- FIG. 4 is a flow diagram of the preferred method embodiment of the invention.
- the method shown in FIG. 4 is used for determining the range of an incoming acoustic signal, which consists of a plurality of data samples generated as output from the analog-to-digital converters 14 shown in FIG. 1.
- the method begins at a Start state (Step 402).
- the incoming data stream is read into a computer memory as a set of samples (Step 404).
- the invention normally would be applied to enhance a "moving window" of data representing portions of a continuous acoustic data stream, such that the entire data stream is processed.
- an acoustic data stream to be enhanced is represented as a series of data "buffers" of fixed length, regardless of the duration of the original acoustic data stream.
- a typical practical buffer length is 1024 data points.
- the data is then optionally applied to a signal detector function (Step 406).
- This step allows the system to discriminate against noises that are not of the same class (i.e., relative volume) as the desired signal.
- a voice detector is applied at this step. Examples of such voice detectors are described in U.S. Patent Application Serial No.
- the data that has passed through the signal detector is then used to determine the distribution of acoustic power as a function of angle (Step 408).
- the preferred embodiment uses the sound localization method described in U.S. Patent No. 5,526,433, the teachings of which are hereby incorporated by reference.
- the output of the sound localization is a distribution P( ⁇ ) of power as a function of time lag ⁇ .
- the DTR is then computed from the acoustic power distribution P(r) (Step 410).
- a preferred embodiment of a DTR estimator is described below.
- DTR P P r .
- the threshold value may be user selected, based on empirical experience. In the alternative, other techniques may be used to generate an indication as to whether the distance of the acoustic signal to the pair of microphones is near or far based on the direct-to- reverberant ratio. For example, a training phase could be used for the system to set the threshold while the user speaks while being alternatively close and far from the microphones.
- Step 418 If any of the input data remains to be processed (Step 418), then the entire process is repeated on a next sample of acoustic data (Step 404). Otherwise, processing ends (Step 420).
- the final output is a decision or indication as to whether the distance of the sound source to the pair of microphones is near or far. Such information is useful for deciding whether to process voiced commands (e.g., at a kiosk in an airport), or transmit voice to a remote listener in a hands-free communication system.
- the invention may be implemented in hardware or software, or a combination of both
- the algorithms included as part of the invention are not inherently related to any particular computer or other apparatus.
- various general-purpose machines may be used with programs written in accordance with the teachings herein, or it may be more convenient to construct more specialized apparatus to perform the required method steps.
- the invention is implemented in one or more computer programs executing on programmable systems each comprising at least one processor, at least one data storage system (including volatile and non-volatile memory and/or storage elements), and at least two microphone inputs.
- the program code is executed on the processors to perform the functions described herein.
- Each such program may be implemented in any desired computer language (including machine, assembly, high level procedural, or object oriented programming languages) to communicate with a computer system.
- the language may be a compiled or interpreted language.
- Each such computer program is preferably stored on a storage media or device (e.g. , solid state, magnetic or optical media) readable by a general or special purpose programmable computer, for configuring and operating the computer when the storage media or device is read by the computer to perform the procedures described herein.
- the inventive system may also be considered to be implemented as a computer-readable storage medium, configured with a computer program, where the storage medium so configured causes a computer to operate in a specific and predefined manner to perform the functions described herein.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Otolaryngology (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
- Circuit For Audible Band Transducer (AREA)
- Measurement Of Mechanical Vibrations Or Ultrasonic Waves (AREA)
Abstract
Description
Claims
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US434781 | 1995-05-04 | ||
US09/434,781 US6243322B1 (en) | 1999-11-05 | 1999-11-05 | Method for estimating the distance of an acoustic signal |
PCT/US2000/029862 WO2001035118A1 (en) | 1999-11-05 | 2000-10-30 | Method to determine whether an acoustic source is near or far from a pair of microphones |
Publications (1)
Publication Number | Publication Date |
---|---|
EP1226454A1 true EP1226454A1 (en) | 2002-07-31 |
Family
ID=23725677
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP00976703A Withdrawn EP1226454A1 (en) | 1999-11-05 | 2000-10-30 | Method to determine whether an acoustic source is near or far from a pair of microphones |
Country Status (6)
Country | Link |
---|---|
US (1) | US6243322B1 (en) |
EP (1) | EP1226454A1 (en) |
JP (1) | JP2003514412A (en) |
AU (1) | AU1443901A (en) |
CA (1) | CA2390287C (en) |
WO (1) | WO2001035118A1 (en) |
Families Citing this family (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2364121B (en) | 2000-06-30 | 2004-11-24 | Mitel Corp | Method and apparatus for locating a talker |
JP4247002B2 (en) * | 2003-01-22 | 2009-04-02 | 富士通株式会社 | Speaker distance detection apparatus and method using microphone array, and voice input / output apparatus using the apparatus |
DE602004020872D1 (en) | 2003-02-25 | 2009-06-10 | Oticon As | T IN A COMMUNICATION DEVICE |
EP1620748A1 (en) * | 2003-04-22 | 2006-02-01 | Koninklijke Philips Electronics N.V. | Object position estimation system, apparatus and method |
US20040213415A1 (en) * | 2003-04-28 | 2004-10-28 | Ratnam Rama | Determining reverberation time |
WO2004097350A2 (en) * | 2003-04-28 | 2004-11-11 | The Board Of Trustees Of The University Of Illinois | Room volume and room dimension estimation |
GB0324536D0 (en) * | 2003-10-21 | 2003-11-26 | Mitel Networks Corp | Detecting acoustic echoes using microphone arrays |
CN1930489A (en) * | 2004-03-09 | 2007-03-14 | 皇家飞利浦电子股份有限公司 | Object position estimation |
EP1581026B1 (en) * | 2004-03-17 | 2015-11-11 | Nuance Communications, Inc. | Method for detecting and reducing noise from a microphone array |
CA2593523C (en) * | 2005-01-05 | 2014-04-08 | Temple University - Of The Commonwealth System Of Higher Education | Treatment of drug-resistant proliferative disorders |
US8180067B2 (en) * | 2006-04-28 | 2012-05-15 | Harman International Industries, Incorporated | System for selectively extracting components of an audio input signal |
US8036767B2 (en) | 2006-09-20 | 2011-10-11 | Harman International Industries, Incorporated | System for extracting and changing the reverberant content of an audio input signal |
JP4854533B2 (en) * | 2007-01-30 | 2012-01-18 | 富士通株式会社 | Acoustic judgment method, acoustic judgment device, and computer program |
ATE453910T1 (en) * | 2007-02-06 | 2010-01-15 | Oticon As | ESTIMATION OF YOUR OWN VOICE ACTIVITY WITH A HEARING AID SYSTEM BASED ON THE RATIO BETWEEN DIRECT SOUND AND REBREAKING |
FR2948484B1 (en) * | 2009-07-23 | 2011-07-29 | Parrot | METHOD FOR FILTERING NON-STATIONARY SIDE NOISES FOR A MULTI-MICROPHONE AUDIO DEVICE, IN PARTICULAR A "HANDS-FREE" TELEPHONE DEVICE FOR A MOTOR VEHICLE |
KR101387195B1 (en) * | 2009-10-05 | 2014-04-21 | 하만인터내셔날인더스트리스인코포레이티드 | System for spatial extraction of audio signals |
CN102549661B (en) * | 2009-10-21 | 2013-10-09 | 松下电器产业株式会社 | Sound processing apparatus, sound processing method and hearing aid |
EP2362238B1 (en) | 2010-02-26 | 2014-06-04 | Honda Research Institute Europe GmbH | Estimating the distance from a sensor to a sound source |
EP2410769B1 (en) | 2010-07-23 | 2014-10-22 | Sony Ericsson Mobile Communications AB | Method for determining an acoustic property of an environment |
US10015589B1 (en) | 2011-09-02 | 2018-07-03 | Cirrus Logic, Inc. | Controlling speech enhancement algorithms using near-field spatial statistics |
JP5834948B2 (en) * | 2012-01-24 | 2015-12-24 | 富士通株式会社 | Reverberation suppression apparatus, reverberation suppression method, and computer program for reverberation suppression |
US8965033B2 (en) * | 2012-08-31 | 2015-02-24 | Sonos, Inc. | Acoustic optimization |
CN104010265A (en) * | 2013-02-22 | 2014-08-27 | 杜比实验室特许公司 | Audio space rendering device and method |
RU2648604C2 (en) * | 2013-02-26 | 2018-03-26 | Конинклейке Филипс Н.В. | Method and apparatus for generation of speech signal |
DE102013207149A1 (en) * | 2013-04-19 | 2014-11-06 | Siemens Medical Instruments Pte. Ltd. | Controlling the effect size of a binaural directional microphone |
US9618618B2 (en) * | 2014-03-10 | 2017-04-11 | Elwha Llc | Systems and methods for ultrasonic position and motion detection |
US9739883B2 (en) | 2014-05-16 | 2017-08-22 | Elwha Llc | Systems and methods for ultrasonic velocity and acceleration detection |
US9995823B2 (en) | 2015-07-31 | 2018-06-12 | Elwha Llc | Systems and methods for utilizing compressed sensing in an entertainment system |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4312053A (en) * | 1971-12-03 | 1982-01-19 | Subcom, Inc. | Range and depth detection system |
US3784747A (en) * | 1971-12-03 | 1974-01-08 | Bell Telephone Labor Inc | Speech suppression by predictive filtering |
JPH087260B2 (en) * | 1990-05-18 | 1996-01-29 | 運輸省船舶技術研究所長 | Method and device for detecting distance using propagation characteristics of fog whistle sound |
JP2985982B2 (en) * | 1991-09-27 | 1999-12-06 | 日本電信電話株式会社 | Sound source direction estimation method |
FR2687496B1 (en) * | 1992-02-18 | 1994-04-01 | Alcatel Radiotelephone | METHOD FOR REDUCING ACOUSTIC NOISE IN A SPEAKING SIGNAL. |
JPH06105400A (en) * | 1992-09-17 | 1994-04-15 | Olympus Optical Co Ltd | Three-dimensional space reproduction system |
US6002776A (en) * | 1995-09-18 | 1999-12-14 | Interval Research Corporation | Directional acoustic signal processor and method therefor |
US5673327A (en) * | 1996-03-04 | 1997-09-30 | Julstrom; Stephen D. | Microphone mixer |
JP3522954B2 (en) * | 1996-03-15 | 2004-04-26 | 株式会社東芝 | Microphone array input type speech recognition apparatus and method |
US5778082A (en) * | 1996-06-14 | 1998-07-07 | Picturetel Corporation | Method and apparatus for localization of an acoustic source |
US6160758A (en) * | 1996-06-28 | 2000-12-12 | Scientific Innovations, Inc. | Utilization of auto and cross-correlation functions in methods for locating a source of a primary signal and for localizing signals |
JPH1164090A (en) * | 1997-08-20 | 1999-03-05 | Fujitsu Ltd | Sound source detector |
JPH11304906A (en) * | 1998-04-20 | 1999-11-05 | Nippon Telegr & Teleph Corp <Ntt> | Sound-source estimation device and its recording medium with recorded program |
-
1999
- 1999-11-05 US US09/434,781 patent/US6243322B1/en not_active Expired - Lifetime
-
2000
- 2000-10-30 CA CA002390287A patent/CA2390287C/en not_active Expired - Lifetime
- 2000-10-30 JP JP2001536596A patent/JP2003514412A/en active Pending
- 2000-10-30 EP EP00976703A patent/EP1226454A1/en not_active Withdrawn
- 2000-10-30 WO PCT/US2000/029862 patent/WO2001035118A1/en active Application Filing
- 2000-10-30 AU AU14439/01A patent/AU1443901A/en not_active Abandoned
Non-Patent Citations (1)
Title |
---|
See references of WO0135118A1 * |
Also Published As
Publication number | Publication date |
---|---|
AU1443901A (en) | 2001-06-06 |
CA2390287C (en) | 2008-03-25 |
WO2001035118A1 (en) | 2001-05-17 |
US6243322B1 (en) | 2001-06-05 |
JP2003514412A (en) | 2003-04-15 |
CA2390287A1 (en) | 2001-05-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2390287C (en) | Acoustic source range detection system | |
RU2596592C2 (en) | Spatial audio processor and method of providing spatial parameters based on acoustic input signal | |
CN102077274B (en) | Multi-microphone voice activity detector | |
CA2011775C (en) | Method of detecting acoustic signal | |
EP1208563B1 (en) | Noisy acoustic signal enhancement | |
US6952672B2 (en) | Audio source position detection and audio adjustment | |
CN110675887B (en) | Multi-microphone switching method and system for conference system | |
EP3757993B1 (en) | Pre-processing for automatic speech recognition | |
TW202147862A (en) | Robust speaker localization in presence of strong noise interference systems and methods | |
US20230026347A1 (en) | Methods for reducing error in environmental noise compensation systems | |
US20080120100A1 (en) | Method For Detecting Target Sound, Method For Detecting Delay Time In Signal Input, And Sound Signal Processor | |
JP2017181899A (en) | Noise suppression device, speech recognition device, noise suppression method, and noise suppression program | |
EP1229517B1 (en) | Method for recognizing speech with noise-dependent variance normalization | |
CN111383629A (en) | Voice processing method and device, electronic equipment and storage medium | |
JP5691224B2 (en) | Parameter setting method and sound processing apparatus | |
JP2913105B2 (en) | Sound signal detection method | |
KR101073632B1 (en) | A zero-crossing-based multiple source localization apparatus in reverberant environments | |
JP6361360B2 (en) | Reverberation judgment device and program | |
KR19980037008A (en) | Remote speech input and its processing method using microphone array | |
WO2020230460A1 (en) | Information processing device, information processing system, information processing method, and program | |
JPH05183997A (en) | Automatic discriminating device with effective sound | |
JP3332144B2 (en) | Target sound source area detection method and apparatus | |
JPH09127982A (en) | Voice recognition device | |
KR20230066056A (en) | Method and device for classification of uncorrelated stereo content, cross-talk detection and stereo mode selection in sound codec | |
JP2009229583A (en) | Signal detection method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20020507 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LI LU MC NL PT SE |
|
AX | Request for extension of the european patent |
Free format text: AL;LT;LV;MK;RO;SI |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: WAVEMAKERS INC. |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: HARMAN BECKER AUTOMOTIVE SYSTEMS-WAVEMAKERS, INC. |
|
17Q | First examination report despatched |
Effective date: 20060915 |
|
RTI1 | Title (correction) |
Free format text: ACOUSTIC SOURCE NEAR/FAR DETERMINATION SYSTEM |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: QNX SOFTWARE SYSTEMS (WAVEMAKERS), INC. |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20090108 |