JP2006227328A - Sound processor - Google Patents

Sound processor Download PDF

Info

Publication number
JP2006227328A
JP2006227328A JP2005041533A JP2005041533A JP2006227328A JP 2006227328 A JP2006227328 A JP 2006227328A JP 2005041533 A JP2005041533 A JP 2005041533A JP 2005041533 A JP2005041533 A JP 2005041533A JP 2006227328 A JP2006227328 A JP 2006227328A
Authority
JP
Japan
Prior art keywords
sound source
sound
signal
unit
band
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2005041533A
Other languages
Japanese (ja)
Other versions
JP4462063B2 (en
JP2006227328A5 (en
Inventor
Masato Togami
真人 戸上
Akio Amano
明雄 天野
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hitachi Ltd
Original Assignee
Hitachi Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hitachi Ltd filed Critical Hitachi Ltd
Priority to JP2005041533A priority Critical patent/JP4462063B2/en
Publication of JP2006227328A publication Critical patent/JP2006227328A/en
Publication of JP2006227328A5 publication Critical patent/JP2006227328A5/ja
Application granted granted Critical
Publication of JP4462063B2 publication Critical patent/JP4462063B2/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Circuit For Audible Band Transducer (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)

Abstract

<P>PROBLEM TO BE SOLVED: To estimate directions of sound sources with high accuracy even when the number of sound sources is lager than the number of microphones. <P>SOLUTION: This sound processor has a sound source duplication judging part which judges whether a band divided signal which is obtained by performing a band division is a signal in which a plurality of sound sources are duplicated or a signal which consists of one sound source, and the sound processor can estimate directions of the sound sources with high accuracy by estimating the directions of the sound sources using only frequency components in which sound sources are not duplicated. <P>COPYRIGHT: (C)2006,JPO&NCIPI

Description

本発明は、例えば複数のマイクロホン素子で観測した音声や音楽や各種雑音が混合した信号から、目的とする音のみを復元する音源分離技術に属する。   The present invention belongs to a sound source separation technique for restoring only a target sound, for example, from a signal obtained by mixing voice, music, and various noises observed with a plurality of microphone elements.

従来より、マイクロホン素子を複数使って、音源の方向を推定する音源定位技術があった。従来の音源定位技術として、死角形成型音源定位技術と、音声のスパース性を利用した音源定位技術の二つが存在する。   Conventionally, there has been a sound source localization technique that estimates the direction of a sound source using a plurality of microphone elements. There are two conventional sound source localization technologies: a blind spot forming type sound source localization technology and a sound source localization technology using the sparseness of speech.

死角形成型音源定位技術は、判定対象の方向以外に存在する音源方向に死角を形成し、判定対象の方向の音のみを抽出することで、方向毎の音のパワーを算出する。そして、その方向毎の音のパワーから音源方向を推定する。死角形成型音源定位技術は、音源数がマイク数を下回る場合、高精度に音源方向を推定できることが知られている。(例えば、非特許文献1参照)。   The blind spot forming type sound source localization technique calculates the power of sound for each direction by forming a blind spot in the direction of the sound source that exists outside the direction of the determination target and extracting only the sound in the direction of the determination target. Then, the sound source direction is estimated from the sound power for each direction. It is known that the blind spot forming type sound source localization technology can estimate the sound source direction with high accuracy when the number of sound sources is less than the number of microphones. (For example, refer nonpatent literature 1).

音声のスパース性を利用した音源定位技術とは、同じ時間に複数の音源が同じ周波数成分を保持する確率は低いという前提に立ち、各帯域分割信号を、ある一つの方向に全て割り当てることで、方向毎の音のパワーを算出し、その方向毎の音のパワーから音源方向を推定する手法である(例えば、特許文献1参照)。   The sound source localization technology using the sparseness of speech is based on the premise that the probability that multiple sound sources hold the same frequency component at the same time is low, and assigns each band division signal all in one direction, This is a method of calculating the sound power for each direction and estimating the sound source direction from the sound power for each direction (see, for example, Patent Document 1).

特開2003−271167号公報JP 2003-271167 A

大賀寿郎, 山崎芳男, 金田豊, ”音響システムとディジタル処理,” 電子情報通信学会,pp.203-209,1995/3/25Toshiro Oga, Yoshio Yamazaki, Yutaka Kaneda, "Acoustic systems and digital processing," IEICE, pp.203-209, 1995/3/25

従来の死角形成型音源定位技術では、音源数がマイク数より多い場合に音源定位性能が劣化するという課題があった。
また従来の音声のスパース性を利用した音源定位技術は、死角形成型音源定位技術と比べ、音源数がマイク数より多い場合の音源定位性能の劣化は少ないが、原理的に同じ時間に同じ周波数成分を複数の音源が共有する確率が低いという前提が成立しない場合に、性能の劣化が生じる。
音源の中に、音楽などの音声以外の音源が含まれる場合、同じ時間に同じ周波数成分を複数の音源が共有する確率が高くなってしまい、性能の劣化が生じるという課題がある。
The conventional blind spot forming type sound source localization technology has a problem that the sound source localization performance deteriorates when the number of sound sources is larger than the number of microphones.
In addition, the conventional sound source localization technology that uses the sparsity of speech has less degradation in sound source localization performance when the number of sound sources is greater than the number of microphones, compared to the blind spot forming type sound source localization technology, but in principle the same frequency at the same time When the premise that the probability that a plurality of sound sources share a component is low is not satisfied, performance degradation occurs.
When a sound source other than sound such as music is included in the sound source, there is a problem that the probability that a plurality of sound sources share the same frequency component at the same time is increased, resulting in performance degradation.

帯域分割して得られる帯域分割信号が複数の音源が重複している信号であるか一つの音源だけからなる信号であるかを判定する音源重複判定部を持つ。
具体的には、チャネル毎に複数の周波数帯域に分割されたマイクロホンアレーからの信号から音源方向を推定する音源定位部と、推定された音源方向毎に上記帯域分割信号を強調する音源分離部と、強調された帯域分割信号と上記推定された音源方向の情報を用いて、帯域毎に複数または単数の音源からの信号であるか判定する音源重複判定部とを有し、単数の音源からの帯域分割信号と判断された信号を用いて音源探索を行う音声処理装置。
A sound source duplication determination unit that determines whether a band division signal obtained by band division is a signal in which a plurality of sound sources overlap or a signal composed of only one sound source.
Specifically, a sound source localization unit that estimates a sound source direction from a signal from a microphone array divided into a plurality of frequency bands for each channel, and a sound source separation unit that emphasizes the band division signal for each estimated sound source direction; A sound source duplication determination unit that determines whether the signal is from a plurality or a single sound source for each band using the emphasized band division signal and the information on the estimated sound source direction, and from a single sound source An audio processing apparatus that performs sound source search using a signal determined to be a band division signal.

本発明では、複数の音源が重複しているかどうかを判定し単一の音源が鳴っている帯域分割信号のみを音源定位に用いることで、複数の音源が重複し音源の方向情報が失われた帯域成分を使わない。よって従来技術と比べ、音声や音楽の鳴っている方向を高精度に知ることができる。   In the present invention, it is determined whether or not a plurality of sound sources are overlapped, and only the band division signal from which a single sound source is sounded is used for sound source localization, so that a plurality of sound sources are overlapped and sound source direction information is lost. Do not use band components. Therefore, compared with the prior art, it is possible to know the direction in which voice or music is being played with high accuracy.

本発明の実施の形態について図面を用いて説明する。図1は、本発明の音声処理装置の基本構成図である。マイクロホンアレイ1は複数のマイク構成され音信号を収録する。マイクロホンアレイ1で収録した音信号は帯域分割部2に送られる。帯域分割部2では、マイク毎に音信号を短時間フーリエ変換し、帯域分割信号に変換する。帯域分割部2で帯域分割された音信号は、音源定位部3に送られる。音源定位部3では、帯域ごとに音源方向を推定し、帯域ごとの音源方向を出力する。音源分離部4では、音源定位部3が出力した音源方向の音を帯域毎に強調し抽出し、出力する。音源重複量推定部5では、音源分離部4が出力する帯域毎の強調信号と帯域分割部2が出力する帯域分割信号から、帯域毎に信号が複数の音源が混合した信号であるか単一音源だけからなる信号であるかを判定する指標である音源重複量を計算する。   Embodiments of the present invention will be described with reference to the drawings. FIG. 1 is a basic configuration diagram of a speech processing apparatus according to the present invention. The microphone array 1 includes a plurality of microphones and records sound signals. The sound signal recorded by the microphone array 1 is sent to the band dividing unit 2. In the band dividing unit 2, the sound signal is subjected to a short-time Fourier transform for each microphone and converted into a band divided signal. The sound signal divided by the band dividing unit 2 is sent to the sound source localization unit 3. The sound source localization unit 3 estimates the sound source direction for each band and outputs the sound source direction for each band. The sound source separation unit 4 emphasizes and extracts the sound in the sound source direction output from the sound source localization unit 3 for each band, and outputs it. In the sound source overlap amount estimation unit 5, the signal is a signal obtained by mixing a plurality of sound sources for each band from the enhancement signal for each band output from the sound source separation unit 4 and the band division signal output from the band division unit 2. A sound source overlap amount, which is an index for determining whether the signal is composed of only sound sources, is calculated.

音源重複判定部6は、音源重複量推定部5が出力する音源重複量尺度から帯域毎に信号が複数の音源が混合した信号であるか単一音源だけからなる信号であるかを判定する。音源方向探索部7では、音源重複判定部6で単一音源だけからなる信号であると判定された帯域分割信号の音源方向だけを用いて、音源方向を探索し、探索した音源方向を出力する。音源方向の探索時に、複数の音源が混合した帯域分割信号を用いると、正しい音源方向を推定することが困難になる。音源方向探索部7は、単一音源だけからなる帯域分割信号のみを用いて、音源方向を探索することができ、高精度に音源方向を推定することが可能となる。尚、図1に示したマイクロホンアレイ以外の各部はコンピュータの制御部にプログラムを読み込むことによって実現される。又は、ハードウェア、ハードとソフトの協調処理によって実現されるものであってもよい。   The sound source overlap determination unit 6 determines whether the signal is a signal obtained by mixing a plurality of sound sources for each band or a signal composed of only a single sound source from the sound source overlap amount scale output by the sound source overlap amount estimation unit 5. The sound source direction search unit 7 searches for the sound source direction using only the sound source direction of the band division signal determined to be a signal consisting of only a single sound source by the sound source overlap determination unit 6, and outputs the searched sound source direction. . If a band division signal in which a plurality of sound sources are mixed is used when searching for the sound source direction, it is difficult to estimate the correct sound source direction. The sound source direction searching unit 7 can search the sound source direction using only the band division signal consisting of only a single sound source, and can estimate the sound source direction with high accuracy. Each unit other than the microphone array shown in FIG. 1 is realized by reading a program into a control unit of a computer. Alternatively, it may be realized by hardware, hardware and software cooperative processing.

各処理部ごとに処理の詳細を説明する。マイクロホンアレイ部1は音圧信号x(t)を観測する。
帯域分割部2はx(t)に短時間フーリエ変換を施し、x(f:τ)を得る。τは短時間フーリエ変換のフレームインデックスで、fは周波数である。音源方向dのみから音が伝達する場合、x(f:τ)=ad(f)S(f:τ)とできる。ad(f)は音の空間伝達特性と呼ばれ、音が空間を伝わってマイクまで伝播する際の、振幅の減衰と位相の遅延を考慮してモデル化することができる。
Details of the processing will be described for each processing unit. The microphone array unit 1 observes the sound pressure signal x (t).
The band dividing unit 2 performs a short-time Fourier transform on x (t) to obtain x (f: τ). τ is a frame index of the short-time Fourier transform, and f is a frequency. When sound is transmitted only from the sound source direction d, x (f: τ) = ad (f) S (f: τ) can be obtained. ad (f) is called a sound spatial transfer characteristic, and can be modeled in consideration of amplitude attenuation and phase delay when sound travels through space to the microphone.

Figure 2006227328
Figure 2006227328

はad(f)を振幅の減衰と位相の遅延を考慮してモデル化したものである。ここで、rd,iは音源dからマイクiまでの距離で、τd,iは、音源dから発せられた音がマイクiに到達するまでにかかる時間である。Dは音源数とする。 Is a model of ad (f), taking into account amplitude attenuation and phase delay. Here, rd, i is the distance from the sound source d to the microphone i, and τd, i is the time taken for the sound emitted from the sound source d to reach the microphone i. D is the number of sound sources.

Figure 2006227328
Figure 2006227328

は、帯域分割信号ごとに音源方向を求める式である。音源定位部3では、数2にしたがい、時間τ、周波数fごとに音源方向jτ:fを求める。ここで、Λは音源をサーチする音源方向集合である。 Is an expression for obtaining the sound source direction for each band-divided signal. The sound source localization unit 3 obtains the sound source direction jτ: f for each time τ and frequency f according to Equation 2. Here, Λ is a sound source direction set for searching for a sound source.

Figure 2006227328
Figure 2006227328

は、音源定位部3が推定した帯域分割信号ごとの音源方向から、その音源方向の信号を抽出する式である。音源分離部4では、数3にしたがい、帯域毎に、音源方向jτ:f方向の音だけ分離し抽出する。 Is an expression for extracting a signal in the sound source direction from the sound source direction for each band division signal estimated by the sound source localization unit 3. The sound source separation unit 4 separates and extracts only sound in the sound source direction jτ: f direction for each band according to Equation 3.

Figure 2006227328
Figure 2006227328

は、音源分離部4が分離出力した帯域分割信号ごとの音源方向と強調処理する前の信号の差を計算することで、抽出した音源以外の音源の重複量を計算する式である。音源重複量推定部5では、数4に従い、τ、fごとに音源重複量を推定する。x(f:τ)が単一の音源からなる場合、数4は−∞になる。またx(f:τ)が複数の音源からなる場合、数4は有限の値を取る。複数の音源が重複するほど、数4は大きい値を取る。音源重複判定部6では、τ、fごとに、音源重複量推定部5で推定した音源重複量が予め定められた値以下となる場合に、音源重複が無いと判定する。 Is an equation for calculating the overlap amount of sound sources other than the extracted sound source by calculating the difference between the sound source direction for each band division signal separated and output by the sound source separation unit 4 and the signal before enhancement processing. The sound source overlap amount estimation unit 5 estimates the sound source overlap amount for each τ and f according to Equation 4. When x (f: τ) consists of a single sound source, Equation 4 becomes −∞. When x (f: τ) is composed of a plurality of sound sources, Equation 4 takes a finite value. As multiple sound sources overlap, Equation 4 takes a larger value. The sound source duplication determination unit 6 determines that there is no sound source duplication when the sound source duplication amount estimated by the sound source duplication amount estimation unit 5 is equal to or less than a predetermined value for each of τ and f.

Figure 2006227328
Figure 2006227328

は、音源重複量が予め定められた値以下となった、帯域分割信号のみ利用した方向毎の音のパワーを計算する式である。j<Λであり、P(j)は方向ごとの音源パワースペクトルである。Pthは、音源重複量推定部5で用いる音源重複量のいき値である。音源方向探索部7では、音源重複判定部6が重複がないと判定したτ、fのx(f:τ)を用いて、数5を使って方向毎の音のパワーを計算し、方向毎の音のパワーから音源方向を推定する。P(j-1)<P(j)<P(j+1)が成立する音源パワースペクトルP(j)だけを取り出し、P(j)を降順に整列させる。その降順に整列させたP(j)のうち大きいものから予め定めた音源数分だけ取り出し、取り出したP(j)のjを音源方向として出力する。 Is a formula for calculating the power of sound in each direction using only the band-divided signal when the sound source overlap amount is equal to or less than a predetermined value. j <Λ, and P (j) is a sound source power spectrum for each direction. Pth is a threshold value of the sound source overlap amount used in the sound source overlap amount estimation unit 5. The sound source direction search unit 7 calculates the sound power for each direction using Equation 5 using x (f: τ) of τ and f determined by the sound source overlap determination unit 6 to be non-overlapping. The sound source direction is estimated from the power of the sound. Only the sound source power spectrum P (j) satisfying P (j-1) <P (j) <P (j + 1) is extracted, and P (j) is arranged in descending order. A predetermined number of sound sources are extracted from the largest P (j) arranged in descending order, and j of the extracted P (j) is output as the sound source direction.

図2は本発明の音声処理装置と妨害音抑圧装置を組み合わせた装置の構成図である。目的音抽出部8では、予め設定しておいた方向、もしくはユーザからの別途入力手段を介した指定方向から到来する音声成分を帯域分割部の出力信号から抽出する。P(j)のjの空間的な音源方向をD(j)とする。予め定める理想的な目的音方向D0とする。目的音抽出部8では、j_0=argmax| D(j)- D0|を推定目的音方向(以後目的音方向)とし、それ以外の音源方向を妨害音方向とする。以後、x(f:τ)=x(f)と短時間フーリエ変換のフレームインデックスを省略する。そして目的音方向に指向性を持つ二つの線形フィルタを入力信号にかける。   FIG. 2 is a block diagram of an apparatus that combines the speech processing apparatus of the present invention and the interference sound suppression apparatus. The target sound extraction unit 8 extracts a speech component coming from a preset direction or a designated direction via a separate input means from the user from the output signal of the band dividing unit. Let D (j) be the spatial sound source direction of P in P (j). An ideal target sound direction D0 determined in advance is used. The target sound extraction unit 8 sets j_0 = argmax | D (j) −D0 | as the estimated target sound direction (hereinafter referred to as the target sound direction) and the other sound source directions as the interference sound direction. Thereafter, x (f: τ) = x (f) and the frame index of the short-time Fourier transform are omitted. Then, two linear filters having directivity in the target sound direction are applied to the input signal.

ここで,Ωを妨害音集合とし、d0を目的音、S0(f)を目的音成分、N0(f)をd番目の妨害音成分とする。目的音方向に指向性を持つ二つの線形フィルタg,hを入力信号にかけた後の出力信号から、目的音のみを抽出した信号と、妨害音のみを抽出した信号を、   Here, Ω is an interference sound set, d0 is a target sound, S0 (f) is a target sound component, and N0 (f) is a d-th interference sound component. From the output signal after applying two linear filters g and h having directivity in the target sound direction to the input signal, a signal that extracts only the target sound and a signal that extracts only the interfering sound,

Figure 2006227328
Figure 2006227328

で算出することができる。y(1)(f)は目的音のみを抽出した信号で、y(2)(f)は妨害音のみを抽出した信号である。g,hは、各妨害音について少なくとも1つの線形フィルタが死角を形成する線形フィルタ対であり、かつ2つの線形フィルタの出力パワーの期待値の積が最も小さくなるような線形フィルタ対であるとする。これら二つの線形フィルタg,hの出力信号を用いて、 Can be calculated. y (1) (f) is a signal obtained by extracting only the target sound, and y (2) (f) is a signal obtained by extracting only the disturbing sound. g and h are a linear filter pair in which at least one linear filter forms a blind spot for each interference sound, and a linear filter pair in which the product of the expected values of the output power of the two linear filters is the smallest. To do. Using the output signals of these two linear filters g and h,

Figure 2006227328
Figure 2006227328

で、目的音だけを分離し、抽出する。数7は、目的音を強調した信号から、妨害音のみを抽出した信号をパワースペクトル領域で引くことで、目的音のパワーを高精度に復元するための式である。目的音抽出部8は、数7を使い、目的音を強調した信号を抽出する。 Then, only the target sound is separated and extracted. Equation 7 is an expression for restoring the power of the target sound with high accuracy by subtracting a signal obtained by extracting only the interfering sound from the signal in which the target sound is emphasized in the power spectrum region. The target sound extraction unit 8 uses Equation 7 to extract a signal that emphasizes the target sound.

波形生成部9では、目的音抽出部8で抽出した音声成分を逆フーリエ変換して波形を生成し、出力する。
この実施例の構成に基づく妨害音抑圧装置は、目的音や妨害音の方向が未知であり、妨害音数がマイク数より多い場合であっても、高精度に妨害音を抑圧することができる。
The waveform generation unit 9 generates a waveform by performing inverse Fourier transform on the voice component extracted by the target sound extraction unit 8 and outputs the waveform.
The interfering sound suppression device based on the configuration of this embodiment can suppress the interfering sound with high accuracy even when the direction of the target sound and the interfering sound is unknown and the number of interfering sounds is larger than the number of microphones. .

本発明の基本構成の一実施例を示す図。The figure which shows one Example of the basic composition of this invention. 本発明と妨害音抑圧手法を組み合わせた一実施例の詳細を表すブロック図。The block diagram showing the detail of one Example which combined this invention and the interference sound suppression method.

符号の説明Explanation of symbols

1・・・マイクロホンアレイ、2・・・帯域分割部、3・・・音源定位部、4・・・音源分離部、5・・・音源重複量推定部、6・・・音源重複判定部、7・・・音源方向探索部、8・・・目的音抽出部、9・・・波形生成部。 DESCRIPTION OF SYMBOLS 1 ... Microphone array, 2 ... Band division part, 3 ... Sound source localization part, 4 ... Sound source separation part, 5 ... Sound source duplication amount estimation part, 6 ... Sound source duplication determination part, 7 ... Sound source direction searching unit, 8 ... Target sound extracting unit, 9 ... Waveform generating unit.

Claims (3)

少なくとも2チャネル以上のマイクロホン素子を持つマイクロホンアレーと、
前記マイクロホンアレーからの信号をチャネル毎に複数の周波数帯域に分割する帯域分割部と、
前記帯域分割された帯域分割信号から音源方向を推定する音源定位部と、
上記推定された音源方向毎に上記帯域分割信号を強調する音源分離部と、
上記強調された帯域分割信号と上記推定された音源方向の情報を用いて、該帯域分割信号が複数または単数の音源からの信号であるか判定する音源重複判定部と、
上記単数の音源からの帯域分割信号と判断された信号を用いて音源探索を行う音源探索部とを有することを特徴とする音声処理装置。
A microphone array having microphone elements of at least two channels;
A band dividing unit for dividing a signal from the microphone array into a plurality of frequency bands for each channel;
A sound source localization unit that estimates a sound source direction from the band-divided band division signal;
A sound source separation unit that emphasizes the band division signal for each estimated sound source direction;
A sound source duplication determination unit that determines whether the band division signal is a signal from a plurality of or a single sound source using the emphasized band division signal and the information on the estimated sound source direction;
A sound processing apparatus comprising: a sound source search unit that performs sound source search using a signal determined to be a band division signal from the single sound source.
前記強調された音源帯域信号と前期帯域分割部から出力される音源帯域信号の差を計算し、該差の大きさを音源分離部が出力する信号の大きさで割った値を算出する音源重複量推定部をさらに有し、
前記音源重複判定部は、前記音源重複判定部で求められる値が予め設定された値以上であるかを判断することで上記判定を行うことを特徴とする請求項1記載の音声処理装置。
Calculate the difference between the emphasized sound source band signal and the sound source band signal output from the previous band dividing unit, and calculate the value obtained by dividing the difference by the magnitude of the signal output from the sound source separating unit A quantity estimation unit;
The sound processing apparatus according to claim 1, wherein the sound source overlap determination unit performs the determination by determining whether a value obtained by the sound source overlap determination unit is equal to or greater than a preset value.
上記音源方向探索部からの出力に基づいて、特定方向からの信号を上記帯域分割部から出力される音源帯域信号から抽出する目的音抽出部と、
上記抽出された音源帯域信号の波形を生成して出力する波形生成部とを用いることを特徴とする請求項1又は2に記載の音声処理装置。
Based on the output from the sound source direction search unit, a target sound extraction unit that extracts a signal from a specific direction from the sound source band signal output from the band dividing unit;
The sound processing apparatus according to claim 1, wherein a waveform generation unit that generates and outputs a waveform of the extracted sound source band signal is used.
JP2005041533A 2005-02-18 2005-02-18 Audio processing device Expired - Fee Related JP4462063B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2005041533A JP4462063B2 (en) 2005-02-18 2005-02-18 Audio processing device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2005041533A JP4462063B2 (en) 2005-02-18 2005-02-18 Audio processing device

Publications (3)

Publication Number Publication Date
JP2006227328A true JP2006227328A (en) 2006-08-31
JP2006227328A5 JP2006227328A5 (en) 2007-08-16
JP4462063B2 JP4462063B2 (en) 2010-05-12

Family

ID=36988748

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2005041533A Expired - Fee Related JP4462063B2 (en) 2005-02-18 2005-02-18 Audio processing device

Country Status (1)

Country Link
JP (1) JP4462063B2 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007233239A (en) * 2006-03-03 2007-09-13 National Institute Of Advanced Industrial & Technology Method, system, and program for utterance event separation
JP2008145610A (en) * 2006-12-07 2008-06-26 Univ Of Tokyo Sound source separation and localization method
JP2010020165A (en) * 2008-07-11 2010-01-28 Fujitsu Ltd Noise suppressing device, mobile phone, noise suppressing method and computer program
JP2010217773A (en) * 2009-03-18 2010-09-30 Yamaha Corp Signal processing device and program
JP2011139378A (en) * 2009-12-28 2011-07-14 Fujitsu Ltd Signal processing apparatus, microphone array device, signal processing method, and signal processing program
WO2023058162A1 (en) 2021-10-06 2023-04-13 マクセル株式会社 Audio augmented reality object playback device and audio augmented reality object playback method

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007233239A (en) * 2006-03-03 2007-09-13 National Institute Of Advanced Industrial & Technology Method, system, and program for utterance event separation
JP4565162B2 (en) * 2006-03-03 2010-10-20 独立行政法人産業技術総合研究所 Speech event separation method, speech event separation system, and speech event separation program
JP2008145610A (en) * 2006-12-07 2008-06-26 Univ Of Tokyo Sound source separation and localization method
JP2010020165A (en) * 2008-07-11 2010-01-28 Fujitsu Ltd Noise suppressing device, mobile phone, noise suppressing method and computer program
US9135924B2 (en) 2008-07-11 2015-09-15 Fujitsu Limited Noise suppressing device, noise suppressing method and mobile phone
JP2010217773A (en) * 2009-03-18 2010-09-30 Yamaha Corp Signal processing device and program
JP2011139378A (en) * 2009-12-28 2011-07-14 Fujitsu Ltd Signal processing apparatus, microphone array device, signal processing method, and signal processing program
WO2023058162A1 (en) 2021-10-06 2023-04-13 マクセル株式会社 Audio augmented reality object playback device and audio augmented reality object playback method

Also Published As

Publication number Publication date
JP4462063B2 (en) 2010-05-12

Similar Documents

Publication Publication Date Title
JP6584930B2 (en) Information processing apparatus, information processing method, and program
EP2355097B1 (en) Signal separation system and method
EP2633697B1 (en) Three-dimensional sound capturing and reproducing with multi-microphones
US20160351179A1 (en) Single-channel, binaural and multi-channel dereverberation
EP3189521B1 (en) Method and apparatus for enhancing sound sources
JP6019969B2 (en) Sound processor
JP2013518477A (en) Adaptive noise suppression by level cue
JP6485711B2 (en) Sound field reproduction apparatus and method, and program
JP4462063B2 (en) Audio processing device
JP6225245B2 (en) Signal processing apparatus, method and program
CN103428609A (en) Apparatus and method for removing noise
CN111863015A (en) Audio processing method and device, electronic equipment and readable storage medium
Chatterjee et al. ClearBuds: wireless binaural earbuds for learning-based speech enhancement
US9966081B2 (en) Method and apparatus for synthesizing separated sound source
JP2014126856A (en) Noise removal device and control method for the same
JP2007047427A (en) Sound processor
JP2007240605A (en) Sound source separating method and sound source separation system using complex wavelet transformation
JP2006227328A5 (en)
JP2004325127A (en) Sound source detection method, sound source separation method, and apparatus for executing them
JP2006178333A (en) Proximity sound separation and collection method, proximity sound separation and collecting device, proximity sound separation and collection program, and recording medium
Miyazaki et al. Theoretical analysis of parametric blind spatial subtraction array and its application to speech recognition performance prediction
JP7347520B2 (en) Sound signal processing device, sound signal processing method, and sound signal processing program
JP4249697B2 (en) Sound source separation learning method, apparatus, program, sound source separation method, apparatus, program, recording medium
JP2010217268A (en) Low delay signal processor generating signal for both ears enabling perception of direction of sound source
JP3787103B2 (en) Speech processing apparatus, speech processing method, speech processing program

Legal Events

Date Code Title Description
A521 Request for written amendment filed

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20070704

A621 Written request for application examination

Free format text: JAPANESE INTERMEDIATE CODE: A621

Effective date: 20070704

A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20100104

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20100126

A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20100208

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20130226

Year of fee payment: 3

R151 Written notification of patent or utility model registration

Ref document number: 4462063

Country of ref document: JP

Free format text: JAPANESE INTERMEDIATE CODE: R151

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20130226

Year of fee payment: 3

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20140226

Year of fee payment: 4

LAPS Cancellation because of no payment of annual fees