US20120310638A1 - Audio signal processing method, audio apparatus therefor, and electronic apparatus therefor - Google Patents
Audio signal processing method, audio apparatus therefor, and electronic apparatus therefor Download PDFInfo
- Publication number
- US20120310638A1 US20120310638A1 US13/483,571 US201213483571A US2012310638A1 US 20120310638 A1 US20120310638 A1 US 20120310638A1 US 201213483571 A US201213483571 A US 201213483571A US 2012310638 A1 US2012310638 A1 US 2012310638A1
- Authority
- US
- United States
- Prior art keywords
- sub
- signals
- channel
- audio
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000005236 sound signal Effects 0.000 title claims abstract description 180
- 238000003672 processing method Methods 0.000 title claims description 24
- 230000002194 synthesizing effect Effects 0.000 claims description 14
- 238000000034 method Methods 0.000 claims 2
- 238000010586 diagram Methods 0.000 description 8
- 238000004590 computer program Methods 0.000 description 2
- 238000013500 data storage Methods 0.000 description 2
- 238000005316 response function Methods 0.000 description 2
- 230000013707 sensory perception of sound Effects 0.000 description 2
- 230000003321 amplification Effects 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 238000003199 nucleic acid amplification method Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000005406 washing Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/008—Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R3/00—Circuits for transducers, loudspeakers or microphones
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/60—Receiver circuitry for the reception of television signals according to analogue transmission standards for the sound signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/20—Speech recognition techniques specially adapted for robustness in adverse environments, e.g. in noise, of stress induced speech
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L2021/02082—Noise filtering the noise being echo, reverberation of the speech
Definitions
- Exemplary embodiments relate to an audio signal processing method, an audio apparatus using the same, and an electronic apparatus using the same, and more particularly, to an audio signal processing method for receiving a multi-channel signal and outputting corresponding acoustic signals, an audio apparatus using the same, and an electronic apparatus using the same.
- Audio apparatuses may output received multi-channel signals through a plurality of speakers.
- the audio apparatuses may receive a voice signal corresponding to a voice of a user, recognize the received voice signal, and perform an operation corresponding to a command, operation, or request based on the recognized voice signal.
- the recognizing of a received voice signal and the performing of an operation corresponding to a command, operation, or request based on the recognized voice signal are called a voice recognition operation.
- an audio apparatus capable of performing the voice recognition operation, when a voice signal is received through a predetermined microphone while acoustic signals are being output through a plurality of speakers, the audio apparatus must acquire only the voice signal by removing the acoustic signals from among signals input to the microphone. Then, the audio apparatus recognizes the acquired voice signal.
- MCAEC Multi-Channel Acoustic Echo Cancellation
- the echo cancellation filter converges to a predetermined value, thereby effectively canceling the acoustic signals.
- the echo cancellation filter diverges without converging. Accordingly, since residual echo remains in a signal for voice recognition, the voice recognition operation cannot be effectively performed.
- the audio apparatus cannot correctly recognize the voice signal and cannot perform an operation or command corresponding to the voice signal.
- an audio apparatus for effectively lowering a correlation between acoustic signals output to a plurality of speakers is necessary.
- the exemplary embodiments provide an audio signal processing method for effectively lowering a correlation between acoustic signals output to a plurality of speakers, an audio apparatus therefor, and an electronic apparatus therefor.
- the exemplary embodiments also provide an audio signal processing method for correctly recognizing an input voice signal and an electronic apparatus therefor.
- an audio apparatus includes: a decorrelator for generating decorrelated signals by applying a phase shifting value adjusted based on a correlation difference between audio signals included in a multi-channel signal to the audio signals; and a speaker set including at least two speakers for outputting acoustic signals corresponding to the decorrelated signals.
- the decorrelator may generate the decorrelated signals by applying a phase shifting value adjusted based on a phase difference between the audio signals to the audio signals.
- the decorrelator may receive the multi-channel signal including first and second channel audio signals.
- the decorrelator may reduce the phase shifting value when a phase difference between the first and second channel audio signals is great.
- the decorrelator may set the phase shifting value to 0 when the phase difference between the first and second channel audio signals is 180°.
- the decorrelator may set the phase shifting value to a predetermined limit value when the phase difference between the first and second channel audio signals is 0°.
- the decorrelator may generate first and second decorrelated signals corresponding to the first and second channel audio signals by subtracting the phase shifting value from a phase of the first channel audio signal and adding the phase shifting value to a phase of the second channel audio signal, respectively.
- the decorrelator may include: a pre-processor for receiving the multi-channel signal, including the first and second channel audio signals, dividing the multi-channel signal into a plurality of sub-bands, and generating first and second channel sub-audio signals in each of the plurality of sub-bands; a decorrelation processor for generating first and second sub-decorrelated signals by applying a phase shifting value adjusted based on a phase difference between the first and second channel sub-audio signals to the first and second channel sub-audio signals; and a synthesizer for generating the first decorrelated signal by synthesizing the first sub-decorrelated signals in the plurality of sub-bands and generating the second decorrelated signal by synthesizing the second sub-decorrelated signals in the plurality of sub-bands.
- the decorrelation processor may generate the first and second sub-decorrelated signals by setting a corrected phase shifting value by multiplying the phase shifting value by a predetermined weighting value based on sub-bands, subtracting the corrected phase shifting value from a phase of the first channel sub-audio signal, and adding the corrected phase shifting value to a phase of the second channel sub-audio signal.
- the decorrelation processor may set the predetermined weighting value based on sub-bands.
- the decorrelation processor may set the predetermined weighting value in one sub-band and the predetermined weighting value in another sub-band adjacent to the one sub-band to have different signs.
- the decorrelation processor may divide the plurality of sub-bands into a plurality of groups and set the predetermined weighting value based on groups.
- the decorrelation processor may set the predetermined weighting value in one group and the predetermined weighting value in another group adjacent to the one group to have different signs.
- the audio apparatus may further include: an echo filter unit for generating echo signals corresponding to the decorrelated signals; a microphone for receiving the acoustic signals and a voice signal; and a processor for processing voice recognition on a subtraction signal obtained by subtracting the echo signals from an output signal of the microphone.
- an electronic apparatus includes: a decorrelator for generating decorrelated signals by applying a phase shifting value adjusted based on a correlation difference between audio signals included in a multi-channel signal to the audio signals; an echo filter unit for generating echo signals corresponding to the decorrelated signals; a speaker set including at least two speakers for outputting acoustic signals corresponding to the decorrelated signals; a microphone for receiving the acoustic signals and a voice signal; and a processor for processing voice recognition on a subtraction signal obtained by subtracting the echo signals from an output signal of the microphone.
- an audio signal processing method includes: receiving a multi-channel signal including audio signals; adjusting a phase shifting value based on a correlation difference between the audio signals; and generating decorrelated signals by applying the phase shifting value to the audio signals.
- FIG. 1 is a perspective view of an electronic apparatus
- FIG. 2 is a block diagram of an audio apparatus according to an aspect of exemplary embodiment
- FIG. 3 is a block diagram of an audio apparatus according to another aspect an exemplary embodiment
- FIG. 4 illustrates graphs for describing a phase shifting value adjustment performed by the audio apparatus according to one or more aspects of exemplary embodiments
- FIGS. 5A and 5B illustrate other graphs for describing a phase shifting value adjustment performed by the audio apparatus according to one or more aspects of exemplary embodiments
- FIG. 6 is a block diagram of an electronic apparatus according to an exemplary embodiment
- FIG. 7 is a block diagram of an electronic apparatus according to an exemplary embodiment
- FIG. 8 is a flowchart illustrating an audio signal processing method according to an exemplary embodiment
- FIG. 9 is a flowchart illustrating an audio signal processing method according to an exemplary embodiment.
- FIG. 10 is a flowchart for describing operation 930 of FIG. 9 in more detail.
- FIG. 1 is a perspective view of an electronic apparatus 100 .
- the electronic apparatus 100 shown in FIG. 1 is, for example, a digital TV.
- the electronic apparatus 100 may be any one of various electronic products, including an audio apparatus for outputting an audio signal.
- the electronic apparatus 100 may be a digital TV, an audio system, a refrigerator, a washing machine, a personal computer, a mobile computer, a pad-type terminal, or a mobile phone.
- the electronic apparatus 100 includes an audio apparatus (not shown) including a decorrelator (not shown) and a speaker set.
- the decorrelator may be included in a circuit block included in the electronic apparatus 100
- the speaker set may include at least two speakers, for example, first and second speakers 110 and 120 .
- the electronic apparatus 100 may further include at least one of a microphone 130 and a display unit 140 .
- the microphone 130 may include a microphone array (not shown) and may receive all audible audio signals. In detail, the microphone 130 may receive a voice signal generated by a user.
- the decorrelator adjusts and outputs a correlation among a plurality of audio signals included in a multi-channel signal.
- the correlation-adjusted audio signals are output through the first and second speakers 110 and 120 .
- the microphone 130 receives the voice signal in addition to the audio signals output from the first and second speakers 110 and 120 .
- the digital TV may output a Picture-In-Picture (PIP) screen on the display unit 140 .
- PIP Picture-In-Picture
- the display unit 140 may output broadcasting images on a main screen and output images of a video call on a sub-screen.
- the electronic apparatus 100 is a digital TV capable of performing a voice recognition control
- the digital TV may perform an operation corresponding to the input voice signal. That is, when the user inputs a voice signal into the microphone 130 , the electronic apparatus 100 may perform an operation corresponding to the input voice signal.
- the electronic apparatus 100 When a voice signal is received while audio signals are being output from the first and second speakers 110 and 120 , the electronic apparatus 100 must detect or discern only the voice signal and not the audio signals output from the first and second speakers 110 and 120 . To extract the voice signal from among all signals input to the microphone 130 , audio signals output from the speakers must be cancelled from among all the signals input to the microphone 130 by using an echo cancellation filter (not shown).
- the audio signals output from the first and second speakers 110 and 120 are input to the microphone 130 through various paths 111 , 112 , 113 , 121 , 122 , and 123 according to a surrounding environment where the electronic apparatus 100 is located.
- paths 111 , 112 , 113 , 121 , 122 , and 123 are called ‘actual echo paths’.
- the actual echo paths may vary according to the surrounding environment and noise when the audio signals are output, so it is difficult to clearly determine the actual echo paths.
- an impulse response function is designed by estimating echo paths corresponding to the actual echo paths (hereinafter, ‘estimated echo paths’) and reflecting the estimated echo paths.
- the echo cancellation filter may filter echo signals corresponding to the audio signals based on the designed impulse response function.
- the echo cancellation filter may be implemented using an echo filter unit 620 and an adder 671 , which will be described with reference to FIG. 6 .
- the electronic apparatus 100 may receive the multi-channel signal including a right channel (R-channel) audio signal and a left channel (L-channel) audio signal from the outside or may itself generate the multi-channel signal.
- the speaker set includes the first and second speakers 110 and 120
- the first speaker 110 may output the R-channel audio signal
- the second speaker 120 may output the L-channel audio signal.
- FIG. 2 is a block diagram of an audio apparatus 200 according to an aspect of an exemplary embodiment.
- the audio apparatus 200 includes a decorrelator 210 and a speaker set 250 .
- the decorrelator 210 generates decorrelated signals Sout corresponding to audio signals included in a multi-channel signal Sm by reducing a correlation between the audio signals
- the decorrelator 210 receives the multi-channel signal Sm and generates the decorrelated signals Sout by applying a phase shifting value adjusted based on a correlation difference between the audio signals included in the multi-channel signal Sm to the audio signals.
- the multi-channel signal Sm includes a plurality of audio signals corresponding to a plurality of channels.
- the multi-channel signal Sm may include an R-channel audio signal and an L-channel audio signal.
- the multi-channel signal Sm may include 5.1-channel audio signals, 7.1-channel audio signals, or 10.2-channel audio signals.
- the speaker set 250 includes at least two speakers for outputting acoustic signals corresponding to the decorrelated signals Sout generated by the decorrelator 210 .
- the number of speakers included in the speaker set 250 may vary according to the number of audio signals included in the multi-channel signal Sm.
- the decorrelator 210 generates the decorrelated signals Sout by applying to the audio signals a phase shifting value adjusted based on a phase difference between the audio signals included in the multi-channel signal Sm.
- the phase shifting value is a value to be subtracted from or added to a phase of an audio signal included in the multi-channel signal Sm.
- a corrected phase shifting value may be set by multiplying the phase shifting value by a predetermined weighting value. The corrected phase shifting value will be described in detail with reference to FIG. 5 below.
- the phase shifting value may be a positive (+) value from 0 to a predetermined limit value.
- the predetermined limit value is a phase shifting value applied when a phase difference between audio signals is 0 or a value close to 0 and may be differently set according to a filtering performance, a target voice recognition ratio, and a product specification of an echo cancellation filter (not shown) included in the audio apparatus 200 .
- the decorrelator 210 may receive the multi-channel signal Sm, including first and second channel audio signals, from an external broadcasting station.
- the first and second channel audio signals may correspond to the R-channel audio signal and the L-channel audio signal, respectively.
- the multi-channel signal Sm including the first and second channel audio signals, is illustrated.
- the decorrelator 210 generates first and second decorrelated signals corresponding to the first and second channel audio signals, respectively.
- the decorrelator 210 sets a phase shifting value based on a phase difference between the R-channel audio signal and the L-channel audio signal. Then, the decorrelator 210 may generate an R-channel decorrelated signal and an L-channel decorrelated signal corresponding to the R-channel audio signal and the L-channel audio signal, respectively, so that the R-channel decorrelated signal and the L-channel decorrelated signal have phases obtained by adding the phase shifting value to a phase of the R-channel audio signal and subtracting the phase shifting value from a phase of the L-channel audio signal, respectively.
- the decorrelator 210 reduces the phase shifting value in inverse proportion to a phase difference between the R-channel audio signal and the L-channel audio signal.
- the decorrelator 210 may set the phase shifting value to 0 when the phase difference between the first and second channel audio signals is 180°. If the phase difference between the first and second channel audio signals is 180°, i.e., ⁇ (pi), the first and second channel audio signals have opposite phases to each other. Thus, the phase shifting value is set to 0 so that the first and second decorrelated signals have a low correlation even though the first and second decorrelated signals are generated.
- the decorrelator 210 may set the phase shifting value to a predetermined limit value when the phase difference between the first and second channel audio signals is 0°.
- the decorrelator 210 may set phase values for the first and second decorrelated signals by subtracting the phase shifting value from a phase of the first channel audio signal and adding the phase shifting value to a phase of the second channel audio signal. Then, the decorrelator 210 may generate the first and second decorrelated signals having the phase values for the first and second decorrelated signals.
- the decorrelator 210 may set phase values for the first and second decorrelated signals by subtracting a corrected phase shifting value from the phase of the first channel audio signal and adding the corrected phase shifting value to the phase of the second channel audio signal.
- FIG. 3 is a block diagram of an audio apparatus 300 according to an aspect of exemplary embodiment.
- the audio apparatus 300 may include a decorrelator 310 and a speaker set 350 . Since the audio apparatus 300 , the decorrelator 310 , and the speaker set 350 correspond to the audio apparatus 200 , the decorrelator 210 , and the speaker set 250 , respectively, a repeated description thereof is omitted.
- an operation of generating decorrelated signals by lowering a correlation between audio signals is called a decorrelation operation.
- the decorrelator 310 may divide audio signals included in a multi-channel signal Sm into a plurality of sub-bands and generate decorrelated signals based on sub-bands.
- a sub-band indicates a frequency band from when a frequency band of the multi-channel signal Sm is divided into a plurality of sub-frequency bands.
- the multi-channel signal Sm includes first and second channel audio signals.
- the first and second channel audio signals may correspond to an R-channel audio signal and an L-channel audio signal, respectively.
- FIG. 3 illustrates that the decorrelator 310 divides the audio signals included in the multi-channel signal Sm into n sub-bands and performs a decorrelation operation for generating decorrelated signals for the n sub-bands.
- the decorrelator 310 includes a pre-processor 320 , a decorrelation processor 330 , and a synthesizer 340 .
- the pre-processor 320 receives the multi-channel signal Sm, including the first and second channel audio signals, and divides the multi-channel signal Sm into first to nth sub-audio signals S_sub 1 to S_subn.
- the pre-processor 320 divides the multi-channel signal Sm into n frequency bands and outputs the first sub-audio signal S_sub 1 of a first sub-band to the nth sub-audio signal S_subn of an nth sub-band.
- the multi-channel signal Sm may include the first and second channel audio signals.
- the first sub-audio signal S_sub 1 of the first sub-band may include first and second channel sub-audio signals of the first sub-band
- the nth sub-audio signal S_subn of the nth sub-band may include first and second channel sub-audio signals of the nth sub-band.
- the decorrelation processor 330 generates first and second sub-decorrelated signals by applying a phase shifting value in a predetermined sub-band, which is adjusted based on a phase difference between first and second channel sub-audio signals in the predetermined sub-band, to the first and second channel sub-audio signals.
- the phase shifting value will be described in detail with reference to FIG. 4 below.
- FIG. 4 illustrates graphs 410 , 420 , and 430 for describing the phase shifting value adjustment performed by the audio apparatus 200 or 300 according to one or more aspects of exemplary embodiments.
- an X-axis indicates a phase difference between first and second channel audio signals
- a Y-axis indicates a phase shifting value.
- the X-axis may indicate a phase difference between first and second channel sub-audio signals in a predetermined sub-band.
- the X-axis indicates a phase difference between first and second channel audio signals.
- the graph 410 shows a phase shifting value applied in a high frequency band
- the graph 420 shows a phase shifting value applied in an intermediate frequency band
- the graph 430 shows a phase shifting value applied in a low frequency band.
- FIG. 4 illustrates that a total frequency band is divided into 3 frequency bands, i.e., the high frequency band, the intermediate frequency band, and the low frequency band, and a phase shifting value is adjusted for each divided frequency band.
- the total frequency band is divided into a various number of frequency bands and phase shifting value graphs are optimized for the divided frequency bands.
- the phase shifting value graphs may be experimentally optimized so that sound quality distortion is minimized while an echo cancellation filter (not shown) does not diverge.
- a phase shifting value is set to a predetermined limit value a, which is the maximum value.
- the phase shifting value may be set to 0 or a value close to 0.
- the phase shifting value increases as the phase difference between the first and second channel audio signals is small, and the phase shifting value decreases as the phase difference between the first and second channel audio signals is large.
- the frequency band is close to an audible frequency band which a user can recognize well. Since distortion of the first or second channel audio signal increases when a phase shifting value is set large in the low frequency band, which is the audible frequency band, the phase shifting value may be set small in the low frequency band, which is the audible frequency band.
- the high frequency band is a frequency band which the user cannot audibly recognize well, the user cannot recognize sound quality distortion well even though the phase shifting value is set large.
- the phase shifting value may be set large in the high frequency band. Accordingly, predetermined limit values a, b, and c may be set differently according to frequency bands.
- the predetermined limit values a, b, and c of a phase shifting value and forms of the phase shifting value graphs 410 , 420 , and 430 shown in FIG. 4 may be experimentally optimized and determined.
- a phase shifting value may be set through sound quality distortion experiments so that the echo cancellation filter does not diverge while the sound quality distortion is minimized.
- the decorrelation processor 330 may include first to nth sub-processors 331 , 332 , and 333 for performing the decorrelation operation for n sub-frequency bands.
- the first sub-processor 331 receives first and second channel sub-audio signals S_sub 1 in a first sub-band and generates first and second channel sub-decorrelated signals S_su 11 and S_su 12 corresponding to the first and second channel sub-audio signals S_sub 1 .
- the second sub-processor 332 receives first and second channel sub-audio signals S_sub 2 in a second sub-band and generates first and second channel sub-decorrelated signals S_su 21 and S_su 22 corresponding to the first and second channel sub-audio signals S_sub 2 .
- the nth sub-processor 333 receives first and second channel sub-audio signals S_subn in an nth sub-band and generates first and second channel sub-decorrelated signals S_sun 1 and S_sun 2 corresponding to the first and second channel sub-audio signals S_subn.
- the decorrelation processor 330 receives the sub-audio signal S_sub 2 in the second sub-band, which includes the first and second channel sub-audio signals.
- the second sub-processor 332 adjusts a phase shifting value based on a phase difference between the first and second channel sub-audio signals.
- the second sub-processor 332 generates the first channel sub-decorrelated signal S_su 21 by applying the phase shifting value to the first channel sub-audio signal of the second sub-band and generates the second channel sub-decorrelated signal S_su 22 by applying the phase shifting value to the second channel sub-audio signal of the second sub-band.
- the synthesizer 340 generates a first decorrelated signal Sc 1 and a second decorrelated signal Sc 2 by synthesizing the first channel sub-decorrelated signals S_su 11 , S_su 21 , . . . , S_sun 1 and the second channel sub-decorrelated signals S_su 12 , S_su 22 , . . . , S_sun 2 , respectively.
- the synthesizer 340 generates the first decorrelated signal Sc 1 corresponding to the total frequency band by synthesizing the first channel sub-decorrelated signals S_su 11 , S_su 21 , . . . , S_sun 1 in the n sub-bands.
- the synthesizer 340 generates the second decorrelated signal Sc 2 corresponding to the total frequency band by synthesizing the second channel sub-decorrelated signals S_su 12 , S_su 22 , . . . , S_sun 2 in the n sub-bands.
- the decorrelation operation of the decorrelator 310 allows the first decorrelated signal Sc 1 and the second decorrelated signal Sc 2 generated by the synthesizer 340 to have a low correlation therebetween.
- the speaker set 350 may include a plurality of speakers, namely, first and second speakers 351 and 352 , for receiving the plurality of decorrelated signals, namely, the first and second decorrelated signals Sc 1 and Sc 2 , output from the decorrelator 310 and outputting output audio signals, namely, first and second audio signals Sout 1 and Sout 2 , corresponding to the decorrelated signals, namely, the first and second decorrelated signals Sc 1 and Sc 2 , respectively.
- FIG. 3 illustrates that the speaker set 350 includes the first speaker 351 and the second speaker 352 .
- the first speaker 351 converts the first decorrelated signal Sc 1 to a first audio signal Sout 1 which the user can audibly recognize and outputs the first audio signal Sout 1 .
- the second speaker 352 converts the second decorrelated signal Sc 2 to a second audio signal Sout 2 which the user can audibly recognize and outputs the second audio signal Sout 2 .
- the decorrelation processor 330 may set a corrected phase shifting value by multiplying a phase shifting value by a predetermined weighting value. Then, the decorrelation processor 330 may generate first and second sub-decorrelated signals by subtracting the corrected phase shifting value from the phase of the first channel sub-audio signal and adding the corrected phase shifting value to the phase of the second channel sub-audio signal, respectively.
- the predetermined weighting value may be set differently based on sub-bands. The predetermined weighting value may be set by the decorrelation processor 330 or may be received from the outside as an experimentally optimized value.
- the predetermined weighting value may be increased.
- FIGS. 5A and 5B illustrate other graphs for describing a phase shifting value adjustment performed by the audio apparatus 200 or 300 according to one or more aspects of exemplary embodiments.
- FIG. 5A is a graph for describing a predetermined weighting value according to an aspect of an exemplary embodiment
- FIG. 5B is a graph for describing a predetermined weighting value according to another aspect of an exemplary embodiment.
- an X-axis indicates the order of sub-bands
- a Y-axis indicates a predetermined weighting value.
- the X-axis indicates the order k of sub-bands in the low frequency band to the high frequency band when the total frequency band of the multi-channel signal Sm is divided into n sub-bands.
- the first sub-processor 331 generates first and second sub-decorrelated signals corresponding to the first and second channel sub-audio signals by using a corrected phase shifting value by multiplying a phase shifting value by a weighting value w 1 corresponding to the first sub-band.
- a weighting value w 2 is a weighting value applied in the second sub-band
- a weighting value w 3 is a weighting value applied in the nth sub-band.
- a positive weighting value may be set in sub-bands of an odd order
- a negative weighting value may be set in sub-bands of an even order.
- an absolute value of a weighting value may be set to increase as an order value of a sub-band increases. For example, as a value of the order k of a sub-band increases, an absolute value of a weighting value increases as w 1 , w 2 , and w 3 .
- the first sub-audio signal S_sub 1 of the first sub-band which includes the first channel sub-audio signal and the second channel sub-audio signal, is input to the first sub-processor 331 .
- a phase of the first channel sub-audio signal may be represented as e j ⁇ 1 (k)
- the second channel sub-audio signal may be represented as e j ⁇ 2 (k) .
- k denotes an order of a sub-band
- ⁇ 1 (k) denotes a phase value of the first channel sub-audio signal in a kth sub-band
- ⁇ 2 (k) denotes a phase value of the second channel sub-audio signal in the kth sub-band.
- the phase shifting value described in FIG. 4 may be represented as e j ⁇ ⁇ (k) .
- the first sub-processor 331 may set a phase of a first channel sub-decorrelated signal in the kth sub-band as e j( ⁇ 1 ⁇ ⁇ )(k) by subtracting the phase shifting value from the phase of the first channel sub-audio signal.
- the first sub-processor 331 may set a phase of a second channel sub-decorrelated signal in the kth sub-band as e j( ⁇ 2 + ⁇ ⁇ )(k) by adding the phase shifting value to the phase of the second channel sub-audio signal.
- a weighting value applied in a predetermined sub-band k may be represented as wk.
- the first sub-processor 331 may set the phase of the second channel sub-decorrelated signal in the kth sub-band as e j( ⁇ 2 +w k ⁇ ⁇ )(k) by adding the corrected phase shifting value to the phase of the second channel sub-audio signal.
- the first channel sub-decorrelated signal is generated by synthesizing first channel sub-decorrelated signals of all sub-bands
- the second channel sub-decorrelated signal is generated by synthesizing second channel sub-decorrelated signals of all sub-bands.
- a magnitude of the first and second channel sub-decorrelated signals may vary according to a magnitude of the first and second channel audio signals, and a product specification, such as a maximum power or amplification efficiency, of the audio apparatus 300 .
- the decorrelator 310 may divide the plurality of sub-bands into a plurality of groups and set a predetermined weighting value having a different value based on groups. In addition, the decorrelator 310 may divide the plurality of sub-bands into a plurality of groups and receive a predetermined weighting value having a different value based on groups from the outside.
- first to third sub-bands may be set as a first group, group 1
- fourth to sixth sub-bands may be set as a second group
- group 2 seventh and eighth sub-bands may be set as a third group
- group 3 seventh and eighth sub-bands may be set as a third group
- ninth and greater sub-bands may be set as a fourth group, group 4 .
- one group e.g., group 1
- another group e.g., group 2
- adjacent to the one group may be set to have different signs.
- a sub-decorrelated signal When a sub-decorrelated signal is generated by adjusting a phase of a sub-audio signal, if a first sub-decorrelated signal is generated by subtracting a phase shifting value from a phase of a first channel sub-audio signal and a second sub-decorrelated signal is generated by adding the phase shifting value to a phase of a second channel sub-audio signal, spatial perception of sound is biased to one side.
- signs of weighting values in sub-bands may be set differently so that the spatial perception of sound is alternately biased to the left and the right. Accordingly, spatial perception which the user feels when the user listens to audio may be prevented from being biased to one side.
- FIG. 6 is a block diagram of an electronic apparatus 600 according to an exemplary embodiment.
- the electronic apparatus 600 may correspond to the audio apparatus 200 or 300 described with reference to FIGS. 2 to 5B .
- the electronic apparatus 600 may be an audio apparatus and may further include an echo filter unit 620 , a microphone 640 , and a processor 670 , in comparison with the audio apparatus 200 or 300 .
- a decorrelator 610 and a speaker set 650 of the electronic apparatus 600 correspond to the decorrelator 210 or 310 and the speaker set 250 or 350 of FIG. 2 or 3 , respectively. Thus, a repeated description thereof is omitted.
- the electronic apparatus 600 includes the decorrelator 610 , the echo filter unit 620 , the microphone 640 , the speaker set 650 , and the processor 670 .
- the decorrelator 610 receives a multi-channel signal Sm including audio signals and outputs decorrelated signals by applying a phase shifting value adjusted based on a correlation difference between the audio signals to the audio signals.
- the echo filter unit 620 generates echo signals corresponding to the decorrelated signals output from the decorrelator 610 .
- the speaker set 650 outputs acoustic signals corresponding to the decorrelated signals.
- the decorrelated signals are a first channel decorrelated signal and a second channel decorrelated signal as described above.
- an acoustic signal corresponding to the first channel decorrelated signal is called an R-channel audio signal
- an acoustic signal corresponding to the second channel decorrelated signal is called an L-channel audio signal.
- the microphone 640 receives a voice signal Sin and the acoustic signals output from the speaker set 650 .
- the voice signal Sin may be a signal due to talking by the user to control an operation of the electronic apparatus 600 by voice recognition.
- the voice signal Sin may be a signal due to talking by the user to input the signal into the electronic apparatus 600 .
- the L-channel audio signal may be output through the paths 121 , 122 , and 123 of FIG. 1
- the R-channel audio signal may be output through the paths 111 , 112 , and 113 of FIG. 1 .
- the microphone 640 may receive the L-channel audio signal transmitted through the paths 121 , 122 , and 123 and receive the R-channel audio signal transmitted through the paths 111 , 112 , and 113 . That is, the microphone 640 may receive the voice signal Sin, the L-channel audio signal, and the R-channel audio signal.
- the echo filter unit 620 estimates the acoustic signals output from the speaker set 650 and input to the microphone 640 , which may vary according to a surrounding environment where the electronic apparatus 600 is located.
- the paths of the acoustic signals input to the microphone 640 may be actual acoustic echo paths and vary according to a surrounding environment, and it is difficult to clearly determine the actual acoustic echo paths.
- the echo filter unit 620 may estimate echo paths corresponding to the actual acoustic echo paths and generate estimated acoustic signals based on the estimated echo paths.
- the estimated acoustic signals correspond to the actual acoustic signals output from the speaker set 650 .
- the processor 670 performs voice recognition processing on a subtraction signal obtained by subtracting the echo signals from an output signal of the microphone 640 .
- the processor 670 may include a controller 673 and the adder 671 .
- the adder 671 generates the subtraction signal by subtracting the echo signals output from the echo filter unit 620 from the output signal of the microphone 640 .
- the controller 673 performs voice recognition processing on the subtraction signal output from the adder 671 .
- the controller 673 recognizes voice corresponding to the subtraction signal and performs an operation corresponding to the recognized voice.
- the echo filter unit 620 may generate a first estimated acoustic signal by estimating an acoustic signal output from a first speaker (not shown) included in the speaker set 650 and generate a second estimated acoustic signal by estimating an acoustic signal output from a second speaker (not shown) included in the speaker set 650 .
- the adder 671 subtracts the first and second estimated acoustic signals output from the echo filter unit 620 from the output signal of the microphone 640 .
- the controller 673 may perform voice recognition processing on an output signal of the adder 671 and recognize a command or data according to the output signal of the adder 671 .
- the above-described echo cancellation filter may consist of the echo filter unit 620 and the adder 671 .
- FIG. 7 is a block diagram of an electronic apparatus 700 according to an exemplary embodiment.
- the electronic apparatus 700 corresponds to the electronic apparatus 600 of FIG. 6 .
- a multi-channel signal Sm input to a decorrelator 710 includes first and second channel audio signals
- a speaker set 750 includes a first speaker 751 for outputting the first channel audio signal and a second speaker 752 for outputting the second channel audio signal
- an echo filter unit 720 includes a first filter unit 721 and a second filter unit 722 .
- a processor 770 may include a controller 773 and two adders 771 and 772 .
- the decorrelator 710 outputs a first channel decorrelated signal Sc 1 corresponding to the first channel audio signal and a second channel decorrelated signal Sc 2 corresponding to the second channel audio signal.
- the first speaker 751 receives the first channel decorrelated signal Sc 1 and outputs an R-channel audio signal Sout 1 , which is an acoustic signal which a user can audibly recognize.
- the second speaker 752 receives the second channel decorrelated signal Sc 2 and outputs an L-channel audio signal Sout 2 , which is an acoustic signal which the user can audibly recognize.
- a microphone 740 receives a voice signal Sin, the R-channel audio signal Sout 1 , and the L-channel audio signal Sout 2 and outputs the voice signal Sin, the R-channel audio signal Sout 1 , and the L-channel audio signal Sout 2 .
- the first filter unit 721 receives the first channel decorrelated signal Sc 1 and estimates echo paths through which the R-channel audio signal Sout 1 is output. Then, the first filter unit 721 generates a first estimated acoustic signal Sp 1 corresponding to the first channel decorrelated signal Sc 1 by applying the first channel decorrelated signal Sc 1 to the estimated echo paths.
- the second filter unit 722 receives the second channel decorrelated signal Sc 2 and estimates echo paths through which the L-channel audio signal Sout 2 is output. Then, the second filter unit 722 generates a second estimated acoustic signal Sp 2 corresponding to the second channel decorrelated signal Sc 2 by applying the second channel decorrelated signal Sc 2 to the estimated echo paths.
- the adder 771 subtracts the second channel decorrelated signal Sc 2 from an output signal of the microphone 740 .
- the adder 772 subtracts the first channel decorrelated signal Sc 1 from an output signal of the adder 771 .
- the two adders 771 and 772 are included in the processor 770 in FIG. 7 , one adder may be used to subtract the first and second channel decorrelated signals Sc 1 and Sc 2 from the output signal of the microphone 740 .
- FIG. 8 is a flowchart illustrating an audio signal processing method 800 according to an exemplary embodiment.
- the audio signal processing method 800 may be performed by the audio apparatus 200 or 300 or the electronic apparatus 600 or 700 .
- the audio signal processing method 800 is described with reference to the audio apparatus 200 .
- operation 810 a multi-channel signal including audio signals is received. Operation 810 may be performed by the decorrelator 210 .
- a phase shifting value is adjusted based on a correlation difference between the audio signals included in the multi-channel signal received in operation 810 .
- Operation 820 may be performed by the decorrelator 210 .
- decorrelated signals are generated by applying the phase shifting value adjusted in operation 820 to the audio signals included in the multi-channel signal. Operation 830 may be performed by the decorrelator 210 .
- the decorrelated signals generated in operation 830 may be output through the plurality of speakers included in the speaker set 250 .
- Operations of the audio signal processing method 800 are the same as an operation of the audio apparatus 200 or 300 or the electronic apparatus 600 or 700 . Thus, a detailed description thereof is omitted.
- FIG. 9 is a flowchart illustrating an audio signal processing method 900 according to an exemplary embodiment.
- the audio signal processing method 900 may be performed by the audio apparatus 300 or the electronic apparatus 600 or 700 .
- the audio signal processing method 900 is described with reference to the audio apparatus 300 .
- operation 910 is the same as operation 810 , a detailed description thereof is omitted.
- a description is provided by assuming that a multi-channel signal, including first and second channel audio signals, is received.
- the multi-channel signal received in operation 910 is divided into a plurality of sub-bands.
- the first and second channel audio signals are divided into the plurality of sub-bands, and first and second channel sub-audio signals are generated in each of the plurality of sub-bands.
- Operations 910 and 920 may be performed by the pre-processor 320 .
- a phase shifting value in a predetermined sub-band is adjusted based on a phase difference between the first and second channel sub-audio signals in the predetermined sub-band. Operation 930 may be performed by the decorrelation processor 330 .
- first and second sub-decorrelated signals are generated by applying the phase shifting value in the predetermined sub-band, which is generated in operation 930 , to the first and second channel sub-audio signals in the predetermined sub-band. Operation 940 may be performed by the decorrelation processor 330 .
- first and second decorrelated signals are generated by synthesizing first and second sub-decorrelated signals generated for the plurality of sub-bands. Operation 950 may be performed by the synthesizer 340 .
- the first and second decorrelated signals generated in operation 950 may be output through the first and second speakers 351 and 352 included in the speaker set 350 .
- Operations of the audio signal processing method 900 are the same as an operation of the audio apparatus 300 or the electronic apparatus 600 or 700 . Thus, a detailed description thereof is omitted.
- FIG. 10 is a flowchart for describing operation 930 of FIG. 9 in more detail.
- operation 930 corresponds to operation 1000 , which may include operations 1010 , 1020 , and 1030 .
- a phase shifting value is set based on sub-bands.
- a weighting value is set based on sub-bands.
- a corrected phase shifting value is set by multiplying the phase shifting value set in operation 1010 by the weighting value set in operation 1020 .
- an exemplary embodiment can be embodied as computer-readable code on a computer-readable recording medium.
- the computer-readable recording medium is any data storage device that can store data that can be thereafter read by a computer system. Examples of the computer-readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices.
- the computer-readable recording medium can also be distributed over network-coupled computer systems so that the computer-readable code is stored and executed in a distributed fashion.
- aspects of an exemplary embodiment may be written as a computer program transmitted over a computer-readable transmission medium, such as a carrier wave, and received and implemented in general-use or special-purpose digital computers that execute the programs.
- one or more units of the audio signal processing apparatus can include a processor or microprocessor executing a computer program stored in a computer-readable medium.
- an audio apparatus using the same, and an electronic apparatus using the same, decorrelated signals corresponding to audio signals included in a multi-channel signal are generated based on a correlation between the audio signals.
- the audio signals along predetermined echo paths without divergence of an echo filter may be correctly estimated. Accordingly, an accuracy of voice recognition may increase.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Circuit For Audible Band Transducer (AREA)
Abstract
Description
- This application claims priority from U.S. Provisional Application No. 61/491,260, filed on May 30, 2011, in the US PTO and Korean Patent Application No. 10-2012-0036799, filed on Apr. 9, 2012, in the Korean Intellectual Property Office, the disclosures of which are incorporated herein in their entirety by reference.
- 1. Field
- Exemplary embodiments relate to an audio signal processing method, an audio apparatus using the same, and an electronic apparatus using the same, and more particularly, to an audio signal processing method for receiving a multi-channel signal and outputting corresponding acoustic signals, an audio apparatus using the same, and an electronic apparatus using the same.
- 2. Description of the Related Art
- Audio apparatuses may output received multi-channel signals through a plurality of speakers. In addition, the audio apparatuses may receive a voice signal corresponding to a voice of a user, recognize the received voice signal, and perform an operation corresponding to a command, operation, or request based on the recognized voice signal. Hereinafter, the recognizing of a received voice signal and the performing of an operation corresponding to a command, operation, or request based on the recognized voice signal are called a voice recognition operation.
- In an audio apparatus capable of performing the voice recognition operation, when a voice signal is received through a predetermined microphone while acoustic signals are being output through a plurality of speakers, the audio apparatus must acquire only the voice signal by removing the acoustic signals from among signals input to the microphone. Then, the audio apparatus recognizes the acquired voice signal.
- To remove acoustic signals as described above, Multi-Channel Acoustic Echo Cancellation (MCAEC) is used. MCAEC may be implemented using an echo cancellation filter.
- If the acoustic signals output through the plurality of speakers have a low correlation, the echo cancellation filter converges to a predetermined value, thereby effectively canceling the acoustic signals. However, if the acoustic signals output through the plurality of speakers have a high correlation, the echo cancellation filter diverges without converging. Accordingly, since residual echo remains in a signal for voice recognition, the voice recognition operation cannot be effectively performed.
- That is, if the echo cancellation filter cannot effectively cancel the acoustic signals, the audio apparatus cannot correctly recognize the voice signal and cannot perform an operation or command corresponding to the voice signal.
- Thus, in audio apparatuses capable of recognizing a voice signal, an audio apparatus for effectively lowering a correlation between acoustic signals output to a plurality of speakers is necessary.
- The exemplary embodiments provide an audio signal processing method for effectively lowering a correlation between acoustic signals output to a plurality of speakers, an audio apparatus therefor, and an electronic apparatus therefor.
- The exemplary embodiments also provide an audio signal processing method for correctly recognizing an input voice signal and an electronic apparatus therefor.
- According to an aspect of an exemplary embodiment, an audio apparatus includes: a decorrelator for generating decorrelated signals by applying a phase shifting value adjusted based on a correlation difference between audio signals included in a multi-channel signal to the audio signals; and a speaker set including at least two speakers for outputting acoustic signals corresponding to the decorrelated signals.
- The decorrelator may generate the decorrelated signals by applying a phase shifting value adjusted based on a phase difference between the audio signals to the audio signals.
- The decorrelator may receive the multi-channel signal including first and second channel audio signals.
- The decorrelator may reduce the phase shifting value when a phase difference between the first and second channel audio signals is great.
- The decorrelator may set the phase shifting value to 0 when the phase difference between the first and second channel audio signals is 180°.
- The decorrelator may set the phase shifting value to a predetermined limit value when the phase difference between the first and second channel audio signals is 0°.
- The decorrelator may generate first and second decorrelated signals corresponding to the first and second channel audio signals by subtracting the phase shifting value from a phase of the first channel audio signal and adding the phase shifting value to a phase of the second channel audio signal, respectively.
- The decorrelator may include: a pre-processor for receiving the multi-channel signal, including the first and second channel audio signals, dividing the multi-channel signal into a plurality of sub-bands, and generating first and second channel sub-audio signals in each of the plurality of sub-bands; a decorrelation processor for generating first and second sub-decorrelated signals by applying a phase shifting value adjusted based on a phase difference between the first and second channel sub-audio signals to the first and second channel sub-audio signals; and a synthesizer for generating the first decorrelated signal by synthesizing the first sub-decorrelated signals in the plurality of sub-bands and generating the second decorrelated signal by synthesizing the second sub-decorrelated signals in the plurality of sub-bands.
- The decorrelation processor may generate the first and second sub-decorrelated signals by setting a corrected phase shifting value by multiplying the phase shifting value by a predetermined weighting value based on sub-bands, subtracting the corrected phase shifting value from a phase of the first channel sub-audio signal, and adding the corrected phase shifting value to a phase of the second channel sub-audio signal.
- The decorrelation processor may set the predetermined weighting value based on sub-bands.
- The decorrelation processor may set the predetermined weighting value in one sub-band and the predetermined weighting value in another sub-band adjacent to the one sub-band to have different signs.
- The decorrelation processor may divide the plurality of sub-bands into a plurality of groups and set the predetermined weighting value based on groups.
- The decorrelation processor may set the predetermined weighting value in one group and the predetermined weighting value in another group adjacent to the one group to have different signs.
- The audio apparatus may further include: an echo filter unit for generating echo signals corresponding to the decorrelated signals; a microphone for receiving the acoustic signals and a voice signal; and a processor for processing voice recognition on a subtraction signal obtained by subtracting the echo signals from an output signal of the microphone.
- According to another aspect of an exemplary embodiment, an electronic apparatus includes: a decorrelator for generating decorrelated signals by applying a phase shifting value adjusted based on a correlation difference between audio signals included in a multi-channel signal to the audio signals; an echo filter unit for generating echo signals corresponding to the decorrelated signals; a speaker set including at least two speakers for outputting acoustic signals corresponding to the decorrelated signals; a microphone for receiving the acoustic signals and a voice signal; and a processor for processing voice recognition on a subtraction signal obtained by subtracting the echo signals from an output signal of the microphone.
- According to another aspect of an exemplary embodiment, an audio signal processing method includes: receiving a multi-channel signal including audio signals; adjusting a phase shifting value based on a correlation difference between the audio signals; and generating decorrelated signals by applying the phase shifting value to the audio signals.
- The above and other features and advantages of the exemplary embodiments will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings in which:
-
FIG. 1 is a perspective view of an electronic apparatus; -
FIG. 2 is a block diagram of an audio apparatus according to an aspect of exemplary embodiment; -
FIG. 3 is a block diagram of an audio apparatus according to another aspect an exemplary embodiment; -
FIG. 4 illustrates graphs for describing a phase shifting value adjustment performed by the audio apparatus according to one or more aspects of exemplary embodiments; -
FIGS. 5A and 5B illustrate other graphs for describing a phase shifting value adjustment performed by the audio apparatus according to one or more aspects of exemplary embodiments; -
FIG. 6 is a block diagram of an electronic apparatus according to an exemplary embodiment; -
FIG. 7 is a block diagram of an electronic apparatus according to an exemplary embodiment; -
FIG. 8 is a flowchart illustrating an audio signal processing method according to an exemplary embodiment; -
FIG. 9 is a flowchart illustrating an audio signal processing method according to an exemplary embodiment; and -
FIG. 10 is a flowchart for describingoperation 930 ofFIG. 9 in more detail. - Expressions such as “at least one of,” when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list.
- An audio signal processing method, an audio apparatus, and an electronic apparatus will now be described in detail with reference to the accompanying drawings.
-
FIG. 1 is a perspective view of anelectronic apparatus 100. Theelectronic apparatus 100 shown inFIG. 1 is, for example, a digital TV. Theelectronic apparatus 100 may be any one of various electronic products, including an audio apparatus for outputting an audio signal. For example, theelectronic apparatus 100 may be a digital TV, an audio system, a refrigerator, a washing machine, a personal computer, a mobile computer, a pad-type terminal, or a mobile phone. - The
electronic apparatus 100 includes an audio apparatus (not shown) including a decorrelator (not shown) and a speaker set. The decorrelator may be included in a circuit block included in theelectronic apparatus 100, and the speaker set may include at least two speakers, for example, first andsecond speakers - The
electronic apparatus 100 may further include at least one of amicrophone 130 and adisplay unit 140. Themicrophone 130 may include a microphone array (not shown) and may receive all audible audio signals. In detail, themicrophone 130 may receive a voice signal generated by a user. - The decorrelator adjusts and outputs a correlation among a plurality of audio signals included in a multi-channel signal. The correlation-adjusted audio signals are output through the first and
second speakers - When the user inputs a voice signal into the
microphone 130 while audio signals are being output from the first andsecond speakers microphone 130 receives the voice signal in addition to the audio signals output from the first andsecond speakers - For example, assuming that the
electronic apparatus 100 is a digital TV capable of performing a video call function, the digital TV may output a Picture-In-Picture (PIP) screen on thedisplay unit 140. In a PIP mode, thedisplay unit 140 may output broadcasting images on a main screen and output images of a video call on a sub-screen. - As another example, assuming that the
electronic apparatus 100 is a digital TV capable of performing a voice recognition control, when the user inputs a predetermined command or request into the digital TV with a voice signal, the digital TV may perform an operation corresponding to the input voice signal. That is, when the user inputs a voice signal into themicrophone 130, theelectronic apparatus 100 may perform an operation corresponding to the input voice signal. - When a voice signal is received while audio signals are being output from the first and
second speakers electronic apparatus 100 must detect or discern only the voice signal and not the audio signals output from the first andsecond speakers microphone 130, audio signals output from the speakers must be cancelled from among all the signals input to themicrophone 130 by using an echo cancellation filter (not shown). - The audio signals output from the first and
second speakers microphone 130 throughvarious paths electronic apparatus 100 is located. Hereinafter,paths echo filter unit 620 and anadder 671, which will be described with reference toFIG. 6 . - The
electronic apparatus 100 may receive the multi-channel signal including a right channel (R-channel) audio signal and a left channel (L-channel) audio signal from the outside or may itself generate the multi-channel signal. When the speaker set includes the first andsecond speakers first speaker 110 may output the R-channel audio signal, and thesecond speaker 120 may output the L-channel audio signal. - Electronic apparatuses and audio apparatuses which may be included in the electronic apparatuses will now be described in detail with reference to
FIGS. 2 to 7 . -
FIG. 2 is a block diagram of anaudio apparatus 200 according to an aspect of an exemplary embodiment. - Referring to
FIG. 2 , theaudio apparatus 200 includes adecorrelator 210 and aspeaker set 250. - The
decorrelator 210 generates decorrelated signals Sout corresponding to audio signals included in a multi-channel signal Sm by reducing a correlation between the audio signals - In detail, the
decorrelator 210 receives the multi-channel signal Sm and generates the decorrelated signals Sout by applying a phase shifting value adjusted based on a correlation difference between the audio signals included in the multi-channel signal Sm to the audio signals. The multi-channel signal Sm includes a plurality of audio signals corresponding to a plurality of channels. For example, the multi-channel signal Sm may include an R-channel audio signal and an L-channel audio signal. As another example, the multi-channel signal Sm may include 5.1-channel audio signals, 7.1-channel audio signals, or 10.2-channel audio signals. - The speaker set 250 includes at least two speakers for outputting acoustic signals corresponding to the decorrelated signals Sout generated by the
decorrelator 210. The number of speakers included in the speaker set 250 may vary according to the number of audio signals included in the multi-channel signal Sm. - In detail, the
decorrelator 210 generates the decorrelated signals Sout by applying to the audio signals a phase shifting value adjusted based on a phase difference between the audio signals included in the multi-channel signal Sm. The phase shifting value is a value to be subtracted from or added to a phase of an audio signal included in the multi-channel signal Sm. A corrected phase shifting value may be set by multiplying the phase shifting value by a predetermined weighting value. The corrected phase shifting value will be described in detail with reference toFIG. 5 below. The phase shifting value may be a positive (+) value from 0 to a predetermined limit value. The predetermined limit value is a phase shifting value applied when a phase difference between audio signals is 0 or a value close to 0 and may be differently set according to a filtering performance, a target voice recognition ratio, and a product specification of an echo cancellation filter (not shown) included in theaudio apparatus 200. - In addition, the
decorrelator 210 may receive the multi-channel signal Sm, including first and second channel audio signals, from an external broadcasting station. The first and second channel audio signals may correspond to the R-channel audio signal and the L-channel audio signal, respectively. Hereinafter, the multi-channel signal Sm, including the first and second channel audio signals, is illustrated. In addition, it is illustrated that thedecorrelator 210 generates first and second decorrelated signals corresponding to the first and second channel audio signals, respectively. - For example, when the multi-channel signal Sm, including the R-channel audio signal and the L-channel audio signal, is received, the
decorrelator 210 sets a phase shifting value based on a phase difference between the R-channel audio signal and the L-channel audio signal. Then, thedecorrelator 210 may generate an R-channel decorrelated signal and an L-channel decorrelated signal corresponding to the R-channel audio signal and the L-channel audio signal, respectively, so that the R-channel decorrelated signal and the L-channel decorrelated signal have phases obtained by adding the phase shifting value to a phase of the R-channel audio signal and subtracting the phase shifting value from a phase of the L-channel audio signal, respectively. - In detail, the
decorrelator 210 reduces the phase shifting value in inverse proportion to a phase difference between the R-channel audio signal and the L-channel audio signal. - In addition, the
decorrelator 210 may set the phase shifting value to 0 when the phase difference between the first and second channel audio signals is 180°. If the phase difference between the first and second channel audio signals is 180°, i.e., π (pi), the first and second channel audio signals have opposite phases to each other. Thus, the phase shifting value is set to 0 so that the first and second decorrelated signals have a low correlation even though the first and second decorrelated signals are generated. - In addition, the
decorrelator 210 may set the phase shifting value to a predetermined limit value when the phase difference between the first and second channel audio signals is 0°. - In addition, the
decorrelator 210 may set phase values for the first and second decorrelated signals by subtracting the phase shifting value from a phase of the first channel audio signal and adding the phase shifting value to a phase of the second channel audio signal. Then, thedecorrelator 210 may generate the first and second decorrelated signals having the phase values for the first and second decorrelated signals. - In addition, the
decorrelator 210 may set phase values for the first and second decorrelated signals by subtracting a corrected phase shifting value from the phase of the first channel audio signal and adding the corrected phase shifting value to the phase of the second channel audio signal. -
FIG. 3 is a block diagram of anaudio apparatus 300 according to an aspect of exemplary embodiment. Referring toFIG. 3 , theaudio apparatus 300 may include adecorrelator 310 and aspeaker set 350. Since theaudio apparatus 300, thedecorrelator 310, and the speaker set 350 correspond to theaudio apparatus 200, thedecorrelator 210, and the speaker set 250, respectively, a repeated description thereof is omitted. Hereinafter, an operation of generating decorrelated signals by lowering a correlation between audio signals is called a decorrelation operation. - The
decorrelator 310 may divide audio signals included in a multi-channel signal Sm into a plurality of sub-bands and generate decorrelated signals based on sub-bands. A sub-band indicates a frequency band from when a frequency band of the multi-channel signal Sm is divided into a plurality of sub-frequency bands. Hereinafter, it is illustrated that the multi-channel signal Sm includes first and second channel audio signals. As described above, the first and second channel audio signals may correspond to an R-channel audio signal and an L-channel audio signal, respectively. -
FIG. 3 illustrates that thedecorrelator 310 divides the audio signals included in the multi-channel signal Sm into n sub-bands and performs a decorrelation operation for generating decorrelated signals for the n sub-bands. - In detail, the
decorrelator 310 includes a pre-processor 320, adecorrelation processor 330, and asynthesizer 340. - The pre-processor 320 receives the multi-channel signal Sm, including the first and second channel audio signals, and divides the multi-channel signal Sm into first to nth sub-audio signals S_sub1 to S_subn.
- That is, the pre-processor 320 divides the multi-channel signal Sm into n frequency bands and outputs the first sub-audio signal S_sub1 of a first sub-band to the nth sub-audio signal S_subn of an nth sub-band. The multi-channel signal Sm may include the first and second channel audio signals. In addition, the first sub-audio signal S_sub1 of the first sub-band may include first and second channel sub-audio signals of the first sub-band, and the nth sub-audio signal S_subn of the nth sub-band may include first and second channel sub-audio signals of the nth sub-band.
- The
decorrelation processor 330 generates first and second sub-decorrelated signals by applying a phase shifting value in a predetermined sub-band, which is adjusted based on a phase difference between first and second channel sub-audio signals in the predetermined sub-band, to the first and second channel sub-audio signals. The phase shifting value will be described in detail with reference toFIG. 4 below. -
FIG. 4 illustratesgraphs 410, 420, and 430 for describing the phase shifting value adjustment performed by theaudio apparatus FIG. 4 , an X-axis indicates a phase difference between first and second channel audio signals, and a Y-axis indicates a phase shifting value. In addition, the X-axis may indicate a phase difference between first and second channel sub-audio signals in a predetermined sub-band. Hereinafter, the X-axis indicates a phase difference between first and second channel audio signals. - Referring to
FIG. 4 , thegraph 410 shows a phase shifting value applied in a high frequency band, the graph 420 shows a phase shifting value applied in an intermediate frequency band, and the graph 430 shows a phase shifting value applied in a low frequency band.FIG. 4 illustrates that a total frequency band is divided into 3 frequency bands, i.e., the high frequency band, the intermediate frequency band, and the low frequency band, and a phase shifting value is adjusted for each divided frequency band. However, it may be designed where the total frequency band is divided into a various number of frequency bands and phase shifting value graphs are optimized for the divided frequency bands. In addition, the phase shifting value graphs may be experimentally optimized so that sound quality distortion is minimized while an echo cancellation filter (not shown) does not diverge. - Referring to the
graph 410 ofFIG. 4 , when a phase difference between first and second channel audio signals corresponding to the high frequency band is 0°, a phase shifting value is set to a predetermined limit value a, which is the maximum value. In addition, when the phase difference between the first and second channel audio signals is 180°, i.e., π (pi), the phase shifting value may be set to 0 or a value close to 0. In detail, the phase shifting value increases as the phase difference between the first and second channel audio signals is small, and the phase shifting value decreases as the phase difference between the first and second channel audio signals is large. - In addition, as a frequency band goes from the high frequency band to the low frequency band, the frequency band is close to an audible frequency band which a user can recognize well. Since distortion of the first or second channel audio signal increases when a phase shifting value is set large in the low frequency band, which is the audible frequency band, the phase shifting value may be set small in the low frequency band, which is the audible frequency band. On the contrary, since the high frequency band is a frequency band which the user cannot audibly recognize well, the user cannot recognize sound quality distortion well even though the phase shifting value is set large. Thus, the phase shifting value may be set large in the high frequency band. Accordingly, predetermined limit values a, b, and c may be set differently according to frequency bands.
- The predetermined limit values a, b, and c of a phase shifting value and forms of the phase shifting
value graphs 410, 420, and 430 shown inFIG. 4 may be experimentally optimized and determined. For example, a phase shifting value may be set through sound quality distortion experiments so that the echo cancellation filter does not diverge while the sound quality distortion is minimized. - The
decorrelation processor 330 may include first tonth sub-processors - In detail, the
first sub-processor 331 receives first and second channel sub-audio signals S_sub1 in a first sub-band and generates first and second channel sub-decorrelated signals S_su11 and S_su12 corresponding to the first and second channel sub-audio signals S_sub1. Thesecond sub-processor 332 receives first and second channel sub-audio signals S_sub2 in a second sub-band and generates first and second channel sub-decorrelated signals S_su21 and S_su22 corresponding to the first and second channel sub-audio signals S_sub2. Thenth sub-processor 333 receives first and second channel sub-audio signals S_subn in an nth sub-band and generates first and second channel sub-decorrelated signals S_sun1 and S_sun2 corresponding to the first and second channel sub-audio signals S_subn. - For example, when the correlation operation in the second sub-band is used as an example, the
decorrelation processor 330 receives the sub-audio signal S_sub2 in the second sub-band, which includes the first and second channel sub-audio signals. Thesecond sub-processor 332 adjusts a phase shifting value based on a phase difference between the first and second channel sub-audio signals. Thesecond sub-processor 332 generates the first channel sub-decorrelated signal S_su21 by applying the phase shifting value to the first channel sub-audio signal of the second sub-band and generates the second channel sub-decorrelated signal S_su22 by applying the phase shifting value to the second channel sub-audio signal of the second sub-band. - Since operations of the first sub-processor 331 to the
nth sub-processor 333 are the same as the operation of the second sub-processor 332 described above, a detailed description thereof is omitted. - The
synthesizer 340 generates a first decorrelated signal Sc1 and a second decorrelated signal Sc2 by synthesizing the first channel sub-decorrelated signals S_su11, S_su21, . . . , S_sun1 and the second channel sub-decorrelated signals S_su12, S_su22, . . . , S_sun2, respectively. - In detail, the
synthesizer 340 generates the first decorrelated signal Sc1 corresponding to the total frequency band by synthesizing the first channel sub-decorrelated signals S_su11, S_su21, . . . , S_sun1 in the n sub-bands. - In addition, the
synthesizer 340 generates the second decorrelated signal Sc2 corresponding to the total frequency band by synthesizing the second channel sub-decorrelated signals S_su12, S_su22, . . . , S_sun2 in the n sub-bands. - The decorrelation operation of the
decorrelator 310 allows the first decorrelated signal Sc1 and the second decorrelated signal Sc2 generated by thesynthesizer 340 to have a low correlation therebetween. - The speaker set 350 may include a plurality of speakers, namely, first and
second speakers decorrelator 310 and outputting output audio signals, namely, first and second audio signals Sout1 and Sout2, corresponding to the decorrelated signals, namely, the first and second decorrelated signals Sc1 and Sc2, respectively.FIG. 3 illustrates that the speaker set 350 includes thefirst speaker 351 and thesecond speaker 352. - In detail, the
first speaker 351 converts the first decorrelated signal Sc1 to a first audio signal Sout1 which the user can audibly recognize and outputs the first audio signal Sout1. Thesecond speaker 352 converts the second decorrelated signal Sc2 to a second audio signal Sout2 which the user can audibly recognize and outputs the second audio signal Sout2. - In addition, the
decorrelation processor 330 may set a corrected phase shifting value by multiplying a phase shifting value by a predetermined weighting value. Then, thedecorrelation processor 330 may generate first and second sub-decorrelated signals by subtracting the corrected phase shifting value from the phase of the first channel sub-audio signal and adding the corrected phase shifting value to the phase of the second channel sub-audio signal, respectively. The predetermined weighting value may be set differently based on sub-bands. The predetermined weighting value may be set by thedecorrelation processor 330 or may be received from the outside as an experimentally optimized value. - In addition, as a frequency band goes from a sub-band of the low frequency band to a sub-band of the high frequency band, the predetermined weighting value may be increased.
-
FIGS. 5A and 5B illustrate other graphs for describing a phase shifting value adjustment performed by theaudio apparatus FIG. 5A is a graph for describing a predetermined weighting value according to an aspect of an exemplary embodiment, andFIG. 5B is a graph for describing a predetermined weighting value according to another aspect of an exemplary embodiment. - Referring to
FIGS. 5A and 5B , an X-axis indicates the order of sub-bands, and a Y-axis indicates a predetermined weighting value. In detail, the X-axis indicates the order k of sub-bands in the low frequency band to the high frequency band when the total frequency band of the multi-channel signal Sm is divided into n sub-bands. - Referring to
FIG. 5A , k=1 indicates the first sub-band, and thefirst sub-processor 331 generates first and second sub-decorrelated signals corresponding to the first and second channel sub-audio signals by using a corrected phase shifting value by multiplying a phase shifting value by a weighting value w1 corresponding to the first sub-band. A weighting value w2 is a weighting value applied in the second sub-band, and a weighting value w3 is a weighting value applied in the nth sub-band. - In detail, a predetermined weighting value in one sub-band (e.g., k=1) and a predetermined weighting value in another sub-band (e.g., k=2) adjacent to the one sub-band may be set to have different signs. For example, a positive weighting value may be set in sub-bands of an odd order, and a negative weighting value may be set in sub-bands of an even order.
- In addition, an absolute value of a weighting value may be set to increase as an order value of a sub-band increases. For example, as a value of the order k of a sub-band increases, an absolute value of a weighting value increases as w1, w2, and w3.
- Referring to
FIGS. 3 and 5A , when the order k of a sub-band is 1, the first sub-audio signal S_sub1 of the first sub-band, which includes the first channel sub-audio signal and the second channel sub-audio signal, is input to thefirst sub-processor 331. A phase of the first channel sub-audio signal may be represented as ejφ1 (k), and the second channel sub-audio signal may be represented as ejφ2 (k). Here, k denotes an order of a sub-band, φ1(k) denotes a phase value of the first channel sub-audio signal in a kth sub-band, and φ2(k) denotes a phase value of the second channel sub-audio signal in the kth sub-band. - The phase shifting value described in
FIG. 4 may be represented as ejφΔ (k). When k=1, the first sub-processor 331 may set a phase of a first channel sub-decorrelated signal in the kth sub-band as ej(φ1 −φΔ )(k) by subtracting the phase shifting value from the phase of the first channel sub-audio signal. In addition, the first sub-processor 331 may set a phase of a second channel sub-decorrelated signal in the kth sub-band as ej(φ2 +φΔ )(k) by adding the phase shifting value to the phase of the second channel sub-audio signal. - When first and second channel sub-decorrelated signals are generated by applying a corrected phase shifting value, a weighting value applied in a predetermined sub-band k may be represented as wk. In this case, the corrected phase shifting value may be represented as wk·φΔ. Accordingly, when k=1, the first sub-processor 331 may set the phase of the first channel sub-decorrelated signal in the kth sub-band as by ej(φ
1 −wk ·φΔ )(k) subtracting the corrected phase shifting value from the phase of the first channel sub-audio signal. In addition, the first sub-processor 331 may set the phase of the second channel sub-decorrelated signal in the kth sub-band as ej(φ2 +wk ·φΔ )(k) by adding the corrected phase shifting value to the phase of the second channel sub-audio signal. - In addition, the first channel sub-decorrelated signal is generated by synthesizing first channel sub-decorrelated signals of all sub-bands, and the second channel sub-decorrelated signal is generated by synthesizing second channel sub-decorrelated signals of all sub-bands. A magnitude of the first and second channel sub-decorrelated signals may vary according to a magnitude of the first and second channel audio signals, and a product specification, such as a maximum power or amplification efficiency, of the
audio apparatus 300. - Referring to
FIG. 5B , thedecorrelator 310 may divide the plurality of sub-bands into a plurality of groups and set a predetermined weighting value having a different value based on groups. In addition, thedecorrelator 310 may divide the plurality of sub-bands into a plurality of groups and receive a predetermined weighting value having a different value based on groups from the outside. - As shown in
FIG. 5B , first to third sub-bands may be set as a first group, group1, fourth to sixth sub-bands may be set as a second group, group2, seventh and eighth sub-bands may be set as a third group, group3, and ninth and greater sub-bands may be set as a fourth group, group4. In addition, one group (e.g., group1) and another group (e.g., group2) adjacent to the one group may be set to have different signs. - When a sub-decorrelated signal is generated by adjusting a phase of a sub-audio signal, if a first sub-decorrelated signal is generated by subtracting a phase shifting value from a phase of a first channel sub-audio signal and a second sub-decorrelated signal is generated by adding the phase shifting value to a phase of a second channel sub-audio signal, spatial perception of sound is biased to one side. Thus, as described with reference to
FIGS. 5A and 5B , signs of weighting values in sub-bands may be set differently so that the spatial perception of sound is alternately biased to the left and the right. Accordingly, spatial perception which the user feels when the user listens to audio may be prevented from being biased to one side. -
FIG. 6 is a block diagram of anelectronic apparatus 600 according to an exemplary embodiment. Referring toFIG. 6 , theelectronic apparatus 600 may correspond to theaudio apparatus FIGS. 2 to 5B . Theelectronic apparatus 600 may be an audio apparatus and may further include anecho filter unit 620, amicrophone 640, and aprocessor 670, in comparison with theaudio apparatus decorrelator 610 and a speaker set 650 of theelectronic apparatus 600 correspond to thedecorrelator FIG. 2 or 3, respectively. Thus, a repeated description thereof is omitted. - The
electronic apparatus 600 includes thedecorrelator 610, theecho filter unit 620, themicrophone 640, the speaker set 650, and theprocessor 670. - The
decorrelator 610 receives a multi-channel signal Sm including audio signals and outputs decorrelated signals by applying a phase shifting value adjusted based on a correlation difference between the audio signals to the audio signals. - The
echo filter unit 620 generates echo signals corresponding to the decorrelated signals output from thedecorrelator 610. - The speaker set 650 outputs acoustic signals corresponding to the decorrelated signals. For example, the decorrelated signals are a first channel decorrelated signal and a second channel decorrelated signal as described above. Hereinafter, an acoustic signal corresponding to the first channel decorrelated signal is called an R-channel audio signal, and an acoustic signal corresponding to the second channel decorrelated signal is called an L-channel audio signal.
- The
microphone 640 receives a voice signal Sin and the acoustic signals output from the speaker set 650. - The voice signal Sin may be a signal due to talking by the user to control an operation of the
electronic apparatus 600 by voice recognition. Alternatively, the voice signal Sin may be a signal due to talking by the user to input the signal into theelectronic apparatus 600. The L-channel audio signal may be output through thepaths FIG. 1 , and the R-channel audio signal may be output through thepaths FIG. 1 . Themicrophone 640 may receive the L-channel audio signal transmitted through thepaths paths microphone 640 may receive the voice signal Sin, the L-channel audio signal, and the R-channel audio signal. - In detail, the
echo filter unit 620 estimates the acoustic signals output from the speaker set 650 and input to themicrophone 640, which may vary according to a surrounding environment where theelectronic apparatus 600 is located. The paths of the acoustic signals input to themicrophone 640 may be actual acoustic echo paths and vary according to a surrounding environment, and it is difficult to clearly determine the actual acoustic echo paths. Thus, theecho filter unit 620 may estimate echo paths corresponding to the actual acoustic echo paths and generate estimated acoustic signals based on the estimated echo paths. The estimated acoustic signals correspond to the actual acoustic signals output from the speaker set 650. - The
processor 670 performs voice recognition processing on a subtraction signal obtained by subtracting the echo signals from an output signal of themicrophone 640. In detail, theprocessor 670 may include acontroller 673 and theadder 671. - The
adder 671 generates the subtraction signal by subtracting the echo signals output from theecho filter unit 620 from the output signal of themicrophone 640. - The
controller 673 performs voice recognition processing on the subtraction signal output from theadder 671. In detail, thecontroller 673 recognizes voice corresponding to the subtraction signal and performs an operation corresponding to the recognized voice. - For example, the
echo filter unit 620 may generate a first estimated acoustic signal by estimating an acoustic signal output from a first speaker (not shown) included in the speaker set 650 and generate a second estimated acoustic signal by estimating an acoustic signal output from a second speaker (not shown) included in the speaker set 650. Theadder 671 subtracts the first and second estimated acoustic signals output from theecho filter unit 620 from the output signal of themicrophone 640. Then, thecontroller 673 may perform voice recognition processing on an output signal of theadder 671 and recognize a command or data according to the output signal of theadder 671. - In addition, the above-described echo cancellation filter (not shown) may consist of the
echo filter unit 620 and theadder 671. -
FIG. 7 is a block diagram of anelectronic apparatus 700 according to an exemplary embodiment. - Referring to
FIG. 7 , theelectronic apparatus 700 corresponds to theelectronic apparatus 600 ofFIG. 6 . In theelectronic apparatus 700 ofFIG. 7 , which is compared with thedevice 600 ofFIG. 6 , a multi-channel signal Sm input to adecorrelator 710 includes first and second channel audio signals, aspeaker set 750 includes afirst speaker 751 for outputting the first channel audio signal and asecond speaker 752 for outputting the second channel audio signal, and anecho filter unit 720 includes afirst filter unit 721 and asecond filter unit 722. In addition, aprocessor 770 may include acontroller 773 and twoadders - The
decorrelator 710 outputs a first channel decorrelated signal Sc1 corresponding to the first channel audio signal and a second channel decorrelated signal Sc2 corresponding to the second channel audio signal. - The
first speaker 751 receives the first channel decorrelated signal Sc1 and outputs an R-channel audio signal Sout1, which is an acoustic signal which a user can audibly recognize. Thesecond speaker 752 receives the second channel decorrelated signal Sc2 and outputs an L-channel audio signal Sout2, which is an acoustic signal which the user can audibly recognize. - A
microphone 740 receives a voice signal Sin, the R-channel audio signal Sout1, and the L-channel audio signal Sout2 and outputs the voice signal Sin, the R-channel audio signal Sout1, and the L-channel audio signal Sout2. - The
first filter unit 721 receives the first channel decorrelated signal Sc1 and estimates echo paths through which the R-channel audio signal Sout1 is output. Then, thefirst filter unit 721 generates a first estimated acoustic signal Sp1 corresponding to the first channel decorrelated signal Sc1 by applying the first channel decorrelated signal Sc1 to the estimated echo paths. - The
second filter unit 722 receives the second channel decorrelated signal Sc2 and estimates echo paths through which the L-channel audio signal Sout2 is output. Then, thesecond filter unit 722 generates a second estimated acoustic signal Sp2 corresponding to the second channel decorrelated signal Sc2 by applying the second channel decorrelated signal Sc2 to the estimated echo paths. - The
adder 771 subtracts the second channel decorrelated signal Sc2 from an output signal of themicrophone 740. Theadder 772 subtracts the first channel decorrelated signal Sc1 from an output signal of theadder 771. Although the twoadders processor 770 inFIG. 7 , one adder may be used to subtract the first and second channel decorrelated signals Sc1 and Sc2 from the output signal of themicrophone 740. -
FIG. 8 is a flowchart illustrating an audiosignal processing method 800 according to an exemplary embodiment. The audiosignal processing method 800 may be performed by theaudio apparatus electronic apparatus signal processing method 800 is described with reference to theaudio apparatus 200. - Referring to
FIG. 8 , inoperation 810, a multi-channel signal including audio signals is received.Operation 810 may be performed by thedecorrelator 210. - In
operation 820, a phase shifting value is adjusted based on a correlation difference between the audio signals included in the multi-channel signal received inoperation 810.Operation 820 may be performed by thedecorrelator 210. - In
operation 830, decorrelated signals are generated by applying the phase shifting value adjusted inoperation 820 to the audio signals included in the multi-channel signal.Operation 830 may be performed by thedecorrelator 210. - The decorrelated signals generated in
operation 830 may be output through the plurality of speakers included in the speaker set 250. - Operations of the audio
signal processing method 800 are the same as an operation of theaudio apparatus electronic apparatus -
FIG. 9 is a flowchart illustrating an audiosignal processing method 900 according to an exemplary embodiment. The audiosignal processing method 900 may be performed by theaudio apparatus 300 or theelectronic apparatus signal processing method 900 is described with reference to theaudio apparatus 300. - Referring to
FIG. 9 , sinceoperation 910 is the same asoperation 810, a detailed description thereof is omitted. Hereinafter, a description is provided by assuming that a multi-channel signal, including first and second channel audio signals, is received. - In
operation 920, the multi-channel signal received inoperation 910 is divided into a plurality of sub-bands. In detail, the first and second channel audio signals are divided into the plurality of sub-bands, and first and second channel sub-audio signals are generated in each of the plurality of sub-bands.Operations pre-processor 320. - In
operation 930, a phase shifting value in a predetermined sub-band is adjusted based on a phase difference between the first and second channel sub-audio signals in the predetermined sub-band.Operation 930 may be performed by thedecorrelation processor 330. - In
operation 940, first and second sub-decorrelated signals are generated by applying the phase shifting value in the predetermined sub-band, which is generated inoperation 930, to the first and second channel sub-audio signals in the predetermined sub-band.Operation 940 may be performed by thedecorrelation processor 330. - In
operation 950, first and second decorrelated signals are generated by synthesizing first and second sub-decorrelated signals generated for the plurality of sub-bands.Operation 950 may be performed by thesynthesizer 340. - The first and second decorrelated signals generated in
operation 950 may be output through the first andsecond speakers - Operations of the audio
signal processing method 900 are the same as an operation of theaudio apparatus 300 or theelectronic apparatus -
FIG. 10 is a flowchart for describingoperation 930 ofFIG. 9 in more detail. - Referring to
FIG. 10 ,operation 930 corresponds tooperation 1000, which may includeoperations - In
operation 1010, a phase shifting value is set based on sub-bands. - In
operation 1020, a weighting value is set based on sub-bands. - In
operation 1030, a corrected phase shifting value is set by multiplying the phase shifting value set inoperation 1010 by the weighting value set inoperation 1020. - Since an operation of setting the corrected phase shifting value has been described in detail with reference to
FIGS. 5A and 5B , a detailed description thereof is omitted. - While not restricted thereto, an exemplary embodiment can be embodied as computer-readable code on a computer-readable recording medium. The computer-readable recording medium is any data storage device that can store data that can be thereafter read by a computer system. Examples of the computer-readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The computer-readable recording medium can also be distributed over network-coupled computer systems so that the computer-readable code is stored and executed in a distributed fashion. Also, aspects of an exemplary embodiment may be written as a computer program transmitted over a computer-readable transmission medium, such as a carrier wave, and received and implemented in general-use or special-purpose digital computers that execute the programs. Moreover, while not required in all aspects, one or more units of the audio signal processing apparatus can include a processor or microprocessor executing a computer program stored in a computer-readable medium.
- As described above, according to an audio signal processing method according to one or more exemplary embodiments, an audio apparatus using the same, and an electronic apparatus using the same, decorrelated signals corresponding to audio signals included in a multi-channel signal are generated based on a correlation between the audio signals. Thus, the audio signals along predetermined echo paths without divergence of an echo filter may be correctly estimated. Accordingly, an accuracy of voice recognition may increase.
- While the inventive concept has been particularly shown and described with reference to exemplary embodiments thereof, it will be understood by those of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope of the inventive concept as defined by the following claims.
Claims (24)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/483,571 US9047862B2 (en) | 2011-05-30 | 2012-05-30 | Audio signal processing method, audio apparatus therefor, and electronic apparatus therefor |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161491260P | 2011-05-30 | 2011-05-30 | |
KR1020120036799A KR101930907B1 (en) | 2011-05-30 | 2012-04-09 | Method for audio signal processing, audio apparatus thereof, and electronic apparatus thereof |
KR10-2012-0036799 | 2012-04-09 | ||
US13/483,571 US9047862B2 (en) | 2011-05-30 | 2012-05-30 | Audio signal processing method, audio apparatus therefor, and electronic apparatus therefor |
Publications (2)
Publication Number | Publication Date |
---|---|
US20120310638A1 true US20120310638A1 (en) | 2012-12-06 |
US9047862B2 US9047862B2 (en) | 2015-06-02 |
Family
ID=47262335
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/483,571 Active 2033-07-16 US9047862B2 (en) | 2011-05-30 | 2012-05-30 | Audio signal processing method, audio apparatus therefor, and electronic apparatus therefor |
Country Status (2)
Country | Link |
---|---|
US (1) | US9047862B2 (en) |
KR (1) | KR101930907B1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150199953A1 (en) * | 2013-02-28 | 2015-07-16 | Google Inc. | Non-linear post-processing control in stereo aec |
CN111739540A (en) * | 2020-07-20 | 2020-10-02 | 天域全感音科技有限公司 | Audio signal acquisition device, computer equipment and method |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020176585A1 (en) * | 2001-01-23 | 2002-11-28 | Egelmeers Gerardus Paul Maria | Asymmetric multichannel filter |
US7181019B2 (en) * | 2003-02-11 | 2007-02-20 | Koninklijke Philips Electronics N. V. | Audio coding |
US8498422B2 (en) * | 2002-04-22 | 2013-07-30 | Koninklijke Philips N.V. | Parametric multi-channel audio representation |
US8538037B2 (en) * | 2006-04-13 | 2013-09-17 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Audio signal decorrelator, multi channel audio signal processor, audio signal processor, method for deriving an output audio signal from an input audio signal and computer program |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5828756A (en) | 1994-11-22 | 1998-10-27 | Lucent Technologies Inc. | Stereophonic acoustic echo cancellation using non-linear transformations |
FR2762467B1 (en) | 1997-04-16 | 1999-07-02 | France Telecom | MULTI-CHANNEL ACOUSTIC ECHO CANCELING METHOD AND MULTI-CHANNEL ACOUSTIC ECHO CANCELER |
US6895093B1 (en) | 1998-03-03 | 2005-05-17 | Texas Instruments Incorporated | Acoustic echo-cancellation system |
-
2012
- 2012-04-09 KR KR1020120036799A patent/KR101930907B1/en active IP Right Grant
- 2012-05-30 US US13/483,571 patent/US9047862B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020176585A1 (en) * | 2001-01-23 | 2002-11-28 | Egelmeers Gerardus Paul Maria | Asymmetric multichannel filter |
US8498422B2 (en) * | 2002-04-22 | 2013-07-30 | Koninklijke Philips N.V. | Parametric multi-channel audio representation |
US7181019B2 (en) * | 2003-02-11 | 2007-02-20 | Koninklijke Philips Electronics N. V. | Audio coding |
US8538037B2 (en) * | 2006-04-13 | 2013-09-17 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Audio signal decorrelator, multi channel audio signal processor, audio signal processor, method for deriving an output audio signal from an input audio signal and computer program |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150199953A1 (en) * | 2013-02-28 | 2015-07-16 | Google Inc. | Non-linear post-processing control in stereo aec |
US9123324B2 (en) * | 2013-02-28 | 2015-09-01 | Google Inc. | Non-linear post-processing control in stereo acoustic echo cancellation |
CN111739540A (en) * | 2020-07-20 | 2020-10-02 | 天域全感音科技有限公司 | Audio signal acquisition device, computer equipment and method |
Also Published As
Publication number | Publication date |
---|---|
KR20120133995A (en) | 2012-12-11 |
KR101930907B1 (en) | 2019-03-12 |
US9047862B2 (en) | 2015-06-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8046219B2 (en) | Robust two microphone noise suppression system | |
US8892432B2 (en) | Signal processing system, apparatus and method used on the system, and program thereof | |
EP3189521B1 (en) | Method and apparatus for enhancing sound sources | |
US9437180B2 (en) | Adaptive noise reduction using level cues | |
US8345890B2 (en) | System and method for utilizing inter-microphone level differences for speech enhancement | |
EP2245861B1 (en) | Enhanced blind source separation algorithm for highly correlated mixtures | |
US9443532B2 (en) | Noise reduction using direction-of-arrival information | |
US20120288100A1 (en) | Method and apparatus for processing multi-channel de-correlation for cancelling multi-channel acoustic echo | |
US8958572B1 (en) | Adaptive noise cancellation for multi-microphone systems | |
US9516411B2 (en) | Signal-separation system using a directional microphone array and method for providing same | |
US20080031466A1 (en) | Multi-channel echo compensation system | |
CN111128210B (en) | Method and system for audio signal processing with acoustic echo cancellation | |
US11671755B2 (en) | Microphone mixing for wind noise reduction | |
US8682006B1 (en) | Noise suppression based on null coherence | |
US20130016854A1 (en) | Microphone array processing system | |
US9886966B2 (en) | System and method for improving noise suppression using logistic function and a suppression target value for automatic speech recognition | |
US20150189436A1 (en) | Method, apparatus, computer program code and storage medium for processing audio signals | |
US9412354B1 (en) | Method and apparatus to use beams at one end-point to support multi-channel linear echo control at another end-point | |
US9047862B2 (en) | Audio signal processing method, audio apparatus therefor, and electronic apparatus therefor | |
US20110051955A1 (en) | Microphone signal compensation apparatus and method thereof | |
US9905247B2 (en) | Signal processing apparatus, medium apparatus, signal processing method, and signal processing program | |
US20230319469A1 (en) | Suppressing Spatial Noise in Multi-Microphone Devices | |
US20210076133A1 (en) | Input signal decorrelation | |
EP3029671A1 (en) | Method and apparatus for enhancing sound sources |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JEONG, JAE-HOON;JEONG, SO-YOUNG;KIM, JEONG-SU;AND OTHERS;REEL/FRAME:028289/0103 Effective date: 20120530 |
|
FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |