CN104347076A - Network audio packet loss concealment method and device - Google Patents
Network audio packet loss concealment method and device Download PDFInfo
- Publication number
- CN104347076A CN104347076A CN201310345063.6A CN201310345063A CN104347076A CN 104347076 A CN104347076 A CN 104347076A CN 201310345063 A CN201310345063 A CN 201310345063A CN 104347076 A CN104347076 A CN 104347076A
- Authority
- CN
- China
- Prior art keywords
- sound signal
- packet loss
- frame
- amplitude spectrum
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Landscapes
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
The invention discloses a network audio packet loss concealment method and device, relating to the field of audio transmission. The method comprises the steps of decoding an audio stream subjected to packet loss, reversing audio signals obtained through decoding, then, carrying out packet loss compensation on the audio signals and the reversed audio signals respectively, reversing the compensated reversed audio signals again, and finally, carrying out magnitude average on the compensated audio signals of two channels, so as to obtain output signals. By the bidirectional-prediction error concealment technology, the quality of audio can be further improved, and particularly, a relatively clear voice effect can be obtained in the network environment that the packet loss rate is relatively high.
Description
Technical field
The present invention relates to audio transmission field, particularly a kind of network audio packet loss covering method and device, thus the discontinuous sense that attenuating network audio stream produces due to packet loss, improve the auditory effect of network audio stream.
Background technology
At VOIP(Voice over Internet Protocol, the networking telephone) in application, due to some reasons such as network qualities, may there is packet loss phenomenon in audio stream, and serious may affect voice quality, causes interference to call.
ILBC(internet Low Bit Rate Codec, Internet Low Bit-rate is encoded) algorithm is a kind of based on CELP(Code Excited Linear Prediction, Qualcomm Code Excited Linear Prediction (QCELP)) low bit rate speech coding algorithm, the voice quality of its excellence, outstanding long-term prediction method and packet loss shelter (Packet Loss Concealment, be called for short PLC) technology, solve the problem of interconnected transfers on network voice well.ILBC designs mainly for packet network, and main advantage is its process for Network Packet Loss.ILBC adopts and carries out direct coding by original state structure adaptive codebook to pumping signal, and achieve the separate coding between speech frame, the impact of frame losing is confined in this frame.When after loss of data, iLBC according to the pumping signal recorded before and speech parameter, can carry out relevant treatment, residual error strengthens, mix the operations such as white noise, generates the voice signal of simulation, thus substitute the voice lost.Therefore, iLBC, when audio stream generation packet loss, can lower the discontinuous sense that packet loss produces.
But, serious packet loss phenomenon may be there is when network environment is poor, adopt existing iLBC packet loss macking technique still cannot obtain satisfied sound effect, therefore, be necessary to propose a kind of packet loss macking technique under the network environment that packet loss is higher, still can obtain sound effect more clearly.
Summary of the invention
An embodiment of the present invention technical matters to be solved is: solve the problem how obtaining sound effect more clearly under the network environment that packet loss is higher.
According to an aspect of the embodiment of the present invention, a kind of network audio packet loss covering method is proposed, comprise: be pulse code modulation (PCM) PCM signal by the audio stream after packet loss, caused by packet loss the PCM data value of the data segment lacked to be set to 0, using the PCM signal after process as sound signal S; Carry out obtaining sound signal S ' time reversal to sound signal S; Sound signal S obtains sound signal Si through Internet Low Bit-rate coding iLBC Discarded Packets compensation process, and sound signal S ' obtains sound signal Si ' through iLBC Discarded Packets compensation process; Carry out obtaining sound signal Sit time reversal to sound signal Si '; By moving window, the sound signal So that amplitude spectrum on average obtains output is carried out to sound signal Si and sound signal Sit.
By moving window, the sound signal So that amplitude spectrum on average obtains output is carried out to sound signal Si and sound signal Sit, specifically comprise: adopt moving window to sound signal Si, sound signal Sit, sound signal So carries out framing, the length 2N of every frame, slip between two frames is spaced apart N, and remembers that the i-th frame data in sound signal Si are vector α
i, the i-th frame data in sound signal Sit are vector beta
i, the i-th frame data in sound signal So are vector O
i, the vector of window function is w; According to window function to α
icarry out windowing operation to obtain
, right
its amplitude spectrum A is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum φ
i; According to window function to β
icarry out windowing operation to obtain
, right
its amplitude spectrum B is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum θ
i; To amplitude spectrum A
iwith amplitude spectrum B
ibe averaged and obtain
; To with
for amplitude spectrum with φ
ifor the signal of phase spectrum carries out inverse fast Fourier transform, the result according to inverse fast Fourier transform calculates O
i.
ILBC Discarded Packets compensation process specifically comprises: rebuild linear predictor coefficient to lost frames, and rebuild residual signals.
According to another aspect of the embodiment of the present invention, a kind of network audio packet loss covering appts is proposed, comprise: decoder module, for being pulse code modulation (PCM) PCM signal by the audio stream after packet loss, the PCM data value of the data segment lacked is caused by packet loss to be set to 0, using the PCM signal after process as sound signal S; Reversal block, for carrying out obtaining sound signal S ' to sound signal S time reversal; Compensating module, obtain sound signal Si for sound signal S through Internet Low Bit-rate coding iLBC Discarded Packets compensation process, sound signal S ' obtains sound signal Si ' through iLBC Discarded Packets compensation process; Reversal block, also for carrying out obtaining sound signal Sit time reversal to sound signal Si '; Amplitude averaging module, for carrying out by moving window the sound signal So that amplitude spectrum on average obtains output to sound signal Si and sound signal Sit.
Amplitude averaging module, specifically for: adopt moving window to sound signal Si, sound signal Sit, sound signal So carry out framing, the length 2N of every frame, and the slip between two frames is spaced apart N, and remember that the i-th frame data in sound signal Si are vector α
i, the i-th frame data in sound signal Sit are vector beta
i, the i-th frame data in sound signal So are vector O
i, the vector of window function is w; According to window function to α
icarry out windowing operation to obtain
, right
its amplitude spectrum A is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum φ
i; According to window function to β
icarry out windowing operation to obtain
, right
its amplitude spectrum B is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum θ
i; To amplitude spectrum A
iwith amplitude spectrum B
ibe averaged and obtain
; To with
for amplitude spectrum with φ
ifor the signal of phase spectrum carries out inverse fast Fourier transform, the result according to inverse fast Fourier transform calculates O
i.
The iLBC Discarded Packets compensation process of compensating module specifically for: linear predictor coefficient is rebuild to lost frames, and rebuilds residual signals.
The present invention is to the audio stream after packet loss, and the sound signal obtained of decoding is reversed, then respectively Discarded Packets compensation is carried out to sound signal and reversion sound signal, and the reversion sound signal after compensating is reversed again, finally the sound signal amplitude of carrying out after two-way compensation is on average outputed signal, this bi-directional predicted error concealment technology can improve audio quality further, especially under the network environment that packet loss is higher, still can obtain sound effect more clearly.
By referring to the detailed description of accompanying drawing to exemplary embodiment of the present invention, further feature of the present invention and advantage thereof will become clear.
Accompanying drawing explanation
In order to be illustrated more clearly in the embodiment of the present invention or technical scheme of the prior art, be briefly described to the accompanying drawing used required in embodiment or description of the prior art below, apparently, accompanying drawing in the following describes is only some embodiments of the present invention, for those of ordinary skill in the art, under the prerequisite not paying creative work, other accompanying drawing can also be obtained according to these accompanying drawings.
Fig. 1 is the schematic flow sheet of a network audio packet loss covering method of the present invention embodiment.
Fig. 2 is the Hanning window function schematic diagram that the present invention adopts.
Fig. 3 is that the present invention adopts moving window to voice signal overlap-add operation schematic diagram.
Fig. 4 is the result schematic diagram that the present invention assesses the voice quality after error covering.
Fig. 5 is the structural representation of a network audio packet loss covering appts of the present invention embodiment.
Embodiment
Below in conjunction with the accompanying drawing in the embodiment of the present invention, be clearly and completely described the technical scheme in the embodiment of the present invention, obviously, described embodiment is only the present invention's part embodiment, instead of whole embodiments.Illustrative to the description only actually of at least one exemplary embodiment below, never as any restriction to the present invention and application or use.Based on the embodiment in the present invention, those of ordinary skill in the art, not making the every other embodiment obtained under creative work prerequisite, belong to the scope of protection of the invention.
In order to obtain sound effect more clearly under the network environment that packet loss is higher, the present invention adopts bi-directional predicted error concealment technology can improve audio quality further, special under the network environment that packet loss is higher, still can obtain sound effect more clearly.
Fig. 1 is the schematic flow sheet of a network audio packet loss covering method of the present invention embodiment.
As shown in Figure 1, the network audio packet loss covering method of the present embodiment comprises the following steps:
Audio stream after packet loss is PCM(pulse code modulation (PCM) by S101) signal, caused by packet loss the PCM data value of the data segment lacked to be set to 0, using the PCM signal after process as sound signal S.
S102, carries out obtaining sound signal S ' to sound signal S time reversal.
The present embodiment can by sound signal S and sound signal S ' respectively stored in two buffer zones.
S103, sound signal S obtain sound signal Si through iLBC Discarded Packets compensation process, and sound signal S ' obtains sound signal Si ' through iLBC Discarded Packets compensation process.
Wherein, the follow-up detailed description of iLBC Discarded Packets compensation processing procedure.
S104, carries out obtaining sound signal Sit to sound signal Si ' time reversal.
S105, carries out by moving window the sound signal So that amplitude spectrum on average obtains output to sound signal Si and sound signal Sit.
The present invention proposes a kind of exemplary amplitude spectrum averaging method, can refer step (1) ~ step (5), specific as follows:
(1) suppose the sound signal Si and the sound signal Sit that there are two sections of equal length, carrying out superposing the specific practice obtaining output audio signal So is:
Adopt moving window b to sound signal Si, sound signal Sit, sound signal So carry out framing, the length 2N of every frame, and the slip between two frames is spaced apart N, namely each frame and two interframe before and after it have respectively 50% overlapping.Do the flatness that can ensure to superpose the voice obtained like this, and remember that the i-th frame data in sound signal Si are vector α
i, the i-th frame data in sound signal Sit are vector beta
i, the i-th frame data in sound signal So are vector O
i, the Hanning(Hamming that 2N point is corresponding) and the vector of window function is w.
Fig. 2 is the Hanning window function schematic diagram that the present invention adopts, and as shown in Figure 2, the length of Hanning window is 2N, N=80.
Fig. 3 is that the present invention adopts moving window to voice signal overlap-add operation schematic diagram, and as shown in Figure 3, the spacing between two moving windows is 1/2 frame.
(2) according to window function to α
icarry out windowing operation to obtain
, right
carry out Fast Fourier Transform (FFT) (being called for short FFT) and calculate its amplitude spectrum A afterwards
iwith phase spectrum φ
i.
Windowing operation can with reference to following formula:
Wherein, represents windowing operation.
Amplitude spectrum calculates can with reference to following formula:
Wherein, abs represents that plural number takes absolute value computing, and F represents Fast Fourier Transform (FFT).
Phase spectrum calculates can with reference to following formula:
Wherein, ang represents and gets phasing degree to plural number, unit radian.
(3) according to window function to β
icarry out windowing operation to obtain
, right
its amplitude spectrum B is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum θ
i.
Correlation computations can join out previous step (2), repeats no more here.
(4) to amplitude spectrum A
iwith amplitude spectrum B
ibe averaged and obtain
, can with reference to following formula:
(5) to
for amplitude spectrum with φ
isignal for phase spectrum carries out inverse fast Fourier transform (being called for short IFFT), and the result according to inverse fast Fourier transform calculates O
i, specifically can with reference to following formula:
Wherein, real (.) represents that getting real part to plural number operates.
It should be noted that, the Discarded Packets compensation processing procedure based on iLBC technology can with reference to prior art, such as, by Andersen, S.V.; Kleijn, W.B.; Hagen, R.; Linden, J.; Murthi, M.N.; The people such as Skoglund, J. deliver " iLBC-a linear predictive coder with robustness to packet losses, " Speech Coding, 2002, IEEE Workshop Proceedings., vol., no., pp.23,25,6-9Oct.2002.As a kind of example, brief description iLBC Discarded Packets compensation processing procedure.
In the process that decoding end is decoded frame by frame to the code stream received, after iLBC Decoder accepts to the code stream of every frame, first to judge that whether present frame is complete, be handled as follows respectively according to complete or generation packet loss:
If complete, then according to normal iLBC decoding process reconstructed speech signal, need the status information of preserving present frame, these status informations comprise linear predictor coefficient (being called for short LPC) information, decoded residual signals etc.If the bit rate of next frame is lost, the information that these are preserved will be used.
If there occurs packet loss, so just carry out packet loss and shelter process (PLC process).PLC is mainly according to the decoded information of former frame, and the lost frames that the method approximate substitution utilizing pitch synchronous to repeat is current, to reach Discarded Packets compensation.If present frame is lost, just need the PLC method reconstructing lost speech data adopting follow-up introduction, if there is continuous multiple frames packet loss, just need the PLC method reconstructing lost speech data repeatedly adopting follow-up introduction.It should be noted that, the frame that face is lost more rearward is more difficult to accurate reconstruction, successively decreases frame by frame so adopt the gain of continual data package dropout, to avoid introducing larger distorted signals.
PLC reconstructing lost speech data mainly comprises two steps:
(1) linear predictor coefficient (being called for short LPC) is rebuild
Due to from spatially or the time, last subframe all has maximum correlation with the linear predictor coefficient of current lost frames, therefore, the present invention: using the linear predictor coefficient of the linear predictor coefficient of last subframe of past frame as lost frames.
(2) residual signals (or claiming pumping signal) is rebuild
Residual signals can be divided into two parts usually: quasi-periodicity composition and noise like composition.Therefore in fact first PLC needs to rebuild this two parts, composition can be similar to according to the pitch period measuring former frame and obtain quasi-periodicity, noise like composition then can obtain by producing random noise, and the energy proportion of the two also can use for reference the proportionate relationship of former frame.Therefore, pitch Detection is carried out to former frame, with the phonological component of pitch synchronous mode reconstructing lost frame, utilize correlativity to obtain the gain of class noise, phonological component is mixed to rebuild residual signals with class noise.
Network audio packet loss covering method of the present invention, to the audio stream after packet loss, and the sound signal obtained of decoding is reversed, then respectively Discarded Packets compensation is carried out to sound signal and reversion sound signal, and the reversion sound signal after compensating is reversed again, finally the sound signal amplitude of carrying out after two-way compensation is on average outputed signal, this bi-directional predicted error concealment technology can improve audio quality further, special under the network environment that packet loss is higher, still can obtain sound effect more clearly.
In order to prove that bi-directional predicted error concealment technology can improve audio quality further, invention has been experiment simulation, below experimentation and simulation result being described.
The sound signal that experiment adopts is a segment length is 100s, and sampling rate is the voice signal of 8k, stores, be designated as X with 16 bit PCM format.
Experimental procedure is as follows:
(1) adopt G.729 mode to encode to X, G.729 frame length is 160 sample points, and the length of X just in time may be partitioned into an integer audio frame, and voice packet presses generation time serial number.
(2) analog network packet loss process.According to the packet loss r pre-set, produce 0/1 random series, sequence length equals the number of speech frame, i.e. the corresponding voice packet of each random number, and the voice packet of corresponding numeral " 1 " is considered to abandon.
(3) order is decoded to voice packet by number, carries out normal decoder, export the PCM data after this frame decoding to the voice packet that packet loss zone bit is " 0 ".Be the voice packet of " 1 " to packet loss zone bit, directly export the PCM data that 20ms value is 0 entirely.The speech data that this decode procedure obtains is designated as S.
(4) use iLBC demoder to encode to S, the frame length of iLBC coding is set to 160 sample points.Again decode to the voice after coding, iLBC except needs number of speech frames certificate, also needs the packet loss zone bit that this frame is corresponding when decoding.Be the speech frame (lost frames) of " 1 " to packet loss zone bit, the phonetic synthesis parameter of this frame of phonetic synthesis parameter prediction that iLBC can obtain according to decoding former frame when decoding, and reconstruct the speech data of loss.
(5) carry out obtaining sound signal S ' time reversal to S, and use iLBC to encode to S '.Time reversal is carried out to the 0/1 packet loss zone bit sequence generated before.And with iLBC, to speech frame and corresponding packet loss zone bit thereof, it is decoded, and obtains the voice Si ' after Discarded Packets compensation, is carried out obtaining signal Sit time reversal by signal Si '.
(6) voice signal Si and voice signal Sit is carried out amplitude spectrum overlap-add operation and obtain output audio signal So.
The present invention adopts subjective speech quality assessment (PESQ) (see ITU-T suggestion P.862) to assess the voice quality after error concealment, and assessment result as shown in Figure 4.Result shows, when packet loss is more than 5%, the method that the present invention proposes uses iLBC algorithm compared to simple, more effectively can improve the voice quality after error concealment.
Fig. 5 is the structural representation of a network audio packet loss covering appts of the present invention embodiment.
As shown in Figure 5, the network audio packet loss covering appts of the present embodiment comprises:
Decoder module 501, for being pulse code modulation (PCM) PCM signal by the audio stream after packet loss, causes the PCM data value of the data segment lacked to be set to 0, using the PCM signal after process as sound signal S by packet loss;
Reversal block 502, for carrying out obtaining sound signal S ' to sound signal S time reversal;
Compensating module 503, obtain sound signal Si for sound signal S through Internet Low Bit-rate coding iLBC Discarded Packets compensation process, sound signal S ' obtains sound signal Si ' through iLBC Discarded Packets compensation process;
Reversal block 502, also for carrying out obtaining sound signal Sit time reversal to sound signal Si ';
Amplitude averaging module 504, for carrying out by moving window the sound signal So that amplitude spectrum on average obtains output to sound signal Si and sound signal Sit.
Wherein, amplitude averaging module 504 specifically for:
Adopt moving window to sound signal Si, sound signal Sit, sound signal So carry out framing, the length 2N of every frame, and the slip between two frames is spaced apart N, and remember that the i-th frame data in sound signal Si are vector α
i, the i-th frame data in sound signal Sit are vector beta
i, the i-th frame data in sound signal So are vector O
i, the vector of window function is w;
According to window function to α
icarry out windowing operation to obtain
, right
its amplitude spectrum A is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum φ
i;
According to window function to β
icarry out windowing operation to obtain
, right
its amplitude spectrum B is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum θ
i;
To amplitude spectrum A
iwith amplitude spectrum B
ibe averaged and obtain
;
To with
for amplitude spectrum with φ
ifor the signal of phase spectrum carries out inverse fast Fourier transform, the result according to inverse fast Fourier transform calculates O
i.
Wherein, the iLBC Discarded Packets compensation process of compensating module 503 specifically for: linear predictor coefficient is rebuild to lost frames, and rebuilds residual signals.
Wherein, compensating module 503 rebuild linear predictor coefficient time specifically for: using the linear predictor coefficient of the linear predictor coefficient of last subframe of past frame as lost frames.
Wherein, compensating module 503 rebuild residual signals time specifically for: pitch Detection is carried out to former frame, with the phonological component of pitch synchronous mode reconstructing lost frame, utilize correlativity to obtain the gain of class noise, phonological component is mixed to rebuild residual signals with class noise.
Network audio packet loss covering appts of the present invention, to the audio stream after packet loss, and the sound signal obtained of decoding is reversed, then respectively Discarded Packets compensation is carried out to sound signal and reversion sound signal, and the reversion sound signal after compensating is reversed again, finally the sound signal amplitude of carrying out after two-way compensation is on average outputed signal, this bi-directional predicted error concealment technology can improve audio quality further, special under the network environment that packet loss is higher, still can obtain sound effect more clearly.
One of ordinary skill in the art will appreciate that all or part of step realizing above-described embodiment can have been come by hardware, the hardware that also can carry out instruction relevant by program completes, described program can be stored in a kind of computer-readable recording medium, the above-mentioned storage medium mentioned can be ROM (read-only memory), disk or CD etc.
The foregoing is only preferred embodiment of the present invention, not in order to limit the present invention, within the spirit and principles in the present invention all, any amendment done, equivalent replacement, improvement etc., all should be included within protection scope of the present invention.
Claims (10)
1. a network audio packet loss covering method, comprising:
Be pulse code modulation (PCM) PCM signal by the audio stream after packet loss, caused by packet loss the PCM data value of the data segment lacked to be set to 0, using the PCM signal after process as sound signal S;
Carry out obtaining sound signal S ' time reversal to sound signal S;
Sound signal S obtains sound signal Si through Internet Low Bit-rate coding iLBC Discarded Packets compensation process, and sound signal S ' obtains sound signal Si ' through iLBC Discarded Packets compensation process;
Carry out obtaining sound signal Sit time reversal to sound signal Si ';
By moving window, the sound signal So that amplitude spectrum on average obtains output is carried out to sound signal Si and sound signal Sit.
2. method according to claim 1, is characterized in that, described to sound signal Si and sound signal Sit by moving window carry out amplitude spectrum on average obtain export sound signal So, specifically comprise:
Adopt moving window to sound signal Si, sound signal Sit, sound signal So carry out framing, the length 2N of every frame, and the slip between two frames is spaced apart N, and remember that the i-th frame data in sound signal Si are vector α
i, the i-th frame data in sound signal Sit are vector beta
i, the i-th frame data in sound signal So are vector O
i, the vector of window function is w;
According to window function to α
icarry out windowing operation to obtain
, right
its amplitude spectrum A is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum φ
i;
According to window function to β
icarry out windowing operation to obtain
, right
its amplitude spectrum B is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum θ
i;
To amplitude spectrum A
iwith amplitude spectrum B
ibe averaged and obtain
;
To with
for amplitude spectrum with φ
ifor the signal of phase spectrum carries out inverse fast Fourier transform, the result according to inverse fast Fourier transform calculates O
i.
3. method according to claim 1, is characterized in that, described iLBC Discarded Packets compensation process specifically comprises: rebuild linear predictor coefficient to lost frames, and rebuild residual signals.
4. method according to claim 3, is characterized in that, described to lost frames rebuild linear predictor coefficient specifically comprise: using the linear predictor coefficient of the linear predictor coefficient of last subframe of past frame as lost frames.
5. method according to claim 3, it is characterized in that, described reconstruction residual signals specifically comprises: carry out pitch Detection to former frame, with the phonological component of pitch synchronous mode reconstructing lost frame, utilize correlativity to obtain the gain of class noise, phonological component is mixed to rebuild residual signals with class noise.
6. a network audio packet loss covering appts, comprising:
Decoder module, for being pulse code modulation (PCM) PCM signal by the audio stream after packet loss, causes the PCM data value of the data segment lacked to be set to 0, using the PCM signal after process as sound signal S by packet loss;
Reversal block, for carrying out obtaining sound signal S ' to sound signal S time reversal;
Compensating module, obtain sound signal Si for sound signal S through Internet Low Bit-rate coding iLBC Discarded Packets compensation process, sound signal S ' obtains sound signal Si ' through iLBC Discarded Packets compensation process;
Reversal block, also for carrying out obtaining sound signal Sit time reversal to sound signal Si ';
Amplitude averaging module, for carrying out by moving window the sound signal So that amplitude spectrum on average obtains output to sound signal Si and sound signal Sit.
7. device according to claim 6, is characterized in that, described amplitude averaging module, specifically for:
Adopt moving window to sound signal Si, sound signal Sit, sound signal So carry out framing, the length 2N of every frame, and the slip between two frames is spaced apart N, and remember that the i-th frame data in sound signal Si are vector α
i, the i-th frame data in sound signal Sit are vector beta
i, the i-th frame data in sound signal So are vector O
i, the vector of window function is w;
According to window function to α
icarry out windowing operation to obtain
, right
its amplitude spectrum A is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum φ
i;
According to window function to β
icarry out windowing operation to obtain
, right
its amplitude spectrum B is calculated after carrying out Fast Fourier Transform (FFT)
iwith phase spectrum θ
i;
To amplitude spectrum A
iwith amplitude spectrum B
ibe averaged and obtain
;
To with
for amplitude spectrum with φ
ifor the signal of phase spectrum carries out inverse fast Fourier transform, the result according to inverse fast Fourier transform calculates O
i.
8. device according to claim 6, is characterized in that, the iLBC Discarded Packets compensation process of described compensating module specifically for: linear predictor coefficient is rebuild to lost frames, and rebuilds residual signals.
9. device according to claim 8, is characterized in that, described compensating module rebuild linear predictor coefficient time specifically for: using the linear predictor coefficient of the linear predictor coefficient of last subframe of past frame as lost frames.
10. device according to claim 8, it is characterized in that, described compensating module rebuild residual signals time specifically for: pitch Detection is carried out to former frame, with the phonological component of pitch synchronous mode reconstructing lost frame, utilize correlativity to obtain the gain of class noise, phonological component is mixed to rebuild residual signals with class noise.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310345063.6A CN104347076B (en) | 2013-08-09 | 2013-08-09 | Network audio packet loss covering method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310345063.6A CN104347076B (en) | 2013-08-09 | 2013-08-09 | Network audio packet loss covering method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104347076A true CN104347076A (en) | 2015-02-11 |
CN104347076B CN104347076B (en) | 2017-07-14 |
Family
ID=52502544
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201310345063.6A Active CN104347076B (en) | 2013-08-09 | 2013-08-09 | Network audio packet loss covering method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104347076B (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106960673A (en) * | 2017-02-08 | 2017-07-18 | 中国人民解放军信息工程大学 | A kind of voice covering method and equipment |
CN107689226A (en) * | 2017-08-29 | 2018-02-13 | 中国民航大学 | A kind of low capacity Methods of Speech Information Hiding based on iLBC codings |
CN111326166A (en) * | 2020-02-25 | 2020-06-23 | 网易(杭州)网络有限公司 | Voice processing method and device, computer readable storage medium and electronic equipment |
CN111640442A (en) * | 2020-06-01 | 2020-09-08 | 北京猿力未来科技有限公司 | Method for processing audio packet loss, method for training neural network and respective devices |
CN111883147A (en) * | 2020-07-23 | 2020-11-03 | 北京达佳互联信息技术有限公司 | Audio data processing method and device, computer equipment and storage medium |
CN111883170A (en) * | 2020-04-08 | 2020-11-03 | 珠海市杰理科技股份有限公司 | Voice signal processing method and system, audio processing chip and electronic equipment |
US10997982B2 (en) | 2018-05-31 | 2021-05-04 | Shure Acquisition Holdings, Inc. | Systems and methods for intelligent voice activation for auto-mixing |
CN113035207A (en) * | 2021-03-03 | 2021-06-25 | 北京猿力未来科技有限公司 | Audio processing method and device |
US11297423B2 (en) | 2018-06-15 | 2022-04-05 | Shure Acquisition Holdings, Inc. | Endfire linear array microphone |
US11297426B2 (en) | 2019-08-23 | 2022-04-05 | Shure Acquisition Holdings, Inc. | One-dimensional array microphone with improved directivity |
US11303981B2 (en) | 2019-03-21 | 2022-04-12 | Shure Acquisition Holdings, Inc. | Housings and associated design features for ceiling array microphones |
US11302347B2 (en) | 2019-05-31 | 2022-04-12 | Shure Acquisition Holdings, Inc. | Low latency automixer integrated with voice and noise activity detection |
US11310596B2 (en) | 2018-09-20 | 2022-04-19 | Shure Acquisition Holdings, Inc. | Adjustable lobe shape for array microphones |
US11310592B2 (en) | 2015-04-30 | 2022-04-19 | Shure Acquisition Holdings, Inc. | Array microphone system and method of assembling the same |
US11438691B2 (en) | 2019-03-21 | 2022-09-06 | Shure Acquisition Holdings, Inc. | Auto focus, auto focus within regions, and auto placement of beamformed microphone lobes with inhibition functionality |
US11445294B2 (en) | 2019-05-23 | 2022-09-13 | Shure Acquisition Holdings, Inc. | Steerable speaker array, system, and method for the same |
US11477327B2 (en) | 2017-01-13 | 2022-10-18 | Shure Acquisition Holdings, Inc. | Post-mixing acoustic echo cancellation systems and methods |
US11523212B2 (en) | 2018-06-01 | 2022-12-06 | Shure Acquisition Holdings, Inc. | Pattern-forming microphone array |
US11552611B2 (en) | 2020-02-07 | 2023-01-10 | Shure Acquisition Holdings, Inc. | System and method for automatic adjustment of reference gain |
US11558693B2 (en) | 2019-03-21 | 2023-01-17 | Shure Acquisition Holdings, Inc. | Auto focus, auto focus within regions, and auto placement of beamformed microphone lobes with inhibition and voice activity detection functionality |
US11678109B2 (en) | 2015-04-30 | 2023-06-13 | Shure Acquisition Holdings, Inc. | Offset cartridge microphones |
US11706562B2 (en) | 2020-05-29 | 2023-07-18 | Shure Acquisition Holdings, Inc. | Transducer steering and configuration systems and methods using a local positioning system |
US11785380B2 (en) | 2021-01-28 | 2023-10-10 | Shure Acquisition Holdings, Inc. | Hybrid audio beamforming system |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101325631A (en) * | 2007-06-14 | 2008-12-17 | 华为技术有限公司 | Method and apparatus for implementing bag-losing hide |
CN102110441A (en) * | 2010-12-22 | 2011-06-29 | 中国科学院声学研究所 | Method for generating sound masking signal based on time reversal |
US20120101814A1 (en) * | 2010-10-25 | 2012-04-26 | Polycom, Inc. | Artifact Reduction in Packet Loss Concealment |
WO2012070340A1 (en) * | 2010-11-26 | 2012-05-31 | 株式会社エヌ・ティ・ティ・ドコモ | Concealment signal generating device, concealment signal generation method and concealment signal generation program |
WO2012158159A1 (en) * | 2011-05-16 | 2012-11-22 | Google Inc. | Packet loss concealment for audio codec |
-
2013
- 2013-08-09 CN CN201310345063.6A patent/CN104347076B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101325631A (en) * | 2007-06-14 | 2008-12-17 | 华为技术有限公司 | Method and apparatus for implementing bag-losing hide |
US20120101814A1 (en) * | 2010-10-25 | 2012-04-26 | Polycom, Inc. | Artifact Reduction in Packet Loss Concealment |
WO2012070340A1 (en) * | 2010-11-26 | 2012-05-31 | 株式会社エヌ・ティ・ティ・ドコモ | Concealment signal generating device, concealment signal generation method and concealment signal generation program |
CN102110441A (en) * | 2010-12-22 | 2011-06-29 | 中国科学院声学研究所 | Method for generating sound masking signal based on time reversal |
WO2012158159A1 (en) * | 2011-05-16 | 2012-11-22 | Google Inc. | Packet loss concealment for audio codec |
Cited By (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11310592B2 (en) | 2015-04-30 | 2022-04-19 | Shure Acquisition Holdings, Inc. | Array microphone system and method of assembling the same |
US11832053B2 (en) | 2015-04-30 | 2023-11-28 | Shure Acquisition Holdings, Inc. | Array microphone system and method of assembling the same |
US11678109B2 (en) | 2015-04-30 | 2023-06-13 | Shure Acquisition Holdings, Inc. | Offset cartridge microphones |
US11477327B2 (en) | 2017-01-13 | 2022-10-18 | Shure Acquisition Holdings, Inc. | Post-mixing acoustic echo cancellation systems and methods |
CN106960673A (en) * | 2017-02-08 | 2017-07-18 | 中国人民解放军信息工程大学 | A kind of voice covering method and equipment |
CN107689226A (en) * | 2017-08-29 | 2018-02-13 | 中国民航大学 | A kind of low capacity Methods of Speech Information Hiding based on iLBC codings |
US11798575B2 (en) | 2018-05-31 | 2023-10-24 | Shure Acquisition Holdings, Inc. | Systems and methods for intelligent voice activation for auto-mixing |
US10997982B2 (en) | 2018-05-31 | 2021-05-04 | Shure Acquisition Holdings, Inc. | Systems and methods for intelligent voice activation for auto-mixing |
US11800281B2 (en) | 2018-06-01 | 2023-10-24 | Shure Acquisition Holdings, Inc. | Pattern-forming microphone array |
US11523212B2 (en) | 2018-06-01 | 2022-12-06 | Shure Acquisition Holdings, Inc. | Pattern-forming microphone array |
US11297423B2 (en) | 2018-06-15 | 2022-04-05 | Shure Acquisition Holdings, Inc. | Endfire linear array microphone |
US11770650B2 (en) | 2018-06-15 | 2023-09-26 | Shure Acquisition Holdings, Inc. | Endfire linear array microphone |
US11310596B2 (en) | 2018-09-20 | 2022-04-19 | Shure Acquisition Holdings, Inc. | Adjustable lobe shape for array microphones |
US11558693B2 (en) | 2019-03-21 | 2023-01-17 | Shure Acquisition Holdings, Inc. | Auto focus, auto focus within regions, and auto placement of beamformed microphone lobes with inhibition and voice activity detection functionality |
US11303981B2 (en) | 2019-03-21 | 2022-04-12 | Shure Acquisition Holdings, Inc. | Housings and associated design features for ceiling array microphones |
US11438691B2 (en) | 2019-03-21 | 2022-09-06 | Shure Acquisition Holdings, Inc. | Auto focus, auto focus within regions, and auto placement of beamformed microphone lobes with inhibition functionality |
US11778368B2 (en) | 2019-03-21 | 2023-10-03 | Shure Acquisition Holdings, Inc. | Auto focus, auto focus within regions, and auto placement of beamformed microphone lobes with inhibition functionality |
US11445294B2 (en) | 2019-05-23 | 2022-09-13 | Shure Acquisition Holdings, Inc. | Steerable speaker array, system, and method for the same |
US11800280B2 (en) | 2019-05-23 | 2023-10-24 | Shure Acquisition Holdings, Inc. | Steerable speaker array, system and method for the same |
US11688418B2 (en) | 2019-05-31 | 2023-06-27 | Shure Acquisition Holdings, Inc. | Low latency automixer integrated with voice and noise activity detection |
US11302347B2 (en) | 2019-05-31 | 2022-04-12 | Shure Acquisition Holdings, Inc. | Low latency automixer integrated with voice and noise activity detection |
US11750972B2 (en) | 2019-08-23 | 2023-09-05 | Shure Acquisition Holdings, Inc. | One-dimensional array microphone with improved directivity |
US11297426B2 (en) | 2019-08-23 | 2022-04-05 | Shure Acquisition Holdings, Inc. | One-dimensional array microphone with improved directivity |
US11552611B2 (en) | 2020-02-07 | 2023-01-10 | Shure Acquisition Holdings, Inc. | System and method for automatic adjustment of reference gain |
CN111326166A (en) * | 2020-02-25 | 2020-06-23 | 网易(杭州)网络有限公司 | Voice processing method and device, computer readable storage medium and electronic equipment |
CN111883170A (en) * | 2020-04-08 | 2020-11-03 | 珠海市杰理科技股份有限公司 | Voice signal processing method and system, audio processing chip and electronic equipment |
CN111883170B (en) * | 2020-04-08 | 2023-09-08 | 珠海市杰理科技股份有限公司 | Voice signal processing method and system, audio processing chip and electronic equipment |
US11706562B2 (en) | 2020-05-29 | 2023-07-18 | Shure Acquisition Holdings, Inc. | Transducer steering and configuration systems and methods using a local positioning system |
CN111640442B (en) * | 2020-06-01 | 2023-05-23 | 北京猿力未来科技有限公司 | Method for processing audio packet loss, method for training neural network and respective devices |
CN111640442A (en) * | 2020-06-01 | 2020-09-08 | 北京猿力未来科技有限公司 | Method for processing audio packet loss, method for training neural network and respective devices |
CN111883147A (en) * | 2020-07-23 | 2020-11-03 | 北京达佳互联信息技术有限公司 | Audio data processing method and device, computer equipment and storage medium |
CN111883147B (en) * | 2020-07-23 | 2024-05-07 | 北京达佳互联信息技术有限公司 | Audio data processing method, device, computer equipment and storage medium |
US11785380B2 (en) | 2021-01-28 | 2023-10-10 | Shure Acquisition Holdings, Inc. | Hybrid audio beamforming system |
CN113035207A (en) * | 2021-03-03 | 2021-06-25 | 北京猿力未来科技有限公司 | Audio processing method and device |
CN113035207B (en) * | 2021-03-03 | 2024-03-22 | 北京猿力未来科技有限公司 | Audio processing method and device |
Also Published As
Publication number | Publication date |
---|---|
CN104347076B (en) | 2017-07-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104347076A (en) | Network audio packet loss concealment method and device | |
CN101471073B (en) | Package loss compensation method, apparatus and system based on frequency domain | |
US7552048B2 (en) | Method and device for performing frame erasure concealment on higher-band signal | |
EP2438592B1 (en) | Method, apparatus and computer program product for reconstructing an erased speech frame | |
US9524721B2 (en) | Apparatus and method for concealing frame erasure and voice decoding apparatus and method using the same | |
CA2332596C (en) | Improved lost frame recovery techniques for parametric, lpc-based speech coding systems | |
JP2012098740A (en) | Frame erasure cancel in voice communications | |
CN100578618C (en) | Decoding method and device | |
WO2013061584A1 (en) | Hybrid sound-signal decoder, hybrid sound-signal encoder, sound-signal decoding method, and sound-signal encoding method | |
TW202044231A (en) | Decoder and decoding method for lc3 concealment including full frame loss concealment and partial frame loss concealment | |
CN108109629A (en) | A kind of more description voice decoding methods and system based on linear predictive residual classification quantitative | |
CN101783142B (en) | Transcoding method, device and communication equipment | |
CN106030704A (en) | Method and apparatus for encoding/decoding an audio signal | |
US9858939B2 (en) | Methods and apparatus for post-filtering MDCT domain audio coefficients in a decoder | |
Gueham et al. | Packet loss concealment method based on interpolation in packet voice coding | |
Gueham et al. | An enhanced insertion packet loss concealment method for voice over IP network services | |
KR101452635B1 (en) | Method for packet loss concealment using LMS predictor, and thereof recording medium | |
Merazka | Packet loss concealment by interpolation for speech over IP network services | |
US20120087231A1 (en) | Packet Loss Recovery Method and Device for Voice Over Internet Protocol | |
CN101533639A (en) | Voice signal processing method and device | |
Merazka | The use of FEC method for packet loss concealment for CELP based coders in packet networks | |
Benamirouche et al. | Low complexity forward error correction for CELP-type speech coding over erasure channel transmission | |
Deng et al. | Phase unwrapping based packet loss concealment using deep neural networks | |
Hellerud et al. | Perceptually controlled error protection for audio streaming over IP networks | |
Merazka | Packet loss concealment using time scale modification for CELP based coders in packet network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |