WO2005086138A1 - エラー隠蔽装置およびエラー隠蔽方法 - Google Patents
エラー隠蔽装置およびエラー隠蔽方法 Download PDFInfo
- Publication number
- WO2005086138A1 WO2005086138A1 PCT/JP2005/003407 JP2005003407W WO2005086138A1 WO 2005086138 A1 WO2005086138 A1 WO 2005086138A1 JP 2005003407 W JP2005003407 W JP 2005003407W WO 2005086138 A1 WO2005086138 A1 WO 2005086138A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- frame
- error
- replacement
- audio
- sub
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 127
- 238000012545 processing Methods 0.000 claims abstract description 40
- 238000001514 detection method Methods 0.000 claims abstract description 26
- 230000001052 transient effect Effects 0.000 claims abstract description 14
- 230000008569 process Effects 0.000 claims description 23
- 238000006467 substitution reaction Methods 0.000 claims description 22
- 238000009499 grossing Methods 0.000 claims description 7
- 238000004364 calculation method Methods 0.000 claims description 5
- 238000004891 communication Methods 0.000 claims description 3
- 230000007717 exclusion Effects 0.000 claims 1
- 238000013213 extrapolation Methods 0.000 abstract description 51
- 238000004458 analytical method Methods 0.000 abstract description 16
- 230000000694 effects Effects 0.000 abstract description 9
- 230000015572 biosynthetic process Effects 0.000 abstract description 5
- 230000005236 sound signal Effects 0.000 abstract description 5
- 238000003786 synthesis reaction Methods 0.000 abstract description 5
- 210000005069 ears Anatomy 0.000 abstract 1
- 230000001788 irregular Effects 0.000 abstract 1
- 230000005540 biological transmission Effects 0.000 description 10
- 238000010586 diagram Methods 0.000 description 8
- 238000005516 engineering process Methods 0.000 description 5
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 5
- 230000001419 dependent effect Effects 0.000 description 4
- 238000012360 testing method Methods 0.000 description 4
- 230000010354 integration Effects 0.000 description 3
- 230000000737 periodic effect Effects 0.000 description 3
- 230000010076 replication Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000012805 post-processing Methods 0.000 description 2
- 101150108962 TEPP gene Proteins 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000009172 bursting Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000003750 conditioning effect Effects 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 230000006866 deterioration Effects 0.000 description 1
- 230000003292 diminished effect Effects 0.000 description 1
- 238000005562 fading Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
- 230000001172 regenerating effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 238000012795 verification Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L1/00—Arrangements for detecting or preventing errors in the information received
- H04L1/004—Arrangements for detecting or preventing errors in the information received by using forward error control
- H04L1/0056—Systems characterized by the type of code used
- H04L1/0061—Error detection codes
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/005—Correction of errors induced by the transmission channel, if related to the coding algorithm
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L1/00—Arrangements for detecting or preventing errors in the information received
- H04L1/004—Arrangements for detecting or preventing errors in the information received by using forward error control
- H04L1/0045—Arrangements at the receiver end
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/78—Detection of presence or absence of voice signals
Definitions
- the present invention relates to techniques for concealing lost or corrupted digital signals (errors) in transmission, delivery, and storage media.
- Error concealment is performed to minimize such deterioration.
- the basic idea is to replace error 'data with some synthetic or generated data.
- the purpose of error concealment is to hide errors resulting from data loss or data corruption during transmission, reception, storage, encoding, or decoding processes, ie to mask this in the best case An error is not recognized, or at least an error is more pronounced.
- sender-assisted embeds error correction information in the transmitted information so that the sender can retransmit the lost or corrupted information or recover the lost data.
- receiver-based approach does not require information from the sender to handle concealment of lost data. This is to work by replacing lost or corrupted data with useful information, which is usually an estimate of lost or corrupted parts. There is. These replacement data are generated at the receiver without the assistance of the transmitter.
- Non-Patent Document 1 Data interpolation techniques (see, for example, Non-Patent Document 1) can provide better results than the simple methods described above. This attempts to reconstruct the replacement data by also interpolating the non-error 'data forces adjacent to the error' data.
- FIG. 1 is a flow chart showing the procedure of an error concealment technique incorporating various error concealment processes of muting, duplication and interpolation.
- ST 101 checks whether this is an error frame. If this is not an error frame, then the received data is stored in memory at ST 102, possibly for future use during the interpolation process of ST 107. Received frame force If detected as an error frame in ST101, a series of decisions are made in ST103 and ST105, and if a delay is allowed in the system, in a previous or future non error frame. Based on it, determine which error concealment technique is best to apply to the current error frame. In ST103, check whether the current error frame is appropriate to mute. If muting is selected, the current error frame is replaced with! /, And zero data frame at ST104. That is, the current error 'frame is replaced by silence.
- Non-Japanese Articles 1 1. Kauppinen et al., "Audio signal Extrapolation-Theory and
- the sender-based error concealment scheme has a problem that retransmission causes longer delay and embedded information for error recovery increases transmission overhead. These states are not suitable for applications such as real time communication or transmission systems.
- Muting causes the received signal to disappear suddenly! /, As a result, there is a problem that the acoustic level changes rapidly.
- bursting 'errors' can cause silence for a long time. In addition, it can also be "juzzy" when constant errors are received.
- Interpolation also causes "click" noise 'artifacts due to boundary discontinuity problems There is a problem of causing The quality of concealment is greatly diminished for single frame 'errors, which is better for force burst' errors than for muting and duplication. Although these techniques are simple and easy to implement, they can not achieve satisfactory concealment results.
- Codec-dependent schemes provide better results, but are not yet sufficient in terms of versatility. Usually, this is only used for certain applications or certain voice or audio 'codecs. Many codec dependent schemes work in the compression domain by predicting the parameters used by a particular codec. Such schemes typically also require higher computational processing costs.
- an object of the present invention is to provide an error concealment technique for a voice signal or an audio signal in which the effect of the error concealment is as inconspicuous as possible.
- the present invention applies appropriate means of classifying voice data or audio data with different characteristics and generating replacement data for concealing error data.
- FIG. 2 shows the main configuration of an error concealment device according to the present invention.
- Voice data or audio data can be supplied via various media, such as being received from a storage medium (or transmission system via a wired or wireless network) 200. These data are organized into frames or packets, usually in compressed form. When a packet is received, an appropriate voice or audio decoder 250 is used to decode the packet into uncompressed audio format frames suitable for playback. Verification is performed at all stages to ensure that received packets are not lost or corrupted and that there is no error in decoding received packets. If an error occurs at any stage, the error concealment processing unit 300 is notified to conceal the error frame.
- the internal configuration of the error concealment processing unit 300 according to the present invention is shown in the block diagram of FIG.
- a signal is sent to indicate whether it is a force that is an error frame or a non-error frame.
- the notification signal detection unit 301 detects this signal. Determine the action to be taken with respect to the frame of input data.
- a non-error frame it is stored in the sample storage unit 302, and a frame of the same data is sent out for reproduction or storage.
- the stored data is used by the voice detection unit 303, the transient signal detection unit 305, and the extrapolation processing unit 307 when a future error frame is generated.
- the speech detection unit 303 makes a determination as to whether it is a speech signal or a non-speech signal on the preceding non-error data frame.
- speech replication processor 304 is used to generate a substitute frame for concealing error 'frames by performing pitch calculation and replication.
- transient signal detection is performed in the transient signal detection unit 305 to determine whether it is an attack signal or not.
- the transient signal detection unit 305 selects an appropriate region (hereinafter referred to as a transient region) as a region including a transient signal for the filter bank analysis unit 306 as the medium power of the leading non-error frame.
- the filter bank analysis unit 306 generates N subbands by filter bank analysis. These subbands may or may not be equal in bandwidth.
- Each of these sub-bands passes through the extrapolation processing unit 307 where the sub-band 'sample for error concealment is extrapolated to generate replacement sub-band' data for the current error 'frame.
- the filter bank synthesis unit 308 reconstructs subband 'data for forming a frame of generated data using the synthesis filter bank together with the subband' data generated by the extrapolation processing unit 307.
- the generated data is sent to the post-processing unit 309 to solve the problem of frame boundary discontinuity and the problem of signal level mismatch before being stored and sent out by the sample storage unit 302.
- the internal configuration of the mantle processing unit 307 is shown in the block diagram of FIG.
- a first test is performed on each subband at the first test section 401. For sub-bands that are considered unimportant, no extravagance is performed, but instead silence substitution 402 replaces zero. For the sub-bands regarded as important, the filter coefficient calculation unit 403 calculates filter coefficients used for the extrapolation filter. In these sub-bands, to what extent the sub-bands are predictable and predictable (hereinafter referred to as “predictability level”).
- a second test is also performed in the second test unit 404 to determine the In the case of a sub-band that is classified as non-predictable and non-predictive, this sub-band is considered to be similar to noise, and so the sub-band replacement section 407 noises this sub-band 'data by means of noise replacement. 'Power to replace with data, or replace this sub-band' data with sub-bands contained in the same preceding frame.
- the filter order adjustment unit 405 adjusts the filter order of the envelope filter based on the predictability level derived by the second inspection unit 404. .
- These subbands are extrapolated in subband section 406 using subband filters commonly used in filtering techniques.
- a lost or damaged frame can be effectively concealed.
- FIG. 1 A flow chart showing the procedure of an error concealment technique incorporating various error concealment processing.
- FIG. 2 A diagram showing the main configuration of an error concealment device according to the present invention.
- FIG. 3 A block diagram showing an internal configuration of an error concealment processing unit according to the present invention
- FIG. 4 A block diagram showing the internal configuration of the extrapolation processing unit according to the present invention
- FIG. 5 A flowchart showing the detailed steps of the error concealment algorithm in the first embodiment.
- FIG. 6 A flow chart showing a detailed procedure of voice sample generation processing in the first embodiment.
- FIG. 7 A flowchart showing a detailed procedure of the audio sample generation process in the first embodiment.
- FIG. 8 A diagram showing frame boundary smoothing processing in the first embodiment.
- FIG. 9 A diagram showing a signal obtained by the audio sample generation process in the first embodiment.
- FIG. 10 A diagram for explaining detection of an attack signal in the first embodiment.
- FIG. 5 is a flowchart showing the detailed steps of the error concealment algorithm in the first embodiment. Received frames are classified into non-error frames, error frames as speech, and error frames as audio.
- ST501 it is detected whether the received audio 'frame is an error frame or a non-error frame.
- Non-error • If a frame is received, store the frame in memory in ST513 and make it available in the future if an error frame is received.
- This memory has a first in, first out (FIFO) structure and can store n frames of past audio 'samples. The standard size of the memory can be set to 2 frames so that the last two frames of the sample are stored in the memory.
- FIFO first in, first out
- the standard size of the memory can be set to 2 frames so that the last two frames of the sample are stored in the memory.
- ST 512 it is stored as an error 'flag before error flag of current frame' flag. This is used in ST 502 to check if the previous frame is an offset of the error 'frame or non-error' frame.
- the type classification of the error 'frame when the error' frame is detected is performed as follows.
- the preceding frame is a non-error frame,! /.
- These past samples are basically the leading non-error 'frames stored at ST513.
- the speech detection algorithm used in the general speech coding technology can be used in ST 503.
- One simple method is to apply autocorrelation to the non-error'frames that precede the error'frames and retrieve periodicity and pitch information. These information will be stored for future use.
- the voice flag force is set.
- the voice flag is used in the case of a burst error where many consecutive frames result in an error. In such a case, only the first error frame of a series of error frames passes the voice detection of force T503, and the subsequent continuous error frames need to pass detection again first at ST503.
- the speech detection in ST 503 is skipped since detection is already performed in the first error frame. It uses an audio flag to determine which type of power, frame, it is either an audio type or an audio type.
- the voice sample generation process is used in ST 506 to generate samples to replace the current error 'frame to conceal or minimize errors. ST 506 will be described later using FIG.
- Processing of error frames classified as audio is performed as follows.
- the current frame is classified as an audio 'type' frame in ST504, it is set to voice flag power ⁇ in ST507.
- the audio sample generation process is used in ST 508, and a sample is generated to replace the current error frame and conceal the error.
- ST 508 will be described in detail later using FIG.
- a process called post process is performed as follows.
- the generated support Signal conditioning is performed on the sample.
- the purpose of this signal processing is to hear "click” noise and "pobbing" noise, etc. due to rapid changes in signal characteristics as it passes between non-error frames and error concealment frames at frame boundaries. It is about reducing artefacts.
- the error concealment samples generated may have significantly different signal levels or power compared to previous non-error frames. Therefore, the generated samples are adjusted in ST 509 to coincide with the signal level preceding the error frame.
- FIG. 6 is a flowchart showing a detailed procedure of speech sample generation processing (ST 506 in FIG. 5) when the error 'frame is classified as a speech type.
- ST 506 in FIG. 5 a simple method is presented as a method of deriving speech samples.
- the signal obtained by the audio sample generation process shown in FIG. 6 is shown in FIG.
- the preceding non-error' frame force also duplicates the segment (part of the frame) in ST 601.
- This segment has a length that is a multiple of the pitch period obtained in ST 503, but is less than one frame length.
- the pitch alignment in ST602, ie, adjustment of the replacement position of the duplicate frame is performed so that the pitch period of the duplicate frame is aligned with the pitch period of the preceding frame. This ensures continuity in the periodicity up to the non-error 'frame force and even duplicate frames It is This duplication and alignment are repeated until one complete replacement frame capable of replacing the error 'frame is obtained in ST603.
- the pitch period of the duplicate frame and the pitch period of the preceding frame are adjusted by adjusting the duplication section, that is, repeating the duplication until obtaining a substitution frame capable of replacing the error 'frame. You may make them
- FIG. 7 is a flowchart showing a detailed procedure of audio 'sample generation processing (ST 508 in FIG. 5) when an error' frame classified as audio 'type is detected.
- the samples generated to conceal the errored frame are derived (from ST 513) mainly from non-errored samples preceding the errored frame. Usually, these perfect samples without loss contain information, and once extracted, they can gain strength on what the subsequent signal might look like. There are two paths to this process. One is used when the frame preceding this error frame is a non-error frame. The other path is used if the preceding frame is also an error 'frame.
- the error condition of the preceding frame is checked. If the preceding frame is not an error frame, then it means that the current error frame is the first error frame following a non-error frame.
- the current error frame is defined as frame E
- the non-error frame preceding the error frame is defined as frame P.
- detection of an attack signal is performed on frame P.
- the detection of this attack signal is described in FIG. If there is no attack signal in frame P, the entire frame is used in the subsequent filter 'band analysis of ST703. If an attack signal is detected in frame P, the start position of the attack signal is confirmed, and in frame P, samples prior to the attack signal are discarded. That is, only samples after the start position of the attack signal are used for filter 'band analysis after force.
- the generated samples are derived from the signal characteristics of frame P. Because the signal to be generated is created to simulate frame P, By not excluding samples (areas) before the start position of the attack signal, the generated signal usually has a reduced version similar to but smaller than the attack signal from frame P (similar attack signal ) Will appear.
- the advantage of performing the above detection is that of the "double attack” of the signal generated by eliminating in the subsequent processing the unwanted part of the signal, ie the area before the start of the attack signal. The point is that the problem can be minimized.
- the signal is divided into a plurality of frequency bands, ie, K subbands in ST 703 using an analysis filter bank.
- the bandwidths of each subband need not be the same. That is, the subband bandwidths may or may not be equal.
- one possible subband configuration is that for the first K / 4 subbands, each subband is L / (2K) samples. It has bandwidth.
- each sub-band will have a bandwidth of L / K samples, and for the last K / 2 sub-bands, each sub-band will have a bandwidth of 2 L / K samples.
- the lower the subband the smaller the subband bandwidth, which results in better frequency resolution.
- the higher the subbands the greater the bandwidth.
- Another possible configuration is to ensure that all subbands have the same L / K sample bandwidth.
- ⁇ is a small value to prevent underflow
- L is the sub-band length (ie bandwidth)
- This parameter is compared to the threshold in ST704. If this threshold is exceeded, the flag and the skip 'flag are set to' 1 'in ST 705 to indicate that this sub-band is excluded from the signal generation process, and that sub-band is replaced with zero in ST 706. Proceed to the next subband at ST716. If the parameter exceeds the threshold, The skip 'flag is set to' 0 'in ST 707 to indicate that this sub-band is included in the signal generation process and is not skipped. This skip 'flag is used in the subsequent frame if the subsequent frame is also an error' frame in ST718.
- ST713 and ST721 of the present invention signal extrapolation is used for audio sample generation processing.
- ST 708 segments of known audio signals are used to derive a set of impulse response coefficients a using Berg's algorithm.
- This error residual is updated for each iteration of the coefficient calculation process, which returns the final prediction error residual E after the last iteration. Between the predictability level Pr and the error residual rate E / E
- the 1 sb 0 1 relationship reduces the predictability level accordingly when the ratio E / E is small
- the order of the extrapolation filter FO is the past sump used to calculate the extrapolation coefficient
- the order of the envelope filter FO is out at ST 710 if the predictability level is high
- the predictability level Pr of each subband is used to conceal the error. It is also used to decide whether to use subtractive processing to generate sub-band 'samples, or to use sub-band replacement methods to replace error' frames.
- the outer flag is set to indicate this decision, and this flag is used in ST 720 when the next frame is also an error frame and the same sub-band uses the same type of signal generation scheme It is supposed to be.
- the extrapolation flag is set to "1" at ST 712 sb
- the level of predictability Pr is also connected to the extrapolation signal sb if the predictability approaches or exceeds the threshold.
- y, n -a y (n-1)-a y (n-2)- ⁇ ⁇ ⁇ -a y, n-p) ⁇ ⁇ )
- y (n) is the current extrapolated sample
- y (ni) is the past output sample
- a is the outer coefficient
- p is the predictor order.
- ⁇ is a factor that controls the amount of influence on the extrapolated signal.
- the extrapolation flag is set to “0” in ST s sb
- the predictability level is also used to determine the substitution method used. If the predictability level is slightly below the threshold, then the same subband from the previous non-error 'frame P is repeated in the subbands. If this is well below the threshold, the sub-bands are simply replaced by random signals.
- the following processing is performed for the subsequent error 'frame. If it is indicated at ST 701 that the preceding frame is also an error frame, the sample generation process continues at the point at which it stopped at the previous frame. The previously calculated parameters, such as filter order and predictability level, are reused. These parameters are not calculated again.
- the force to be tested whether a particular subband needs to be generated is indicated by the skip flag in ST 718 and the outer flag in ST 720. These flags are determined in ST704, ST705, ST707, ST711, ST712, and ST714 during the first error frame after the non-error frame.
- subband replacement method is used in ST722.
- the level of predictability determines whether to use the noise replacement or the V, offset replacement method for subband iterations as described above.
- a single or a plurality of lost or damaged frames can be effectively concealed.
- This is a receiver-based error concealment scheme that is suitable for real-time communications or transmission systems because it does not introduce retransmission delays and has low transmission overhead. It is suitable for a wide range of audio applications as it can be used as a stand-alone 'module' that is not codec dependent. Noise 'artifacts such as jerky, echo effects, clicks, etc. It is greatly reduced.
- the input signal is classified into voice and audio so that appropriate error concealment methods can be applied to each signal classification.
- selecting appropriate regions of the filter bank analysis frame may cause similar transients of the regenerated signal to be regenerated, which can be annoying. Is reduced.
- By using a filter bank to decompose the signal into smaller frequency bandwidths, regenerating different signals in each bank based on signal characteristics such as predictability level and importance of contribution to each sub-band overall signal The scheme can be applied.
- the extrapolation process results in a continuous signal that solves the problem of boundary discontinuities at the start of the error 'frame, which is one of the main causes of noise artifacts. Discontinuity problems at the end of the error 'frame are solved using backward extrapolation, signal mixing and signal level adjustment.
- the present invention has been described by way of example in the case of being configured by node software, the present invention can also be realized by software.
- a first aspect of the present invention is a method of error concealment, which classifies input data into various characteristics and generates substitution data for concealing error data by applying appropriate measures. , Said method,
- a second aspect of the present invention includes the repetition of one or more of the future capacity steps (lb) to (lg) described in steps (lb) and (lh). It is a method characterized by
- a third aspect of the present invention is that the classification described in step (lc) calculates the autocorrelation of the preceding non-error 'frame before the error' frame to search for the presence of a pitch period. Is a method characterized in that it is performed by
- the error frame when the pitch period can detect an acceptable pitch period, the error frame is classified as the voice frame.
- the method is characterized in that it is used to classify the error frame as the audio frame if the pitch period can not be detected.
- step (Id) further including:
- a sixth aspect of the present invention is the method as recited in step (le), wherein the method comprises: generating the non-error frame force prior to the replacement audio 'frame force error described above frame; Furthermore,
- a seventh aspect of the present invention is the method according to the step (wherein the signal level is made close to the preceding frame in order to reduce an artifact caused by an abrupt change of the signal processing level described in ID). It is characterized in that it is carried out by adjusting the signal.
- the boundary smoothing described in the step (lg) generates a short, segment of a subsequent frame force sample using reverse direction, and
- the method is characterized in that it is achieved by cross-fading and mixing with the permuted frame to reduce the artifacts caused by the boundary discontinuity problem.
- a ninth aspect of the present invention is the force according to the step (6b), wherein the appropriate area is the area until the end of the preceding frame, or the starting force of the attack signal in the preceding frame.
- the method is characterized in that the appropriate region is the entire preceding frame when an attack signal is not detected.
- the tenth aspect of the present invention is characterized in that the bandwidth of the sub-band according to step (6c) is high.
- the method is characterized in that it may or may not be equal to have a lower frequency domain force, such bandwidth as compared to the frequency domain or any possible configuration.
- the importance of each of the error subbands described in step (6d) is determined by how important the contribution of the subbands to the preceding non-error 'frame is. It is a method characterized in that it depends on whether there is any.
- a twelfth aspect of the present invention is the filter order force according to step (6e), wherein the length of the region selected in step (6b) of the sixth aspect of the present invention It is a method characterized by being calculated by dividing by the number of subbands.
- a thirteenth aspect of the present invention is a method characterized in that it is calculated by the Berg's algorithm using the preceding sub-band as an input.
- the filter coefficients are used by the AR filter to extrapolate samples using samples of the preceding subbands.
- a fourteenth aspect of the present invention provides the predictability level Pr force Pr ⁇ E according to step (6 g)
- the method is characterized in that the algorithm power of the above-mentioned Berg and the prediction error residual power to be calculated are also derived.
- a fifteenth aspect of the present invention is characterized in that the extrapolation filter order FO described in step (6h) is adjusted based on the predictability level Pr, and the extrapolation filter order power prediction
- the probability level is inversely proportional to the value of the predictability level, as shown by FO (l / Pr), such that the higher the probability level, the smaller the extrapolation filter's order will be or vice versa. It is the method to characterize.
- the substitution subband described in (60) is generated according to the condition of the importance and the predictability level of the preceding subband, and Furthermore,
- said error sub-band power said substitution of said error sub-bands by the operation of replacing the error 'sub-bands with zero, called silence substitution, when it is classified as unimportant in step (6 d) Generating subbands;
- a seventeenth aspect of the present invention is characterized in that the substitution subband described in step (16a) is further influenced by the predictability level Pr, and the extrapolation value is as follows: It is a method characterized by being adjusted by the reduced version of the predictability level Pr.
- a eighteenth aspect of the present invention is a method of generating a replacement frame of an error 'frame by performing an error' frame prior to the error 'frame also by covering,
- a nineteenth aspect of the present invention is the force according to the step (18b), which is an area until the end of the preceding frame, and the starting force of an attack signal in the preceding frame.
- the appropriate area is the entire preceding frame when the attack signal is not detected.
- a twentieth aspect of the present invention is the bandwidth of the sub-band described in the step (18c) has a bandwidth in which the region of low frequency is smaller than the region of high frequency or any possible configuration. So that they may or may not be equal!
- a twenty-first aspect of the present invention relates to how important the contribution of each of the above-mentioned error.
- Subbands described in step (18d) to the preceding non-error 'frame is. It is a method characterized by being determined by
- a twenty-second aspect of the present invention is the filter order force according to step (18e).
- the method is characterized by being calculated by dividing the length of the region selected in (18b) by the number of subbands of filter bank analysis.
- a method characterized in that it is calculated by the Berg's algorithm using the preceding sub-band as an input. Note that the filter coefficients are used by the AR filter to extrapolate samples using the samples of the preceding subbands.
- a twenty-fourth aspect of the present invention is the prediction residual E calculated after the Berg algorithm, as indicated by the predictability level Pr force E ⁇ E described in step (18 g) Ratio to the ratio of the predicted remainder E calculated before the Berg's algorithm for
- the algorithm power of the Berg and the prediction error residual power to be calculated are also derived.
- a twenty-fifth aspect of the present invention is characterized in that the extrapolation filter order FO described in step (18h) is adjusted based on the predictability level Pr, and the extrapolation filter order force prediction
- the value of the predictability level is inversely proportional to the value of the predictability level, as indicated by FO (l / Pr), such that the higher the likelihood level, the less the order of the extrapolation filter will decrease accordingly or vice versa.
- a twenty-sixth aspect of the present invention is the method according to the step (a), wherein the substitution subband described in (180 is generated according to the condition of importance and the predictability level of the preceding subband, and the method further comprises ,
- step (16a) If the error 'subband is classified as important in step (16d) and the predictability level exceeds a predetermined threshold, the error' subband extrapolation using signal extrapolation. Generating the permutation subbands of the subbands;
- step (16c) If the error 'sub-band is classified as important in step (16 d) and the predictability level is some other value, the operation of replacing the error' sub-band with noise, called noise replacement, Generating the permutation sub-band of the error 'sub-band;
- step (16d) The error sub-band power If it is classified as unimportant in step (16 d), the replacement of the error sub-bands by the operation of replacing the error 'sub-bands with zero, called silence substitution. Generating subbands;
- the method in the substitution subband described in step (26a), is characterized in that it is influenced by the measureability level Pr, and the extrapolated value is adjusted by the reduced version of the predictability level Pr as follows.
- y (n) (-a y (n-l)-a y (n-2) a y (n-p)) X a log (Pr) ⁇ ⁇ ⁇ ⁇ (Expression 4)
- a twenty-eighth aspect of the present invention is a method of generating a substitution sub-band of an error 'frame based on the condition of importance and the predictability level Pr of the preceding sub-band, said method comprising the steps of: Furthermore,
- the error sub-band is classified as important and the predictability level is slightly below but close to a predetermined threshold, the error sub-band is referred to as the sub-band iteration. Generating the replacement subbands of the error 'subbands by the operation of replacing the non-error's frame with the corresponding preceding subbands;
- noise replacement is an operation of replacing the error sub-band with noise.
- a twenty-ninth aspect of the present invention is characterized in that the substitution subband force described in step (28a) is further affected by the predictability level Pr, and the extrapolation value is determined as follows: A method characterized in that it is adjusted by a reduced version of the measureability level Pr.
- a thirtieth aspect of the present invention is a method of selecting a non-error 'frame force appropriate region for filter bank analysis, said method comprising (30a) searching for a sudden start of an attack signal in said non-error 'frame;
- a thirty-first aspect of the present invention is that, in the case where the start of the attack signal is present in the non-error 'frame, in step (30b), the region starts from the start of the attack signal. Select the entire area of the non-error frame in step (30b) if there is no error signal in the non-error frame that is selected until the end of the error frame or the end of the frame. Is a method characterized by
- a thirty-second aspect of the present invention is a method of calculating the filter order of an extrapolation filter, wherein
- a thirty-third aspect of the present invention is that, in the step (32b), when the start of an attack signal is present in the non-error 'frame, the length is the starting force of the attack signal. If the force selected until the end of the frame or there is no attack signal in the non-error frame, then in step (32b), select the entire area of the non-error frame as the length Is a method characterized by
- a thirty-fourth aspect of the present invention is a method of adjusting the order of an extrapolation filter, said method comprising
- a thirty-fifth aspect of the present invention is a method characterized in that the extrapolated filter coefficient described in step (34b) is calculated by Berg's algorithm using samples of the sub band.
- a thirty-sixth aspect of the present invention is the prediction remainder E calculated after the Berg's algorithm, as indicated by the predictability level Pr force E ⁇ E described in step (34c).
- the algorithm power of the Berg and the prediction error residual power to be calculated are also derived.
- a thirty-seventh aspect of the present invention is characterized in that the extrapolation filter order FO described in the step (34d) is adjusted based on the predictability level Pr, and the extrapolation filter order force prediction
- the value of the predictability level is inversely proportional to the value of the predictability level, as indicated by FO (l / Pr), such that the higher the likelihood level, the less the order of the extrapolation filter will decrease accordingly or vice versa.
- a thirty-eighth aspect of the present invention is a subband selection method for generating a substitute frame, the method comprising:
- a thirty-ninth aspect of the present invention is characterized in that the importance of each of the sub-bands described in step (38a) is determined depending on how important the sub-band contribution to the frame is. It is a way to
- the fortieth aspect of the present invention provides the predictability level Pr force Pr ⁇ described in step (38b).
- the predicted remainder E calculated after the Berg's algorithm, as indicated by E / E
- the algorithm power of the Berg and the prediction error residual power to be calculated are also derived.
- the replacement method according to step (38c) is selected based on the condition of importance and the predictability level of the sub-band.
- step (41a) selecting the signal extrapolation method if the sub-bands are classified as important in step (38a) and the predictability level exceeds a predetermined threshold
- step (41b) selecting the subband repetition method if the error subband is classified as important in step (38a) and the predictability level is slightly below but close to a predetermined threshold value ,
- step (41c) selecting a noise replacement method if said error 'sub-band is classified as important in step (38a) and said predictability level is another value;
- Each of the above functional blocks is typically implemented as an LSI, which is an integrated circuit.
- the force is taken to be LSI, depending on the degree of integration, IC, system LSI, super L
- circuit integration is not limited to LSI's, and implementation using dedicated circuitry or general purpose processors is also possible.
- FPGA Field can be programmed after LSI fabrication
- the present invention is useful as a concealment technique for lost or corrupted digital audio signals (errors) in transmission, delivery, and storage media.
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Detection And Prevention Of Errors In Transmission (AREA)
- Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
- Maintenance And Management Of Digital Transmission (AREA)
Abstract
Description
Claims
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
AT05719724T ATE523876T1 (de) | 2004-03-05 | 2005-03-01 | Fehlerverbergungseinrichtung und fehlerverbergungsverfahren |
EP05719724A EP1722359B1 (en) | 2004-03-05 | 2005-03-01 | Error conceal device and error conceal method |
JP2006510676A JP4744438B2 (ja) | 2004-03-05 | 2005-03-01 | エラー隠蔽装置およびエラー隠蔽方法 |
US10/591,601 US7809556B2 (en) | 2004-03-05 | 2005-03-01 | Error conceal device and error conceal method |
CN2005800071542A CN1930607B (zh) | 2004-03-05 | 2005-03-01 | 差错隐藏装置以及差错隐藏方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004061797 | 2004-03-05 | ||
JP2004-061797 | 2004-03-05 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005086138A1 true WO2005086138A1 (ja) | 2005-09-15 |
Family
ID=34918086
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/003407 WO2005086138A1 (ja) | 2004-03-05 | 2005-03-01 | エラー隠蔽装置およびエラー隠蔽方法 |
Country Status (6)
Country | Link |
---|---|
US (1) | US7809556B2 (ja) |
EP (1) | EP1722359B1 (ja) |
JP (1) | JP4744438B2 (ja) |
CN (1) | CN1930607B (ja) |
AT (1) | ATE523876T1 (ja) |
WO (1) | WO2005086138A1 (ja) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2007043642A1 (ja) * | 2005-10-14 | 2007-04-19 | Matsushita Electric Industrial Co., Ltd. | スケーラブル符号化装置、スケーラブル復号装置、およびこれらの方法 |
WO2007143953A1 (fr) * | 2006-06-08 | 2007-12-21 | Huawei Technologies Co., Ltd. | Dispositif et procédé pour dissimulation de trames perdues |
JP2008182380A (ja) * | 2007-01-23 | 2008-08-07 | Alpine Electronics Inc | オーディオ装置 |
JP2009175693A (ja) * | 2007-11-05 | 2009-08-06 | Huawei Technologies Co Ltd | 減衰率を取得する方法および装置 |
JP2009290567A (ja) * | 2008-05-29 | 2009-12-10 | Canon Inc | データ送信装置及びデータ受信装置 |
JP2010511201A (ja) * | 2006-11-28 | 2010-04-08 | サムスン エレクトロニクス カンパニー リミテッド | フレームエラー隠匿方法及び装置、これを利用した復号化方法及び装置 |
JP2013519920A (ja) * | 2010-02-11 | 2013-05-30 | クゥアルコム・インコーポレイテッド | サブ帯域コード化復号器における損失パケットの隠蔽 |
JP2013527479A (ja) * | 2009-06-29 | 2013-06-27 | オーディエンス,インコーポレイテッド | 破損したオーディオ信号の修復 |
JP2016510432A (ja) * | 2013-02-05 | 2016-04-07 | テレフオンアクチーボラゲット エル エム エリクソン(パブル) | オーディオフレーム損失のコンシールメントを制御する方法及び装置 |
US9536540B2 (en) | 2013-07-19 | 2017-01-03 | Knowles Electronics, Llc | Speech signal separation and synthesis based on auditory scene analysis and speech modeling |
US9820042B1 (en) | 2016-05-02 | 2017-11-14 | Knowles Electronics, Llc | Stereo separation and directional suppression with omni-directional microphones |
US9838784B2 (en) | 2009-12-02 | 2017-12-05 | Knowles Electronics, Llc | Directional audio capture |
US9978388B2 (en) | 2014-09-12 | 2018-05-22 | Knowles Electronics, Llc | Systems and methods for restoration of speech components |
US10762907B2 (en) | 2016-01-29 | 2020-09-01 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus and method for improving a transition from a concealed audio signal portion to a succeeding audio signal portion of an audio signal |
JP2021502608A (ja) * | 2017-11-10 | 2021-01-28 | フラウンホーファー−ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン | 異なる損失隠蔽ツールのセットをサポートするオーディオデコーダ |
US11127408B2 (en) | 2017-11-10 | 2021-09-21 | Fraunhofer—Gesellschaft zur F rderung der angewandten Forschung e.V. | Temporal noise shaping |
US11217261B2 (en) | 2017-11-10 | 2022-01-04 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Encoding and decoding audio signals |
US11315583B2 (en) | 2017-11-10 | 2022-04-26 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio encoders, audio decoders, methods and computer programs adapting an encoding and decoding of least significant bits |
US11380341B2 (en) | 2017-11-10 | 2022-07-05 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Selecting pitch lag |
US11462226B2 (en) | 2017-11-10 | 2022-10-04 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Controlling bandwidth in encoders and/or decoders |
US11545167B2 (en) | 2017-11-10 | 2023-01-03 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Signal filtering |
US11562754B2 (en) | 2017-11-10 | 2023-01-24 | Fraunhofer-Gesellschaft Zur F Rderung Der Angewandten Forschung E.V. | Analysis/synthesis windowing function for modulated lapped transformation |
US12033646B2 (en) | 2017-11-10 | 2024-07-09 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Analysis/synthesis windowing function for modulated lapped transformation |
Families Citing this family (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8798172B2 (en) * | 2006-05-16 | 2014-08-05 | Samsung Electronics Co., Ltd. | Method and apparatus to conceal error in decoded audio signal |
EP2153436B1 (en) * | 2007-05-14 | 2014-07-09 | Freescale Semiconductor, Inc. | Generating a frame of audio data |
CN101308655B (zh) * | 2007-05-16 | 2011-07-06 | 展讯通信(上海)有限公司 | 一种音频编解码方法与装置 |
CN101325537B (zh) * | 2007-06-15 | 2012-04-04 | 华为技术有限公司 | 一种丢帧隐藏的方法和设备 |
KR101403340B1 (ko) * | 2007-08-02 | 2014-06-09 | 삼성전자주식회사 | 변환 부호화 방법 및 장치 |
US20090048828A1 (en) * | 2007-08-15 | 2009-02-19 | University Of Washington | Gap interpolation in acoustic signals using coherent demodulation |
CN100524462C (zh) | 2007-09-15 | 2009-08-05 | 华为技术有限公司 | 对高带信号进行帧错误隐藏的方法及装置 |
US8126578B2 (en) * | 2007-09-26 | 2012-02-28 | University Of Washington | Clipped-waveform repair in acoustic signals using generalized linear prediction |
DE602007004504D1 (de) * | 2007-10-29 | 2010-03-11 | Harman Becker Automotive Sys | Partielle Sprachrekonstruktion |
CN101488344B (zh) * | 2008-01-16 | 2011-09-21 | 华为技术有限公司 | 一种量化噪声泄漏控制方法及装置 |
MX2010009932A (es) | 2008-03-10 | 2010-11-30 | Fraunhofer Ges Forschung | Metodo y dispositivo para manipular una señal de audio que tiene un evento transitorio. |
CN100550133C (zh) * | 2008-03-20 | 2009-10-14 | 华为技术有限公司 | 一种语音信号处理方法及装置 |
US20100185441A1 (en) * | 2009-01-21 | 2010-07-22 | Cambridge Silicon Radio Limited | Error Concealment |
US8676573B2 (en) * | 2009-03-30 | 2014-03-18 | Cambridge Silicon Radio Limited | Error concealment |
US8316267B2 (en) | 2009-05-01 | 2012-11-20 | Cambridge Silicon Radio Limited | Error concealment |
IL200694A (en) * | 2009-09-02 | 2014-04-30 | Gil Levy | A method and device for transmitting frames in a communication network |
JP2011237753A (ja) * | 2010-04-14 | 2011-11-24 | Sony Corp | 信号処理装置および方法、並びにプログラム |
PL3518234T3 (pl) | 2010-11-22 | 2024-04-08 | Ntt Docomo, Inc. | Urządzenie i sposób kodowania audio |
EP2464145A1 (en) | 2010-12-10 | 2012-06-13 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus and method for decomposing an input signal using a downmixer |
AU2012217215B2 (en) * | 2011-02-14 | 2015-05-14 | Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. | Apparatus and method for error concealment in low-delay unified speech and audio coding (USAC) |
TWI483245B (zh) | 2011-02-14 | 2015-05-01 | Fraunhofer Ges Forschung | 利用重疊變換之資訊信號表示技術 |
TR201903388T4 (tr) | 2011-02-14 | 2019-04-22 | Fraunhofer Ges Forschung | Bir ses sinyalinin parçalarının darbe konumlarının şifrelenmesi ve çözülmesi. |
EP2676268B1 (en) | 2011-02-14 | 2014-12-03 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus and method for processing a decoded audio signal in a spectral domain |
EP2676270B1 (en) | 2011-02-14 | 2017-02-01 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Coding a portion of an audio signal using a transient detection and a quality result |
MY165853A (en) | 2011-02-14 | 2018-05-18 | Fraunhofer Ges Forschung | Linear prediction based coding scheme using spectral domain noise shaping |
TWI610296B (zh) | 2011-10-21 | 2018-01-01 | 三星電子股份有限公司 | 訊框錯誤修補裝置及音訊解碼裝置 |
EP2954516A1 (en) | 2013-02-05 | 2015-12-16 | Telefonaktiebolaget LM Ericsson (PUBL) | Enhanced audio frame loss concealment |
HUE045991T2 (hu) | 2013-02-05 | 2020-01-28 | Ericsson Telefon Ab L M | Hangkeret-elvesztés elrejtése |
WO2014202770A1 (en) | 2013-06-21 | 2014-12-24 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Method and apparatus for obtaining spectrum coefficients for a replacement frame of an audio signal, audio decoder, audio receiver and system for transmitting audio signals |
MY181026A (en) | 2013-06-21 | 2020-12-16 | Fraunhofer Ges Forschung | Apparatus and method realizing improved concepts for tcx ltp |
KR102132326B1 (ko) * | 2013-07-30 | 2020-07-09 | 삼성전자 주식회사 | 통신 시스템에서 오류 은닉 방법 및 장치 |
WO2015174912A1 (en) | 2014-05-15 | 2015-11-19 | Telefonaktiebolaget L M Ericsson (Publ) | Audio signal classification and coding |
PL3367380T3 (pl) * | 2014-06-13 | 2020-06-29 | Telefonaktiebolaget Lm Ericsson (Publ) | Obsługa sekwencji błędów ramki |
US20180213294A1 (en) * | 2017-01-23 | 2018-07-26 | Ramp Holdings, Inc. | Recovering from gaps in video transmission for web browser-based players |
US10763885B2 (en) | 2018-11-06 | 2020-09-01 | Stmicroelectronics S.R.L. | Method of error concealment, and associated device |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001203648A (ja) * | 2000-01-20 | 2001-07-27 | Sony Corp | 送信方法及び装置、受信方法及び装置、配信方法及びシステム、再生方法及び装置、記録媒体 |
JP2003218932A (ja) * | 2001-11-15 | 2003-07-31 | Matsushita Electric Ind Co Ltd | エラー隠蔽装置及び方法 |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CA1321645C (en) * | 1988-09-28 | 1993-08-24 | Akira Ichikawa | Method and system for voice coding based on vector quantization |
US5384891A (en) * | 1988-09-28 | 1995-01-24 | Hitachi, Ltd. | Vector quantizing apparatus and speech analysis-synthesis system using the apparatus |
IT1241358B (it) * | 1990-12-20 | 1994-01-10 | Sip | Sistema di codifica del segnale vocale con sottocodice annidato |
ATE294441T1 (de) * | 1991-06-11 | 2005-05-15 | Qualcomm Inc | Vocoder mit veränderlicher bitrate |
SE502244C2 (sv) | 1993-06-11 | 1995-09-25 | Ericsson Telefon Ab L M | Sätt och anordning för avkodning av ljudsignaler i ett system för mobilradiokommunikation |
US5574825A (en) * | 1994-03-14 | 1996-11-12 | Lucent Technologies Inc. | Linear prediction coefficient generation during frame erasure or packet loss |
US5450449A (en) * | 1994-03-14 | 1995-09-12 | At&T Ipm Corp. | Linear prediction coefficient generation during frame erasure or packet loss |
US5615298A (en) * | 1994-03-14 | 1997-03-25 | Lucent Technologies Inc. | Excitation signal synthesis during frame erasure or packet loss |
KR970011728B1 (ko) * | 1994-12-21 | 1997-07-14 | 김광호 | 음향신호의 에러은닉방법 및 그 장치 |
GB9520445D0 (en) * | 1995-10-06 | 1995-12-06 | British Telecomm | Convolutional codes |
JP3617503B2 (ja) * | 1996-10-18 | 2005-02-09 | 三菱電機株式会社 | 音声復号化方法 |
JP2000267699A (ja) * | 1999-03-19 | 2000-09-29 | Nippon Telegr & Teleph Corp <Ntt> | 音響信号符号化方法および装置、そのプログラム記録媒体、および音響信号復号装置 |
US6597961B1 (en) * | 1999-04-27 | 2003-07-22 | Realnetworks, Inc. | System and method for concealing errors in an audio transmission |
DE19921122C1 (de) * | 1999-05-07 | 2001-01-25 | Fraunhofer Ges Forschung | Verfahren und Vorrichtung zum Verschleiern eines Fehlers in einem codierten Audiosignal und Verfahren und Vorrichtung zum Decodieren eines codierten Audiosignals |
US6850559B1 (en) * | 1999-06-28 | 2005-02-01 | At&T Corp. | System and methods for transmitting data |
EP1199709A1 (en) * | 2000-10-20 | 2002-04-24 | Telefonaktiebolaget Lm Ericsson | Error Concealment in relation to decoding of encoded acoustic signals |
DE60222445T2 (de) * | 2001-08-17 | 2008-06-12 | Broadcom Corp., Irvine | Verfahren zum verbergen von bitfehlern für die sprachcodierung |
US7143032B2 (en) | 2001-08-17 | 2006-11-28 | Broadcom Corporation | Method and system for an overlap-add technique for predictive decoding based on extrapolation of speech and ringinig waveform |
CN1323532C (zh) * | 2001-11-15 | 2007-06-27 | 松下电器产业株式会社 | 错误隐蔽装置和方法 |
US7835916B2 (en) * | 2003-12-19 | 2010-11-16 | Telefonaktiebolaget Lm Ericsson (Publ) | Channel signal concealment in multi-channel audio systems |
-
2005
- 2005-03-01 JP JP2006510676A patent/JP4744438B2/ja not_active Expired - Fee Related
- 2005-03-01 US US10/591,601 patent/US7809556B2/en active Active
- 2005-03-01 AT AT05719724T patent/ATE523876T1/de not_active IP Right Cessation
- 2005-03-01 WO PCT/JP2005/003407 patent/WO2005086138A1/ja active Application Filing
- 2005-03-01 CN CN2005800071542A patent/CN1930607B/zh not_active Expired - Fee Related
- 2005-03-01 EP EP05719724A patent/EP1722359B1/en not_active Not-in-force
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001203648A (ja) * | 2000-01-20 | 2001-07-27 | Sony Corp | 送信方法及び装置、受信方法及び装置、配信方法及びシステム、再生方法及び装置、記録媒体 |
JP2003218932A (ja) * | 2001-11-15 | 2003-07-31 | Matsushita Electric Ind Co Ltd | エラー隠蔽装置及び方法 |
Non-Patent Citations (1)
Title |
---|
GOTO M. ET AL: "Audio Frame Sonshitsu Hosho Hoho no kento.", THE INSTITUTE OF ELECTRONICS, INFORMATION AND COMMUNICATION ENGINEERS., 8 March 2004 (2004-03-08), pages 143, XP008114506 * |
Cited By (38)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8069035B2 (en) | 2005-10-14 | 2011-11-29 | Panasonic Corporation | Scalable encoding apparatus, scalable decoding apparatus, and methods of them |
WO2007043642A1 (ja) * | 2005-10-14 | 2007-04-19 | Matsushita Electric Industrial Co., Ltd. | スケーラブル符号化装置、スケーラブル復号装置、およびこれらの方法 |
JP5142723B2 (ja) * | 2005-10-14 | 2013-02-13 | パナソニック株式会社 | スケーラブル符号化装置、スケーラブル復号装置、およびこれらの方法 |
WO2007143953A1 (fr) * | 2006-06-08 | 2007-12-21 | Huawei Technologies Co., Ltd. | Dispositif et procédé pour dissimulation de trames perdues |
US7778824B2 (en) | 2006-06-08 | 2010-08-17 | Huawei Technologies Co., Ltd. | Device and method for frame lost concealment |
JP2014160280A (ja) * | 2006-11-28 | 2014-09-04 | Samsung Electronics Co Ltd | フレームエラー隠匿方法及び装置、これを利用した復号化方法及び装置 |
US10096323B2 (en) | 2006-11-28 | 2018-10-09 | Samsung Electronics Co., Ltd. | Frame error concealment method and apparatus and decoding method and apparatus using the same |
JP2017078870A (ja) * | 2006-11-28 | 2017-04-27 | サムスン エレクトロニクス カンパニー リミテッド | フレームエラー隠匿装置 |
JP2010511201A (ja) * | 2006-11-28 | 2010-04-08 | サムスン エレクトロニクス カンパニー リミテッド | フレームエラー隠匿方法及び装置、これを利用した復号化方法及び装置 |
US9424851B2 (en) | 2006-11-28 | 2016-08-23 | Samsung Electronics Co., Ltd. | Frame error concealment method and apparatus and decoding method and apparatus using the same |
US8843798B2 (en) | 2006-11-28 | 2014-09-23 | Samsung Electronics Co., Ltd. | Frame error concealment method and apparatus and decoding method and apparatus using the same |
JP2008182380A (ja) * | 2007-01-23 | 2008-08-07 | Alpine Electronics Inc | オーディオ装置 |
US8320265B2 (en) | 2007-11-05 | 2012-11-27 | Huawei Technologies Co., Ltd. | Method and apparatus for obtaining an attenuation factor |
JP2010176142A (ja) * | 2007-11-05 | 2010-08-12 | Huawei Technologies Co Ltd | 減衰率を取得する方法および装置 |
JP2009175693A (ja) * | 2007-11-05 | 2009-08-06 | Huawei Technologies Co Ltd | 減衰率を取得する方法および装置 |
JP2009290567A (ja) * | 2008-05-29 | 2009-12-10 | Canon Inc | データ送信装置及びデータ受信装置 |
JP2013527479A (ja) * | 2009-06-29 | 2013-06-27 | オーディエンス,インコーポレイテッド | 破損したオーディオ信号の修復 |
US9838784B2 (en) | 2009-12-02 | 2017-12-05 | Knowles Electronics, Llc | Directional audio capture |
JP2013519920A (ja) * | 2010-02-11 | 2013-05-30 | クゥアルコム・インコーポレイテッド | サブ帯域コード化復号器における損失パケットの隠蔽 |
JP2016510432A (ja) * | 2013-02-05 | 2016-04-07 | テレフオンアクチーボラゲット エル エム エリクソン(パブル) | オーディオフレーム損失のコンシールメントを制御する方法及び装置 |
JP2017097365A (ja) * | 2013-02-05 | 2017-06-01 | テレフオンアクチーボラゲット エルエム エリクソン(パブル) | オーディオフレーム損失のコンシールメントを制御する方法及び装置 |
US9536540B2 (en) | 2013-07-19 | 2017-01-03 | Knowles Electronics, Llc | Speech signal separation and synthesis based on auditory scene analysis and speech modeling |
US9978388B2 (en) | 2014-09-12 | 2018-05-22 | Knowles Electronics, Llc | Systems and methods for restoration of speech components |
US10762907B2 (en) | 2016-01-29 | 2020-09-01 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Apparatus and method for improving a transition from a concealed audio signal portion to a succeeding audio signal portion of an audio signal |
US9820042B1 (en) | 2016-05-02 | 2017-11-14 | Knowles Electronics, Llc | Stereo separation and directional suppression with omni-directional microphones |
JP2021502608A (ja) * | 2017-11-10 | 2021-01-28 | フラウンホーファー−ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン | 異なる損失隠蔽ツールのセットをサポートするオーディオデコーダ |
US11127408B2 (en) | 2017-11-10 | 2021-09-21 | Fraunhofer—Gesellschaft zur F rderung der angewandten Forschung e.V. | Temporal noise shaping |
US11217261B2 (en) | 2017-11-10 | 2022-01-04 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Encoding and decoding audio signals |
US11315580B2 (en) | 2017-11-10 | 2022-04-26 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio decoder supporting a set of different loss concealment tools |
US11315583B2 (en) | 2017-11-10 | 2022-04-26 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio encoders, audio decoders, methods and computer programs adapting an encoding and decoding of least significant bits |
JP7073492B2 (ja) | 2017-11-10 | 2022-05-23 | フラウンホーファー-ゲゼルシャフト・ツール・フェルデルング・デル・アンゲヴァンテン・フォルシュング・アインゲトラーゲネル・フェライン | 異なる損失隠蔽ツールのセットをサポートするオーディオデコーダ |
US11380341B2 (en) | 2017-11-10 | 2022-07-05 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Selecting pitch lag |
US11380339B2 (en) | 2017-11-10 | 2022-07-05 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio encoders, audio decoders, methods and computer programs adapting an encoding and decoding of least significant bits |
US11386909B2 (en) | 2017-11-10 | 2022-07-12 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Audio encoders, audio decoders, methods and computer programs adapting an encoding and decoding of least significant bits |
US11462226B2 (en) | 2017-11-10 | 2022-10-04 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Controlling bandwidth in encoders and/or decoders |
US11545167B2 (en) | 2017-11-10 | 2023-01-03 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Signal filtering |
US11562754B2 (en) | 2017-11-10 | 2023-01-24 | Fraunhofer-Gesellschaft Zur F Rderung Der Angewandten Forschung E.V. | Analysis/synthesis windowing function for modulated lapped transformation |
US12033646B2 (en) | 2017-11-10 | 2024-07-09 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Analysis/synthesis windowing function for modulated lapped transformation |
Also Published As
Publication number | Publication date |
---|---|
EP1722359A4 (en) | 2009-09-02 |
EP1722359B1 (en) | 2011-09-07 |
EP1722359A1 (en) | 2006-11-15 |
JPWO2005086138A1 (ja) | 2008-01-24 |
US7809556B2 (en) | 2010-10-05 |
US20070198254A1 (en) | 2007-08-23 |
CN1930607B (zh) | 2010-11-10 |
ATE523876T1 (de) | 2011-09-15 |
CN1930607A (zh) | 2007-03-14 |
JP4744438B2 (ja) | 2011-08-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4744438B2 (ja) | エラー隠蔽装置およびエラー隠蔽方法 | |
KR101290425B1 (ko) | 소거된 스피치 프레임을 복원하는 시스템 및 방법 | |
KR101246991B1 (ko) | 오디오 신호 처리 방법 | |
JP5289320B2 (ja) | ピッチ周期訂正を用いたデジタルオーディオ信号の損失ブロックの合成 | |
JP5587405B2 (ja) | スピーチフレーム内の情報のロスを防ぐためのシステムおよび方法 | |
JP4603091B2 (ja) | 高帯域信号にフレーム消失の隠蔽を行う方法および装置 | |
JP4864201B2 (ja) | 音声信号の量子化ノイズをマスクするためのシステムと方法 | |
WO2002021515A1 (fr) | Dissimulation d'erreurs de transmission dans un signal audio | |
EP2277172A1 (fr) | Dissimulation d'erreur de transmission dans un signal audionumerique dans une structure de decodage hierarchique | |
EP2458585B1 (en) | Error concealment for sub-band coded audio signals | |
CN1418361A (zh) | 宽带信号传输系统 | |
JP2013527479A (ja) | 破損したオーディオ信号の修復 | |
JPWO2005036527A1 (ja) | スペクトル包絡線符号化のための時間境界及び周波数分解能の決定方法 | |
RU2328775C2 (ru) | Улучшенное маскирование ошибки в области частот | |
WO1999040573A1 (fr) | Procede de decodage d'un signal audio avec correction des erreurs de transmission | |
RU2437170C2 (ru) | Ослабление чрезмерной тональности, в частности, для генерирования возбуждения в декодере при отсутствии информации | |
JP4825944B2 (ja) | レート判定誤りとそのアーティファクトの低減方法及び装置 | |
JP5390690B2 (ja) | 音声コーデックの品質向上装置およびその方法 | |
KR100792209B1 (ko) | 디지털 오디오 패킷 손실을 복구하기 위한 방법 및 장치 | |
KR101452635B1 (ko) | Lms 예측기를 이용한 패킷 손실 은닉 방법 및 이를 기록한 기록매체 | |
CN105632504B (zh) | Adpcm编解码器及adpcm解码器丢包隐藏的方法 | |
Ofir | Packet loss concealment for audio streaming | |
KR101551236B1 (ko) | 패킷 손실 은닉에서의 적응형 뮤팅 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006510676 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005719724 Country of ref document: EP Ref document number: 200580007154.2 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2007198254 Country of ref document: US Ref document number: 10591601 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 2005719724 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 10591601 Country of ref document: US |