WO2005119950A1 - 音声データ送信/受信装置および音声データ送信/受信方法 - Google Patents
音声データ送信/受信装置および音声データ送信/受信方法 Download PDFInfo
- Publication number
- WO2005119950A1 WO2005119950A1 PCT/JP2005/009252 JP2005009252W WO2005119950A1 WO 2005119950 A1 WO2005119950 A1 WO 2005119950A1 JP 2005009252 W JP2005009252 W JP 2005009252W WO 2005119950 A1 WO2005119950 A1 WO 2005119950A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- data sequence
- voice
- decoding
- channel
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 51
- 238000012545 processing Methods 0.000 claims description 51
- 230000005540 biological transmission Effects 0.000 claims description 28
- 238000012937 correction Methods 0.000 claims description 18
- 230000003111 delayed effect Effects 0.000 claims description 17
- 238000004364 calculation method Methods 0.000 claims description 9
- 238000004891 communication Methods 0.000 abstract description 14
- 230000001934 delay Effects 0.000 abstract description 5
- 238000000926 separation method Methods 0.000 description 17
- 230000005236 sound signal Effects 0.000 description 15
- 238000010586 diagram Methods 0.000 description 14
- 238000001514 detection method Methods 0.000 description 7
- 230000000694 effects Effects 0.000 description 7
- CBGUOGMQLZIXBE-XGQKBEPLSA-N clobetasol propionate Chemical compound C1CC2=CC(=O)C=C[C@]2(C)[C@]2(F)[C@@H]1[C@@H]1C[C@H](C)[C@@](C(=O)CCl)(OC(=O)CC)[C@@]1(C)C[C@@H]2O CBGUOGMQLZIXBE-XGQKBEPLSA-N 0.000 description 6
- 229940069205 cormax Drugs 0.000 description 6
- 230000003044 adaptive effect Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 4
- 238000007429 general method Methods 0.000 description 3
- 230000010354 integration Effects 0.000 description 3
- 230000006978 adaptation Effects 0.000 description 1
- 230000015572 biosynthetic process Effects 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 235000021156 lunch Nutrition 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 238000013139 quantization Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000003595 spectral effect Effects 0.000 description 1
- 238000003786 synthesis reaction Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/005—Correction of errors induced by the transmission channel, if related to the coding algorithm
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/008—Multichannel audio signal coding or decoding using interchannel correlation to reduce redundancy, e.g. joint-stereo, intensity-coding or matrixing
Definitions
- the present invention relates to an audio data transmission Z receiver and an audio data transmission Z reception method, and more particularly to an audio data transmission used in an audio communication system in which compensation processing is performed for erroneous audio data and lost audio data.
- the present invention relates to a Z receiver and an audio data transmission Z reception method.
- IP Internet Protocol
- radio transmission error etc.
- voice data For voice communication in an IP (Internet Protocol) network or a wireless communication network, loss of IP packets, radio transmission error, etc. can prevent voice data from being received on the receiving side. Some voice data may be received. Therefore, in a voice communication system, processing is generally performed to compensate for erroneous voice data or lost voice data.
- a voice signal as an input source signal is encoded as voice data, multiplexed (packetized), and transmitted to a destination device.
- multiplexing is performed with one speech frame as one transmission unit.
- the voice codec methods of 3GPP (3rd Generation Partnership Project) standard, AMR (Adaptive Multi-Rate) and AMR—AW—Adaptive Multi-Rate Wideband (WB) are IP packet networks Defines the format of audio data.
- Non-Patent Document 2 discloses a frame compensation method of AMR regarding the compensation process of a voice frame.
- the speech processing operation in the above-described speech communication system will be outlined using FIG. .
- the sequence numbers ( ⁇ , n ⁇ 2, n ⁇ 1, n, n + 1, n + 2, ⁇ ) in FIG. 1 are frame numbers assigned to each speech frame.
- the voice signal is decoded according to the order of the frame numbers, and the decoded voice is output as a sound wave.
- encoding, multiplexing, transmission, separation and decoding are performed for each speech frame. For example, when the nth frame is lost, a speech frame (for example, the n-1th frame or the n-2th frame) received in the past is referred to, and the frame compensation process for the nth frame is performed.
- Non-Patent Document 1 describes a rule regarding multiplexing when audio data is multi-channel data (for example, stereo audio data). According to the same document, when audio data is, for example, data of two channels, audio data of the left channel (L ch) and audio data of the right channel (R-ch) corresponding to the same time are multiplexed. .
- Non-Patent Document 1 Real-time Transfer Protocol (RTP) Payload Format and File Storage Format for the Adaptive Multi-Rate (AMR) and Adaptive Multi-Rate Wideband (AMR-WB) Audio Codecs ", IETF RFC 3267
- RTP Real-time Transfer Protocol
- AMR Adaptive Multi-Rate
- AMR-WB Adaptive Multi-Rate Wideband Audio Codecs
- Non-Patent Document 2 "Mandatory Speech Codec speech processing lunches; AMR Speech Codecs; Error concealment of lost frames", 3rd Generation Partnership Project, TS 26.091
- the present invention has been made in view of the situation, and it is an object of the present invention to provide an audio data transmission Z receiver and an audio data transmission Z reception method that can realize high quality frame compensation. I assume.
- An audio data transmitting apparatus transmits a multi-channel audio data series including a first data series corresponding to a first channel and a second data series corresponding to a second channel.
- delay means for performing delay processing for delaying the first data sequence from the second data sequence by a predetermined delay amount, and the voice data sequence subjected to the delay processing.
- a configuration is adopted that includes: multiplexing means for multiplexing; and transmission means for transmitting the multiplexed voice data sequence.
- a voice data receiving apparatus is a multi-channel voice data sequence including a first data sequence corresponding to a first channel and a second data sequence corresponding to a second channel, and the first data sequence Means for receiving the audio data sequence multiplexed in a state delayed by a predetermined delay amount from the second data sequence, separation means for separating the received audio data sequence for each channel, and separation Decoding means for decoding the voiced data series for each channel, and the decoding means, when there is a loss or an error in the voice data series separated, the first data series or the like.
- a configuration is adopted that includes a compensation unit that compensates for the loss or error in the other data sequence using one data sequence of the second data sequence.
- the audio data transmission method of the present invention is an audio data transmission method for transmitting a multi-channel audio data sequence including a first data sequence corresponding to a first channel and a second data sequence corresponding to a second channel.
- a multiplexing step of multiplexing an audio data sequence and a transmitting step of transmitting the multiplexed audio data sequence are provided.
- the audio data receiving method of the present invention is a multi-channel audio data sequence including a first data sequence corresponding to a first channel and a second data sequence corresponding to a second channel.
- a decoding step of decoding the separated voice data sequence for each channel wherein the decoding step includes the first step when there is a loss or error in the separated voice data sequence. It is arranged to have a compensation step of compensating for the loss or error in the other data series using one data series out of the data series and the second data series.
- FIG. 1 A diagram for explaining an example of voice processing operation in a conventional voice communication system
- FIG. 2A A block diagram showing a configuration of a voice data transmitting apparatus according to a first embodiment of the present invention
- FIG. 2A Block diagram showing the configuration of the voice data receiving apparatus according to Embodiment 1 of the present invention
- Figure 3 A block diagram showing the internal configuration of the voice decoding unit in the voice data receiving apparatus according to Embodiment 1 of the present invention
- FIG. 4 A diagram for explaining the operation of an audio data transmitting apparatus and an audio data receiving apparatus according to Embodiment 1 of the present invention.
- FIG. 5 A block diagram showing an internal configuration of a voice decoding unit in a voice data receiving apparatus according to a second embodiment of the present invention.
- FIG. 6 A block diagram showing an internal configuration of a voice decoding unit in a voice data receiving apparatus according to a third embodiment of the present invention
- FIG. 7 A block diagram showing a modification of the internal configuration of the speech decoding unit in the speech data receiving apparatus according to Embodiment 3 of the present invention.
- FIGS. 2A and 2B are block diagrams respectively showing configurations of an audio data transmitting apparatus and an audio data receiving apparatus according to Embodiment 1 of the present invention.
- the source-side power input multi-channel audio signal has two channels including the left channel (L-ch) and the right channel (R-ch), ie this audio signal is a stereo signal . Therefore, the voice data transmitting apparatus 10 and the voice data receiving apparatus 20 respectively shown in FIGS. 2A and 2B are provided with two processing systems for the left and right channels.
- the number of channels of the audio signal is not limited to two. When the number of channels is three or more, by providing three or more processing systems on the transmission side and the reception side, respectively, the same function and effect as those of the present embodiment can be realized.
- Speech data transmitting apparatus 10 shown in FIG. 2A has speech coding section 102, delay section 104, multiplexing section 106, and transmission section 108.
- Speech coding section 102 codes the input multi-channel speech signal and outputs code data. This encoding is performed independently for each channel.
- L-ch coded data is referred to as “L-ch coded data”
- R-ch coded data is referred to as “R-ch coded data”.
- Delay section 104 delays the L-ch code data from voice code section 102 by one voice frame and outputs the delayed data to multiplexing section 106. That is, the delay unit 104 is disposed downstream of the speech coding unit 102. As described above, since the delay processing is arranged in the latter stage of the speech coding processing, the delay processing can be performed on the data after the coding, and the delay processing is arranged in the former stage of the speech coding process. Processing can be simplified compared to when
- the delay amount in the delay processing performed by the delay unit 104 is preferably set in units of voice frames, but is not limited to one voice frame.
- the voice communication system including the voice data transmitting apparatus 10 and the voice data receiving apparatus 20 according to the present embodiment is premised on the main application being real-time voice communication as well as streaming of audio data, for example. . Therefore, in order to prevent undesired effects from being exerted on the communication quality by setting the delay amount to a large value, in the present embodiment, the delay amount is preset to the minimum value, that is, one voice frame. There is.
- the delay unit 104 delays only L ch code data only.
- the way of performing delay processing on the force voice data is not limited to this.
- delay The unit 104 may have a configuration in which not only L ch code data but also R ch code data is delayed and the difference between the delay amounts is set in units of speech frames. Also, instead of delaying L-ch, only R-ch may be delayed.
- Multiplexing section 106 is configured to transmit the L-ch code data from delay section 104 and the R-ch code data of speech code section 102 in a predetermined format (for example, a format similar to that in the prior art). Multi-channel voice data by multiplexing. That is, in this embodiment, for example, L-ch code data having frame number N is multiplexed with R-ch code data having frame number N + 1.
- Transmitting section 108 performs transmission processing determined in advance according to the transmission path to audio data receiving apparatus 20 on the audio data from multiplexing section 106, and transmits the audio data to audio data receiving apparatus 20. .
- the voice data receiving apparatus 20 shown in FIG. 2B includes a receiving unit 110, a voice data loss detecting unit 112, a separating unit 114, a delay unit 116, and a voice decoding unit 118.
- the speech decoding unit 118 has a frame compensation unit 120.
- FIG. 3 is a block diagram showing a more detailed configuration of the speech decoding unit 118. As shown in FIG.
- the speech decoding unit 118 shown in FIG. 3 includes an L-ch decoding unit 122 and an R-ch decoding unit 124 in addition to the frame compensation unit 120.
- frame compensation section 120 has switch section 126 and superposition addition section 128, and superposition addition section 128 includes L-ch superposition addition section 130 and R-ch superposition addition. It has part 132.
- Reception unit 110 performs predetermined reception processing on the received voice data received from voice data transmitting apparatus 10 via the transmission path.
- Voice data loss detection unit 112 generates a loss or an error (hereinafter, “loss or error” is collectively referred to as “loss”) in the received voice data subjected to the reception processing by reception unit 110, Detect whether or not When the occurrence of a loss is detected, a loss flag is output to the separation unit 114, the switch unit 126, and the superposition addition unit 128.
- the loss flag indicates which speech frame has been lost in the series of speech frames constituting each of L-ch code data and R-ch code data.
- Separation section 114 determines whether the loss flag has been input from voice data loss detection section 112 or not. Accordingly, the received voice data from the receiver 110 is separated for each channel.
- the L ch code data and R ch code data obtained by the separation are output to the L ch decoding unit 122 and the delay unit 116, respectively.
- the delay unit 116 corresponds to delaying L-ch on the transmission side, and in order to match (return) the time relationship between L-ch and R-ch, R- from the separation unit 114
- the ch encoded data is delayed by one audio frame and output to the R-ch decoding unit 124.
- the amount of delay in the delay processing performed by the delay unit 116 is preferably performed in units of speech frames, but is not limited to one speech frame.
- the delay amount in delay section 116 is set to the same value as the delay amount in delay section 104 in audio data transmitting apparatus 10.
- delay processing section 116 is a process that matches only the time relationship between L-ch and R-ch, which delays only R-ch code data
- audio data will be processed.
- the delay unit 116 may have a configuration in which not only R-ch code data but also L ch code data are delayed, and the difference between the delay amounts is set in units of speech frames. . Also, when R-ch is delayed on the transmitting side, L-ch is delayed on the receiving side.
- the audio decoding unit 118 performs processing for decoding multi-channel audio data for each channel.
- the L ch decoding unit 122 decodes the L ch coded data from the separation unit 114, and the L ch decoded speech signal obtained by the decoding is output. Since the output terminal of the L-ch decoding unit 122 and the input terminal of the L-ch superposition addition unit 130 are always connected V, the output of the L-ch decoded speech signal to the L-ch superposition addition unit 130 is always a line. It will be.
- the R-ch decoding unit 124 decodes the R-ch coded data from the delay unit 124, and the R-ch decoded speech signal obtained by the decoding is output. Since the output end of the R-ch decoding unit 124 and the input end of the R-ch superposition addition unit 132 are always connected, the output of the R-ch decoded speech signal to the R-ch superposition addition unit 132 is always constant. To be done.
- switch unit 126 According to the information content indicated by the loss flag, the connection state of the Lch decoding unit 122 and the R-ch overlapping addition unit 132 and the connection state of the R-ch decoding unit 124 and the Lch overlapping addition unit 130 are switched.
- R-ch decoding unit 124 R Of the ch decoded speech signal, the R-ch decoded speech signal obtained by decoding the speech frame corresponding to frame number K is not only the R-ch superposition adder 132 but also the L-ch superposition adder The output end of the R-ch decoding unit 124 is connected to the input end of the L-ch superposition and addition unit 130 so that the output is also output to 130.
- an audio frame that belongs to R-ch code data and corresponds to frame number K.
- the speech frame corresponding to the frame number K is decoded when the loss flag indicating that the frame loss has occurred is input.
- the output end of the L-ch decoding unit 122 is R-ch so that the L-ch decoded speech signal obtained by the above is output not only to the L-ch superposition addition unit 130 but also to the R-ch superposition addition unit 132. Connect with the input terminal of the superposition adder 132.
- the superposition addition unit 128 performs superposition addition processing described later on the multi-channel decoded speech signal in accordance with the loss flag from the speech data loss detection unit 112. More specifically, the loss flag from the voice data loss detection unit 112 is input to both the L-ch superposition and addition unit 130 and the R-ch superposition and addition unit 132.
- the L-ch superposition addition unit 130 When the loss flag is not input, the L-ch superposition addition unit 130 outputs the L-ch decoded speech signal from the L-ch decoding unit 122 as it is.
- the L-ch decoded speech signal to be output is converted to a sound wave and output by, for example, speech output processing in a later stage not shown.
- L-ch superposition addition section 130 receives, for example, a loss flag indicating that the speech frame corresponding to frame number K belonging to the R-ch code data has been lost.
- the Lch decoded speech signal is output as it is.
- the L channel decoded speech signal to be output is output to, for example, the above-mentioned speech output processing stage.
- the L-ch superposition addition unit 130 belongs to, for example, L-ch code data. If a loss flag indicating that the voice frame corresponding to the frame number has been lost is input, the Lch decoding unit 122 uses the encoded data of the voice frame up to the frame number K 1 or the decoded voice signal to use The compensation signal (L ch compensation signal) obtained by performing compensation of the frame of frame number K in a general method of decoding and the speech frame corresponding to frame number K in the R-ch decoding unit 124 are decoded.
- the R-ch decoded speech signal obtained by The superposition is performed, for example, such that the weight of the Lch compensation signal is large near both ends of the frame of frame number K, and the weight of the R-ch decoded signal is large otherwise.
- the L-ch decoded speech signal corresponding to the frame number K is restored, and the frame compensation process for the speech frame of the frame number K (L ch code data) is completed.
- the restored L-ch decoded speech signal is output to the above-mentioned speech output processing stage, for example.
- the superposition addition unit instead of using the Lch compensation signal and R-ch decode signal as described above, one end of the decoded signal of frame number K 1 of L-ch is used.
- the superposition addition is performed by using the rear end of the decoded signal of the frame number K-1 and the R-ch !, and the result is the frame number as the rear end signal of the decoded signal of the L ch frame number K 1
- the K frame may output the R-ch decoded signal as it is.
- the R-ch superposition addition unit 132 When the loss flag is not input, the R-ch superposition addition unit 132 outputs the R-ch decoded speech signal from the R-ch decoding unit 124 as it is.
- the output R-ch decoded voice signal is output to, for example, the above-mentioned voice output processing stage.
- the R-ch superposition addition unit 132 R-Outputs the ch decoded speech signal as it is.
- the output R-ch decoded speech signal is output to, for example, the above-mentioned speech output processing stage.
- the R-ch superposition addition unit 132 receives, for example, a loss flag indicating that the speech frame corresponding to the frame number K belonging to the R-ch code data has been lost.
- R-ch decoding unit 124 encoded data of speech frame up to frame number K 1
- R-ch compensation signal and L-ch decoding unit 122 correspond to frame number K And an L-ch decoded speech signal obtained by decoding the speech frame.
- the superposition is performed, for example, in the vicinity of both ends of the frame of frame number K: R ⁇ c
- the weight of the h compensation signal is large so that the weight of the L ch decoded signal is large otherwise. In this way, the R-ch decoded speech signal corresponding to frame number K is restored and
- the restored R-ch decoded speech signal is output to the above-mentioned speech output processing stage, for example.
- the audio decoding unit 118 uses the state data to decode the next audio frame depending on the decoding state of the audio frame in the past.
- the case where the flooding method is adopted will be described.
- the state data obtained when the R-ch decoding unit 124 decodes the used R-ch encoded data may be acquired and used for decoding the next audio frame. By doing this, discontinuities between frames can be avoided.
- the normal decoding process means a decoding process to be performed on a speech frame due to loss!
- the R-ch decoding unit 124 when the R-ch decoding unit 124 performs a normal decoding process on the voice frame following (immediately after) the voice frame in which the loss occurred, the voice frame in which the loss occurs.
- the state data obtained when the Lch coded data is decoded by the Lch decoding unit 122 may be used for the compensation for the next voice frame. By doing this, discontinuities between frames can be avoided.
- CELP Code
- the decoded waveform is superimposed and added between adjacent frames, and the final decoded speech waveform is Even if there is previous frame decoding in the configuration to be obtained, waveform data, etc., normal speech decoding is performed on the next (immediately) speech frame of the lossy speech frame using these state data. Good.
- FIG. 4 is a diagram for explaining the operation of the voice data transmitting apparatus 10 and the voice data receiving apparatus 20 according to the present embodiment.
- the multi-channel audio signal input to speech code selector 102 also includes a sequence of L-ch audio signals and a sequence power of R-ch audio signals.
- each of L-ch and R-ch audio signals (for example, L-ch audio signal SL (n) and R-ch audio signal SR (n)) corresponding to the same frame number correspond to each other.
- L-ch audio signal SL (n) and R-ch audio signal SR (n) corresponding to the same frame number correspond to each other.
- it is input to the speech code input unit 102.
- Each audio signal corresponding to the same frame number as each other is an audio signal to be output as an acoustic wave at the same time.
- the multi-channel voice signal is subjected to each processing by voice coding unit 102, delay unit 104, and multiplexing unit 106 to become transmission voice data.
- the transmission voice data is multiplexed with L ch code data delayed by one voice frame from R ch code data.
- L ch code data delayed by one voice frame from R ch code data.
- 1 ⁇ 1 code data 0 ⁇ (11-1) is multiplexed with R-ch code data CR (n).
- voice data is packetized.
- the generated transmission voice data is transmitted from the transmitter to the receiver.
- the received voice data received by voice data receiving apparatus 20 is multiplexed in a state where the L ch encoded data is delayed by one voice frame than the R-ch code data. It has been For example, L-ch code data CL '(n-1) is multiplexed with R-ch code data CR, (n).
- Such multi-channel received speech data is subjected to each processing by separation section 114, delay section 116 and speech decoding section 118 to become a decoded speech signal.
- the received voice data received by voice data receiving apparatus 20 is divided into L ch encoded data CL ′ (n ⁇ 1) and R ⁇ ch encoded data CR, (n). There was a loss Suppose.
- the corresponding decoded voice signal SR' (n) will also be lost, but L-with the same frame number as the coded data CR, (n) Since the encoded data CL (n) of ch is received without loss, frame compensation is performed using the decoded speech signal SL '(n) decoded by the encoded data CL' (n). As a result, the decoded speech signal SR ′ (n) is restored. By performing such frame compensation, it is possible to improve the sound quality to be restored.
- multi-channel audio data multiplexed in a state delayed by one audio frame from L ch encoded data strength data is separated and separated for each channel.
- loss or error occurs in the encoded data
- one of the L-ch code data and R-ch code data is used to transmit the loss or error in the other data sequence. To compensate. For this reason, at the receiving end, even when loss or error occurs in the voice frame, at least one channel of multi-channel can be correctly received, and frame compensation of another channel can be performed using that channel. It is possible to realize high quality frame compensation.
- voice frames of one channel can be recovered using voice frames of another channel, frame compensation performance of each channel included in multi-channel is improved. It can be raised.
- the present embodiment has been described by way of example of a configuration in which data of one channel is delayed in the latter stage of speech code processing unit 102, the configuration capable of realizing the effects of the present embodiment is not limited to this. It is not limited to.
- the data of one channel may be delayed before the voice code processing unit 102.
- the delay amount to be set is not limited to the unit of voice frame. For example, it is also possible to make the delay amount shorter than one audio frame. For example, assuming that one audio frame is 20 ms, the delay amount can be set to 0.5 audio frame (1 Oms).
- FIG. 5 is a block diagram showing a configuration of an audio decoding unit in an audio data receiving apparatus according to Embodiment 2 of the present invention.
- the voice data transmitting apparatus and voice data receiving apparatus according to the present embodiment have the same basic configuration as that described in the first embodiment, and thus the same or corresponding components are used. Are given the same reference numerals and the detailed description thereof is omitted.
- the difference between the present embodiment and the first embodiment is only the internal configuration of the speech decoding unit.
- the speech decoding unit 118 shown in FIG. 5 has a frame compensation unit 120.
- the frame compensation unit 120 includes a switch unit 202, an L-ch decoding unit 204, and an R-ch decoding unit 206.
- switch unit 202 When a loss flag is input from voice data loss detection unit 112, switch unit 202 connects the connection state of delay unit 114 and R-ch decoding unit 206 according to the information content indicated by the loss flag, and also delays the delay unit. 116 and L—Switch the connection state of the ch decoding unit 204.
- the L ⁇ ch encoded data from the demultiplexing unit 114 is output to the L ⁇ ch decoding unit 204 so that the L ⁇ ch encoded data is output to the L ⁇ ch decoding unit 204 only.
- the output end of ch is connected to the input end of L-ch decoder 204.
- the output end of the delay unit 116 is R-ch decoding unit so that the R-ch code data from the delay unit 116 is output only to the R-ch decoding unit 206. Connect with the input end of 206.
- the R-ch encoded data from delay section 116 is input, of the R-ch encoded data from delay section 116
- the output end of the delay unit 116 is output to the L-ch decoding unit 204 and the R-ch so that they are output to the L-ch decoding unit 204 as well as the speech frame strength R-ch decoding unit 206 corresponding to the frame number K. It connects with both input ends of the decoding unit 206.
- an audio frame corresponding to a frame number K that belongs to R-ch code data is also known as a frame number K that belongs to R-ch code data.
- the L-ch output terminal of the separation unit 114 is connected to both the input terminals of the R-ch decoding unit 206 and the L-ch decoding unit 204 so that the output is also output to the R-ch decoding unit 206 instead.
- the L-ch decoding unit 204 decodes the L-ch coded data. This decoded result is output as an L ch decoded speech signal. That is, this decoding process is a normal speech decoding process.
- the L-ch decoding unit 204 decodes the R-ch coded data. By decoding the R-ch code data by the Lch decoding unit 204 in this way, it is possible to restore the voice signal corresponding to the L-ch coded data in which the loss occurs.
- the restored voice signal is output as an L-ch decoded voice signal. That is, this decoding process is a speech decoding process for frame compensation.
- the R-ch decoding unit 206 decodes the R-ch encoded data. This decoding result is output as an R-ch decoded speech signal. That is, this decoding process is a normal speech decoding process.
- the R-ch decoding unit 206 decodes the L-ch encoded data. As described above, the L-ch code data is decoded by the R-ch decoding unit 206, so that it is possible to restore the voice signal corresponding to the loss-encoded R-ch encoded data. The restored speech signal is output as an R-ch decoded speech signal. That is, this decoding process is a speech decoding process for frame compensation. As described above, according to the present embodiment, on the transmitting side, a multichannel in which delay processing is performed to delay L ch encoded data from R ch encoded data by one speech frame is performed.
- multi-channel audio data multiplexed in a state delayed by one audio frame from L ch encoded data strength data is separated and separated for each channel.
- one of the L-ch code data and R-ch code data is used to transmit the loss or error in the other data sequence. To compensate. For this reason, at the receiving end, even when loss or error occurs in the voice frame, at least one channel of multi-channel can be correctly received, and frame compensation of another channel can be performed using that channel. It is possible to realize high quality frame compensation.
- FIG. 6 is a block diagram showing a configuration of a speech decoding unit in a speech data receiving apparatus according to Embodiment 3 of the present invention.
- the voice data transmitting apparatus and voice data receiving apparatus according to the present embodiment have the same basic configuration as that described in the first embodiment, and thus the same or corresponding components are used. Are given the same reference numerals and the detailed description thereof is omitted.
- the difference between the present embodiment and the first embodiment is only the internal configuration of the speech decoding unit.
- the speech decoding unit 118 shown in FIG. 6 has a frame compensation unit 120.
- the frame compensation unit 120 includes a switch unit 302, an L-ch frame compensation unit 304, an L-ch decoding unit 306, an R-ch decoding unit 308, an R-ch frame compensation unit 310, and a correlation degree determination unit 312.
- the switch unit 302 receives the input of the loss flag from the voice data loss detection unit 112, the information content indicated by the input loss flag, and the presence or absence of the instruction signal from the correlation degree determination unit 312. , And switches the connection state between the L-ch decoding unit 306 and the R-ch decoding unit 308. Similarly, the connection relationship between the delay unit 116 and the L-ch decoding unit 306 and the R-ch decoding unit 308 is switched.
- the L-ch code data from the separating unit 114 is output to the L-ch decoding unit 306 only, so that L — Ch
- the output end of the L-ch decoder 306 is connected to the output end of
- the output end of the delay unit 116 is R-ch decoding unit so that the R-ch code data from the delay unit 116 is output only to the R-ch decoding unit 308. Connect to the input end of 308.
- connection relationship does not depend on the instruction signal from the correlation degree determination unit 312, but when the loss flag is input, the connection relationship also depends on the instruction signal.
- L-ch of separation section 114 is used. The connection between the output end and the L-ch decoding unit 306 and the R-ch decoding unit 308 is open.
- the L-ch output terminal of the separation unit 114 is connected to both the input terminals of the R-ch decoding unit 308 and the L-ch decoding unit 306 so that the signal is also output.
- L-ch frame compensation unit 304 and R-ch frame compensation unit 310 When a loss flag indicating that L-ch or R-ch coded data has been lost is input to L-ch frame compensation unit 304 and R-ch frame compensation unit 310, an instruction signal is input. When there is no frame, frame compensation is performed using the information of the previous frame of the same channel as in the conventional general method, compensation data (coded data or decoded signal), L-ch decoding unit 306 And R-ch decoding unit 308 respectively.
- L-ch decoding section 306 Decode the L ch encoded data. This decoded result is output as an L ch decoded speech signal. That is, this decoding process is a normal speech decoding process.
- the L-ch decoding unit 306 When the L-ch decoding unit 306 receives the loss flag and the R-ch code data from the delay unit 116 is input, the L-ch decoding unit 306 receives the R-ch coded data. Decrypt. By decoding the R-ch encoded data by the L-ch decoding unit 306 in this way, it is possible to restore the speech signal corresponding to the L-ch encoded data in which the loss occurs. The restored voice signal is output as an L ch decoded voice signal. That is, this decoding process is an audio decoding process for frame compensation.
- L-ch decoding section 306 performs the following decoding processing when compensation data from L-ch frame compensation section 304 is input when a loss flag is input. That is, when encoded data is input as the compensation data, the encoded data is decoded, and when a compensation decoding signal is input, the signal is used as an output signal as it is. Also in this case, it is possible to restore the audio signal corresponding to the Lch code data which has been lost. The restored speech signal is output as an L ch decoded speech signal.
- the R-ch decoding unit 308 decodes the R-ch encoded data. This decoding result is output as an R-ch decoded speech signal. That is, this decoding process is a normal speech decoding process.
- the R-ch decoding unit 308 decodes the L-ch coded data. Do. As described above, the L-ch code data is decoded by the R-ch decoding unit 308, so that it is possible to restore the voice signal corresponding to the loss-caused R-ch encoded data. The restored voice signal is output as an R-ch decoded voice signal. That is, this decoding process is an audio decoding process for frame compensation.
- the R-ch decoding unit 308 performs the following decoding processing. That is, when encoded data is input as the compensation data, the encoded data is decoded, and when a compensation decoding signal is input, the signal is used as an output signal as it is. Also in this case, the speech signal corresponding to the loss R-ch code data is restored. can do. The restored speech signal is output as an R-ch decoded speech signal.
- Correlation degree determination section 312 calculates a correlation degree Cor between the Lch decoded speech signal and the R-ch decoded speech signal using the following equation (1).
- sL ′ (i) and sR ′ (i) are an L-ch decoded speech signal and an R-ch decoded speech signal, respectively. According to the above equation (1), the degree of correlation Cor in the section up to the voice sample value one sample before (that is, immediately before) the voice sample value L samples before the compensation frame is calculated.
- the correlation degree determination unit 312 compares the calculated correlation degree Cor with a predetermined threshold. As a result of the comparison, if the correlation degree Cor is higher than a predetermined threshold value, it is determined that the correlation between the L ch decoded speech signal and the R — ch decoded speech signal is high. Then, when a loss occurs, an instruction signal for instructing to use code data of the mutually different channels is output to the switch unit 302.
- the correlation degree determination unit 312 compares the calculated correlation degree Cor with the above threshold value, if the correlation degree Cor is less than or equal to the threshold value, between the L ch decoded speech signal and the R-ch decoded speech signal It is determined that the correlation of is low. Then, in order to use code data of the same channel when a loss occurs, the instruction signal is not output to the switch unit 302.
- the degree of correlation Cor between the Lch decoded speech signal and the R-ch decoded speech signal is compared with a predetermined threshold, and each other is compared according to the result of the comparison.
- a predetermined threshold In order to decide whether or not to perform frame compensation using the code data of the channel, it is possible to perform compensation based on voice data of the channels of each other only when the correlation between the channels is high. It is possible to prevent degradation of compensation quality by performing frame compensation using voice data of low and low correlation channels. Further, in the present embodiment, since the compensation based on the audio data of the same channel is performed when the correlation is low, the quality of the frame compensation can be maintained continuously.
- the correlation degree determination unit 312 is not required to perform coding at the time of frame compensation.
- the case where it is provided in the frame compensation unit 120 in the second embodiment using a data set is described as an example.
- the configuration of the frame compensation unit 120 provided with the correlation degree determination unit 312 is not limited to this.
- the correlation degree determination unit 312 is provided in the frame compensation unit 120 (Embodiment 1) using decoded speech at the time of frame compensation, the same operation and effect can be realized.
- FIG. 1 A configuration diagram of this case is shown in FIG.
- the operation in this case is mainly different from the operation in the configuration in FIG. 3 in the first embodiment in the operation of the switch unit 126. That is, the connection state in the switch unit 126 is switched according to the loss flag and the result of the instruction signal output from the correlation degree determination unit 312. For example, when a loss flag indicating that L-ch coded data has been lost is input and an instruction signal is input, the compensation signal obtained by L-ch frame compensation unit 304 and R- The decoded signal of ch is input to the L-ch superposition addition unit 130 and superposition addition is performed.
- L-ch frame compensation section 304 uses the information of L-ch up to the previous frame of the lost frame and uses the same frame as the conventional general method.
- the compensation processing is performed.
- the compensation data (coded data or decoded signal) is output to the L ch decoding unit 122, and the L-ch decoding unit 122 outputs the compensation signal of the compensation frame.
- the encoded data is input as the compensation data
- the encoded data is used for decoding
- the compensation decoded signal is input, the signal is used as the output signal as it is.
- the decoded signal up to the previous frame in the L-ch decoding unit 122 or state data is used, or the previous frame of the L-ch superposition addition unit 130 is used. Even if you use the output signal up to,.
- the operation of the R-ch frame compensation unit 310 is similar to that of the L-ch.
- correlation degree determination unit 312 performs calculation processing of correlation degree Cor of a predetermined section, but the correlation degree calculation processing method in correlation degree determination unit 312 is not limited to this! ,
- the correlation degree calculation processing method in correlation degree determination unit 312 is not limited to this! .
- the maximum value Cor-max is compared with a predetermined threshold, and if the maximum value Cor-max exceeds the threshold, it is determined that the correlation between channels is high. By doing this, it is possible to realize the same effects as the above.
- the decoded speech of the other channel used for frame compensation may be shifted after being shifted by the shift amount (ie, the number of speech samples) for obtaining the maximum value Cor_max.
- the shift amount ⁇ max of the speech sample to be the maximum value Cor ⁇ max is calculated by using the following equation (3).
- L-ch frame compensation a signal obtained by shifting the R-ch decoded signal in the positive time direction by a shift amount of -max is used.
- R-ch frame compensation a signal is used in which the L-ch decoded signal is shifted in the negative time direction by _max as the shift amount.
- Cor _ max max ⁇ J s '(--) ⁇ sR' (-i-M-k) ⁇ (k: — M ⁇ M)
- x _ max arg max ⁇ sL '(-i-M) ⁇ sR' (-i-M-k) ⁇ (k:-M to M)
- sL, (i) and sR, (i) are an L-ch decoded speech signal and an R-ch decoded speech signal, respectively.
- the L sample portion in the section from the sound sample value before L + M samples to the sound sample value one sample before (that is, immediately before) is the calculation target section.
- the shift amount of the audio sample for M sample is the calculation target range.
- the shift amount ⁇ -max may be an integer value in the number of voice samples, or may be a decimal value obtained by increasing the resolution between voice sample values.
- the configuration may be such that an amplitude correction value calculation unit that calculates an amplitude correction value for the result is provided internally.
- the speech decoding unit 118 is provided with an amplitude correction unit that corrects the amplitude of the decoding result of the speech data of the other data series by using the calculated amplitude correction value. Then, when performing frame compensation using audio data of another channel, the amplitude of the decoded signal may be corrected using the correction value.
- the arrangement of the amplitude correction value calculation unit is not limited to the inside of the correlation degree determination unit 312 as long as it is inside the speech decoding unit 118.
- a signal obtained by multiplying the decoded signal of R ch by the amplitude correction value g_opt is used.
- R-ch frame compensation a signal obtained by multiplying the inverse of the amplitude correction value lZg_opt by the L-ch decoded signal is used.
- ⁇ max is the shift amount of the voice sample when the degree of correlation obtained by equation (3) is maximized.
- the method of calculating the amplitude correction value is not limited to equation (4).
- A) Let g be the amplitude correction value that minimizes D (g) in equation (5), b ) Determine the shift amounts k and g that minimize D (g, k) in equation (6), and let g at that time be the amplitude correction value, and c) L for a predetermined interval before the compensation frame ch and R—square root of power of decoded signal of ch (or average The ratio of width values) may be used as a correction value.
- Each function block employed in the description of each of the aforementioned embodiments may typically be implemented as an LSI constituted by an integrated circuit. These may be individually made into one chip, or may be made into one chip so as to include some or all of them.
- LSI is used to refer to “IC,” “system LSI,” “super LSI,” and “monolithor LSI” depending on the degree of integration.
- circuit integration is not limited to LSI's, and implementation using dedicated circuitry or general purpose processors is also possible.
- FPGA Field can be programmed after LSI fabrication
- the audio data transmission Z receiving apparatus and the audio data transmission Z reception method of the present invention are useful in an audio communication system or the like in which compensation processing is performed for erroneous audio data and lost audio data.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Detection And Prevention Of Errors In Transmission (AREA)
- Time-Division Multiplex Systems (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
- Circuits Of Receivers In General (AREA)
- Communication Control (AREA)
- Input Circuits Of Receivers And Coupling Of Receivers And Audio Equipment (AREA)
- Stereo-Broadcasting Methods (AREA)
- Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
Abstract
Description
Claims
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US11/628,045 US8209168B2 (en) | 2004-06-02 | 2005-05-20 | Stereo decoder that conceals a lost frame in one channel using data from another channel |
AT05741618T ATE444613T1 (de) | 2004-06-02 | 2005-05-20 | Vorrichtung und verfahren zum empfangen von audiodaten |
JP2006514064A JP4456601B2 (ja) | 2004-06-02 | 2005-05-20 | 音声データ受信装置および音声データ受信方法 |
EP05741618A EP1746751B1 (en) | 2004-06-02 | 2005-05-20 | Audio data receiving apparatus and audio data receiving method |
DE602005016916T DE602005016916D1 (de) | 2004-06-02 | 2005-05-20 | Vorrichtung und verfahren zum empfangen von audiodaten |
CN2005800178145A CN1961511B (zh) | 2004-06-02 | 2005-05-20 | 声音数据接收装置及声音数据接收方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004-165016 | 2004-06-02 | ||
JP2004165016 | 2004-06-02 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2005119950A1 true WO2005119950A1 (ja) | 2005-12-15 |
Family
ID=35463177
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/009252 WO2005119950A1 (ja) | 2004-06-02 | 2005-05-20 | 音声データ送信/受信装置および音声データ送信/受信方法 |
Country Status (7)
Country | Link |
---|---|
US (1) | US8209168B2 (ja) |
EP (1) | EP1746751B1 (ja) |
JP (1) | JP4456601B2 (ja) |
CN (1) | CN1961511B (ja) |
AT (1) | ATE444613T1 (ja) |
DE (1) | DE602005016916D1 (ja) |
WO (1) | WO2005119950A1 (ja) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2048658A1 (en) * | 2006-08-04 | 2009-04-15 | Panasonic Corporation | Stereo audio encoding device, stereo audio decoding device, and method thereof |
WO2009084226A1 (ja) * | 2007-12-28 | 2009-07-09 | Panasonic Corporation | ステレオ音声復号装置、ステレオ音声符号化装置、および消失フレーム補償方法 |
JP2009181050A (ja) * | 2008-01-31 | 2009-08-13 | Panasonic Corp | Ip電話装置およびそのパケットロス補償方法 |
JP2009296497A (ja) * | 2008-06-09 | 2009-12-17 | Fujitsu Telecom Networks Ltd | ステレオ音声信号伝送システム |
JP2010072364A (ja) * | 2008-09-18 | 2010-04-02 | Toshiba Corp | オーディオデータ補間装置及びオーディオデータ補間方法 |
JP2010102042A (ja) * | 2008-10-22 | 2010-05-06 | Ntt Docomo Inc | 音声信号出力装置、音声信号出力方法および音声信号出力プログラム |
JP2014032411A (ja) * | 2013-09-17 | 2014-02-20 | Ntt Docomo Inc | 音声信号出力装置、音声信号出力方法および音声信号出力プログラム |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070280209A1 (en) * | 2006-06-02 | 2007-12-06 | Yahoo! Inc. | Combining selected audio data with a voip stream for communication over a network |
JP5302190B2 (ja) * | 2007-05-24 | 2013-10-02 | パナソニック株式会社 | オーディオ復号装置、オーディオ復号方法、プログラム及び集積回路 |
CN102301748B (zh) * | 2009-05-07 | 2013-08-07 | 华为技术有限公司 | 检测信号延迟的方法、检测装置及编码器 |
CN102810314B (zh) * | 2011-06-02 | 2014-05-07 | 华为终端有限公司 | 音频编码方法及装置、音频解码方法及装置、编解码系统 |
WO2014108738A1 (en) * | 2013-01-08 | 2014-07-17 | Nokia Corporation | Audio signal multi-channel parameter encoder |
RU2648632C2 (ru) | 2014-01-13 | 2018-03-26 | Нокиа Текнолоджиз Ой | Классификатор многоканального звукового сигнала |
CN106328154B (zh) * | 2015-06-30 | 2019-09-17 | 芋头科技(杭州)有限公司 | 一种前端音频处理系统 |
CN106973355B (zh) * | 2016-01-14 | 2019-07-02 | 腾讯科技(深圳)有限公司 | 环绕立体声实现方法和装置 |
US10224045B2 (en) * | 2017-05-11 | 2019-03-05 | Qualcomm Incorporated | Stereo parameters for stereo decoding |
US10043523B1 (en) * | 2017-06-16 | 2018-08-07 | Cypress Semiconductor Corporation | Advanced packet-based sample audio concealment |
US20190005974A1 (en) * | 2017-06-28 | 2019-01-03 | Qualcomm Incorporated | Alignment of bi-directional multi-stream multi-rate i2s audio transmitted between integrated circuits |
CN108777596B (zh) * | 2018-05-30 | 2022-03-08 | 上海惠芽信息技术有限公司 | 一种基于声波的通信方法、通信系统及计算机可读存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10327116A (ja) * | 1997-05-22 | 1998-12-08 | Tadayoshi Kato | タイムダイバシティシステム |
JPH11284582A (ja) * | 1998-03-27 | 1999-10-15 | Toshiba Corp | ディジタル信号伝送システムとその信号伝送装置 |
JPH11313011A (ja) * | 1998-02-27 | 1999-11-09 | Toshiba Corp | ディジタル信号伝送システム及びその信号伝送装置 |
JP2001144733A (ja) * | 1999-11-15 | 2001-05-25 | Nec Corp | 音声伝送装置及び音声伝送方法 |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE3642982A1 (de) * | 1986-12-17 | 1988-06-30 | Thomson Brandt Gmbh | System zur uebertragung |
JP2746033B2 (ja) * | 1992-12-24 | 1998-04-28 | 日本電気株式会社 | 音声復号化装置 |
SE503547C2 (sv) * | 1993-06-11 | 1996-07-01 | Ericsson Telefon Ab L M | Anordning och förfarande för döljande av förlorade ramar |
SE9500858L (sv) * | 1995-03-10 | 1996-09-11 | Ericsson Telefon Ab L M | Anordning och förfarande vid talöverföring och ett telekommunikationssystem omfattande dylik anordning |
JPH08254993A (ja) * | 1995-03-16 | 1996-10-01 | Toshiba Corp | 音声合成装置 |
US5917835A (en) * | 1996-04-12 | 1999-06-29 | Progressive Networks, Inc. | Error mitigation and correction in the delivery of on demand audio |
JP2927242B2 (ja) * | 1996-06-28 | 1999-07-28 | 日本電気株式会社 | 音声符号データの誤り処理装置及び誤り処理方法 |
JP3974712B2 (ja) * | 1998-08-31 | 2007-09-12 | 富士通株式会社 | ディジタル放送用送信・受信再生方法及びディジタル放送用送信・受信再生システム並びにディジタル放送用送信装置及びディジタル放送用受信再生装置 |
GB9820655D0 (en) * | 1998-09-22 | 1998-11-18 | British Telecomm | Packet transmission |
US6327689B1 (en) * | 1999-04-23 | 2001-12-04 | Cirrus Logic, Inc. | ECC scheme for wireless digital audio signal transmission |
US6728924B1 (en) * | 1999-10-21 | 2004-04-27 | Lucent Technologies Inc. | Packet loss control method for real-time multimedia communications |
US6549886B1 (en) * | 1999-11-03 | 2003-04-15 | Nokia Ip Inc. | System for lost packet recovery in voice over internet protocol based on time domain interpolation |
KR100591350B1 (ko) * | 2001-03-06 | 2006-06-19 | 가부시키가이샤 엔.티.티.도코모 | 오디오 데이터 보간장치 및 방법, 오디오 데이터관련 정보작성장치 및 방법, 오디오 데이터 보간 정보 송신장치 및방법, 및 그 프로그램 및 기록 매체 |
JP4016709B2 (ja) * | 2002-04-26 | 2007-12-05 | 日本電気株式会社 | オーディオデータの符号変換伝送方法と符号変換受信方法及び装置とシステムならびにプログラム |
JP4157340B2 (ja) * | 2002-08-27 | 2008-10-01 | 松下電器産業株式会社 | 送信装置、受信装置を含む放送システム、受信装置、及びプログラム。 |
US6985856B2 (en) * | 2002-12-31 | 2006-01-10 | Nokia Corporation | Method and device for compressed-domain packet loss concealment |
US7411985B2 (en) * | 2003-03-21 | 2008-08-12 | Lucent Technologies Inc. | Low-complexity packet loss concealment method for voice-over-IP speech transmission |
-
2005
- 2005-05-20 JP JP2006514064A patent/JP4456601B2/ja active Active
- 2005-05-20 AT AT05741618T patent/ATE444613T1/de not_active IP Right Cessation
- 2005-05-20 US US11/628,045 patent/US8209168B2/en active Active
- 2005-05-20 EP EP05741618A patent/EP1746751B1/en active Active
- 2005-05-20 DE DE602005016916T patent/DE602005016916D1/de active Active
- 2005-05-20 WO PCT/JP2005/009252 patent/WO2005119950A1/ja not_active Application Discontinuation
- 2005-05-20 CN CN2005800178145A patent/CN1961511B/zh active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10327116A (ja) * | 1997-05-22 | 1998-12-08 | Tadayoshi Kato | タイムダイバシティシステム |
JPH11313011A (ja) * | 1998-02-27 | 1999-11-09 | Toshiba Corp | ディジタル信号伝送システム及びその信号伝送装置 |
JPH11284582A (ja) * | 1998-03-27 | 1999-10-15 | Toshiba Corp | ディジタル信号伝送システムとその信号伝送装置 |
JP2001144733A (ja) * | 1999-11-15 | 2001-05-25 | Nec Corp | 音声伝送装置及び音声伝送方法 |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2048658A1 (en) * | 2006-08-04 | 2009-04-15 | Panasonic Corporation | Stereo audio encoding device, stereo audio decoding device, and method thereof |
EP2048658A4 (en) * | 2006-08-04 | 2012-07-11 | Panasonic Corp | STEREOAUDIO CODING DEVICE, STEREOAUDIO DECODING DEVICE AND METHOD THEREFOR |
WO2009084226A1 (ja) * | 2007-12-28 | 2009-07-09 | Panasonic Corporation | ステレオ音声復号装置、ステレオ音声符号化装置、および消失フレーム補償方法 |
US8359196B2 (en) | 2007-12-28 | 2013-01-22 | Panasonic Corporation | Stereo sound decoding apparatus, stereo sound encoding apparatus and lost-frame compensating method |
JP5153791B2 (ja) * | 2007-12-28 | 2013-02-27 | パナソニック株式会社 | ステレオ音声復号装置、ステレオ音声符号化装置、および消失フレーム補償方法 |
JP2009181050A (ja) * | 2008-01-31 | 2009-08-13 | Panasonic Corp | Ip電話装置およびそのパケットロス補償方法 |
JP2009296497A (ja) * | 2008-06-09 | 2009-12-17 | Fujitsu Telecom Networks Ltd | ステレオ音声信号伝送システム |
JP2010072364A (ja) * | 2008-09-18 | 2010-04-02 | Toshiba Corp | オーディオデータ補間装置及びオーディオデータ補間方法 |
JP2010102042A (ja) * | 2008-10-22 | 2010-05-06 | Ntt Docomo Inc | 音声信号出力装置、音声信号出力方法および音声信号出力プログラム |
JP2014032411A (ja) * | 2013-09-17 | 2014-02-20 | Ntt Docomo Inc | 音声信号出力装置、音声信号出力方法および音声信号出力プログラム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2005119950A1 (ja) | 2008-04-03 |
ATE444613T1 (de) | 2009-10-15 |
CN1961511B (zh) | 2010-06-09 |
DE602005016916D1 (de) | 2009-11-12 |
US8209168B2 (en) | 2012-06-26 |
EP1746751A4 (en) | 2007-09-12 |
EP1746751B1 (en) | 2009-09-30 |
US20080065372A1 (en) | 2008-03-13 |
CN1961511A (zh) | 2007-05-09 |
JP4456601B2 (ja) | 2010-04-28 |
EP1746751A1 (en) | 2007-01-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4456601B2 (ja) | 音声データ受信装置および音声データ受信方法 | |
US6985856B2 (en) | Method and device for compressed-domain packet loss concealment | |
CN108701464B (zh) | 多个音频信号的编码 | |
US7797162B2 (en) | Audio encoding device and audio encoding method | |
US20070299669A1 (en) | Audio Encoding Apparatus, Audio Decoding Apparatus, Communication Apparatus and Audio Encoding Method | |
US8359196B2 (en) | Stereo sound decoding apparatus, stereo sound encoding apparatus and lost-frame compensating method | |
KR20070092240A (ko) | 음성 부호화 장치 및 음성 부호화 방법 | |
KR20190137181A (ko) | 다수의 오디오 신호들의 인코딩 | |
US8768691B2 (en) | Sound encoding device and sound encoding method | |
JP2004509367A (ja) | 複数チャネル信号の符号化及び復号化 | |
US8787490B2 (en) | Transmitting data in a communication system | |
KR20230035430A (ko) | 다수의 오디오 신호들의 디코딩 | |
CN110168637B (zh) | 多个音频信号的译码 | |
CN110100280B (zh) | 信道间相位差参数的修改 | |
JP2005094356A (ja) | 音声信号伝送システム、及び音声信号伝送方法 | |
JP2004102074A (ja) | 音声符号化装置、音声復号化装置、音声信号伝送方法及びプログラム | |
US10242683B2 (en) | Optimized mixing of audio streams encoded by sub-band encoding | |
JP4065383B2 (ja) | 音声信号送信装置、音声信号受信装置及び音声信号伝送システム | |
JP2002196795A (ja) | 音声復号装置及び音声符号化・復号装置 | |
Rein et al. | Voice quality evaluation for wireless transmission with ROHC (extended version) | |
Hellerud et al. | Robust Transmission of Lossless Audio with Low Delay over IP Networks | |
JP2003092634A (ja) | 音声多重化装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006514064 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 11628045 Country of ref document: US Ref document number: 2005741618 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 200580017814.5 Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: DE |
|
WWP | Wipo information: published in national office |
Ref document number: 2005741618 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 11628045 Country of ref document: US |