USRE35057E - Speech coding using sparse vector codebook and cyclic shift techniques - Google Patents

Speech coding using sparse vector codebook and cyclic shift techniques Download PDF

Info

Publication number
USRE35057E
USRE35057E US08/012,947 US1294793A USRE35057E US RE35057 E USRE35057 E US RE35057E US 1294793 A US1294793 A US 1294793A US RE35057 E USRE35057 E US RE35057E
Authority
US
United States
Prior art keywords
excitation
frame
iaddend
iadd
frames
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US08/012,947
Inventor
Daniel K. Freeman
Ivan Boyd
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
British Telecommunications PLC
Original Assignee
British Telecommunications PLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from GB878720389A external-priority patent/GB8720389D0/en
Priority claimed from GB878721667A external-priority patent/GB8721667D0/en
Application filed by British Telecommunications PLC filed Critical British Telecommunications PLC
Priority to US08/012,947 priority Critical patent/USRE35057E/en
Application granted granted Critical
Publication of USRE35057E publication Critical patent/USRE35057E/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/08Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters
    • G10L19/10Determination or coding of the excitation function; Determination or coding of the long-term prediction parameters the excitation function being a multipulse excitation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L2019/0001Codebooks
    • G10L2019/0007Codebook element generation
    • G10L2019/0008Algebraic codebooks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L2019/0001Codebooks
    • G10L2019/0013Codebook search algorithms

Definitions

  • a common technique for speech coding is the so-called LPC coding in which at a coder, an input speech signal is divided into time intervals and each interval is analysed to determine the parameters of a synthesis filter whose response is representative of the frequency spectrum of the signal during that interval.
  • the parameters are transmitted to a decoder where theiy periodically update the parameters of a synthesis filter which, when fed with a suitable excitation signal, produces a synthetic speech output which approximates the original input.
  • the coder has also to transmit to the decoder information as to the nature of the excitation which is to be employed.
  • a number of options have been proposed for achieving this, falling into two main categories, viz.
  • Residual excited linear predictive coding where the input signal is passed through a filter which is the inverse of the synthesis filter to produce a residual signal which can be quantised and sent (possibly after filtering) to be used as the excitation, or may be analysed, e.g. to obtain voicing and pithc parameters for transmission to an excitation generator in the decoder.
  • RELP Residual excited linear predictive coding
  • codebooks may compiled using random sequence generation; however another variant is the so-called ⁇ sparse vector ⁇ codebook in which a frame contains only a small number of pulses (e.g. 4 or 5 pulses out of 32 possible positions with a frame).
  • a CELP coder may typically have a 1024-entry codebook.
  • FIGS. 1(a-c) illustrate three typical members of a set of cyclically related excitations to be used in the invention
  • FIG. 1(d) shows a single excitation representing the excitations shown in FIGS. 1(a-c);
  • FIG. 2 is a block diagram of one form of speech coder according to the invention.
  • FIG. 3 is a block diagram of a suitable decoder.
  • the coder now to be described is similar to CELP in that it employs a sparse vector codebook which is, however much smaller than that conventionally used; perhaps 32 or 64 entries.
  • Each entry represents one excitation from which can be derived other members of a set of excitations which differ from the one excitation--and from each other--only by a cyclic shift.
  • Three such members of the set are shown in FIGS. 1a, 1b and 1c for a 32 position frame with five pulses, where it is seen that 1b can be formed from 1a by cyclically shifting the entry to the left, and likewise 1c from 1a.
  • the amount of shift is indicated in the figure by a double-headed arrow.
  • Cyclic shifting means that pulses shifted out of the left-hand end wrap around and reenter from the right.
  • the entry representing the set is stored with the largest pulse in position 1, i.e. as shown in FIG. 1d.
  • the magnitude of the largest pulse need not be stored if the others are normalised by it.
  • the excitation selected can be represented by a 5-bit codeword identifying the entry and a further 5 bits giving the number of shifts from the stored position (if all 32 possible shifts are allowed).
  • FIG. 2 is a block diagram of a speech coder.
  • Speech signals received at an input 1 are converted into samples by a sampler 2 and then into digital form in an analogue-to-digital converter 3.
  • An analysis unit 4 computes, for each successive group of samples, the coefficients of a synthesis filter having a response corresponding to the spectral content of the speech. Derivation of LPC coefficients is well known and will not be described further here.
  • the coefficients are supplied to an output multiplexer 5, andd also to a local synthesis filter 6.
  • the filter update rate may typically be once every 20 ms.
  • the coder has also a codebook store 7 containing the thirty-two codebook entries discussed above.
  • the manner in which the entries are stored is not material to the present invention but it is assumed that each entry (for a five pulse excitation in a 32 sample period frame) contains the positions within the frame and the amplitudes of the four pulses after the first.
  • This information when read from the store is supplied to an excitation generator 8 which produces an actual excitation frame--i.e., 32 values (of which 27 are zero, of course). Its output is supplied via a controllable shifting unit 9 to the input of the synthesis filter 6.
  • the filter output is compared by a subtractor 10 with the input speech samples supplied via a buffer 11 (so that a number of comparisons can be made between one 32-sample speech frame and different filtered excitations).
  • multipulse coding In order to ascertain the appropriate shift value, certain techniques are borrowed from multipulse coding.
  • multipulse coding a ccommon method of deriving the pulse positions and amplitudes is an iterative one, in which one pulse is calculated which minimises the error between the synthetic and actual speech. A further pulse is then found which, in combination with the first, minimises the error and so on. Analysis of the statistics of MP-LPC pulses show that the first pulse to be derived usually has the largest amplitude.
  • This embodiment of the invention makes use of this by carrying out a multipulse search to find the location of this first pulse only. Any of the known methods for this may be employed, for example that described in B. S. Atal and J. R. Remde, ⁇ A New Model of LPC Excitation for producing Natural Sounding Speech at Low Bit Rates, ⁇ Proc. IEEE Int. Conf. ASSP, Paris, 1982, p. 614.
  • a search unit 12 is shown in FIG. 2 for this purpose its output feeds the shifter 9 to determine the rotational shift applied to the excitation generated by the generator 8. Effectively this selects, from 1024 excitations, allowed by the codebook, a particular class of excitations, namely those with the largest pulse occupying the particular position determined by the search unit 13.
  • the output of the subtractor 10 feeds a control unit 13 which also supplies addresses to the store 7 and sift values to the shifting unit 9.
  • the purpose of the control unit is to ascertain which of the 32 possible excitations represented by the selected class gives the smallest subtractor output (usually the mean square value of the differences, over a frame).
  • the finally determined entry and shift are output in the form of a codeword C and shift value S to the output multiplexer 5.
  • the entry determination by the control unit for a given frame of speech available at the output of the buffer 11 is as follows:
  • the above process may also include excitations which are shifted a few positions before and after the position found by the search.
  • the generation of the codebook remains to be mentioned. This can be generated by Gaussian noise techniques, in the manner already proposed in "Scholastic Coding of Speech Signals at very low Bit Rates". B. S. Atal & M. R. Schroeder, Proc IEEE Int Conf on Communications, 1984, pp 1610-1613.
  • a further advantage can be gained however by generating the codebook by statistical anaylsis of the results produced by a multipulse coder. This can remove the approximation involved in the assumption that the first pulse derived by the "multipusle search ⁇ is the largest, since the codebook entries can then be stored with the first obtained pulse in a standard position, and shifted such that this this pulse is brought to the position derived by the unit.
  • DSP digital signal processing
  • the ⁇ multipulse search ⁇ option has been described in the context of shifted codebook entries, it can also be applied to other situations where the allowed excitations can be divided into classes within which all the excitations have the largest, or most significant, pulse in a particular position within the frame. The position of the derived pulse is then used to select the appropriate class and only the codebook entries in that class need to be tested.
  • FIG. 3 shows a decoder for reproducing signals encoded by the apparatus of FIG. 2.
  • An input 30 supplies a demultiplexer 31 which (a) supplies filter coefficients to a synthesis filter 32; (b) supplies codewords to the address input of a codebook store 33; (c) supplies shift values to a shifter 34 which conveys the output of an exccitation generator 35 connected to the store 33 to the input of the synthesis filter 32. Speech output from the filter 32 is supplied via a digital-to-analogue converter 36 to an output 37.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

Speech is analyzed to derive the parameters of a synthesis filter and the parameters of a suitable excitation which is selected from a codebook of excitation frames. The selection of the codebook entry is facilitated by determining a single-pulse excitation (e.g., using conventional multipulse excitation techniques), and using the position of this pulse to narrow the codebook search.

Description

BACKGROUND AND SUMMARY OF THE INVENTION
A common technique for speech coding is the so-called LPC coding in which at a coder, an input speech signal is divided into time intervals and each interval is analysed to determine the parameters of a synthesis filter whose response is representative of the frequency spectrum of the signal during that interval. The parameters are transmitted to a decoder where theiy periodically update the parameters of a synthesis filter which, when fed with a suitable excitation signal, produces a synthetic speech output which approximates the original input.
Clearly the coder has also to transmit to the decoder information as to the nature of the excitation which is to be employed. A number of options have been proposed for achieving this, falling into two main categories, viz.
(i) Residual excited linear predictive coding (RELP) where the input signal is passed through a filter which is the inverse of the synthesis filter to produce a residual signal which can be quantised and sent (possibly after filtering) to be used as the excitation, or may be analysed, e.g. to obtain voicing and pithc parameters for transmission to an excitation generator in the decoder.
(ii) Analysis by synthesis methods in which an excitation is derived such that, when passed through the synthesis filter, the difference between the output obtained and the input speech is minimised. In this category there are two distinct approaches: One is multipulse excitation (MP-LPC) in which a time frame corresponding to a number of speech samples contains a, somewhat smaller, limited number of excitation pulses whose amplitudes and positions are coded. The other approach is stochastic coding or coded excited linear prediction (CELP). The coder and decoder each have a stored list of standard frames of excitations. For each frame of speech, that one of the codebook entries which, when passed through the synthesis filter, produces synthetic speech closet to the actual speech is identified and a codeword assigned to it is sent to the decoder which can then retrieve the same entry from its stored list. Such codebooks may compiled using random sequence generation; however another variant is the so-called `sparse vector ` codebook in which a frame contains only a small number of pulses (e.g. 4 or 5 pulses out of 32 possible positions with a frame). A CELP coder may typically have a 1024-entry codebook.
The present invention is defined in the appended claims.
Some embodiments of the invention will now be described, by way of example, with reference to the accompanying drawings, in which:
BRIEF DESCRIPTION OF THE DRAWING
FIGS. 1(a-c) illustrate three typical members of a set of cyclically related excitations to be used in the invention;
FIG. 1(d) shows a single excitation representing the excitations shown in FIGS. 1(a-c);
FIG. 2 is a block diagram of one form of speech coder according to the invention; and
FIG. 3 is a block diagram of a suitable decoder.
DESCRIPTION OF THE PREFERRED EMBODIMENTS
It will be appreciated from the introduction that multipulse coders and sparse vector CELP coders have in common the features that the exciation employed is in both cases a frame containing a number of pulses significantly smaller than the number of allowable positions within the frame.
The coder now to be described is similar to CELP in that it employs a sparse vector codebook which is, however much smaller than that conventionally used; perhaps 32 or 64 entries. Each entry represents one excitation from which can be derived other members of a set of excitations which differ from the one excitation--and from each other--only by a cyclic shift. Three such members of the set are shown in FIGS. 1a, 1b and 1c for a 32 position frame with five pulses, where it is seen that 1b can be formed from 1a by cyclically shifting the entry to the left, and likewise 1c from 1a. The amount of shift is indicated in the figure by a double-headed arrow. Cyclic shifting means that pulses shifted out of the left-hand end wrap around and reenter from the right. The entry representing the set is stored with the largest pulse in position 1, i.e. as shown in FIG. 1d. The magnitude of the largest pulse need not be stored if the others are normalised by it.
If the number of codebook entries is 32, then the excitation selected can be represented by a 5-bit codeword identifying the entry and a further 5 bits giving the number of shifts from the stored position (if all 32 possible shifts are allowed).
FIG. 2 is a block diagram of a speech coder. Speech signals received at an input 1 are converted into samples by a sampler 2 and then into digital form in an analogue-to-digital converter 3. An analysis unit 4 computes, for each successive group of samples, the coefficients of a synthesis filter having a response corresponding to the spectral content of the speech. Derivation of LPC coefficients is well known and will not be described further here. The coefficients are supplied to an output multiplexer 5, andd also to a local synthesis filter 6. The filter update rate may typically be once every 20 ms.
The coder has also a codebook store 7 containing the thirty-two codebook entries discussed above. The manner in which the entries are stored is not material to the present invention but it is assumed that each entry (for a five pulse excitation in a 32 sample period frame) contains the positions within the frame and the amplitudes of the four pulses after the first. This information, when read from the store is supplied to an excitation generator 8 which produces an actual excitation frame--i.e., 32 values (of which 27 are zero, of course). Its output is supplied via a controllable shifting unit 9 to the input of the synthesis filter 6. The filter output is compared by a subtractor 10 with the input speech samples supplied via a buffer 11 (so that a number of comparisons can be made between one 32-sample speech frame and different filtered excitations).
In order to ascertain the appropriate shift value, certain techniques are borrowed from multipulse coding. In multipulse coding, a ccommon method of deriving the pulse positions and amplitudes is an iterative one, in which one pulse is calculated which minimises the error between the synthetic and actual speech. A further pulse is then found which, in combination with the first, minimises the error and so on. Analysis of the statistics of MP-LPC pulses show that the first pulse to be derived usually has the largest amplitude.
This embodiment of the invention makes use of this by carrying out a multipulse search to find the location of this first pulse only. Any of the known methods for this may be employed, for example that described in B. S. Atal and J. R. Remde, `A New Model of LPC Excitation for producing Natural Sounding Speech at Low Bit Rates,` Proc. IEEE Int. Conf. ASSP, Paris, 1982, p. 614.
A search unit 12 is shown in FIG. 2 for this purpose its output feeds the shifter 9 to determine the rotational shift applied to the excitation generated by the generator 8. Effectively this selects, from 1024 excitations, allowed by the codebook, a particular class of excitations, namely those with the largest pulse occupying the particular position determined by the search unit 13.
The output of the subtractor 10 feeds a control unit 13 which also supplies addresses to the store 7 and sift values to the shifting unit 9. The purpose of the control unit is to ascertain which of the 32 possible excitations represented by the selected class gives the smallest subtractor output (usually the mean square value of the differences, over a frame). The finally determined entry and shift are output in the form of a codeword C and shift value S to the output multiplexer 5.
The entry determination by the control unit for a given frame of speech available at the output of the buffer 11 is as follows:
(i) apply successive code (codebook addresses ) to the store 7
(ii) apply to each codebook entry a shift such as to move the largest pulse to the position indicated by the `multipulse` search.
(iii) monitor the output of the subtractor 10 for all 32 entries to ascertain which gives rise to the lowest mean square difference.
(iv) output the codeword and shift value to the multiplexer.
Compared with a conventional CELP coder using a 1024 entry codebook, there is a small reduction in the signal-to-noise ratio obtained due to the constraints placed on the excitation (i.e. that they fall into 32 mutually shiftable classes). However there is a reduction in the codebook size and hence the storage requirement for the store 7. Moreover, the amount of computation to be carried out by the control unit 13 is significantly reduced since only 32 tests rather than 1024 need to be carried out.
To allow for the sub-optimal selection, inherent in the `multipulse search`, the above process may also include excitations which are shifted a few positions before and after the position found by the search.
This could be achieved by the control unit adding/subtracting appropriate values from the shift value suplied to the shifting unit 9, as indicated by the dotted line connection. However, since the filtered output of a time shifted version of a given excitation is a time shifted version of the filter's response to the given excitation, these shifts could instead be performed by a second shifter 14 placed after the synthesis filter 6. Once wrap-around occurs, however, the result is no longer correct: this problem may be accommodated by (a) not performing shifts which cause wrap around (b) performing the shift but allowing pulses to be lost rather than wrapped around (and informing the decoder) or (c) permitting wraparound but performing a correction to account for the error.
The generation of the codebook remains to be mentioned. This can be generated by Gaussian noise techniques, in the manner already proposed in "Scholastic Coding of Speech Signals at very low Bit Rates". B. S. Atal & M. R. Schroeder, Proc IEEE Int Conf on Communications, 1984, pp 1610-1613. A further advantage can be gained however by generating the codebook by statistical anaylsis of the results produced by a multipulse coder. This can remove the approximation involved in the assumption that the first pulse derived by the "multipusle search` is the largest, since the codebook entries can then be stored with the first obtained pulse in a standard position, and shifted such that this this pulse is brought to the position derived by the unit.
Although the various function elements shown in FIG. 2 are indicated separately, in practice some or all of them might be performed by the same hardware. One of the commercially available digital signal processing (DSP) integrated circuits, suitably programmed, might be employed, for example.
Although the `multipulse search` option has been described in the context of shifted codebook entries, it can also be applied to other situations where the allowed excitations can be divided into classes within which all the excitations have the largest, or most significant, pulse in a particular position within the frame. The position of the derived pulse is then used to select the appropriate class and only the codebook entries in that class need to be tested.
FIG. 3 shows a decoder for reproducing signals encoded by the apparatus of FIG. 2.
An input 30 supplies a demultiplexer 31 which (a) supplies filter coefficients to a synthesis filter 32; (b) supplies codewords to the address input of a codebook store 33; (c) supplies shift values to a shifter 34 which conveys the output of an exccitation generator 35 connected to the store 33 to the input of the synthesis filter 32. Speech output from the filter 32 is supplied via a digital-to-analogue converter 36 to an output 37.

Claims (10)

We claim:
1. A speech coder comprising:
means for generating filter information from frames of input speech signals, said means for generating filter information defining successive representations of a synthesis filter response, and outputting said filter information; and
means for generating frames of excitation information for successive frames of said input speech signals, .[.eahc.]. .Iadd.each .Iaddend.of said excitation frames including a series of pulses, said means for generating frames receiving said input speech frames and said filter information and comprising:
(a) a store of data defining a plurality of .[.representative.]. excitation frames.[., each.]. having a plurality of pulses and .[.each representative frame.]. representing .Iadd.plural classes .Iaddend.of .[.member.]. excitation frames;
(b) means for selecting one of said .[.member.]. excitation frames, said selected excitation frame when applied to the input of a filter having said filter information producing a frame of synthetic speech resembling said input speech, and outputting data identifying said selected excitation frame, said means for selecting including:
(i) means for identifying the position within .[.said input speech.]. .Iadd.an excitation .Iaddend.frame of a single pulse which meets a preselected criterion,
(ii) selecting one of said stored .[.representative.]. .Iadd.classes of .Iaddend.excitation frames depending on the position of said identified single pulse, and
(iii) determining which of said .[.member.]. excitation frames within the .Iadd.selected .Iaddend.class .[.of said selected representative excitation frame.]. that .Iadd.best .Iaddend.matches said input speech frame .Iadd.when used in conjunction with said filter information.Iaddend..
2. A speech coder according to claim 1 in which .[.each.]. .Iadd.said data in the store defines one .Iaddend.of said classes .Iadd.which class .Iaddend.comprises a plurality of member excitation frames each .[.member.]. .Iadd.other class of excitation frames .Iaddend.being a rotationally shifted version of .[.any other member of the same class.]. .Iadd.the stored class.Iaddend..
3. A speech coder according to claim 2 in which said store contains a list of .[.one.]. .Iadd.all .Iaddend.representative .[.member of.]. .Iadd.members of .Iaddend..[.each.]. .Iadd.one .Iaddend.of said classes, and further comprising shifting means controllable to generate other .[.class members.]. .Iadd.classes .Iaddend.from said .Iadd.stored .Iaddend.representative .[.member.]. .Iadd.members.Iaddend..
4. A speech coder according to claim 3 in which said generating means further comprises shifting means for shifting each of said representative members by an amount corresponding to said identified pulse position.
5. A speech coder according to claim 4 in which said shifting means brings the largest pulse of each of said representative members into the same position within the frame as is said single pulse.
6. A speech coder according to claim 4 in which said stored representative excitation frames are generated by a training sequence comprising identification of the position within the frame of a single, first, pulse meeting said predetermined criterion followed by determination of further pulses, and said amount of shift applied by said shifting means is that shift which brings said first pulse of said representative excitation frame into the same position within the frame as said determined single pulse.
7. A speech coder according to claim 3 in which each of said classes comprises .[.a member which has.]. .Iadd.members which have .Iaddend.been shifted by an amount corresponding to said identified single pulse, and members shifted by amounts which are small variations, relative to the frame size of said amount corresponding to said identified single pulse .Iadd.position.Iaddend..
8. A speech coder comprising:
means for generating, from input speech signals, filter information defining successive representations of a synthesis filter response, and outputting said filter information; and
means for generating, form said input speech signals and filter information excitation information for successive frames of said speech signals, comprising:
(a) a store of data defining a plurality of representative excitation frames each consisting of a plurality of pulses;
(b) means for selecting one of said representative excitation frames and the amount of rotational shift to be applied to said selected frame which would when applied to the input of a filter having said filter information produce a frame of synthetic speech resembling said input speech signals, and outputting data identifying said selected frame and said amount of rotational shift;
said means for selecting comprising means for:
(i) determining the position within said .[.framed speech signal.]. .Iadd.excitation frame .Iaddend.of a single pulse which meets a preselected criterion, and
(ii) selecting the one of said excitation frames which when rotationally shifted by an amount derived from the determined position of said single pulse .Iadd.is used to generate a speech signal which .Iaddend.most nearly matches said .[.frame.]. .Iadd.input .Iaddend.speech signal.
9. A speech coder including:
filter means for generating synthesis filter response representations from an input speech signal; and
excitation means for generating excitation frames from said input speech signal and said synthesis filter response representations, said excitation means comprising:
means for identifying the frame position of a single pulse within said .[.input speech signal.]. .Iadd.excitation frame .Iaddend.which meets a preselected criterion;
a codebook store containing a list of standard excitation frames;
.[.means for selecting one of said standard excitation frames using the frame position of said identified pulse;.].
means for cyclically shifting said standard excitation frames to align said standard frame with said identified pulse; and
comparator means for selecting the one of said standard excitation frames which, when aligned and applied to an input filter having said filter response representations, produces synthetic speech most nearly resembling said input speech signal.
10. A method for speech coding using a speech coder having a codebook store containing a list of standard excitation frames .[.each being.]. representative of a class of excitation frames, said method comprising the steps of:
(a) framing a digital input speech signal;
(b) forming filter information defining a synthesis filter response indicative of the framed digital input speech signal;
(c) identifying the position of a pulse in .[.the framed input speech signal.]. .Iadd.an excitation frame .Iaddend.which satisfies a preselected criterion;
(d) .[.selecting a standard excitation frame from the codebook depending on the pulse frame position identified in step (c);
(e).]. determining the amount of shift to apply to the .[.selected.]. standard excitation .[.frame.]. .Iadd.frames .Iaddend.to match the .[.framed input speech signal; and.]. .Iadd.pulse position identified in step (c);
(e) selecting a shifted standard excitation frame from step (d) to match the framed input speech signal when used with said filter information to synthesis a speech signal; and .Iaddend.
(f) outputting data indicative of the selected standard excitation frame and the determined amount of shift.
US08/012,947 1987-08-28 1993-02-03 Speech coding using sparse vector codebook and cyclic shift techniques Expired - Lifetime USRE35057E (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US08/012,947 USRE35057E (en) 1987-08-28 1993-02-03 Speech coding using sparse vector codebook and cyclic shift techniques

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
GB8720389 1987-08-28
GB878720389A GB8720389D0 (en) 1987-08-28 1987-08-28 Speech coding
GB8721667 1987-09-15
GB878721667A GB8721667D0 (en) 1987-09-15 1987-09-15 Speech coding
US35835088A 1988-08-26 1988-08-26
US08/012,947 USRE35057E (en) 1987-08-28 1993-02-03 Speech coding using sparse vector codebook and cyclic shift techniques

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US35835088A Reissue 1987-08-28 1988-08-26

Publications (1)

Publication Number Publication Date
USRE35057E true USRE35057E (en) 1995-10-10

Family

ID=27449976

Family Applications (1)

Application Number Title Priority Date Filing Date
US08/012,947 Expired - Lifetime USRE35057E (en) 1987-08-28 1993-02-03 Speech coding using sparse vector codebook and cyclic shift techniques

Country Status (1)

Country Link
US (1) USRE35057E (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6502068B1 (en) * 1999-09-17 2002-12-31 Nec Corporation Multipulse search processing method and speech coding apparatus

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4038495A (en) * 1975-11-14 1977-07-26 Rockwell International Corporation Speech analyzer/synthesizer using recursive filters
EP0195487A1 (en) * 1985-03-22 1986-09-24 Koninklijke Philips Electronics N.V. Multi-pulse excitation linear-predictive speech coder
USRE32580E (en) * 1981-12-01 1988-01-19 American Telephone And Telegraph Company, At&T Bell Laboratories Digital speech coder
US4776015A (en) * 1984-12-05 1988-10-04 Hitachi, Ltd. Speech analysis-synthesis apparatus and method
US4809330A (en) * 1984-04-23 1989-02-28 Nec Corporation Encoder capable of removing interaction between adjacent frames
US4821324A (en) * 1984-12-24 1989-04-11 Nec Corporation Low bit-rate pattern encoding and decoding capable of reducing an information transmission rate
US4847905A (en) * 1985-03-22 1989-07-11 Alcatel Method of encoding speech signals using a multipulse excitation signal having amplitude-corrected pulses
US4864621A (en) * 1986-09-11 1989-09-05 British Telecommunications Public Limited Company Method of speech coding

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4038495A (en) * 1975-11-14 1977-07-26 Rockwell International Corporation Speech analyzer/synthesizer using recursive filters
USRE32580E (en) * 1981-12-01 1988-01-19 American Telephone And Telegraph Company, At&T Bell Laboratories Digital speech coder
US4809330A (en) * 1984-04-23 1989-02-28 Nec Corporation Encoder capable of removing interaction between adjacent frames
US4776015A (en) * 1984-12-05 1988-10-04 Hitachi, Ltd. Speech analysis-synthesis apparatus and method
US4821324A (en) * 1984-12-24 1989-04-11 Nec Corporation Low bit-rate pattern encoding and decoding capable of reducing an information transmission rate
EP0195487A1 (en) * 1985-03-22 1986-09-24 Koninklijke Philips Electronics N.V. Multi-pulse excitation linear-predictive speech coder
US4847905A (en) * 1985-03-22 1989-07-11 Alcatel Method of encoding speech signals using a multipulse excitation signal having amplitude-corrected pulses
US4932061A (en) * 1985-03-22 1990-06-05 U.S. Philips Corporation Multi-pulse excitation linear-predictive speech coder
US4864621A (en) * 1986-09-11 1989-09-05 British Telecommunications Public Limited Company Method of speech coding

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
D. Lin, "Speech Coding Using Efficient Pseudo-Stochastic Block Codes," ICASSP 87 (International Conference on Acoustics, Speech and Signal Processing, 4-6 Apr. 1987, Dallas, Tex., U.S.), vol. 3, IEEE (New York, U.S.), pp. 1354-1357.
D. Lin, Speech Coding Using Efficient Pseudo Stochastic Block Codes, ICASSP 87 (International Conference on Acoustics, Speech and Signal Processing, 4 6 Apr. 1987, Dallas, Tex., U.S.), vol. 3, IEEE (New York, U.S.), pp. 1354 1357. *
Juang, "Design and Performance of Trellis Vector Quantizers for Speech Signals," IEEE Trans. on ASSP, vol. 36, No. 9, Sep. 1988, pp. 1423-1431.
Juang, Design and Performance of Trellis Vector Quantizers for Speech Signals, IEEE Trans. on ASSP, vol. 36, No. 9, Sep. 1988, pp. 1423 1431. *
Kroon et al., "Quantization Procedures for the Excitation in CELP Coders," IEEE Int'l Conf. on Acoustics, Speech & Signal Processing, Apr. 6-9, 1987, pp. 1649-1652.
Kroon et al., Quantization Procedures for the Excitation in CELP Coders, IEEE Int l Conf. on Acoustics, Speech & Signal Processing, Apr. 6 9, 1987, pp. 1649 1652. *
L. A. Hernandez Gomez et al., On the Behavior of Reduced Complexity Code Excited Linear Prediction, (CELP), ICASSP 86 (IEEE IECEJ ASJ International Conf. on Acoustics, Speech and Signal Processing, 7 11 Apr. 1986, Tokyo, JP), vol. 1, IEEE (New York, US), pp. 469 472. *
L. A. Hernandez-Gomez et al., "On the Behavior of Reduced Complexity Code-Excited Linear Prediction, (CELP)," ICASSP 86 (IEEE IECEJ-ASJ International Conf. on Acoustics, Speech and Signal Processing, 7-11 Apr. 1986, Tokyo, JP), vol. 1, IEEE (New York, US), pp. 469-472.

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6502068B1 (en) * 1999-09-17 2002-12-31 Nec Corporation Multipulse search processing method and speech coding apparatus

Similar Documents

Publication Publication Date Title
US4991214A (en) Speech coding using sparse vector codebook and cyclic shift techniques
US5138661A (en) Linear predictive codeword excited speech synthesizer
US5602961A (en) Method and apparatus for speech compression using multi-mode code excited linear predictive coding
US7363220B2 (en) Method for speech coding, method for speech decoding and their apparatuses
EP0504627B1 (en) Speech parameter coding method and apparatus
CA1336455C (en) Code excited linear predictive vocoder using virtual searching
DE60126149T2 (en) METHOD, DEVICE AND PROGRAM FOR CODING AND DECODING AN ACOUSTIC PARAMETER AND METHOD, DEVICE AND PROGRAM FOR CODING AND DECODING SOUNDS
EP0766232B1 (en) Speech coding apparatus
EP0833305A2 (en) Low bit-rate pitch lag coder
EP0232456A1 (en) Digital speech processor using arbitrary excitation coding
US7146311B1 (en) CELP encoding/decoding method and apparatus
EP0397628B1 (en) Excitation pulse positioning method in a linear predictive speech coder
US5243685A (en) Method and device for the coding of predictive filters for very low bit rate vocoders
JP3137176B2 (en) Audio coding device
EP0578436B1 (en) Selective application of speech coding techniques
US6199040B1 (en) System and method for communicating a perceptually encoded speech spectrum signal
US5797119A (en) Comb filter speech coding with preselected excitation code vectors
EP0500094A2 (en) Speech signal coding and decoding system with transmission of allowed pitch range information
US5839098A (en) Speech coder methods and systems
USRE35057E (en) Speech coding using sparse vector codebook and cyclic shift techniques
EP0745972B1 (en) Method of and apparatus for coding speech signal
JPH08292797A (en) Voice encoding device
GB2199215A (en) A stochastic coder
Akamine et al. CELP coding with an adaptive density pulse excitation model
JP3471889B2 (en) Audio encoding method and apparatus

Legal Events

Date Code Title Description
FPAY Fee payment

Year of fee payment: 8

FPAY Fee payment

Year of fee payment: 12