EP0071716A2 - Allophone vocoder - Google Patents

Allophone vocoder Download PDF

Info

Publication number
EP0071716A2
EP0071716A2 EP82105168A EP82105168A EP0071716A2 EP 0071716 A2 EP0071716 A2 EP 0071716A2 EP 82105168 A EP82105168 A EP 82105168A EP 82105168 A EP82105168 A EP 82105168A EP 0071716 A2 EP0071716 A2 EP 0071716A2
Authority
EP
European Patent Office
Prior art keywords
phoneme
allophone
indicia
sequence
signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
EP82105168A
Other languages
German (de)
French (fr)
Other versions
EP0071716B1 (en
EP0071716A3 (en
Inventor
Granville E. Ott
Kun-Shan Lin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Texas Instruments Inc
Original Assignee
Texas Instruments Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US06/289,603 external-priority patent/US4424415A/en
Priority claimed from US06/289,604 external-priority patent/US4661915A/en
Application filed by Texas Instruments Inc filed Critical Texas Instruments Inc
Publication of EP0071716A2 publication Critical patent/EP0071716A2/en
Publication of EP0071716A3 publication Critical patent/EP0071716A3/en
Application granted granted Critical
Publication of EP0071716B1 publication Critical patent/EP0071716B1/en
Expired legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/0018Speech coding using phonetic or linguistical decoding of the source; Reconstruction using text-to-speech synthesis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis

Definitions

  • This invention relates generally to speech and more particularly to speech recognition, compression, and transmission.
  • This type of device is generally referred to as a "vocoder”.
  • a vocoder was discussed by Richard Schwartz et al in his paper entitled "A Preliminary Design of a Phonic Vocoder Based on a Diphone Model" published in the IEEE International Conference on Accoustics, Speech and Signal Processing (ICASSP 80) proceedings of April 9, 10, 11, 1980 in Denver, Colorado (ICASSP 80 vol. 1, pg. 32 - 35).
  • the diphone model of Schwartz et al entails a phonic vocoder operating at 100 b/s. Each phonome of the speech generates a duration and single pitch value.
  • the inventory of diphone templates are used to synthesize the phoneme string. Additionally the diphone templates are utilized to initially establish which phonemes are being transmitted in the analog speech.
  • a diphone exists from the middle of one phoneme to the middle of the next phoneme due to the structure and stringing ability of a diphone, it is highly cumbersome in use and is generally ineffective in speech synthesis.
  • Diphone synthesis requires the use of an elaborate accoustic -to- phonic rule algorithm so as to create intelligible speech. This extensive accoustic -to- phonic rule algorithm requires a great deal of time and hardware to be effective.
  • Intrinsic to the recognition of an analog speech is the use of a methodology which breaks the analog speech into its component parts which may be compared to some library for identification. Numerous methods and apparatus have evolved so as to approximate the human speech and to model it. These modeling techniques include the vocoder, the voder, linear predictive filters, and other devices.
  • Flanagan discusses two electronic devices which automatically extract the first three formant frequencies from the continuous speech. These devices yield a continuous DC output voltage which are the magnitudes as functions of time, so as to represent the formant frequency of the speech.
  • the formant frequencies are in a analog form, use of an Analog/Digital (AD) converter readily transform these formant frequencies into digital form which is more suitable for use in a electronic environment.
  • AD Analog/Digital
  • the present embodiment employs means to separate the analog speech signal into phoneme parts.
  • a comparison means establishes our match with a phoneme template.
  • a reference code representative of the template is selected by an appropriate means.
  • This invention achieves a data rate of 80 bits per second or less. The technique in which this rate is received and still produce quality speech is through the use of a phoneme-toallophone translation.
  • the input data is normalized as to its speed, pitch, and other indicia; this is compared to a set of phoneme templates, within a set or Library of Templates. An optimal match is made.
  • the input pitch and variations are retained in a stored allophone string or sequence for replay or transmission.
  • Some applications which this device is put, are in a digital dictating machine, a store and play telephone, voice memos, multi-channel voice communications, voice recorded exams, etc.
  • the erroneous matching of the phonemes is more visible than in the synthesized speech situation; but it provides a rough draft or first cut to the document so as to be edited later.
  • An embodiment of the invention allows the apparatus to accept an initialization from the user so as to allow a normalization of the pitch and time parameters. This also allows the apparatus to create a library of phonemes templates which more closely approximates the actual user's phoneme structure.
  • the signal becomes less expensive and more efficient in the use of transmission time and hardware specifications for storage.
  • This invention uses a phoneme-to-allophone matching algorithm, the quality of synthesized speech is vastly improved since allophones more closely map the human utterances.
  • This vocoder accepts the analog speech input and matches it to a set of phoneme templates; the phonemes each contain a phoneme code which is compressed into a sequence of phoneme codes and communicated via a channel. This channel should be as noise free as possible so as to provide accurate transmission.
  • the sequence of phonemes is received and then translated to an analogous allophone sequence and synthesized through known electronic synthesis means.
  • the phoneme recognizer contains an automatic gain control (AGC), a format tracker, templates for the phonemes, and a recognition algorithm.
  • AGC automatic gain control
  • the phoneme recognizer receives the voice input and automatically controls the gain of the voice and sends a signal to the format tracker for analysis and formant extraction.
  • the algorithm operates on the formants and features of the utterance requiring the detection of the phoneme boundary within the speech.
  • the detected phoneme is matched to a phoneme in a library of phoneme templates.
  • Each phoneme template has a corresponding identification code.
  • the selected identification code is sequentially packed and transmitted via a transmission channel to a receiver.
  • the transmission channel may be either wired or wireless communication medium. Ideally the transmission channel is as noiseless as possible so as to reduce errors.
  • the phoneme-to-allophone synthesizer receives the phoneme codes from the channel.
  • the algorithm converts the phoneme sequence into an analagous allophone sequence and thereby produces quality speech.
  • a control means sequentially directs a library of allophone characteristics to be commuicated to a speech synthesizer.
  • a formant is a frequency component in the spectrum of speech which has large amplitude energy. It also has a resonant frequency of the pitch and a voiced sound. This resonant frequency is a multiple of a fundamental frequency.
  • the first formant occurs between 200 to 850 Hertz (Hz)
  • the second formantant occurs between 850 and 2,500 Hz
  • the third formantant occurs between 2,500 and 3,500 Hz.
  • This invention creates a formant tracker which keys upon the strong energy component in each frequency band.
  • the invention utilizes the technique of convolving the spectrum of the speech signal of interest with a sinusoidal signal which frequency is an integer multiple of the fundamental frequency. By varying the frequency of the sinusoidal signal and detecting the amplitude of the convolution, the formant is found in the selected frequency band.
  • the formant tracker is constructed using a pitch tracker together with additional logic around it so as to determine the sinusoidal oscillation and to convolve the two functions over the chosen spectrum frequency.
  • a set of integers is generated so that when each is multiplied by the fundamental frequency, the product lies within the formant range of interest. These three integer sets, one for each formant frequency range, should overlap sufficiently so as to allow the formant center to be sufficiently determined.
  • the integers within each integer set are used to generate a sinusoidal signal based at the product of the integer with the fundamental frequency.
  • the sinusoidal signal and the analog speech signal are integrated over a short time interval or frame.
  • the integration of the two time signals yield a convolution of their spectra.
  • the selected formant centers are, determined by multiplying the optimal integer by the fundamental frequency.
  • Each formant has associated therewith a bandwidth which is another indicia of the received analog speech data.
  • This indicia is combined with other indicia such as a pause or no pause, voice or unvoiced, a slope of the signal, and any other chosen data to generate a data value which is used to match to the library templates for phonemes.
  • One method of encoding the formant is to determine the distance between each formant and thereby achieve a reduction in the number of bits necessary to describe the formant selected.
  • an algorithm is used to match it to a particular approximated phoneme.
  • a tree algorithm is used which strips away the infeasible possibilities so as to reduce the total number of computations required for matching.
  • cycles in the decisional tree are strictly prohibited. A cycle in the decisional tree would allow the possiblity of an ever cycling situation such that a decision is never reached.
  • Any algorithm which matches the perceived phoneme to a phoneme template is permittable so long as it does a best approximation. This includes the algorithm which generates a comparision value for each phoneme template relative to the received phoneme and then chooses the optimal comparison value.
  • the code is transmitted to a storage means, a printer means, or a synthesizer.
  • the phoneme string is mapped into its component allophone set and used to synthesize the speech. This mapping of a phoneme to an allophone set is discussed in the article by Kun-Shan Lin, Gene A. Frantz, and Kathy Goudie in their article "Software Rules Give Personal Computer Real Word Power” appearing in Electronics, February 10, 1981, pg. 122-125, incorporated hereinto by reference. This article discusses the use of software to analyze text and determine its component elements and thereafter to pronouce them on a speech synthesis chip.
  • allophones are extremely powerful since it permits any spoken speech to be recreated without being dependent upon language or a fixed library.
  • the expanse of the allophonic and phoneme matching algorithm is the only limiting factor of the vocoder's ability.
  • mapping sciences such as but not limited to phoneme-to-diphone, are also applicable.
  • FIGURE 1 illustrates in block diagram the capabilities of an embodiment of the invention.
  • Analog speech 101 is picked up by the microphone 102 and transmitted in analog form to the analog to digital (A/D) converter 103. Once the signal has been translated into digital form, it is converted to perceived phoneme via the conversion means 104. Each perceived phoneme is communicated to the comparator 105 and referenced to templates in the library 106 so that a match is obtained. Once a matched phoneme is determined, its code is communicated via the bus 107 to either the phoneme sequencer 108, the storage means 109, or the transmitter 110.
  • A/D analog to digital
  • sequence code which matches to the phoneme sequence totally identifies the analog speech 101.
  • This code sequence is more suseptible to being packed or for storage than the original analog speech 101 due to its digital nature.
  • the phoneme sequencer 108 utilizes the code communicated via the bus 107 to obtain the appropriate phoneme from the library 106.
  • This phoneme from the library 106 has associated with it a set of allophone characteristics which are communicated to the synthesizer 114.
  • the synthesizer 114 communicates an analog signal to operate speaker 115 in the generation of speech 116.
  • a more intelligible and higher quality speech 116 is generated. This translation ability permits the encoding of the data in a phoneme base so as to facilite a lower bit per second transmission rate and thus require less time and storage medium for the recordation of the original analog speech 101.
  • the phoneme codes are stored via storage means 109 for later retrieval. This later retrieval is optionally used by the phoneme sequencer 108, synthesizer 114, and speaker 115 sequence to again synthesize the phoneme sequence in allophone form for generation of speech 116.
  • the storage means 109 communicates the phoneme codes to the phoneme to alphabet coverter 111 which translates the phonemes to their equivalent alphanumeric parts. Once the phonemes have been translated to the alphanumeric parts, such as in ASCII code, they are readily transmitted to the printer 112 so as to produce a paper copy 113 of the original analog speech 101.
  • the storage means 109 allows the invention to generate a speech to printed text so as to permit an automatic dictating device.
  • Another alternative is for the phoneme codes from the bus 107 to be communicated to a transmitter 110.
  • the transmitter generates signals 117 representative of the phoneme codes which are perceived by a remote unit 120 at its receiver 118.
  • the remote unit 120 contains the same capabilities as the transmitting unit 121. This entails the transmission of the phoneme code via a bus l19 from the receiver 118. Again, once the phoneme code is transmitted via the bus via 119, it is susceptible for the remote storage means 109' or the remote sequencer 108'. In another embodiment of the invention the phoneme codes transmitted via the bus 119 are also communicatable to a remote transmitter, not shown.
  • the remote unit 120 utilizes the phoneme codes in the same manner as the local unit 121.
  • the phoneme codes are utilized by the remote sequencer 108' in conjunction with the data in the remote library 106' to generate an analogous allophone sequence which is communicated to the remote synthesizer 114'.
  • the remote synthesizer 114' controls the operation of the remote speaker 115' in generating the speech 116'.
  • the remote unit 120 also has the option of storing the phoneme code at the remote storage means 109' for later use by the remote sequencer 108' or the phoneme to alphabet converter 111'.
  • the phoneme-to-alphabet converter 111' translates the phoneme code to its analagous alphanumeric symbols which are communicated to the printer 112' to generate a paper copy 113 prime.
  • the analog speech is translated to a phoneme code which is more susceptible to storage or for manipulation as a data string.
  • the phoneme code permits easy storage, transmission, generation of a printed copy or eventual syntheses by translation to an analogous allophone sequence.
  • FIGURE 2a illustrates, in block form, an embodiment of the invention which receives the analog speech input and results in a speech output.
  • the original analog speech input 201 is communiated to a phoneme recognizer 202 which generates a sequence of phonemes 203 via a communiation channel 204.
  • the sequence of phoneme 205 is communicated to a phoneme-toallophone synthesizer 206 which translates the phoneme sequence into its analogous allophone sequence so as to generate the speech output 207.
  • the phoneme recognizer 202 and the phoneme-to-allophone synthesizer 206 are alternatively in the same unit or are remote one from the other.
  • the communication channel 204 is either a hard wired device such as bus or a telephone line or a radio transmitter with receiver.
  • FIGURE 2b illustrates an embodiment of the phoneme recognizer 202 illustrated in FIGURE 2a.
  • the analog speech input 201 is communicated to an automatic gain control (AGC) 208 so as to regulate the speech signal into a certain desirable balance.
  • AGC automatic gain control
  • the formant tracker 209 breaks the analog signal into its formant components which are stored in a Random Access Memory (RAM) 210. Although in this embodiment the use of a RAM 210 is illustrated, it is comtemplated that any storage means is satisfactory.
  • the formants stored in RAM 210 are communicated to the phoneme boundary detection means 211 so as to group the formants into perceived phoneme components.
  • Each perceived phoneme is communicated to the recognition algorithm 212 which utilizes the phoneme templates from the Library 213 which is comprised of known phonemes. A best match is made between the perceived phoneme from the phoneme boundary detection means 211 and the templates found in the phoneme template library 213 by the recognition algorithm 212 so as to generate a recognized phoneme 214.
  • the recognition algorithm 212 provides a continuous sequence of phoneme codes so that a blank or non-recognized phoneme does not exist in the sequence. A blank for a non-recognition determination only results in an increase in the noise of the invention.
  • FIGURE 2c illustrates an embodiment of the phoneme-to-allophone synthesizer 206.
  • the sequence of phoneme codes 205 is communicated to the controller 215.
  • the controller 215 utilizes these codes and its prompting of the Read Only Memory (ROM) 217 to communicate to the speech synthesizer 216 the appropriate bit sequence indicitive of the analagous allophone sequence.
  • This data communicated from the ROM 217 to the speech synthesizer 216 establishes the parameters necessary for the modulation of the speaker 218 in the generation of the synthesized speech.
  • the speech synthesizer is chosen from a wide variety of speech synthesis means, including, but not limited to, the use of a linear predictive filter.
  • FIGURE 3 is a block diagram of an embodiment of the invention which generates indicia representative of the analog speech.
  • the automatic gain control (AGC) 301 communicates an analog speech signal to the pitch tracker 302 and the integration means 304, 314, and 324.
  • the pitch tracker 302 generates a fundamental frequency FO.
  • a respective set of integers is determined for which the fundamental frequency FO, when multiplied by the integer falls within the formant range.
  • the respective sets of integers are broadened to include an overlap in the sets so that the entire formant is defined.
  • the integer set for the first formant may contain (0,1,2,3,4) ; the second formant integer set contains (4,5,6,7); the third formant integer set contains (7,8,9).
  • the formant determiner 308 accepts the fundmental frequency FO and utilizes it with an integer value from the integer set for n in the sinusoidal oscillator 303.
  • the sinusoidal oscillator 303 generates a sinusoidal signal, s(t), which is centered at the product n and the fundamental frequency.
  • the sinusoidal signal is communicated to the integrater 304 which integrates the product of the sinusoidal signal s(t) and the analog speech signal, f(t) over the chosen frequency of the formant. This integration by the integrater 304 creates a convolution of the analog speech signal f(t).
  • This operation of generation of a sinusoidal signal 303 and integrating 304 is continued for all integer values within the integer set by the integrater 306.
  • the value of n which generates the maximum amplitude from the integrater 304 is chosen by the determinator 305.
  • This product additionally is determinative of the bandwidth BW1, of the first formant and the pair Fl and BW1 is communicated via channel 307.
  • the formant determinators 318 and 328 generate a sinusoidal signal via the sinusoidal oscillators 313 and 323 respectively and subsequently integrate by the integrators 314 and 324 so as to obtain the optimal M' and K', 315 and 325 respectively.
  • the indicia BW1, Fl, BW2, F2, BW3, F3, and F0 represent the perceived phoneme indicia from the analog speech from the AGC 301. This perceived indicia is used to match the perceived phoneme to a phoneme template in a library so as to obtain a best match.
  • FIGURE 4 indicates the relationship of the bandwidth to the optimal formant.
  • the optimal integer value N' is determined, its amplitude is plotted relative to the surrounding intergers.
  • the independent axes 402 contains the frequencies as dictated by the product of the integer value with the fundamental frequency.
  • the dependent axis 403 contains the amplitude generated by the product in the convolution with the analog speech signal. As illustrated, the optimal value N' generates an amplitude 404.
  • a bandwidth BW1 is determined for the appropriate optimal N'.
  • this bandwidth forms another indicia for determining the perceived phoneme relationship to the phoneme templates of the library. Similar analysis is done for each formant.
  • FIGURE 5 is a flow chart of an embodiment for determining the optimal formant positions.
  • the algorithm is started at 501 and a fundamental frequency, F0, 502 is determined. This fundamental frequency is utilized to optimize on N 503.
  • the optimization on N 503 entails the initializaton of the N value 504 followed by the sinusoidal oscillation based at the product of N FO 505.
  • the frequency convolver 506 generates the convolution of the fundamental frequency FO and the inputted analog speech signal over the chosen frequency of the formant.
  • the convolution is optimized at 507 wherein if it is not the optimal value, the N value is incremented 508 and the process is repeated into an optimal N value is determined.
  • the algorithm proceeds to optimize on the value of M 513 and then to optimize on the value K 523.
  • the optimization on N 503, the optimization of M 513, and the optimization of K 523 are identical in structure and performance.
  • three formant frequencies ranges are utilized to define the human language. It has been found that three ranges accurately describe the human speech, but this methodology is either extendable or contractable at the will of the designer. No loss in generality is encountered when the algorithm is extended to apply to a single formant or similiarly to extend to more than three formants.
  • FIGURE 6 graphically illustrates another methodology for the encoding of the analog speech signal in the formants.
  • the analog speech signal 608 is plotted over the independent axis 601 of frequency.
  • the dependent axis 602 is the amplitude.
  • the frequency range lies between 200 and 700 Hz.
  • the second formant has a frequency range of 850 to 2500 Hz; and the third formant 605 has a frequency range of 2700 to 3500 H z.
  • a method similiar to the methodology discussed in FIGURES 3 and FIGURE 5 is used to determine the location of the maximum amplitude within the formant range. These maxima yield a distance between maxima, 606 and 607 respectively. This distance, d l ,between the optimal formant is used to characterize the perceived phoneme for matching to a phoneme template. This methodology allows two integer values di and d 2 to describe what previously necessitated the use of three integer values (for the first, second and third formant).
  • Figure 7 is an embodiment of th encoding scheme for establishing a word for matching to the phoneme template.
  • the data word 701 in this example is an 8 bit word but any length of word which is capable of adequately describing the perceived phoneme is acceptable.
  • the 8 bits are broken up into four basic components, 702,703, 704, and 705.
  • the first component 702 is indicitive of a pause or no pause situation. Hence if bo is set to a value of 1, a pause has been perceived and the appropriate steps will therefore be taken; similiarly a 0 at bo indicates lack of a pause. A similiar relationship exists at this bl , 703, which indicates a voiced or unvoiced phoneme. B 2 -B 3 , 704, indicate the contour of the analog speech signal; its assigned value indicates a level slope, a positive slope or a negative slope.
  • B 4 -B 7 , 705 indicate a mixture of the relative energy, relative pitch, first distance, and second distance.
  • B 4 -B 7 , 705 are encoded so that their value indicates the characteristics of the perceived phoneme relating to the formant distances.
  • Bits b 4 -b 7 are encoded to communicate the distances between the maximums within each formant as illustrated in Figure 6. From table 706, each value within the range of bits b 4 -b 7 absolutely defines the two distances.
  • FIGURE 8 illustrates the translation of the phoneme code sequence into its appropriate allophone sequence or alternately its alphanumeric counterpart.
  • the phoneme sequence 801 is broken into its phoneme codes such as 802.
  • the phoneme code 802 distinctly describes a particular phoneme 807.
  • This phoneme 807 is either printed 805 in its ASCII alphanumeric character or it is translated to its analagous allophone sequence when it is taken in conjunction with the surrounding phoneme codes 803 and 804.
  • the allophone sequence 806 is generated through the knowledge of the target phoneme 807 and its relationship to its surrounding phonemes.
  • the phonemes which precede, 803, and follow, 804, the target phoneme 802 are retained in memory so as to generate the appropriate allophone sequence 806.
  • FIGURE 9 illustrates the characteristics of an embodiment of a decisional tree which determines the best approximation of the phoneme template in matching the perceived phoneme.
  • the decisional tree is broken up into multiple stages 901, 902, etc. Each stage of the tree breaks the perceived phoneme into a feasible and infeasible matches. As the perceived phoneme is further broken into feasible and infeasible states, the infeasible state becomes absorbing and the feasible state decreases so that eventually a single phoneme template is the only possible choice. Hence, the final stage of the tree must consist of as many nodes as there are templates.
  • the original decision 903 is made on whether the first bit, b 0 , is either set or not set. If the first bit is set, transition is made to node 905; the nodes which follow node 904, Bl, are ignored. This determination on the b o level result in translating the available phoneme templates into an infeasible set, those lying exclusivley behind node 904 and a feasilbe set, those lying behind node B2, 905. A similiar determination is made for each component part of the indicia. In this example, another separation is made on b l and then on the value of b2 - b3- This separation into nodes is continued until a final or terminating node is encountered which uniquely identifies the phoneme template chosen.
  • Movement is acceptable laterally between nodes such as between nodes El, 908, and E2, 909 via the ray 907. This movement is permissable so long as a cycle is not thereby created.
  • ray 910 indicates a cycle between Dl and Cl.
  • a sequence containing Cl-Dl-Cl-Dl-Cl is not acceptable since it is a cycle. This sequence causes a never ending cycle which results in a decision never being made.
  • the one qualification of the tree illustrated in this embodiment is that a decision must eventually be reached.
  • the algorithm illustrated in FIGURE 9 is but one embodiment to identify the best match between the perceived phoneme and the phoneme template. Another approach is to generate a comparison value for each phoneme template relative to the perceived phoneme and then choose the optimal value accordingly. This approach requires more computataion and a longer time for its operation.
  • FIGURES 10a and 10b illustrate a phoneme to allophone transformation wherein a phoneme is translated to its analagous allophone sequence.
  • FIGURE 10a a list of the rules in defining the allophone are made.
  • 1001 illustrates a blank or a word boundary.
  • the different symbols illustrated indicate a different allophonic characteristics which are attachable to a phoneme.
  • the syllables are broken by a period ".”, 1002.
  • These allophonic rules are combined with the phonemes to generate the appropriate allophone sequence.
  • FIGURE lOb illustrates how the phoneme "CH", 1003, translates into an appropriate allophone sequence.
  • the phoneme "CH” is either a “b CH", 1004, as in “chain” or lies within a word as illustrated by "CH", 1005, as in "bewitching".
  • Each phoneme maps into a unique allophone sequence. This allophone sequence is determined through knowledge of the preceeding phoneme and the following phoneme within the phoneme sequence.
  • the invention as described herein details the use of a voice recognition system which translates the analog speech signal into a phoneme sequence which is more susceptible to compaction, storage, transmission, or translation to an analagous allophone sequence for speech synthesis.
  • the phoneme perception allows for an unlimitable vocabulary to be used and also for a best match to be generated.
  • the use of a best match is acceptable since the human ear acts as a filtering mechanism and the human brain ignores random noise so as to also filter the synthesized speech.
  • the synthesized speech is enhanced dramatically through the translation of the phoneme sequence to an analagous allophone sequence.
  • the stored phoneme sequence is suseptible to being translated to an alphanumeric sequence or for transmission via the radio or telephone lines.
  • This invention permits the direct speech to text dictating machines and also for a highly efficient data transmission rate.

Abstract

An allophone vocoder which utilizes the inherent redundancy of the spoken language together with the automatic human filtering of speech so as to obtain a speech compression and recognition system. An analog speech signal is broken up into its phoneme components (105) and encoded for transmission (110). The encoded phoneme sequence has a much higher compression rate than the analog speech signal. The phonemes are then either transmitted (110), stored (109), or used to generate directly an analogous allophone sequence (114) so as to approximate the original speech signal. Due to the inherent redundancy of the spoken language, and the filtering effect of the human, variations or errors in the approximations of the phonemes received from the original speech signal are inconsequential to the comprehension ability of the final allophone synthesized speech.

Description

    BACKGROUND
  • This invention relates generally to speech and more particularly to speech recognition, compression, and transmission.
  • It has long been recognized that analog speech signals contain numerous redundant sounds so as to make it not suitable for efficient data transmission. In a direct human interaction situation this inefficiency is tolerable. The technical requirements to cope with inefficient speech though becomes infeasible due to cost, time, and storage necessary.
  • A need exists for a system which can take analog speech signals and translate it into a digital form which is reconstructable after transmission or storage. This type of device is generally referred to as a "vocoder".
  • A vocoder was discussed by Richard Schwartz et al in his paper entitled "A Preliminary Design of a Phonic Vocoder Based on a Diphone Model" published in the IEEE International Conference on Accoustics, Speech and Signal Processing (ICASSP 80) proceedings of April 9, 10, 11, 1980 in Denver, Colorado (ICASSP 80 vol. 1, pg. 32 - 35). The diphone model of Schwartz et al entails a phonic vocoder operating at 100 b/s. Each phonome of the speech generates a duration and single pitch value. The inventory of diphone templates are used to synthesize the phoneme string. Additionally the diphone templates are utilized to initially establish which phonemes are being transmitted in the analog speech. A diphone exists from the middle of one phoneme to the middle of the next phoneme due to the structure and stringing ability of a diphone, it is highly cumbersome in use and is generally ineffective in speech synthesis.
  • Diphone synthesis requires the use of an elaborate accoustic -to- phonic rule algorithm so as to create intelligible speech. This extensive accoustic -to- phonic rule algorithm requires a great deal of time and hardware to be effective.
  • Intrinsic to the recognition of an analog speech is the use of a methodology which breaks the analog speech into its component parts which may be compared to some library for identification. Numerous methods and apparatus have evolved so as to approximate the human speech and to model it. These modeling techniques include the vocoder, the voder, linear predictive filters, and other devices.
  • One such method of analyzing the analog speech was discussed by James L. Flanagan in the article "Automatic Extraction of Formant for Continuous Speech" first printed in J. Acoust. Soc. Am., line 28, pp. 110 - 118, January 1956, incorporated hereinto by reference.
  • In the article, Flanagan discusses two electronic devices which automatically extract the first three formant frequencies from the continuous speech. These devices yield a continuous DC output voltage which are the magnitudes as functions of time, so as to represent the formant frequency of the speech. Although the formant frequencies are in a analog form, use of an Analog/Digital (AD) converter readily transform these formant frequencies into digital form which is more suitable for use in a electronic environment.
  • Another method was discussed by H. K. Dunn in his article "Methods of Measuring Vowal Formant Bandwidths" J. Acoust. Soc. Am., Vol. 33, pp. 1737-1746, December 1961, incorporated hereinto by reference. In the article, Dunn discloses the use of spectrums of real speech and the use of an artificial larynx in an application.
  • It is clear therefore that an efficient methodology and apparatus for transformation of an analog speech signal to a approximating digital form does not exist. The mere recognition of formants or the use of diphones in the synthesis of the perceived speech is inaccurate and does not allow for quality recordation and transmission of data representative of the original speech signal.
  • DESCRIPTION OF THE INVENTION
  • The present embodiment employs means to separate the analog speech signal into phoneme parts. A comparison means establishes our match with a phoneme template. A reference code representative of the template is selected by an appropriate means. This invention achieves a data rate of 80 bits per second or less. The technique in which this rate is received and still produce quality speech is through the use of a phoneme-toallophone translation. The input data is normalized as to its speed, pitch, and other indicia; this is compared to a set of phoneme templates, within a set or Library of Templates. An optimal match is made. The input pitch and variations are retained in a stored allophone string or sequence for replay or transmission.
  • Since the human ear acts as a filtering mechanism and also due to the inherent redundancy of the spoken language, any errors which are generated in the selection of the optimal phoneme match is minimized. As example, assume that the phoneme recognizer incorrectly matched the spoken phoneme "SH" to the phoneme "CH" in the phrase "we will be taking a cruise on the ship". This results in the phrase becoming "we will be taking a cruise on the chip". Although the transmitted phoneme sequence is not a perfect match, the total phrase is still intelligible to the listener since the human ear and the mental process filters out this incorrect phoneme. The human ear and mental process has developed over the years to compensate for variations in pronunciations and the incorrect usage of words.
  • Some applications, which this device is put, are in a digital dictating machine, a store and play telephone, voice memos, multi-channel voice communications, voice recorded exams, etc. In the situation of a dictating machine, the erroneous matching of the phonemes is more visible than in the synthesized speech situation; but it provides a rough draft or first cut to the document so as to be edited later.
  • An embodiment of the invention allows the apparatus to accept an initialization from the user so as to allow a normalization of the pitch and time parameters. This also allows the apparatus to create a library of phonemes templates which more closely approximates the actual user's phoneme structure.
  • At the compression rate of 80 b/sec, the signal becomes less expensive and more efficient in the use of transmission time and hardware specifications for storage.
  • This invention uses a phoneme-to-allophone matching algorithm, the quality of synthesized speech is vastly improved since allophones more closely map the human utterances.
  • This vocoder accepts the analog speech input and matches it to a set of phoneme templates; the phonemes each contain a phoneme code which is compressed into a sequence of phoneme codes and communicated via a channel. This channel should be as noise free as possible so as to provide accurate transmission. The sequence of phonemes is received and then translated to an analogous allophone sequence and synthesized through known electronic synthesis means.
  • One such means is discussed in Patent No. 4, 209,836 issued to Wiggins, Jr, et al on June 24, 1980, incorporated hereinto by reference. This speech synthesis integrated circuit device uses a linear predictive filter in its generation of the synthesized speech.
  • The control of the data within the synthesizer is well known in the art. One such method for communicating digital speech data and control of the memory for storing the data is disclosed in Patent No. 4,234,761 issued to Wiggins, Jr., et al on November 18, 1980, incorporated hereinto by reference.
  • In the invention, the phoneme recognizer contains an automatic gain control (AGC), a format tracker, templates for the phonemes, and a recognition algorithm. The phoneme recognizer receives the voice input and automatically controls the gain of the voice and sends a signal to the format tracker for analysis and formant extraction. The algorithm operates on the formants and features of the utterance requiring the detection of the phoneme boundary within the speech. The detected phoneme is matched to a phoneme in a library of phoneme templates. Each phoneme template has a corresponding identification code. The selected identification code is sequentially packed and transmitted via a transmission channel to a receiver.
  • The transmission channel may be either wired or wireless communication medium. Ideally the transmission channel is as noiseless as possible so as to reduce errors.
  • The phoneme-to-allophone synthesizer receives the phoneme codes from the channel. The algorithm converts the phoneme sequence into an analagous allophone sequence and thereby produces quality speech. In the phoneme-to-allophone synthesizer, a control means sequentially directs a library of allophone characteristics to be commuicated to a speech synthesizer.
  • The use of an efficient formant tracker is beneficial. A formant is a frequency component in the spectrum of speech which has large amplitude energy. It also has a resonant frequency of the pitch and a voiced sound. This resonant frequency is a multiple of a fundamental frequency. The first formant occurs between 200 to 850 Hertz (Hz), the second formantant occurs between 850 and 2,500 Hz, the third formantant occurs between 2,500 and 3,500 Hz. This invention creates a formant tracker which keys upon the strong energy component in each frequency band.
  • The invention utilizes the technique of convolving the spectrum of the speech signal of interest with a sinusoidal signal which frequency is an integer multiple of the fundamental frequency. By varying the frequency of the sinusoidal signal and detecting the amplitude of the convolution, the formant is found in the selected frequency band.
  • In one embodiment of the formant tracker, it is constructed using a pitch tracker together with additional logic around it so as to determine the sinusoidal oscillation and to convolve the two functions over the chosen spectrum frequency.
  • A set of integers is generated so that when each is multiplied by the fundamental frequency, the product lies within the formant range of interest. These three integer sets, one for each formant frequency range, should overlap sufficiently so as to allow the formant center to be sufficiently determined. The integers within each integer set are used to generate a sinusoidal signal based at the product of the integer with the fundamental frequency. The sinusoidal signal and the analog speech signal are integrated over a short time interval or frame. Mathematically, the integration of the two time signals yield a convolution of their spectra. By performing the integration for each integer, the maximum or largest magnitude becomes evident and the associated optimal integer determines a formant. The selected formant centers are, determined by multiplying the optimal integer by the fundamental frequency. Each formant has associated therewith a bandwidth which is another indicia of the received analog speech data.
  • This indicia is combined with other indicia such as a pause or no pause, voice or unvoiced, a slope of the signal, and any other chosen data to generate a data value which is used to match to the library templates for phonemes.
  • One method of encoding the formant is to determine the distance between each formant and thereby achieve a reduction in the number of bits necessary to describe the formant selected.
  • The use of formant analysis in voice speech is discussed by Schaffer and Rabiner in their article "System For Automatic Formant Analysis of Voice Speech" appearing in the J. Aclust. Soc. Am., Vol. 47, pg. 634 - 648, February, 1970, incorporated hereinto by reference. Schaffer and Rabiner utilized a gain control which varies with time and controls intensity of the output. The use of a cascaded network is used to approximate a combination of the glottal - source spectrum and a radiation load spectrum. The analysis system determines, as a function of time the lowest three formants, the pitch period, and the gain.
  • Once the indicia is determined, an algorithm is used to match it to a particular approximated phoneme. In the preferred embodiment, a tree algorithm is used which strips away the infeasible possibilities so as to reduce the total number of computations required for matching. In this algorithm, since it is a tree approach, cycles in the decisional tree are strictly prohibited. A cycle in the decisional tree would allow the possiblity of an ever cycling situation such that a decision is never reached.
  • Any algorithm which matches the perceived phoneme to a phoneme template is permittable so long as it does a best approximation. This includes the algorithm which generates a comparision value for each phoneme template relative to the received phoneme and then chooses the optimal comparison value.
  • Once the optimal phoneme has been matched to a code, the code is transmitted to a storage means, a printer means, or a synthesizer. Before synthesis, the phoneme string is mapped into its component allophone set and used to synthesize the speech. This mapping of a phoneme to an allophone set is discussed in the article by Kun-Shan Lin, Gene A. Frantz, and Kathy Goudie in their article "Software Rules Give Personal Computer Real Word Power" appearing in Electronics, February 10, 1981, pg. 122-125, incorporated hereinto by reference. This article discusses the use of software to analyze text and determine its component elements and thereafter to pronouce them on a speech synthesis chip.
  • Another algorithm was discussed by Kun-Shan Lin, Kathy Goudie, Gene Frantz, and George Brandingham in their article "Text-to-Speech Using LPC Allophone Stringing" appearing in IEEE Transctions on Consumer Electronics, Vol. CE 27, May 1981, pg. 144-152, incorporated hereinto by reference. This article discusses a response system for a text - to -speech conversion of any English text. The system utilizes an LPC synthesizer chip and a microprocesser. The system converts an input ASCII characters into an alphonic code with their synthesis.
  • The use of allophones is extremely powerful since it permits any spoken speech to be recreated without being dependent upon language or a fixed library. The expanse of the allophonic and phoneme matching algorithm is the only limiting factor of the vocoder's ability.
  • Although the preferred embodiment is a phoneme-to-allophone mapping, other mapping sciences such as but not limited to phoneme-to-diphone, are also applicable.
  • The invention together with its particular embodiments and ramifications will be more fully explained by the following drawings and their accompanying descriptions.
  • DRAWINGS IN BRIEF
    • FIGURE 1 is a block diagram of an embodiment of the invention illustrating the data compression and transmission capabilities of the invention.
    • FIGURE 2a is a block diagram of the communication relationship of the invention.
    • FIGURE 2b and 2c illustrate the recognition side and the synthesis side respectively of the embodiment illustrated in FIGURE 2a.
    • FIGURE 3 is an embodiment of the invention utilized to generate indicia representative of the analog speech signal.
    • FIGURE 4 is illustrative of the determination of the bandwidth associated with a particular format.
    • FIGURE 5 is a flow chart of an embodiment determining the formant of the analog speech.
    • FIGURE 6 illustrates a method of determining indicia so as to define a particular formant structure of an analog speech.
    • FIGURE 7 illustrates an encoding scheme for the idicia.
    • FIGURE 8 illustrates a translational operation of the of the phoneme to either allophone or alphanumeric characters.
    • FIGURE 9 is an example of a decisional tree operating upon the encoded indicia as representated in FIGURE 7.
    • FIGURES 10a and 10b illustrate the translation of phonemes-to-allophones.
    DRAWINGS IN DETAIL
  • FIGURE 1 illustrates in block diagram the capabilities of an embodiment of the invention.
  • Analog speech 101 is picked up by the microphone 102 and transmitted in analog form to the analog to digital (A/D) converter 103. Once the signal has been translated into digital form, it is converted to perceived phoneme via the conversion means 104. Each perceived phoneme is communicated to the comparator 105 and referenced to templates in the library 106 so that a match is obtained. Once a matched phoneme is determined, its code is communicated via the bus 107 to either the phoneme sequencer 108, the storage means 109, or the transmitter 110.
  • The sequence code which matches to the phoneme sequence totally identifies the analog speech 101. This code sequence is more suseptible to being packed or for storage than the original analog speech 101 due to its digital nature.
  • The phoneme sequencer 108 utilizes the code communicated via the bus 107 to obtain the appropriate phoneme from the library 106. This phoneme from the library 106 has associated with it a set of allophone characteristics which are communicated to the synthesizer 114. The synthesizer 114 communicates an analog signal to operate speaker 115 in the generation of speech 116. Through the use of the phoneme-toallophone translation as effectuated by the phoneme sequencer 108, with the aid of library 106, a more intelligible and higher quality speech 116 is generated. This translation ability permits the encoding of the data in a phoneme base so as to facilite a lower bit per second transmission rate and thus require less time and storage medium for the recordation of the original analog speech 101.
  • Alternately, the phoneme codes are stored via storage means 109 for later retrieval. This later retrieval is optionally used by the phoneme sequencer 108, synthesizer 114, and speaker 115 sequence to again synthesize the phoneme sequence in allophone form for generation of speech 116. Optionally, the storage means 109 communicates the phoneme codes to the phoneme to alphabet coverter 111 which translates the phonemes to their equivalent alphanumeric parts. Once the phonemes have been translated to the alphanumeric parts, such as in ASCII code, they are readily transmitted to the printer 112 so as to produce a paper copy 113 of the original analog speech 101.
  • This branch of the operation, the storage means 109, phoneme-to-alphabet converter 111, and printer 112, allows the invention to generate a speech to printed text so as to permit an automatic dictating device.
  • Another alternative is for the phoneme codes from the bus 107 to be communicated to a transmitter 110. The transmitter generates signals 117 representative of the phoneme codes which are perceived by a remote unit 120 at its receiver 118.
  • The remote unit 120 contains the same capabilities as the transmitting unit 121. This entails the transmission of the phoneme code via a bus l19 from the receiver 118. Again, once the phoneme code is transmitted via the bus via 119, it is susceptible for the remote storage means 109' or the remote sequencer 108'. In another embodiment of the invention the phoneme codes transmitted via the bus 119 are also communicatable to a remote transmitter, not shown.
  • The remote unit 120 utilizes the phoneme codes in the same manner as the local unit 121. The phoneme codes are utilized by the remote sequencer 108' in conjunction with the data in the remote library 106' to generate an analogous allophone sequence which is communicated to the remote synthesizer 114'. The remote synthesizer 114' controls the operation of the remote speaker 115' in generating the speech 116'. The remote unit 120 also has the option of storing the phoneme code at the remote storage means 109' for later use by the remote sequencer 108' or the phoneme to alphabet converter 111'. The phoneme-to-alphabet converter 111' translates the phoneme code to its analagous alphanumeric symbols which are communicated to the printer 112' to generate a paper copy 113 prime.
  • It is clear from this embodiment of the invention that the analog speech is translated to a phoneme code which is more susceptible to storage or for manipulation as a data string. The phoneme code permits easy storage, transmission, generation of a printed copy or eventual syntheses by translation to an analogous allophone sequence.
  • FIGURE 2a illustrates, in block form, an embodiment of the invention which receives the analog speech input and results in a speech output.
  • In the embodiment of FIGURE 2a, the original analog speech input 201 is communiated to a phoneme recognizer 202 which generates a sequence of phonemes 203 via a communiation channel 204. The sequence of phoneme 205 is communicated to a phoneme-toallophone synthesizer 206 which translates the phoneme sequence into its analogous allophone sequence so as to generate the speech output 207. It should be noted that the phoneme recognizer 202 and the phoneme-to-allophone synthesizer 206 are alternatively in the same unit or are remote one from the other. In this context the communication channel 204 is either a hard wired device such as bus or a telephone line or a radio transmitter with receiver.
  • FIGURE 2b illustrates an embodiment of the phoneme recognizer 202 illustrated in FIGURE 2a.
  • The analog speech input 201 is communicated to an automatic gain control (AGC) 208 so as to regulate the speech signal into a certain desirable balance. The formant tracker 209 breaks the analog signal into its formant components which are stored in a Random Access Memory (RAM) 210. Although in this embodiment the use of a RAM 210 is illustrated, it is comtemplated that any storage means is satisfactory. The formants stored in RAM 210 are communicated to the phoneme boundary detection means 211 so as to group the formants into perceived phoneme components. Each perceived phoneme is communicated to the recognition algorithm 212 which utilizes the phoneme templates from the Library 213 which is comprised of known phonemes. A best match is made between the perceived phoneme from the phoneme boundary detection means 211 and the templates found in the phoneme template library 213 by the recognition algorithm 212 so as to generate a recognized phoneme 214.
  • As noted earlier, a best match, even if not a perfect recognition, since the natural filtering of the human ear and the error correction of the mental processes of the listener minimizes any errors generated by the recognition algorithm 212. The recognition algorithm 212 provides a continuous sequence of phoneme codes so that a blank or non-recognized phoneme does not exist in the sequence. A blank for a non-recognition determination only results in an increase in the noise of the invention.
  • FIGURE 2c illustrates an embodiment of the phoneme-to-allophone synthesizer 206.
  • The sequence of phoneme codes 205 is communicated to the controller 215. The controller 215 utilizes these codes and its prompting of the Read Only Memory (ROM) 217 to communicate to the speech synthesizer 216 the appropriate bit sequence indicitive of the analagous allophone sequence. This data communicated from the ROM 217 to the speech synthesizer 216 establishes the parameters necessary for the modulation of the speaker 218 in the generation of the synthesized speech.
  • The speech synthesizer is chosen from a wide variety of speech synthesis means, including, but not limited to, the use of a linear predictive filter.
  • FIGURE 3 is a block diagram of an embodiment of the invention which generates indicia representative of the analog speech.
  • This indicia is representative of the perceived phoneme and is used in finding a best match or optimal match with the template in the Library. The automatic gain control (AGC) 301 communicates an analog speech signal to the pitch tracker 302 and the integration means 304, 314, and 324. The pitch tracker 302, generates a fundamental frequency FO.
  • For each formant determiner 308, 318, and 328 a respective set of integers is determined for which the fundamental frequency FO, when multiplied by the integer falls within the formant range. The respective sets of integers are broadened to include an overlap in the sets so that the entire formant is defined. As an example, if the fundamental frequency FO is 200 Hz, the integer set for the first formant may contain (0,1,2,3,4) ; the second formant integer set contains (4,5,6,7); the third formant integer set contains (7,8,9).
  • The formant determiner 308 accepts the fundmental frequency FO and utilizes it with an integer value from the integer set for n in the sinusoidal oscillator 303. The sinusoidal oscillator 303 generates a sinusoidal signal, s(t), which is centered at the product n and the fundamental frequency. The sinusoidal signal is communicated to the integrater 304 which integrates the product of the sinusoidal signal s(t) and the analog speech signal, f(t) over the chosen frequency of the formant. This integration by the integrater 304 creates a convolution of the analog speech signal f(t).
  • This operation of generation of a sinusoidal signal 303 and integrating 304 is continued for all integer values within the integer set by the integrater 306. The value of n which generates the maximum amplitude from the integrater 304 is chosen by the determinator 305. This optimal value, N', is used to generate the first formant Fl defined by Fl=N'x F0. This product additionally is determinative of the bandwidth BW1, of the first formant and the pair Fl and BW1 is communicated via channel 307.
  • In like fashion the formant determinators 318 and 328 generate a sinusoidal signal via the sinusoidal oscillators 313 and 323 respectively and subsequently integrate by the integrators 314 and 324 so as to obtain the optimal M' and K', 315 and 325 respectively.
  • The indicia BW1, Fl, BW2, F2, BW3, F3, and F0, represent the perceived phoneme indicia from the analog speech from the AGC 301. This perceived indicia is used to match the perceived phoneme to a phoneme template in a library so as to obtain a best match.
  • FIGURE 4 indicates the relationship of the bandwidth to the optimal formant.
  • Once the optimal integer value N' is determined, its amplitude is plotted relative to the surrounding intergers. The independent axes 402 contains the frequencies as dictated by the product of the integer value with the fundamental frequency. The dependent axis 403 contains the amplitude generated by the product in the convolution with the analog speech signal. As illustrated, the optimal value N' generates an amplitude 404. By utilizing the surrounding data points 405, 406, 407, and 408, a bandwidth BW1 is determined for the appropriate optimal N'.
  • The use of this bandwidth forms another indicia for determining the perceived phoneme relationship to the phoneme templates of the library. Similar analysis is done for each formant.
  • FIGURE 5 is a flow chart of an embodiment for determining the optimal formant positions.
  • The algorithm is started at 501 and a fundamental frequency, F0, 502 is determined. This fundamental frequency is utilized to optimize on N 503. The optimization on N 503 entails the initializaton of the N value 504 followed by the sinusoidal oscillation based at the product of N FO 505. The frequency convolver 506 generates the convolution of the fundamental frequency FO and the inputted analog speech signal over the chosen frequency of the formant. The convolution is optimized at 507 wherein if it is not the optimal value, the N value is incremented 508 and the process is repeated into an optimal N value is determined. At the optimization of N, the algorithm proceeds to optimize on the value of M 513 and then to optimize on the value K 523. The optimization on N 503, the optimization of M 513, and the optimization of K 523 are identical in structure and performance.
  • In this embodiment three formant frequencies ranges are utilized to define the human language. It has been found that three ranges accurately describe the human speech, but this methodology is either extendable or contractable at the will of the designer. No loss in generality is encountered when the algorithm is extended to apply to a single formant or similiarly to extend to more than three formants.
  • FIGURE 6 graphically illustrates another methodology for the encoding of the analog speech signal in the formants.
  • The analog speech signal 608 is plotted over the independent axis 601 of frequency. The dependent axis 602 is the amplitude. Within the first formant 603, the frequency range lies between 200 and 700 Hz. The second formant has a frequency range of 850 to 2500 Hz; and the third formant 605 has a frequency range of 2700 to 3500 Hz. A method similiar to the methodology discussed in FIGURES 3 and FIGURE 5 is used to determine the location of the maximum amplitude within the formant range. These maxima yield a distance between maxima, 606 and 607 respectively. This distance, dl,between the optimal formant is used to characterize the perceived phoneme for matching to a phoneme template. This methodology allows two integer values di and d2 to describe what previously necessitated the use of three integer values (for the first, second and third formant).
  • Figure 7 is an embodiment of th encoding scheme for establishing a word for matching to the phoneme template.
  • The data word 701 in this example is an 8 bit word but any length of word which is capable of adequately describing the perceived phoneme is acceptable. In this embodiment the 8 bits are broken up into four basic components, 702,703, 704, and 705.
  • The first component 702 is indicitive of a pause or no pause situation. Hence if bo is set to a value of 1, a pause has been perceived and the appropriate steps will therefore be taken; similiarly a 0 at bo indicates lack of a pause. A similiar relationship exists at this bl, 703, which indicates a voiced or unvoiced phoneme. B 2-B3, 704, indicate the contour of the analog speech signal; its assigned value indicates a level slope, a positive slope or a negative slope.
  • B 4-B7, 705, indicate a mixture of the relative energy, relative pitch, first distance, and second distance. B4-B7, 705, are encoded so that their value indicates the characteristics of the perceived phoneme relating to the formant distances. Bits b4-b7 are encoded to communicate the distances between the maximums within each formant as illustrated in Figure 6. From table 706, each value within the range of bits b4-b7 absolutely defines the two distances.
  • FIGURE 8 illustrates the translation of the phoneme code sequence into its appropriate allophone sequence or alternately its alphanumeric counterpart.
  • The phoneme sequence 801 is broken into its phoneme codes such as 802. The phoneme code 802 distinctly describes a particular phoneme 807. This phoneme 807 is either printed 805 in its ASCII alphanumeric character or it is translated to its analagous allophone sequence when it is taken in conjunction with the surrounding phoneme codes 803 and 804.
  • The allophone sequence 806 is generated through the knowledge of the target phoneme 807 and its relationship to its surrounding phonemes. In this context, the phonemes which precede, 803, and follow, 804, the target phoneme 802 are retained in memory so as to generate the appropriate allophone sequence 806.
  • FIGURE 9 illustrates the characteristics of an embodiment of a decisional tree which determines the best approximation of the phoneme template in matching the perceived phoneme.
  • The decisional tree is broken up into multiple stages 901, 902, etc. Each stage of the tree breaks the perceived phoneme into a feasible and infeasible matches. As the perceived phoneme is further broken into feasible and infeasible states, the infeasible state becomes absorbing and the feasible state decreases so that eventually a single phoneme template is the only possible choice. Hence, the final stage of the tree must consist of as many nodes as there are templates.
  • The original decision 903 is made on whether the first bit, b0, is either set or not set. If the first bit is set, transition is made to node 905; the nodes which follow node 904, Bl, are ignored. This determination on the bo level result in translating the available phoneme templates into an infeasible set, those lying exclusivley behind node 904 and a feasilbe set, those lying behind node B2, 905. A similiar determination is made for each component part of the indicia. In this example, another separation is made on bl and then on the value of b2-b3- This separation into nodes is continued until a final or terminating node is encountered which uniquely identifies the phoneme template chosen.
  • Movement is acceptable laterally between nodes such as between nodes El, 908, and E2, 909 via the ray 907. This movement is permissable so long as a cycle is not thereby created. In this context ray 910 indicates a cycle between Dl and Cl. As example, a sequence containing Cl-Dl-Cl-Dl-Cl is not acceptable since it is a cycle. This sequence causes a never ending cycle which results in a decision never being made. The one qualification of the tree illustrated in this embodiment is that a decision must eventually be reached.
  • The algorithm illustrated in FIGURE 9 is but one embodiment to identify the best match between the perceived phoneme and the phoneme template. Another approach is to generate a comparison value for each phoneme template relative to the perceived phoneme and then choose the optimal value accordingly. This approach requires more computataion and a longer time for its operation.
  • FIGURES 10a and 10b illustrate a phoneme to allophone transformation wherein a phoneme is translated to its analagous allophone sequence.
  • In FIGURE 10a, a list of the rules in defining the allophone are made. As illustrated "b", 1001 illustrates a blank or a word boundary. The different symbols illustrated indicate a different allophonic characteristics which are attachable to a phoneme. The syllables are broken by a period ".", 1002. These allophonic rules are combined with the phonemes to generate the appropriate allophone sequence.
  • FIGURE lOb illustrates how the phoneme "CH", 1003, translates into an appropriate allophone sequence. Depending upon the preceeding and the following phoneme, the phoneme "CH" is either a "b CH", 1004, as in "chain" or lies within a word as illustrated by "CH", 1005, as in "bewitching".
  • Each phoneme maps into a unique allophone sequence. This allophone sequence is determined through knowledge of the preceeding phoneme and the following phoneme within the phoneme sequence.
  • The invention as described herein details the use of a voice recognition system which translates the analog speech signal into a phoneme sequence which is more susceptible to compaction, storage, transmission, or translation to an analagous allophone sequence for speech synthesis. The phoneme perception allows for an unlimitable vocabulary to be used and also for a best match to be generated. The use of a best match is acceptable since the human ear acts as a filtering mechanism and the human brain ignores random noise so as to also filter the synthesized speech. The synthesized speech is enhanced dramatically through the translation of the phoneme sequence to an analagous allophone sequence. The stored phoneme sequence is suseptible to being translated to an alphanumeric sequence or for transmission via the radio or telephone lines.
  • This invention permits the direct speech to text dictating machines and also for a highly efficient data transmission rate.

Claims (8)

1. A recognition system for receiving analog speech signals comprising;
a. means,
1. for separating said analog signal into human sound component parts, and
2. for generating perceived indicia representative of each of said human sound component parts; and
b. means for selecting for each human sound component part, based on said perceived indicia a reference data from reference a data set.
2. The recognition system according to claim 1, further comprised of means
.1. for receiving said analog speech signal, and
2. for transmitting said analog speech signal to said means for separating.
3. The recognition system according to claim 1 wherein said reference data has reference indicia approximating said perceived indicia.
4. A device for compression of analog speech data into equivalent digital data comprised of:
a. means for translating said analog speech data into electronic signals;
b. means for separating said electronic signal into a sequence of component human sound;
c. means for assigning each of said component human sound a representative indicia;
d. means for determining, for each of said representative indicia, a selected human sound from a reference set of human sound, said selected human sounds having indicia being the closest approximation to said representative indicia within said reference set of human sounds; and
e. means for digitally storing each of said selected human sounds.
5. A vocoder system comprised of:
a. means
1. for accepting an analog speech signal, and
2. for identifying allophone parts of said analog speech signal, and
3. for determining, for each of the allophone parts, a representative indicia;
b. means
1. for determining, for each indicia, an optimal comparision value, each optimal comparision value being representative of the similiarity between said representative indicia and reference indicia within a library of reference indicia, and
2. for determining, for each optimal comparision value, an associated allophone code; and
c. means
1. for creating an allophone sequence of said associated allophone codes, and
2. for synthesizing an auditory signal from said allophone sequence.
6. A formant tracker receiving an analog signal, said formant tracker comprising;
a. means for tracking a pitch associated with said analog signal and yielding a fundamental frequency therefrom; and
b. means for iteratively establishing for each selected frequency range within a set of selected frequency ranges,
1. an integer set associated with the selected frequency range and
2. an optimal integer for said frequency range from said integer set, and
3. a bandwidth associated with said optimal integer.
7. The formant tracker according to claim 6 wherein said means for establishing an optimal integer includes;
a. means for iteratively generating, for each integer within said integer set,
1. a sinusoidal signal around a chosen center point, and
2. a convolution value being the convolution of said sinusoidal and said analog signal over said frequency range; and
b. means for choosing the optimal integer based on said convolution value.
8. A vocoder network receiving an analog speech signal, said vocoder network comprising:
a. means for recognizing a sequence of phonemes within said analog speech signal; and
b. means,
1. for translating said phoneme sequence into an analogous allophone sequence, and
2. for synthesizing said allophone sequence into an auditory signal.
EP19820105168 1981-08-03 1982-06-14 Allophone vocoder Expired EP0071716B1 (en)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US28969081A 1981-08-03 1981-08-03
US06/289,603 US4424415A (en) 1981-08-03 1981-08-03 Formant tracker
US289604 1981-08-03
US289690 1981-08-03
US06/289,604 US4661915A (en) 1981-08-03 1981-08-03 Allophone vocoder
US289603 1981-08-03

Publications (3)

Publication Number Publication Date
EP0071716A2 true EP0071716A2 (en) 1983-02-16
EP0071716A3 EP0071716A3 (en) 1983-05-11
EP0071716B1 EP0071716B1 (en) 1987-08-26

Family

ID=27403910

Family Applications (1)

Application Number Title Priority Date Filing Date
EP19820105168 Expired EP0071716B1 (en) 1981-08-03 1982-06-14 Allophone vocoder

Country Status (3)

Country Link
EP (1) EP0071716B1 (en)
JP (1) JPS5827200A (en)
DE (1) DE3277095D1 (en)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2547146A1 (en) * 1983-06-02 1984-12-07 Texas Instruments France METHOD AND DEVICE FOR AUDITING SYNTHETIC VOICE MESSAGES AND FOR DISPLAYING CORRESPONDING GRAPHICS MESSAGES
DE3513243A1 (en) * 1985-04-13 1986-10-16 Telefonbau Und Normalzeit Gmbh, 6000 Frankfurt Method for speech transmission and speech storage
FR2642882A1 (en) * 1989-02-07 1990-08-10 Ripoll Jean Louis SPEECH PROCESSING APPARATUS
EP0706172A1 (en) * 1994-10-04 1996-04-10 Hughes Aircraft Company Low bit rate speech encoder and decoder
EP0718819A3 (en) * 1994-12-21 1996-07-10 Hughes Aircraft Co
WO1999040568A1 (en) * 1998-02-03 1999-08-12 Siemens Aktiengesellschaft Method for voice data transmission
US6125284A (en) * 1994-03-10 2000-09-26 Cable & Wireless Plc Communication system with handset for distributed processing
US7353172B2 (en) * 2003-03-24 2008-04-01 Sony Corporation System and method for cantonese speech recognition using an optimized phone set
US7353173B2 (en) * 2002-07-11 2008-04-01 Sony Corporation System and method for Mandarin Chinese speech recognition using an optimized phone set
US7353174B2 (en) * 2003-03-31 2008-04-01 Sony Corporation System and method for effectively implementing a Mandarin Chinese speech recognition dictionary
EP4064280A4 (en) * 2019-11-20 2023-01-11 Vivo Mobile Communication Co., Ltd. Interaction method and electronic device

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4707858A (en) * 1983-05-02 1987-11-17 Motorola, Inc. Utilizing word-to-digital conversion
JPS62231300A (en) * 1986-03-31 1987-10-09 郵政省通信総合研究所長 Automatic zoning of voice processing unit and processing therefor

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5326761A (en) * 1976-08-26 1978-03-13 Babcock Hitachi Kk Injecting device for reducing agent for nox

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
IBM TECHNICAL DISCLOSURE BULLETIN, vol.22, no.4, September 1979, NEW YORK (US) H.R. GRUB: "Dictation System", pages 1327,1328 *
IEEE INTERNATIONAL CONFERENCE ON ACOUSTIC, Speech and Signal Processing, April 10-12, 1978, Tulsa NEW YORK (US) V.N. GUPTA et al.: "Speaker-independent vowel identification in continuous speech", pages 546-548 *
IEEE TRANSACTIONS ON ACOUSTICS, Speech and Signal Processing, vol.ASSP-24, no.2, April 1976 NEW YORK (US) R.B. THOSAR et al.: "An approach towards a synthesis-based speech recognition system", pages 194-196 *
RADIO FERNSEHEN ELEKTRONIK, vol. 29, no.2, 1980 BERLIN (DE) W. TSCHESCHNER et al.: "Technische Systeme der automatischen Sprachverarbeitung", pages 83-87 *
THE JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, vol.42, no.2, August 1967 NEW YORK (US) D.R. REDDY: "Computer Recognition of Connected Speech", pages 329-347 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR2547146A1 (en) * 1983-06-02 1984-12-07 Texas Instruments France METHOD AND DEVICE FOR AUDITING SYNTHETIC VOICE MESSAGES AND FOR DISPLAYING CORRESPONDING GRAPHICS MESSAGES
EP0128093A1 (en) * 1983-06-02 1984-12-12 TEXAS INSTRUMENTS FRANCE Société dite: A general technique to add multilingual speech to videotext systems at a low data rate
DE3513243A1 (en) * 1985-04-13 1986-10-16 Telefonbau Und Normalzeit Gmbh, 6000 Frankfurt Method for speech transmission and speech storage
FR2642882A1 (en) * 1989-02-07 1990-08-10 Ripoll Jean Louis SPEECH PROCESSING APPARATUS
WO1990009656A1 (en) * 1989-02-07 1990-08-23 Alcept Speech processing machine
US6125284A (en) * 1994-03-10 2000-09-26 Cable & Wireless Plc Communication system with handset for distributed processing
US5832425A (en) * 1994-10-04 1998-11-03 Hughes Electronics Corporation Phoneme recognition and difference signal for speech coding/decoding
EP0706172A1 (en) * 1994-10-04 1996-04-10 Hughes Aircraft Company Low bit rate speech encoder and decoder
EP0718819A3 (en) * 1994-12-21 1996-07-10 Hughes Aircraft Co
WO1999040568A1 (en) * 1998-02-03 1999-08-12 Siemens Aktiengesellschaft Method for voice data transmission
US6304845B1 (en) 1998-02-03 2001-10-16 Siemens Aktiengesellschaft Method of transmitting voice data
US7353173B2 (en) * 2002-07-11 2008-04-01 Sony Corporation System and method for Mandarin Chinese speech recognition using an optimized phone set
US7353172B2 (en) * 2003-03-24 2008-04-01 Sony Corporation System and method for cantonese speech recognition using an optimized phone set
US7353174B2 (en) * 2003-03-31 2008-04-01 Sony Corporation System and method for effectively implementing a Mandarin Chinese speech recognition dictionary
EP4064280A4 (en) * 2019-11-20 2023-01-11 Vivo Mobile Communication Co., Ltd. Interaction method and electronic device

Also Published As

Publication number Publication date
DE3277095D1 (en) 1987-10-01
JPH0576040B2 (en) 1993-10-21
EP0071716B1 (en) 1987-08-26
JPS5827200A (en) 1983-02-17
EP0071716A3 (en) 1983-05-11

Similar Documents

Publication Publication Date Title
US4661915A (en) Allophone vocoder
US4424415A (en) Formant tracker
Zwicker et al. Automatic speech recognition using psychoacoustic models
EP0140777B1 (en) Process for encoding speech and an apparatus for carrying out the process
US7035794B2 (en) Compressing and using a concatenative speech database in text-to-speech systems
US6529866B1 (en) Speech recognition system and associated methods
EP0085543B1 (en) Speech recognition apparatus
EP0504927B1 (en) Speech recognition system and method
US4754485A (en) Digital processor for use in a text to speech system
EP0302663B1 (en) Low cost speech recognition system and method
EP0175752B1 (en) Multipulse lpc speech processing arrangement
Syrdal et al. Applied speech technology
EP0071716A2 (en) Allophone vocoder
JP2001166789A (en) Method and device for voice recognition of chinese using phoneme similarity vector at beginning or end
Abe et al. Statistical analysis of bilingual speaker’s speech for cross‐language voice conversion
WO1983002190A1 (en) A system and method for recognizing speech
JPH0215080B2 (en)
Bu et al. Perceptual speech processing and phonetic feature mapping for robust vowel recognition
Wang et al. An experimental analysis on integrating multi-stream spectro-temporal, cepstral and pitch information for mandarin speech recognition
JP3531342B2 (en) Audio processing device and audio processing method
CN111199747A (en) Artificial intelligence communication system and communication method
Wong On understanding the quality problems of LPC speech
Dutono et al. Effects of compound parameters on speaker-independent word recognition
Atal Speech technology in 2001: New research directions
EP0138954B1 (en) Speech pattern processing utilizing speech pattern compression

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

AK Designated contracting states

Designated state(s): DE FR GB NL

PUAL Search report despatched

Free format text: ORIGINAL CODE: 0009013

AK Designated contracting states

Designated state(s): DE FR GB NL

17P Request for examination filed

Effective date: 19831104

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): DE FR GB NL

REF Corresponds to:

Ref document number: 3277095

Country of ref document: DE

Date of ref document: 19871001

ET Fr: translation filed
PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

26N No opposition filed
PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: NL

Payment date: 20010319

Year of fee payment: 20

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20010502

Year of fee payment: 20

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20010531

Year of fee payment: 20

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20010627

Year of fee payment: 20

REG Reference to a national code

Ref country code: GB

Ref legal event code: IF02

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: GB

Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION

Effective date: 20020613

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF EXPIRATION OF PROTECTION

Effective date: 20020614

NLV7 Nl: ceased due to reaching the maximum lifetime of a patent

Effective date: 20020614