JPH04147300A - Speaker's voice quality conversion and processing system - Google Patents

Speaker's voice quality conversion and processing system

Info

Publication number
JPH04147300A
JPH04147300A JP2273088A JP27308890A JPH04147300A JP H04147300 A JPH04147300 A JP H04147300A JP 2273088 A JP2273088 A JP 2273088A JP 27308890 A JP27308890 A JP 27308890A JP H04147300 A JPH04147300 A JP H04147300A
Authority
JP
Japan
Prior art keywords
speaker
voice quality
frequency
voice
spectral envelope
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP2273088A
Other languages
Japanese (ja)
Inventor
Toru Sanada
真田 徹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujitsu Ltd
Original Assignee
Fujitsu Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujitsu Ltd filed Critical Fujitsu Ltd
Priority to JP2273088A priority Critical patent/JPH04147300A/en
Publication of JPH04147300A publication Critical patent/JPH04147300A/en
Pending legal-status Critical Current

Links

Abstract

PURPOSE:To enable high quality voice conversion by executing vectorization of plural rhyme in a frequency axis, obtaining a corresonding path by means of matching according to a dynamic planning method and utilizing this correspondence for frequency conversion. CONSTITUTION:Plural rhymes vectorization means 1 and 2 obtains a spectral envelope at maximum power for each of plural number of rhyme included in voice produced by each speaker. A corresponding path extracting means 3 obtains the corresponding path according to the dynamic planning method by matching the spectral envelope for the two rhyme corresponding to each other by means of the dynamic planning method. A frequency conversion means 4 executes the frequency conversion for each rhyme included in the sound of a speaker A by utilizing the corresponding path. Thus, since nonlinear conversion is executed for the spectral envelope for plural number of rhyme, the voice conversion in higher quality can be obtained.

Description

【発明の詳細な説明】 〔概 要〕 声質の異なる2名の話者の音声について周波数軸上で動
的計西法によるマツチング(以下DPマツチングという
)を行って、対応する動的計画法抽出バス(以下DPバ
バスいう)にもとづいて。
[Detailed Description of the Invention] [Summary] The voices of two speakers with different voice qualities are matched on the frequency axis using the dynamic programming method (hereinafter referred to as DP matching), and the corresponding dynamic programming extraction bus is (hereinafter referred to as DP Babasu).

一方の話者の声質を他方の話者の声質に変換するように
した話者の声質変換処理方式に関し。
This invention relates to a speaker voice quality conversion processing method that converts the voice quality of one speaker to the voice quality of another speaker.

話者の声質の変換に当たって、より高品質での声質変換
を行い得るようにすることを目的とし。
The purpose of this invention is to enable higher-quality voice quality conversion when converting the voice quality of a speaker.

周波数軸上で複数音韻のスペクトルパターンのベクトル
系列でDPマツチングを行い、DPパスを話者間の周波
数軸上での対応関係とし、この対応関係を用いて声質変
換を行うようにし、特に複数音韻のスペクトルパターン
を周波数領域でのベクトル系列として用いる構成として
いる。
DP matching is performed on a vector sequence of spectral patterns of multiple phonemes on the frequency axis, the DP path is defined as a correspondence between speakers on the frequency axis, and voice quality conversion is performed using this correspondence. The spectral pattern is used as a vector sequence in the frequency domain.

〔産業上の利用分野〕[Industrial application field]

声質の異なる2名の話者の音声について周波数軸上でD
Pマツチングを行って、対応するDPババスもとづいて
、一方の話者の声質を他方の話者の声質に変換するよう
にした話者の声質変換処理方式に関する。
D on the frequency axis for the voices of two speakers with different voice qualities
The present invention relates to a speaker voice quality conversion processing method that performs P matching to convert the voice quality of one speaker to the voice quality of another speaker based on the corresponding DP Babas.

このような声質変換処理は、音声認識における話者適応
や音声合成による声質変換について利用される。
Such voice quality conversion processing is used for speaker adaptation in speech recognition and voice quality conversion by voice synthesis.

〔従来の技術〕[Conventional technology]

第2図は従来の場合の構成を示す0図中の符号11は制
御部であってスペクトル包絡を得る処理やDPマツチン
グを行う処理を実行させるべく制御を行うもの、12.
14は夫々パワー最大時点スペクトル包絡抽出部、13
.15は夫々包絡配憶部であって得られたスペクトル包
絡を記憶しておくもの 16は周波数軸DP(ダイナミ
ック・プログラミング)部であって2つのスペクトル包
絡についてDPマツチングを行ってDPババス得るもの
、17は周波数変換表であって得られたDPババスもと
づいて周波数変換を行う上での情報を保持するもの51
8は周波数変換部であって話者Aの音声のスペクトル包
絡を非線形変換するものを表している。
FIG. 2 shows the configuration of a conventional case. Reference numeral 11 in FIG. 2 is a control unit that performs control to execute the process of obtaining a spectrum envelope and the process of performing DP matching; 12.
14 are maximum power point spectrum envelope extraction units; 13
.. 15 is an envelope storage unit that stores the obtained spectrum envelope; 16 is a frequency axis DP (dynamic programming) unit that performs DP matching on two spectrum envelopes to obtain a DP bus; 17 is a frequency conversion table that holds information for frequency conversion based on the obtained DP Babas 51
Reference numeral 8 denotes a frequency conversion unit that nonlinearly converts the spectrum envelope of speaker A's voice.

まず、制御部11で話者Aを選択する0話者Aの単母音
(例えば/a/)が読み込まれて、パワー最大時点スペ
クトル包絡抽出部AI2が単母音(例えば/a/)の発
声のパワー最大時点でのスペクトル包絡を抽出する。こ
のスペクトル包絡が包絡記憶部A13に記憶される。
First, the control unit 11 selects speaker A. A single vowel (for example, /a/) of speaker A is read, and the maximum power point spectrum envelope extraction unit AI2 extracts the utterance of the single vowel (for example, /a/). Extract the spectral envelope at the maximum power point. This spectral envelope is stored in the envelope storage section A13.

次に、制御部11で話者Bを選択する0話者Bの単母音
(例えば/a/)が読み込まれて、パワー最大時点スペ
クトル包絡抽出部B14が話者Aと同じ単母音(例えば
/a/)の発声のパワー最大時点でのスペクトル包絡を
抽出する。このスペクトル包絡が包絡記憶部B15に記
憶される。パワー最大時点スペクトル包絡抽出部B14
と包絡記憶部B15との動作はパワー最大時点スペクト
ル包絡抽出部A12と包絡記憶部A13とに同じである
Next, the control unit 11 reads the monophthong of speaker B (e.g. /a/) for selecting speaker B, and extracts the same monophthong as speaker A (e.g. /a/) at the maximum power point spectral envelope extraction unit B14. The spectrum envelope at the time when the power of the utterance a/) is maximum is extracted. This spectral envelope is stored in the envelope storage section B15. Maximum power point spectrum envelope extraction unit B14
The operations of the envelope storage section B15 and the maximum power point spectrum envelope extraction section A12 and the envelope storage section A13 are the same.

次に制御部11でDPを選択する。この指禾により1周
波数軸DP部16は、包絡記憶部A13と包絡記憶部B
15とから話者Aと話者Bとの同一単母音のスペクトル
包絡を読み込み9周波数領域においてDPマツチングを
行う、このDPパスにより1話者Aと話者Bとの周波数
領域での対応関係が求まる。この周波数領域での対応関
係は周波数変換表17に書き込まれる。
Next, the control unit 11 selects DP. As a result of this instruction, the one frequency axis DP section 16 has an envelope storage section A13 and an envelope storage section B.
15, the spectral envelope of the same single vowel of speaker A and speaker B is read and DP matching is performed in 9 frequency domains. Through this DP pass, the correspondence relationship between speaker A and speaker B in the frequency domain Seek. This correspondence relationship in the frequency domain is written into the frequency conversion table 17.

周波数変換部18は9話者Aのスペクトル包絡が入力さ
れると1周波数変換表17に書き込まれている周波数領
域での対応関係を参照して9話者へのスペクトル包絡を
非線形に伸縮して、声質変換後のスペクトル包絡を出力
する。
When the spectral envelope of the 9 speakers A is input, the frequency conversion unit 18 nonlinearly expands and contracts the spectral envelope of the 9 speakers by referring to the correspondence relationship in the frequency domain written in the 1-frequency conversion table 17. , outputs the spectral envelope after voice quality conversion.

〔発明が解決しようとする課題〕[Problem to be solved by the invention]

従来の用いる音韻が単一音韻(例えば/a/)であった
ために他の音韻での周波数領域での対応関係が大きく異
なる可能性が高かった。そこで。
Since the phoneme used in the past was a single phoneme (for example, /a/), there was a high possibility that the correspondence relationships in the frequency domain for other phonemes would be significantly different. Therefore.

他の音調でも周波数領域での対応関係が適切になる声質
変換方式が必要とされていた。
There was a need for a voice quality conversion method that would provide appropriate correspondence in the frequency domain for other tones.

本発明は5話者の声質の変換”に当たって、より高品質
の声質変換を行い得るようにすることを目的としている
An object of the present invention is to perform voice quality conversion of five speakers with higher quality.

〔課題を解決するための手段〕[Means to solve the problem]

第1図は本発明の原理構成図を示す0図中の符号1.2
は夫々複数音韻スペクトルのベクトル化手段、3は周波
数軸上でのDPマツチングによって対応バスを求める対
応バス抽出手段、4は対応バスによる周波数変換手段を
表す。
Figure 1 shows the principle configuration diagram of the present invention with reference numeral 1.2 in Figure 0.
3 represents a vectorization means for a plurality of phoneme spectra, 3 represents a corresponding bus extraction means for obtaining a corresponding bus by DP matching on the frequency axis, and 4 represents a frequency conversion means using the corresponding bus.

複数音韻スペクトルのベクトル化手段1.2は夫々の話
者の発した音声における複数の音#Iについて夫々パワ
ー最大時点スペクトル包絡を求める。
The plurality of phoneme spectrum vectorization means 1.2 obtains the spectral envelope at the maximum power point for each of the plurality of sounds #I in the speech uttered by each speaker.

対応バス抽出手段は、夫々対応する2つの音韻について
のスペクトル包絡をDPマツチングして。
The corresponding bus extraction means performs DP matching on the spectrum envelopes of two corresponding phonemes.

対応するDPババス求める。そして対応バスによる周波
数変換手段4は2話者Aの音声の中における各音韻につ
いて、対応パスを用いて周波数変換を行う。
Find the corresponding DP bus. Then, the frequency conversion means 4 using the corresponding bus performs frequency conversion on each phoneme in the speech of the two speakers A using the corresponding path.

〔作 用〕[For production]

話者Aの音声を複数音韻スペクトルのベクトル化手段1
に入力してスペクトルパターンによる周波数領域のベク
トル系列を生成する0話者Bの音声を複数音韻スペクト
ルのベクトル化手段2に入力してスペクトルパターンに
よる周波数領域のベクトル系列を生成する0両者の周波
数領域でのベクトル系列を1周波数軸上でのDPマツチ
ングによって対応バスを求める手段3により両者の音声
の周波数II域での対応関係を求める。この周波数領域
での対応関係を用いて、対応パスによる周波数変換手段
4が話者Aの周波数特徴を読み込んで。
Means 1 for vectorizing the speech of speaker A into a plurality of phoneme spectra
The speech of speaker B is input to the multi-phonetic spectrum vectorization means 2 to generate a vector sequence in the frequency domain according to the spectral pattern. The means 3 for determining a corresponding bus by performing DP matching on the vector series on one frequency axis determines the correspondence relationship in the frequency II range of both voices. Using this correspondence relationship in the frequency domain, the frequency conversion means 4 according to the corresponding path reads the frequency characteristics of speaker A.

声質変換後の周波数特徴を出力する。Outputs frequency features after voice quality conversion.

即ち1話者Aと話者Bとの夫々の音韻AI とBiとが
周波数軸上で図示の対応101,102,103・・・
・・・の如きものであったとするとき2話者Aの音fi
Ai上の点Pの周波数を話者Bの音11Bi上の点P′
の周波数に変換される。
That is, the respective phonemes AI and Bi of speaker A and speaker B correspond to each other on the frequency axis as shown in the figure 101, 102, 103, . . .
If the sound is as follows, then the sound fi of speaker A is
The frequency of point P on Ai is expressed as point P' on speaker B's sound 11Bi.
is converted to the frequency of

C実施例〕 第3図は本発明の実施例構成を示す0図中の符号21は
制御部であって第2図図示の制御部11に対応するもの
、22.28は夫々パワー最大時点スペクトル包絡抽出
部、23.29は夫々記憶部選択部であって抽出した夫
々の音韻対応のスペクトル包絡を記憶せしめる場所を選
択するもの。
C Embodiment] FIG. 3 shows the configuration of an embodiment of the present invention. The reference numeral 21 in FIG. 3 is a control section corresponding to the control section 11 shown in FIG. Envelope extraction sections 23 and 29 are storage section selection sections for selecting locations where the extracted spectral envelopes corresponding to each phoneme are to be stored.

24ないし26と30ないし32とは夫々包絡記憶部で
あって第2図図示の包絡記憶部13.15に対応するも
の、27.33は夫々ベクトル生成部であって各包絡記
憶部24ないし26あるいは30ないし32に記憶され
ているスペクトル包絡を取り出して周波数軸DP部34
に供給するもの34は周波数軸DP部であって与えられ
たスペクトル包絡について周波数軸上でのDPマツチン
グを行うもの、35は周波数変換表であって複数の音韻
に対応した変換表を保持するもの、36は周波数変換部
であって複数の音韻ごとに周波数変換を行うものを表し
ている。
Reference numerals 24 to 26 and 30 to 32 are respectively envelope storage units corresponding to the envelope storage units 13 and 15 shown in FIG. 2, and 27 and 33 are vector generation units, respectively. Alternatively, the spectrum envelope stored in 30 to 32 is extracted and the frequency axis DP section 34
34 is a frequency axis DP unit that performs DP matching on the frequency axis for a given spectrum envelope, and 35 is a frequency conversion table that holds conversion tables corresponding to a plurality of phonemes. , 36 represents a frequency conversion unit that performs frequency conversion for each of a plurality of phonemes.

まず、制御部21で話者Aを選択する0話者Aの単母音
の一つ(例えば/a/)が読み込まれてパワー最大時点
スペクトル包絡抽出部A22が。
First, one of the monophthongs of speaker A (for example, /a/) for selecting speaker A is read by the control unit 21, and the spectrum envelope extraction unit A22 selects it at the maximum power point.

この単母音(例えば/a/)の発声のパワー最大時点で
のスペクトル包絡を抽出する。このスペクトル包絡が配
憶部選択部A23により、入力された母音(例えば/a
/)に対応する包絡記憶部A124に言己憶される0話
者Aの単母音の次の一つ(例えば/i/)が読み込まれ
て、パワー最大時点スペクトル包絡抽出部A22が、こ
の単母音(例えば/l/)の発声のパワー最大時点での
スベクトル包絡を抽出する。このスペクトル包絡が記憶
部選択部A23により、入力された母音(例えば/l/
)に対応する包絡記憶部A225に記憶される。同様に
して、他の単母音がn個まで入力されて、包絡記憶部A
n  26までに記憶される。なおここでnは使用する
言語の基本母音数である。
The spectrum envelope at the time when the power of the utterance of this monophthong (for example, /a/) is maximum is extracted. This spectral envelope is selected by the storage section selection section A23 for the input vowel (for example, /a
The next single vowel (for example, /i/) of speaker A stored in the envelope storage unit A124 corresponding to /) is read, and the maximum power point spectrum envelope extraction unit A22 extracts this single vowel. The svector envelope at the time of maximum vocalization power of a vowel (for example, /l/) is extracted. This spectral envelope is determined by the memory selection unit A23 for the input vowel (for example, /l/
) is stored in the envelope storage unit A225. In the same way, up to n other single vowels are input, and the envelope memory unit A
Stored by n 26. Note that n here is the basic number of vowels of the language used.

次に、制御部21で話者Bを選択する。読み込まれる単
母音は話者Bが発声した音声であるが。
Next, the control unit 21 selects speaker B. The monophthong that is read is the voice uttered by speaker B.

これ以降は話者Aが選択された場合の動作と同様である
。各構成部28〜32は上記構成部22〜26に対応す
る。
The subsequent operations are the same as those when speaker A is selected. Each of the components 28 to 32 corresponds to the components 22 to 26 described above.

次に、制御部21でDPマツ7チングを行うことを選択
する。この指示により1周波数軸DP部34は、ベクト
ル生成部A27を通して包絡記憶部A124〜包絡記憶
部An  26からn個のスペクトルパターンを周波数
領域のn次元ベクトル系列に変換して読み込み、ベクト
ル生成部B33を通して包絡記憶部B130〜包絡記憶
部Bn32からn個のスペクトルパターンを周波数領域
のn次元ベクトル系列に変換して読み込み2周波数領域
において夫々DPマツチングを行う、このDPパスによ
り1話者Aと話者Bとの周波数領域での対応関係が求ま
る。この周波数領域での対応関係は周波数変換表35に
書き込まれる。
Next, the control unit 21 selects to perform DP pine seventing. In response to this instruction, the 1-frequency axis DP section 34 converts and reads n spectral patterns from the envelope storage section A124 to the envelope storage section An 26 through the vector generation section A27 into an n-dimensional vector series in the frequency domain, Through this DP path, n spectral patterns are converted into an n-dimensional vector sequence in the frequency domain from the envelope storage unit B130 to the envelope storage unit Bn32, and DP matching is performed in each of the two frequency domains. The correspondence relationship with B in the frequency domain is determined. This correspondence relationship in the frequency domain is written into the frequency conversion table 35.

周波数変換部36は9話者Aのスペクトル包絡が入力さ
れると9周波数変換表35に書き込まれている周波数領
域での対応関係を参照して1話者Aのスペクトル包絡を
非線形に伸縮して、声質変換後のスペクトル包絡を出力
する。
When the spectral envelopes of 9 speakers A are input, the frequency converter 36 non-linearly expands and contracts the spectral envelopes of 1 speaker A with reference to the correspondence relationships in the frequency domain written in the 9 frequency conversion table 35. , outputs the spectral envelope after voice quality conversion.

上記説明において、複数の音韻についてスペクトル包絡
を抽出するとしたが1本発明の場合には(i)複数の定
常母御についてスペクトル包絡を求めたり、(it)複
数の基本母−についてスペクトル包絡を求めたりするこ
とが、より効果的である。しかし勿論それに限られるも
のではない。
In the above explanation, it is assumed that the spectral envelopes are extracted for a plurality of phonemes, but in the case of the present invention, (i) the spectral envelopes are found for a plurality of stationary bases, or (it) the spectral envelopes are found for a plurality of basic bases. It is more effective to However, it is of course not limited to this.

〔発明の効果〕〔Effect of the invention〕

以上説明した如く5本発明によれば、複敞個の音韻につ
いてのスペクトル包絡を非線形変換するようにしている
ために、従来の場合にくらべてより高品質の声質変換を
行うことが可能となる。
As explained above, according to the present invention, since the spectral envelopes of multiple phonemes are nonlinearly transformed, it is possible to perform voice quality conversion with higher quality than in the conventional case. .

【図面の簡単な説明】[Brief explanation of drawings]

第1図は本発明の原理構成図、第2図は従来の場合の構
成、第3図は本発明の実施例構成を示す。 図中、l、2は複数音韻スペクトルのベクトル化手段、
3は周波数軸上でのDP?シチングによフて対応バスを
求める手段、4は対応バスによる周波数変換手段を表す
。 特許出馴人畜十這株式会社 代理人弁理士森1)寛(外2名) 3L来の11八 r屯例檎成
FIG. 1 shows the principle configuration of the present invention, FIG. 2 shows the conventional configuration, and FIG. 3 shows the configuration of an embodiment of the present invention. In the figure, l and 2 are means for vectorizing multiple phoneme spectra;
Is 3 the DP on the frequency axis? 4 represents means for determining a corresponding bus by searching, and 4 represents a frequency conversion means using the corresponding bus. Patent expert Chikujuhi Co., Ltd. agent patent attorney Mori 1) Hiroshi (and 2 others) 3L 118r ton example history

Claims (4)

【特許請求の範囲】[Claims] (1)話者Aの発した読みに対応して当該音韻について
のスペクトル包絡を抽出すると共に、話者Bの発した同
じ読みに対応して当該音韻についてのスペクトル包絡を
抽出するよう構成してなり、 両話者の同一音韻についてのスペクトル包絡を、周波数
領域で動的計画法によるマッチングを行い、当該マッチ
ングにおいて得られた動的計画法抽出パスにもとづいて
、両話者の周波数領域での対応を求めておき、 話者Aの発した音声のスペクトル包絡を、非線形に伸縮
して、話者Bの発した音声のスペクトル包絡に対応した
形に声質を変換する話者の声質変換処理方式において、 上記話者Aと上記話者Bとに対応して複数の音韻につい
ての夫々のスペクトル包絡を抽出する複数音韻スペクト
ルのベクトル化手段(1、2)と、 抽出された夫々の音韻ごとに、周波数領域で動的計画法
によるマッチングを行い、上記動的計画法抽出パスを求
める対応パス抽出手段(3)と、 話者Aの発した音声における上記夫々の音韻に対応する
スペクトル包絡を、上記動的計画法抽出パスにもとづい
て非線形に伸縮する対応パスによる周波数変換手段(4
)とをそなえ、話者Aの発した音声の声質を話者Bの声
質に変換するようにした ことを特徴とする話者の声質変換処理方式。
(1) The spectral envelope of the phoneme is extracted in response to the pronunciation pronounced by speaker A, and the spectral envelope of the phoneme is extracted in response to the same pronunciation pronounced by speaker B. Then, the spectral envelopes of the same phoneme for both speakers are matched using dynamic programming in the frequency domain, and based on the dynamic programming extraction path obtained in the matching, the spectral envelopes of both speakers in the frequency domain are A speaker voice quality conversion processing method that calculates the correspondence, non-linearly expands and contracts the spectral envelope of the voice uttered by speaker A, and converts the voice quality into a form that corresponds to the spectral envelope of the voice uttered by speaker B. a plurality of phoneme spectrum vectorization means (1, 2) for extracting respective spectral envelopes for a plurality of phonemes corresponding to the speaker A and the speaker B; and for each of the extracted phonemes. , a corresponding path extraction means (3) that performs dynamic programming matching in the frequency domain and obtains the dynamic programming extraction path; and a spectral envelope corresponding to each of the phonemes in the speech uttered by speaker A. Frequency conversion means (4
), and converts the voice quality of the voice uttered by speaker A to the voice quality of speaker B.
(2)上記対応パス抽出手段(3)によって得られた各
音韻ごとの動的計画法抽出パスは、周波数変換表(35
)にまとめて格納され保持されてなり、 上記周波数変換手段(4)は、当該周波数変換表(35
)の内容を読み出して利用するようにした ことを特徴とする請求項(1)記載の話者の声質変換処
理方式。
(2) The dynamic programming extraction path for each phoneme obtained by the corresponding path extraction means (3) is calculated using the frequency conversion table (35
), and the frequency conversion means (4) has the frequency conversion table (35
2. The speaker's voice quality conversion processing method according to claim 1, wherein the contents of the speaker's voice are read out and used.
(3)上記周波数変換手段(4)は、話者Aの発した音
声における夫々の音韻に対応したスペクトル包絡を、上
記周波数変換表(35)の内容にもとづいて、非線形に
伸縮するようにしたことを特徴とする請求項(2)記載
の話者の声質変換処理方式。
(3) The frequency conversion means (4) non-linearly expands and contracts the spectral envelope corresponding to each phoneme in the speech uttered by speaker A, based on the contents of the frequency conversion table (35). The speaker's voice quality conversion processing method according to claim (2).
(4)上記複数の音韻が複数個の母音に対応するもので
あることを特徴とする請求項(1)記載の話者の声質変
換処理方式。
(4) The speaker's voice quality conversion processing method according to claim (1), wherein the plurality of phonemes correspond to a plurality of vowels.
JP2273088A 1990-10-11 1990-10-11 Speaker's voice quality conversion and processing system Pending JPH04147300A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2273088A JPH04147300A (en) 1990-10-11 1990-10-11 Speaker's voice quality conversion and processing system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2273088A JPH04147300A (en) 1990-10-11 1990-10-11 Speaker's voice quality conversion and processing system

Publications (1)

Publication Number Publication Date
JPH04147300A true JPH04147300A (en) 1992-05-20

Family

ID=17522976

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2273088A Pending JPH04147300A (en) 1990-10-11 1990-10-11 Speaker's voice quality conversion and processing system

Country Status (1)

Country Link
JP (1) JPH04147300A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001078064A1 (en) * 2000-04-03 2001-10-18 Sharp Kabushiki Kaisha Voice character converting device
WO2002063610A1 (en) * 2001-02-02 2002-08-15 Nec Corporation Voice code sequence converting device and method
JP2006251375A (en) * 2005-03-10 2006-09-21 Yamaha Corp Voice processor and program

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2001078064A1 (en) * 2000-04-03 2001-10-18 Sharp Kabushiki Kaisha Voice character converting device
WO2002063610A1 (en) * 2001-02-02 2002-08-15 Nec Corporation Voice code sequence converting device and method
US7505899B2 (en) 2001-02-02 2009-03-17 Nec Corporation Speech code sequence converting device and method in which coding is performed by two types of speech coding systems
JP2006251375A (en) * 2005-03-10 2006-09-21 Yamaha Corp Voice processor and program
JP4645241B2 (en) * 2005-03-10 2011-03-09 ヤマハ株式会社 Voice processing apparatus and program
US7945446B2 (en) 2005-03-10 2011-05-17 Yamaha Corporation Sound processing apparatus and method, and program therefor

Similar Documents

Publication Publication Date Title
CN112102811B (en) Optimization method and device for synthesized voice and electronic equipment
JPS62231998A (en) Voice synthesization method and apparatus
JPH04147300A (en) Speaker's voice quality conversion and processing system
KR100259777B1 (en) Optimal synthesis unit selection method in text-to-speech system
JP2583074B2 (en) Voice synthesis method
JPH08335096A (en) Text voice synthesizer
JPH06318094A (en) Speech rule synthesizing device
JPH11249679A (en) Voice synthesizer
EP1589524B1 (en) Method and device for speech synthesis
JPH09319394A (en) Voice synthesis method
JP2980382B2 (en) Speaker adaptive speech recognition method and apparatus
JP3503862B2 (en) Speech recognition method and recording medium storing speech recognition program
US6502074B1 (en) Synthesising speech by converting phonemes to digital waveforms
JP3241582B2 (en) Prosody control device and method
JP2839488B2 (en) Speech synthesizer
JP3438293B2 (en) Automatic Word Template Creation Method for Speech Recognition
JP2002358091A (en) Method and device for synthesizing voice
JP2003108170A (en) Method and device for voice synthesis learning
JPH01211799A (en) Regular synthesizing device for multilingual voice
JPH11282484A (en) Voice synthesizer
JP2001117577A (en) Voice synthesizing device
JPH037999A (en) Voice output device
JP2003108180A (en) Method and device for voice synthesis
JPH04298794A (en) Voice data correction system
JP2001249678A (en) Device and method for outputting voice, and recording medium with program for outputting voice