JPH0990972A - Synthesis unit generating method for voice synthesis - Google Patents
Synthesis unit generating method for voice synthesisInfo
- Publication number
- JPH0990972A JPH0990972A JP7248143A JP24814395A JPH0990972A JP H0990972 A JPH0990972 A JP H0990972A JP 7248143 A JP7248143 A JP 7248143A JP 24814395 A JP24814395 A JP 24814395A JP H0990972 A JPH0990972 A JP H0990972A
- Authority
- JP
- Japan
- Prior art keywords
- phoneme
- synthesis
- environment
- synthesis unit
- speech
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Abstract
Description
【0001】[0001]
【発明の属する技術分野】本発明は、音声を人工的に生
成する音声合成における音声合成用合成単位作成方法に
関する。BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a method for creating a speech synthesis unit in speech synthesis for artificially generating speech.
【0002】[0002]
【従来の技術】従来より、規則音声合成方式において
は、「C」を子音、「V」を母音とすると、合成単位と
して、VCV,CV,CVC等が用いられることが多か
った。このような合成単位を使用する方式に比較して、
音声の最も基本的な単位である音素を合成単位とする規
則音声合成方式は、音素を合成単位にすることが自然で
あるにも関わらず、合成時の接続規則が複雑になること
から、あまり利用されていなかった。2. Description of the Related Art Conventionally, in a regular voice synthesis system, when "C" is a consonant and "V" is a vowel, VCV, CV, CVC, etc. are often used as a synthesis unit. Compared to the method using such a composition unit,
The rule-based speech synthesis method, which uses the phoneme, which is the most basic unit of speech, as the synthesis unit, makes the connection rules at the time of synthesis complicated even though it is natural to use phonemes as the synthesis unit. It was not used.
【0003】従来、音素を規則音声合成の合成単位に利
用する場合、単位数が膨大になることから、前後の音韻
環境を全く考慮しないか、あるいは、COC(context-
oriented-clustering )法(詳しくは特願昭62−23
4358号の願書に添付の明細書および図面参照)を適
用するかのいずれかを選択することになる。Conventionally, when a phoneme is used as a synthesis unit for regular speech synthesis, the number of units becomes enormous. Therefore, the preceding and following phonological environments are not considered at all, or the COC (context-
oriented-clustering) method (see Japanese Patent Application No. 62-23 for details)
The application of No. 4358 is to be applied either (see attached description and drawings).
【0004】[0004]
【発明が解決しようとする課題】ところで、前後の音韻
環境を考慮しない音素を規則音声合成の合成単位とした
場合、合成単位数が少ないことや言語による表現上の差
異も少ないことから、多くの言語において利用できると
いう利点があるが、音声特性の変化が特に激しい子音
(C)から母音(V)への移り変わり部分をも規則で接
続するために、高品質な合成音声を得ることは困難であ
った。By the way, when a phoneme that does not consider the preceding and following phonological environments is used as the synthesis unit for the regular speech synthesis, the number of synthesis units is small and the difference in expression depending on the language is small. Although it has the advantage that it can be used in a language, it is difficult to obtain high-quality synthesized speech because the transition part from the consonant (C) to the vowel (V), which has a particularly large change in the speech characteristics, is also connected by a rule. there were.
【0005】一方、COC法は、統計的手法により大量
の音声データの母集団から音韻環境(調音結合)を考慮
した音素をクラスタリングし、そのセントロイドを合成
単位として蓄積する方法である。COC法では、クラス
タリングを自動的に行うため、客観的尺度に基づいて合
成単位が生成されるという利点があるが、クラスタリン
グ母集団の整備に多大な工数がかかる上、合成単位作成
用のテキスト(音声データ)が有限であることから、母
集団中に出現しない音韻環境が存在する可能性がある。On the other hand, the COC method is a method of clustering phonemes considering a phonological environment (articulatory coupling) from a large population of voice data by a statistical method and accumulating the centroid as a synthesis unit. The COC method has an advantage that a synthetic unit is generated based on an objective scale because clustering is automatically performed. However, it takes a lot of man-hours to maintain a clustering population, and a text for creating a synthetic unit ( There is a possibility that there is a phonological environment that does not appear in the population because the (voice data) is finite.
【0006】また、音声合成時に最適な合成単位を大量
の合成単位テーブルから検索するため、テーブル検索の
時間がかかるという欠点がある。さらに、自然な発話状
況から合成単位を生成し蓄積して母集団を構成している
ため、セントロイドの出現頻度によるばらつきが大き
く、音声合成時には、音素環境に対する適合度が小であ
っても最適な合成単位として候補が選択される場合があ
る。したがって、合成音声の品質が不安定であるという
欠点があった。これらの欠点を解決するために、前後の
音韻環境全てを考慮した音素を単位とする方法も考えら
れるが、もちろん、合成単位数が膨大になることから現
実的ではない。Further, since a large amount of synthesis unit tables are searched for the optimum synthesis unit at the time of speech synthesis, there is a disadvantage that the table search takes time. Furthermore, since a synthesis unit is generated from natural utterances and accumulated to form a population, there are large variations due to the frequency of occurrence of centroids, and during speech synthesis it is optimal even if the fitness to the phoneme environment is small. A candidate may be selected as a new synthesis unit. Therefore, there is a drawback that the quality of synthesized speech is unstable. In order to solve these drawbacks, it is possible to consider a method in which a phoneme is taken into consideration in consideration of all the phonetic environments before and after, but it is not realistic because the number of synthesis units becomes huge.
【0007】このように、規則音声合成において、合成
単位を音素とした場合、前後の音韻環境を考慮しない音
素を合成単位とすれば単位数を低減できるが、滑らかな
合成音声を得ることができなかった。また、COC法を
用いた場合には、常時安定した合成音声の生成を期待す
ることはできなかった。さらに、前後の音韻環境の全て
を考慮した音素を合成単位とすれば、滑らかさおよび安
定性は補償されるかもしれないが、合成単位数が膨大に
なり現実的ではない。As described above, in regular speech synthesis, when the synthesis unit is a phoneme, the number of units can be reduced by using a phoneme that does not consider the preceding and following phoneme environments as the synthesis unit, but a smooth synthesized speech can be obtained. There wasn't. Further, when the COC method is used, it is not possible to expect stable generation of synthetic speech at all times. Furthermore, if a phoneme that takes into consideration all of the phonemic environments before and after is used as a synthesis unit, smoothness and stability may be compensated, but the number of synthesis units becomes huge, which is not realistic.
【0008】本発明は上述した事情に鑑みて為されたも
のであり、その第1の目的は、前後の音韻環境を考慮し
た音素の全てを略均一に網羅する合成単位を作成して、
滑らかで安定した合成音声を得ることができる音声合成
用合成単位作成方法を提供することにある。加えて、本
発明の第2の目的は、得られる合成音声の品質を劣化さ
せずに合成単位数を削減することができる音声合成用合
成単位作成方法を提供することにある。The present invention has been made in view of the above circumstances, and a first object of the present invention is to create a synthesis unit that substantially uniformly covers all phonemes in consideration of preceding and following phoneme environments,
It is an object of the present invention to provide a method for creating a synthesis unit for speech synthesis that can obtain smooth and stable synthesized speech. In addition, a second object of the present invention is to provide a synthesis unit creation method for speech synthesis which can reduce the number of synthesis units without deteriorating the quality of the obtained synthesized speech.
【0009】[0009]
【課題を解決するための手段】本発明による音声合成用
合成単位作成方法は、音素を含む音声セグメントを合成
単位として作成する音声合成用合成単位作成方法におい
て、音韻環境に基づいた調音様式の影響を考慮して音声
セグメントを作成し、作成された音声セグメントにおい
て音声学的に等価な音韻環境を共通化して合成単位を作
成することを特徴としている。According to the method for creating a synthesis unit for speech synthesis according to the present invention, in the method for creating a synthesis unit for speech synthesis, which creates a speech segment including a phoneme as a synthesis unit, the influence of the articulation style based on the phonological environment. It is characterized in that a voice segment is created in consideration of the above, and a phoneme environment that is phonetically equivalent is shared in the created voice segment to create a synthesis unit.
【0010】[0010]
【発明の実施の形態】本発明の実施形態について説明す
る前に、本発明の基本的な考え方について説明する。本
発明は、日本語等の言語に現れる前後の音韻環境を考慮
した音声セグメント(実際には音素(トライホン))の
全てを考慮し、それらの中から、音素素性を考慮して音
韻環境を同一視できるものをグループ化して単位数の削
減を図るものである。したがって、音韻環境を考えな
い、あるいは、母集団から統計的手法により合成単位を
生成するといった従来の技術とは明らかに異なってい
る。DESCRIPTION OF THE PREFERRED EMBODIMENTS Before describing the embodiments of the present invention, the basic concept of the present invention will be described. The present invention considers all speech segments (actually, phonemes (triphones)) in consideration of phonemic environments before and after appearing in a language such as Japanese, and selects the phoneme environment from among them in consideration of phoneme characteristics. It aims to reduce the number of units by grouping visible ones. Therefore, it is obviously different from the conventional technique in which the phonological environment is not considered or the synthesis unit is generated from the population by a statistical method.
【0011】以下、図面を参照して、本発明の実施形態
について説明する。図1は本発明の一実施形態による音
声合成用合成単位作成方法を説明するための図であり、
この図に示されるように本実施形態は、ステップS1〜
S4を順に実行することにより為される。なお、図1に
示される音声合成用合成単位作成方法は、日本語の音声
合成に用いられるものであり、以下、各ステップについ
て順に説明する。ただし、以下の説明において、c,
c′,C,C′,Q,v,v′,V,V′,V″はそれ
ぞれ音素を表している。具体的には、v,v′は長母
音、C,C′,c,c′は子音であり、C′はCの無声
化子音、c,c′は無声破裂子音である。また、Qは促
音、V,V′,V″は母音であり、V,V′はそれぞれ
v,v′の短母音を表す。Embodiments of the present invention will be described below with reference to the drawings. FIG. 1 is a diagram for explaining a method for creating a speech synthesis unit according to an embodiment of the present invention.
As shown in this figure, the present embodiment includes steps S1 to S1.
This is done by sequentially executing S4. The method for creating a synthesis unit for speech synthesis shown in FIG. 1 is used for Japanese speech synthesis, and each step will be described below in order. However, in the following description, c,
c ', C, C', Q, v, v ', V, V', V "respectively represent phonemes. Specifically, v, v'are long vowels, C, C ', c, c'is a consonant, C'is an unvoiced consonant of C, c, c'is an unvoiced explosive consonant, Q is a consonant, V, V ', V "are vowels, and V, V'are These represent v and v'short vowels, respectively.
【0012】[ステップS1]ステップS1では、着目
している音素の先行音素が長母音の場合に、当該音素の
音韻環境を短母音が先行音素(直前の音素)である音韻
環境であるものとみなす。すなわち、先行音素の音韻環
境において長母音を短母音と同一視する。日本語に存在
するトライホンの全てを合成単位とすると、促音を含め
てその単位数は約15,000種であり、膨大な数にな
ってしまう。そこで、ステップS1において、音素素性
を考慮して、調音結合などの音韻環境について同一視可
能なものを共通化(同一視)し、合成単位数の削減を図
る。以下に、ステップS1で行われる同一視の例を示
す。 例1:vCV′のCをVCV′のCと同一の合成単位と
みなす。 例2:vCC′のCをVCC′のCと同一の合成単位と
みなす。 例3:v′vCのvをV′vCのvと同一の合成単位と
みなす。 上記例1〜3のような規則によって、約3000個のト
ライホンが削減される。[Step S1] In step S1, when the preceding phoneme of the phoneme of interest is a long vowel, the phoneme environment of the phoneme is a phoneme environment in which a short vowel is a preceding phoneme (previous phoneme). I reckon. That is, long vowels are identified with short vowels in the phoneme environment of the preceding phoneme. If all the triphones existing in Japanese are used as synthesis units, the number of units including consonants is about 15,000, which is an enormous number. Therefore, in step S1, in consideration of phoneme features, common identifiable phoneme environments such as articulatory combinations are identified (identified) to reduce the number of synthesis units. Below, the example of the same identification performed in step S1 is shown. Example 1: C of vCV 'is considered to be the same synthetic unit as C of VCV'. Example 2: Consider the C of vCC 'as the same synthetic unit as the C of VCC'. Example 3: Consider the v of v'vC as the same synthetic unit as the v of V'vC. The rules of Examples 1-3 above reduce about 3000 triphones.
【0013】[ステップS2]ステップS2では、当該
音素の後続音素(直後の音素)が長母音の場合に、当該
音素の音韻環境を短母音が後続音素である音韻環境であ
るものとみなす。すなわち、後続音素の音韻環境におい
て長母音を短母音と同一視する。以下に、ステップS2
で行われる同一視の例を示す。 例4:V′CvのCをV′CVのCと同一の合成単位と
みなす。 例5:CVv′のVをCVV′のVと同一の合成単位と
みなす。 例6:V″vv′のvをV″vV′のvと同一の合成単
位とみなす。 上記例4〜6のような規則によって、約2000個のト
ライホンが削減される。[Step S2] In step S2, when the subsequent phoneme (immediate phoneme) of the phoneme is a long vowel, the phoneme environment of the phoneme is regarded as a phoneme environment in which a short vowel is a subsequent phoneme. That is, long vowels are identified with short vowels in the phoneme environment of the subsequent phoneme. Below, step S2
An example of the same identification performed in FIG. Example 4: Consider the C of V'Cv as the same synthetic unit as the C of V'CV. Example 5: V of CVv 'is considered to be the same synthetic unit as V of CVV'. Example 6: The v of V "vv 'is considered to be the same synthetic unit as the v of V"vV'. The rules of Examples 4-6 above reduce about 2000 triphones.
【0014】[ステップS3]ステップS3では、当該
音素が無声破裂音の場合にその先行音素を一つの短母音
で代表させる。以下に、ステップS3で行われる同一視
の例を示す。 例7:VcV″のcをV′CV″のcと同一の合成単位
とみなす。 例8:VcC(cの無声化)のcをV′cCのcと同一
の合成単位とみなす。 例9:QcVのcをV′cVのcと同一の合成単位とみ
なす。 なお、例7〜9において、V′は代表させる短母音を表
す。上記例7〜9のような規則によって、約1000個
のトライホンが削減される。[Step S3] In step S3, when the phoneme is an unvoiced plosive, the preceding phoneme is represented by one short vowel. Below, an example of the identification performed in step S3 is shown. Example 7: c of VcV ″ is considered to be the same synthetic unit as c of V′CV ″. Example 8: C in VcC (unvoiced c) is considered to be the same synthesis unit as c in V'cC. Example 9: c of QcV is regarded as the same synthetic unit as c of V'cV. In Examples 7 to 9, V'represents a representative short vowel. Rules such as Examples 7-9 above reduce about 1000 triphones.
【0015】[ステップS4]ステップS4では、当該
音素の後続音素が無声破裂音の場合にその後続音素を一
つの無声破裂音で代表させる。 例10:CVcのVをCVc′のVと同じ合成単位とみ
なす。 例11.VCc(Cの無声化)のCをVCc′のCと同
一の合成単位とみなす。 なお、例10,11において、c′は代表させる無声破
裂子音を表す。上記例10,11のような規則によっ
て、約3000個のトライホンが削減される。[Step S4] In step S4, when the succeeding phoneme of the phoneme is an unvoiced plosive sound, the succeeding phoneme is represented by one unvoiced plosive sound. Example 10: V of CVc is considered to be the same synthetic unit as V of CVc '. Example 11 FIG. C of VCc (unvoiced C) is regarded as the same synthesis unit as C of VCc '. In Examples 10 and 11, c'represents a representative unvoiced explosive consonant. Rules such as Examples 10 and 11 above reduce about 3000 triphones.
【0016】結局、上述したステップS1〜S4の全て
によってグループ化を行なうと、約9000個のトライ
ホンが削減され、合成単位となるトライホンは約600
0種になる。以上、本発明の実施形態を図面を参照して
詳述してきたが、具体的な同一視規則は、上述した実施
形態に例示されたものに限られるものではなく、本発明
の要旨を逸脱しない範囲での変更等があっても本発明に
含まれる。After all, when grouping is performed by all of the above-mentioned steps S1 to S4, about 9000 triphones are reduced, and about 600 triphones are combined.
It becomes 0 species. Although the embodiments of the present invention have been described in detail above with reference to the drawings, the specific identification rules are not limited to those exemplified in the above-described embodiments, and do not depart from the gist of the present invention. Even if there is a change in the range, it is included in the present invention.
【0017】[0017]
【発明の効果】以上説明したように、本発明によれば、
所定の言語に現れる全てのトライホンを考慮するように
したため、出現頻度の大小に関わらず、均一で安定した
合成単位を蓄積・構成することができる。また、隣接す
る(前後の)音韻環境をグループ化することにより、合
成単位数を大幅に削減することができる。例えば、日本
語においては、音素素性によって合成単位数を2/5に
大幅削減でき、しかも日本語に存在する全てのトライホ
ンにほぼ均一に対応することができる。また、隣接する
(前後の)音韻環境を考慮した音声セグメント(音素)
を合成単位とするため、音声合成時に滑らかな連結を可
能とし、自然性に優れた音声を合成することができる。As described above, according to the present invention,
Since all triphones appearing in a predetermined language are taken into consideration, uniform and stable synthesis units can be accumulated and configured regardless of the frequency of appearance. In addition, the number of synthesis units can be significantly reduced by grouping adjacent (front and back) phoneme environments. For example, in Japanese, the number of synthesis units can be greatly reduced to 2/5 depending on the phoneme characteristics, and moreover, all triphones existing in Japanese can be handled substantially uniformly. Also, speech segments (phonemes) that take into consideration adjacent (preceding and following) phonological environments.
Is used as a synthesizing unit, it is possible to smoothly connect the voices when synthesizing the voices, and it is possible to synthesize a voice having excellent naturalness.
【図1】本発明の一実施形態による音声合成用合成単位
作成方法を説明するための図である。FIG. 1 is a diagram illustrating a method of creating a synthesis unit for speech synthesis according to an embodiment of the present invention.
───────────────────────────────────────────────────── フロントページの続き (72)発明者 伊藤 憲三 東京都千代田区内幸町一丁目1番6号 日 本電信電話株式会社内 ─────────────────────────────────────────────────── ─── Continued Front Page (72) Inventor Kenzo Ito 1-1-6 Uchisaiwaicho, Chiyoda-ku, Tokyo Nihon Telegraph and Telephone Corporation
Claims (1)
して作成する音声合成用合成単位作成方法において、 音韻環境に基づいた調音様式の影響を考慮して音声セグ
メントを作成し、作成された音声セグメントにおいて音
声学的に等価な音韻環境を共通化して合成単位を作成す
ることを特徴とする音声合成用合成単位作成方法。1. A method for creating a synthesis unit for speech synthesis, which creates a speech segment containing a phoneme as a synthesis unit, creates a speech segment in consideration of the influence of an articulation style based on a phoneme environment, and creates a speech segment in the created speech segment. A method for creating a synthesis unit for speech synthesis, which comprises creating a synthesis unit by commonizing phonetically equivalent phonological environments.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP24814395A JP3275940B2 (en) | 1995-09-26 | 1995-09-26 | Creating synthesis units for speech synthesis |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP24814395A JP3275940B2 (en) | 1995-09-26 | 1995-09-26 | Creating synthesis units for speech synthesis |
Publications (2)
Publication Number | Publication Date |
---|---|
JPH0990972A true JPH0990972A (en) | 1997-04-04 |
JP3275940B2 JP3275940B2 (en) | 2002-04-22 |
Family
ID=17173871
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
JP24814395A Expired - Lifetime JP3275940B2 (en) | 1995-09-26 | 1995-09-26 | Creating synthesis units for speech synthesis |
Country Status (1)
Country | Link |
---|---|
JP (1) | JP3275940B2 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004139033A (en) * | 2002-09-25 | 2004-05-13 | Nippon Hoso Kyokai <Nhk> | Voice synthesizing method, voice synthesizer, and voice synthesis program |
US7603278B2 (en) | 2004-09-15 | 2009-10-13 | Canon Kabushiki Kaisha | Segment set creating method and apparatus |
-
1995
- 1995-09-26 JP JP24814395A patent/JP3275940B2/en not_active Expired - Lifetime
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004139033A (en) * | 2002-09-25 | 2004-05-13 | Nippon Hoso Kyokai <Nhk> | Voice synthesizing method, voice synthesizer, and voice synthesis program |
JP4532862B2 (en) * | 2002-09-25 | 2010-08-25 | 日本放送協会 | Speech synthesis method, speech synthesizer, and speech synthesis program |
US7603278B2 (en) | 2004-09-15 | 2009-10-13 | Canon Kabushiki Kaisha | Segment set creating method and apparatus |
Also Published As
Publication number | Publication date |
---|---|
JP3275940B2 (en) | 2002-04-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7418389B2 (en) | Defining atom units between phone and syllable for TTS systems | |
JP4302788B2 (en) | Prosodic database containing fundamental frequency templates for speech synthesis | |
US8566099B2 (en) | Tabulating triphone sequences by 5-phoneme contexts for speech synthesis | |
US6778962B1 (en) | Speech synthesis with prosodic model data and accent type | |
US6978239B2 (en) | Method and apparatus for speech synthesis without prosody modification | |
Donovan et al. | A hidden Markov-model-based trainable speech synthesizer | |
EP1721311A1 (en) | Text-to-speech method and system, computer program product therefor | |
Black et al. | Optimal data selection for unit selection synthesis. | |
Balyan et al. | Speech synthesis: a review | |
Nose et al. | Sentence selection based on extended entropy using phonetic and prosodic contexts for statistical parametric speech synthesis | |
Olive | A new algorithm for a concatenative speech synthesis system using an augmented acoustic inventory of speech sounds. | |
Pradhan et al. | Building speech synthesis systems for Indian languages | |
Mullah | A comparative study of different text-to-speech synthesis techniques | |
JP2583074B2 (en) | Voice synthesis method | |
US20080077407A1 (en) | Phonetically enriched labeling in unit selection speech synthesis | |
Mustafa et al. | Developing an HMM-based speech synthesis system for Malay: a comparison of iterative and isolated unit training | |
JP3275940B2 (en) | Creating synthesis units for speech synthesis | |
US6847932B1 (en) | Speech synthesis device handling phoneme units of extended CV | |
Anushiya Rachel et al. | A small-footprint context-independent HMM-based synthesizer for Tamil | |
JP5175422B2 (en) | Method for controlling time width in speech synthesis | |
Lobanov et al. | Development of multi-voice and multi-language TTS synthesizer (languages: Belarussian, Polish, Russian) | |
Suzić et al. | Style-code method for multi-style parametric text-to-speech synthesis | |
JP3318290B2 (en) | Voice synthesis method and apparatus | |
Mustafa et al. | A Cross-Lingual Approach to the Development of an HMM-Based Speech Synthesis System for Malay. | |
EP1777697B1 (en) | Method for speech synthesis without prosody modification |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20090208 Year of fee payment: 7 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20090208 Year of fee payment: 7 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20100208 Year of fee payment: 8 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20110208 Year of fee payment: 9 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20110208 Year of fee payment: 9 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20120208 Year of fee payment: 10 |
|
FPAY | Renewal fee payment (event date is renewal date of database) |
Free format text: PAYMENT UNTIL: 20130208 Year of fee payment: 11 |
|
S531 | Written request for registration of change of domicile |
Free format text: JAPANESE INTERMEDIATE CODE: R313531 |
|
R350 | Written notification of registration of transfer |
Free format text: JAPANESE INTERMEDIATE CODE: R350 |
|
EXPY | Cancellation because of completion of term |