JPH0612091A - Japanese speech recognizing method - Google Patents

Japanese speech recognizing method

Info

Publication number
JPH0612091A
JPH0612091A JP4170898A JP17089892A JPH0612091A JP H0612091 A JPH0612091 A JP H0612091A JP 4170898 A JP4170898 A JP 4170898A JP 17089892 A JP17089892 A JP 17089892A JP H0612091 A JPH0612091 A JP H0612091A
Authority
JP
Japan
Prior art keywords
character
likelihood
kana
kanji
japanese
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP4170898A
Other languages
Japanese (ja)
Inventor
Shoichi Matsunaga
昭一 松永
Toshiaki Tsuboi
俊明 坪井
Tomokazu Yamada
智一 山田
Kiyohiro Kano
清宏 鹿野
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nippon Telegraph and Telephone Corp
Original Assignee
Nippon Telegraph and Telephone Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nippon Telegraph and Telephone Corp filed Critical Nippon Telegraph and Telephone Corp
Priority to JP4170898A priority Critical patent/JPH0612091A/en
Publication of JPH0612091A publication Critical patent/JPH0612091A/en
Pending legal-status Critical Current

Links

Abstract

PURPOSE:To convert an input speech to a character string mixed with KANA (Japanese syllabary) and KANJI (Chinese characters) at the high recognition ratio. CONSTITUTION:Concerning the first character of each word, similarity between the input speech and the standard pattern of phonemes is obtained by using Japanese clause syntax 6 and a word dictionary 7 showing the reading, tolerance to first generate that character is obtained, and the sum of the similarity and the tolerance is defined as total tolerance. While remaining characters having the high total tolerance, the second character having the high sum of the similarity and tolerance is similarily searched and concerning the obtained word having the high total tolerance, the similarity with the input speech and the tolerance of the generation are respectively obtained from the structure of noun and postpositional particle concerning the respective candidates of the postpositional particle coming next to the noun and afterwards, the obtained clause having the high total tolerance as a whole is obtained and defined as the recognized result.

Description

【発明の詳細な説明】Detailed Description of the Invention

【0001】[0001]

【産業上の利用分野】この発明は、例えば隠れマルコフ
モデル(例えば、中川聖一「確率モデルによる音声認
識」電子情報通信学会編(1988))、統計的言語モ
デル(例えば、Bahl,L.他“A Statist
ical Approach to Continuo
us Speech Recognition”IEE
E Trans.on PAMI(1983))のよう
な標準パターンと、日本語統計モデル(例えば、山田他
「かな漢字の文字連鎖情報を利用した統計的言語モデ
ル」電子情報通信学会技術報告、SP91−26(19
91))とを用いた日本語音声認識方法に関するもので
ある。
BACKGROUND OF THE INVENTION The present invention is applicable to, for example, a hidden Markov model (for example, Seiichi Nakagawa "Speech recognition by a probabilistic model" edited by The Institute of Electronics, Information and Communication Engineers (1988)), a statistical language model (for example, Bahl, L. et al. "A Statist
iCal Approach to Continuo
us Speech Recognition "IEE
E Trans. on PAMI (1983)) and a Japanese statistical model (for example, Yamada et al., "Statistical Language Model Using Kana-Kanji Character Chain Information," IEICE Technical Report, SP91-26 (19).
91)) and a Japanese speech recognition method.

【0002】[0002]

【従来の技術】従来の隠れマルコフモデル及び統計的言
語モデルを用いた音声認識方法として、学習用テキスト
データベースより、音素の生起順序に関する統計的言語
モデルと、隠れマルコフモデルの音素標準パターンとを
予め作成しておき、入力音声に対し、統計的言語モデル
を用いて、既に認識した直前の複数の音素から、次に生
起する確率の高い複数の音素候補を選出し、これら選出
した音素候補のそれぞれについて、その音素標準パター
ンと入力音声とを照合して、生起尤度と、標準パターン
との類似度との総合的尤度の最も高い音素を認識結果と
して出力することが提案されている。
2. Description of the Related Art As a conventional speech recognition method using a hidden Markov model and a statistical language model, a statistical language model regarding a phoneme occurrence order and a phoneme standard pattern of the hidden Markov model are previously stored in a learning text database. For the input speech, a statistical language model is used to select a plurality of phoneme candidates that have a high probability of occurring next from the plurality of phonemes that have already been recognized, and each of the selected phoneme candidates is selected. It has been proposed that the phoneme standard pattern is collated with the input speech, and the phoneme with the highest total likelihood of the occurrence likelihood and the similarity with the standard pattern is output as the recognition result.

【0003】[0003]

【発明が解決しようとする課題】しかし、この認識方法
は認識結果が、音素単位の系列として出力されるから、
入力音声を日本語文として出力したい場合は、その認識
結果の音素単位の系列を、仮名漢字変換する必要があ
る。つまり、入力音声を音素単位の系列への変換と、そ
の音素系列の仮名漢字系列への変換との2回の変換を行
うため、全体として正しい変換結果が得られる変換性能
が比較的低いものとならざるを得ない。
However, in this recognition method, since the recognition result is output as a sequence of phoneme units,
To output the input speech as a Japanese sentence, it is necessary to convert the phoneme-based sequence of the recognition result into kana-kanji conversion. That is, since the input speech is converted into a sequence of phoneme units and the phoneme sequence is converted into a kana-kanji sequence, the conversion performance is relatively low in that a correct conversion result can be obtained as a whole. I have no choice.

【0004】[0004]

【課題を解決するための手段】この発明によれば、学習
用テキストデータベースから読みを振った仮名及び漢字
の生起順序に関する統計的言語モデルを作成しておき、
また、文節の構文と、読みを振った単語辞書、読みを振
った文字を音素記号に変換する文字音素変換規則、及び
音素標準パターンを作成しておき、この構文と単語辞書
から導かれる認識候補の、仮名漢字の統計モデルと音素
標準パターンを用いて、入力音声を一挙に、仮名漢字混
じり文字系列に変換する。
According to the present invention, a statistical language model relating to the occurrence order of kana and kanji whose readings are changed is created from a learning text database,
In addition, bunsetsu syntax, word dictionary with pronunciation, phoneme conversion rules for converting characters with pronunciation into phoneme symbols, and phoneme standard patterns are created in advance, and recognition candidates derived from this syntax and word dictionary are created. Using the statistical model of Kana-Kanji and the phoneme standard pattern, the input voice is converted into a mixed Kana-Kanji character sequence.

【0005】[0005]

【実施例】図1にこの発明の実施例を示す。入力端子1
から入力された音声は、特徴抽出部2においてディジタ
ル信号に変換され、更にLPCケプストラム分析された
後、1フレーム(1時点、例えば10ミリ秒)ごとに特
徴パラメータに変換される。この特徴パラメータは例え
ばLPCケプストラム係数である。
FIG. 1 shows an embodiment of the present invention. Input terminal 1
The voice input from is converted into a digital signal in the feature extraction unit 2, further subjected to LPC cepstrum analysis, and then converted into a feature parameter every one frame (one time point, for example, 10 milliseconds). This characteristic parameter is, for example, an LPC cepstrum coefficient.

【0006】学習用音声データベースより、上記特徴パ
ラメータと同一形式で、隠れマルコフモデルの音素標準
パターンを作り、標準パターンメモリ4に記憶してお
き、また文字の読みを振った学習用テキストデータベー
スより、読みを振った文字の生起順序に関する統計的言
語モデルを作り統計的言語モデルメモリ5に記憶してあ
る。同様に、上記学習用テキストデータベースから、日
本語文節の構造が単語の遷移規則として記述されている
文節構文6、読みを振った単語辞書7(例えば図2
A)、読みを振った仮名及び漢字を音素の記号列に変換
する文字音素変換規則8(例えば図2B)を作成してそ
れぞれ記憶しておく。なお、仮名に対する読みの振りと
は例えば「は」は「ha」または「wa」の何れを振る
認識部3では、文節構文6と読みを振った単語辞書7と
を用いて選出した複数の文字候補について、その文字候
補の文字音素変換規則8により得られた音素の標準パタ
ーンを標準パターンメモリ4から読みだし、入力音声の
特徴パラメータの類似度(尤度)を求める。また上記文
字候補の生起尤度を統計的言語モデルより求める。
From the learning voice database, a phoneme standard pattern of the hidden Markov model is created in the same format as the above-mentioned characteristic parameters, stored in the standard pattern memory 4, and from the learning text database in which the reading of the characters is given, A statistical language model relating to the order of occurrence of read characters is created and stored in the statistical language model memory 5. Similarly, from the above learning text database, a bunsetsu syntax 6 in which the structure of a Japanese bunsetsu is described as a word transition rule, and a word dictionary 7 with reading (for example, FIG. 2).
A), a character-phoneme conversion rule 8 (for example, FIG. 2B) that converts kana and kanji with pronunciations into phoneme symbol strings is created and stored. It is to be noted that the pronunciation of the pronunciation with respect to the kana is, for example, “ha” is either “ha” or “wa”. In the recognition unit 3, a plurality of characters selected by using the clause syntax 6 and the word dictionary 7 with the pronunciation For a candidate, the standard pattern of phonemes obtained by the character-phoneme conversion rule 8 of the character candidate is read from the standard pattern memory 4, and the similarity (likelihood) of the characteristic parameter of the input speech is obtained. Also, the likelihood of occurrence of the above character candidates is obtained from a statistical language model.

【0007】つまり例えば入力音声のi番目の文字を認
識するには、統計的言語モデルから読みを振った文字
(仮名、漢字)の出現順序に関するトライグラムを用い
て(i−2)番目と(i−1)番目との各文字の認識結
果を基に、i番目に出現する文字の生起尤度を各文字に
ついて求め、その文字の音素列の標準パターンとの類似
性を示す尤度とその文字の生起尤度との和を総合尤度と
する。
That is, for example, in order to recognize the i-th character of the input speech, a trigram concerning the appearance order of the characters (kana and kanji) whose pronunciation is changed from the statistical language model is used to identify the (i-2) -th character. Based on the recognition result of each of the i-1) th character, the occurrence likelihood of the i-th character that appears is calculated for each character, and the likelihood indicating the similarity with the standard pattern of the phoneme sequence of that character and its The sum of the likelihood of occurrence of a character is the total likelihood.

【0008】この読みを振った仮名・漢字候補の選出
と、それらについての標準パターンとの照合と、その総
合尤度から認識結果文字を得る操作とを音声区間が終わ
るまで繰り返し、最後に、それまで得られた認識結果文
字を、認識結果出力部9に送り、その順に仮名、漢字系
列として出力する。次に認識手順の具体例を述べる。例
えば単語辞書7の各単語の最初の文字について入力音声
の音素列と標準パターンとの類似性を求める。即ち、 1.例えば単語辞書7中の単語の最初の文字 東(と
う)より /tou/の標準パターンと入力音声とのマ
ッチングをとり類似性尤度P1を得る。最初に文字 東
(とう)が生起する尤度L1より 東(とう)の総合尤
度Q1=P1+L1を得る。 2.単語辞書7中の単語の最初の文字 山(やま)より
/yama/の標準パターンと入力音声とのマッチン
グをとり類似性尤度P2を得る。最初に文字山(やま)
が生起する尤度L2より 山(やま)の総合尤度Q2=
P2+L2を得る。 3.このようにして各単語の最初の文字との総合尤度を
求め、総合尤度の高い順次に例えば10個を残しそれら
についてその単語の2番目の文字について、例えば東の
次の文字として 京(きょう)より /kyou/の標
準パターンと入力音声とのマッチングをとり類似性尤度
P3を得る。文字 東(とう)の後に 京(きょう)が
生起する尤度L3より 東(とう)京(きょう)の総合
尤度Q3=Q1+P3+L3を得る。 4.山の次の文字として 形(がた)をとり /gat
a/の標準パターンと入力音声とのマッチングをとり類
似性尤度P4を得る。文字 山(やま)の後に形(が
た)が生起する尤度L4より 山(やま)形(がた)の
総合尤度Q4=Q2+P4+L4を得る。 5.そのようにして総合尤度の高いものを成し、日本語
の構文6における名詞+助詞の構造の関係から、その助
詞の候補を辞書から求め、例えば は(わ)より /w
a/の標準パターンと入力音声とのマッチングをとり類
似性尤度P5を得る。文字 東(とう)京(きょう)の
後に は(わ)が生起する尤度L5より文節 東(と
う)京(きょう)は(わ)の総合尤度Q5=Q3+P5
+L5を得る。 6.同様に候補 の より /no/の標準パターンと
入力音声とのマッチングをとり類似性尤度P6を得る。
文字 東(とう)京(きょう)の後に の が生起する
尤度L6より文節 東(とう)京(きょう)の の総合
尤度Q6=Q3+P6+L6を得る。 7.同様に候補 が より /ga/との類似性尤度P
7を得る。文字 山(やま)形(がた)の後に が が
生起する尤度L7より文節 山(やま)形(がた)が
の総合尤度Q7=Q4+P7+L7を得る。 8.また の より /no/との類似性尤度P8を得
る。文字 山(やま)形(がた)の後に の が生起す
る尤度L8より文節 山(やま)形(がた)のの総合尤
度Q8=Q4+P8+L8を得る。 9.このようにして得られた文節の候補 東(とう)京(きょう)は(わ) 東(とう)京(きょう)の 山(やま)形(がた)が 山(やま)形(がた)の の総合尤度Q5,Q6,Q7,Q8を比較する。 10.例えばQ6が最も高ければ認識結果は 東(と
う)京(きょう)の となり、東京の を認識結果とし
て出力する。なお、特徴抽出部2、認識部3、認識結果
出力部9はそれぞれ、専用、または兼用のマイクロプロ
セッサにより処理することができる。
The selection of kana / kanji candidates with this reading, the matching with the standard patterns for them, and the operation of obtaining the recognition result character from the total likelihood thereof are repeated until the end of the voice section, and finally, that The recognition result characters obtained up to this point are sent to the recognition result output unit 9, and are output in that order as a kana and kanji sequence. Next, a specific example of the recognition procedure will be described. For example, for the first character of each word in the word dictionary 7, the similarity between the phoneme sequence of the input speech and the standard pattern is obtained. That is, 1. For example, the similarity likelihood P1 is obtained by matching the standard pattern of / tou / with the input voice from the first character east of the word in the word dictionary 7. First, the total likelihood Q1 = P1 + L1 of the eastern part is obtained from the likelihood L1 of the character eastal part. 2. From the first character mountain of the word in the word dictionary 7, a standard pattern of / yama / is matched with the input voice to obtain the similarity likelihood P2. First, the mountain
Than the likelihood L2 that occurs in the mountain total likelihood Q2 =
Obtain P2 + L2. 3. In this way, the total likelihood of each word with the first character is obtained, and, for example, 10 pieces are left in order of high total likelihood, and the second character of that word, for example, the next character in the east, is set to K ( Today, the standard pattern of / kyou / and the input voice are matched to obtain the similarity likelihood P3. The total likelihood Q3 = Q1 + P3 + L3 of the eastern Kyo is obtained from the likelihood L3 that the Kyo occurs after the character east. 4. Take the shape as the next character of the mountain / gat
The similarity pattern P4 is obtained by matching the standard pattern of a / with the input voice. From the likelihood L4 that the shape (gata) occurs after the character mountain (yama), the overall likelihood Q4 = Q2 + P4 + L4 of the mountain shape (gata) is obtained. 5. In this way, one with a high overall likelihood is obtained, and from the relationship of the structure of the noun + particle in Japanese syntax 6, the candidate for that particle is obtained from the dictionary.
The similarity pattern P5 is obtained by matching the standard pattern of a / with the input voice. From the likelihood L5 that (wa) occurs after the letter east Kyoto, the total likelihood of the sentence east (to) Kyoto is (wa) Q5 = Q3 + P5
Get + L5. 6. Similarly, the similarity pattern P6 is obtained by matching the standard pattern of / no / of the candidate and the input voice.
From the likelihood L6 that occurs after the character Tokyo, we obtain the overall likelihood Q6 = Q3 + P6 + L6 of the phrase Tokyo. 7. Similarly, the candidate is better than / ga /
Get 7. From the likelihood L7 that a character appears after a mountain shape, the Bunshi mountain shape is
To obtain the total likelihood Q7 = Q4 + P7 + L7. 8. Moreover, the similarity likelihood P8 with / no / is obtained. From the likelihood L8 that occurs after the character mountain shape, the total likelihood Q8 = Q4 + P8 + L8 of the mountain shape is obtained. 9. Candidates for the bunsetsu obtained in this way are in the eastern Kyo (wa) eastern Kyo (mountain) shape. The overall likelihoods Q5, Q6, Q7, and Q8 of are compared. 10. For example, if Q6 is the highest, the recognition result will be East, and Tokyo will be output as the recognition result. The feature extraction unit 2, the recognition unit 3, and the recognition result output unit 9 can be processed by a dedicated or shared microprocessor.

【0009】[0009]

【発明の効果】以上述べたように、この発明によれば、
読みを振った仮名・漢字の出現順序に関する統計的言語
モデルと、その読みに対応した音素標準パターンを用い
るため、高い認識性能が予期される。また、入力音声
を、仮名漢字系列に一挙に変換することができ、2回に
分けて変換する場合よりも高い変換性能が予期される。
As described above, according to the present invention,
High recognition performance is expected by using a statistical language model for the appearance order of kana and kanji with pronunciation and phoneme standard patterns corresponding to the pronunciation. In addition, the input voice can be converted into a kana-kanji series at once, and higher conversion performance than that in the case of conversion in two steps is expected.

【0010】文節単位に発声した500文節に対して変
換率による評価を行うと、従来法により、音節の統計情
報(音節のトライグラム)を用いて音素の認識を行った
後に、音素列の認識結果に対して仮名・漢字変換を行う
と70%(正解文節数/総文節数×100)が正しく変
換された。これに対して、前記実施例の手法を用いる
と、同じ音素の認識率でも、読み付きの仮名・漢字統計
情報(読み付きの仮名・漢字群のトライグラム)の効果
により、変換性能は94%に向上する。また、文節構文
を用いることで処理速度は10倍に高速化される。
When the conversion rate is evaluated for 500 syllables uttered in syllable units, the phoneme recognition is performed using the syllable statistical information (syllable trigram) by the conventional method, and then the phoneme string recognition is performed. When the kana / kanji conversion was performed on the result, 70% (correct phrase / total phrase × 100) was correctly converted. On the other hand, when the method of the embodiment is used, the conversion performance is 94% due to the effect of kana / kanji statistical information with reading (trigram of kana / kanji group with reading) even with the same phoneme recognition rate. Improve to. Moreover, the processing speed is increased 10 times by using the clause syntax.

【0011】なお、この発明は上記実施例に限るわけで
はない。たとえば、認識標準パターンのユニットは音素
だけでなく、音節や連音節であってもよい。認識手法は
隠れマルコフモデルに限らず、DPマッチングを用いて
も良い。統計的言語モデルもトライグラムに限らず、バ
イグラムの統計量でも良い。
The present invention is not limited to the above embodiment. For example, the unit of the recognition standard pattern may be not only a phoneme but also a syllable or a continuous syllable. The recognition method is not limited to the hidden Markov model, and DP matching may be used. The statistical language model is not limited to the trigram, but may be a bigram statistic.

【図面の簡単な説明】[Brief description of drawings]

【図1】この発明による音声認識方法を実施した装置の
一例を示すブロック図。
FIG. 1 is a block diagram showing an example of an apparatus that implements a voice recognition method according to the present invention.

【図2】Aは図1中の単語辞書7の記述例を示す図、B
は図1中の文字音素変換規則8の例を示す図である。
FIG. 2A is a diagram showing a description example of a word dictionary 7 in FIG.
FIG. 3 is a diagram showing an example of a character phoneme conversion rule 8 in FIG. 1.

───────────────────────────────────────────────────── フロントページの続き (72)発明者 鹿野 清宏 東京都千代田区内幸町1丁目1番6号 日 本電信電話株式会社内 ─────────────────────────────────────────────────── ─── Continuation of front page (72) Inventor Kiyohiro Kano 1-1-6 Uchisaiwaicho, Chiyoda-ku, Tokyo Nihon Telegraph and Telephone Corporation

Claims (1)

【特許請求の範囲】[Claims] 【請求項1】 入力音声を特徴パラメータの時系列と
し、日本語の構文、単語辞書、文字音素変換規則、及び
テキストデータベースより作成された生起順序に関する
統計的言語モデルを用いて、上記入力音声の特徴パラメ
ータ時系列について、複数の音声認識候補を選出し、こ
れら各音声認識候補について、標準パターンと上記入力
音声の特徴パラメータ時系列とをそれぞれ照合して、生
起の尤度と類似の尤度との総合尤度の高い候補を認識結
果とする日本語音声認識方法において、 上記統計的言語モデルとして、漢字の読みを振ったテキ
ストデータベースから作成された、読みを振った仮名、
及び漢字の生起順序に関する統計的言語モデルを用い、 上記日本語の構文として、日本語文節の構造を単語の遷
移規則として記述した文法を用い、 上記単語辞書として、読みを振った仮名、及び漢字系列
の単語の集合を用い、 上記文字音素変換規則として、読みを振った仮名、及び
漢字を音素の記号列に変換する規則を用い、 上記標準パターンとして、音素標準パターンを用いる、 ことを特徴とする日本語音声認識方法。
1. The input speech is a time series of characteristic parameters, and a statistical language model relating to the occurrence order created from a Japanese syntax, a word dictionary, a graphoneme conversion rule, and a text database is used to input the input speech. For the characteristic parameter time series, a plurality of speech recognition candidates are selected, and for each of these speech recognition candidates, the standard pattern is compared with the characteristic parameter time series of the input speech, respectively, and the likelihood of occurrence and the likelihood of similarity are calculated. In the Japanese speech recognition method in which the candidate with a high total likelihood of is used as the recognition result, as the above-mentioned statistical language model, a kana with a read pronunciation, which is created from a text database with a read kanji,
, And a statistical language model for the occurrence order of kanji, a grammar that describes the structure of Japanese bunsetsu as word transition rules as the Japanese syntax, and a kana with kana and kanji with the reading as the word dictionary. A set of words in a sequence is used, the phoneme standard pattern is used as the standard pattern, using a kana with a reading and a rule for converting kanji into a phoneme symbol string as the character phoneme conversion rule. Japanese speech recognition method.
JP4170898A 1992-06-29 1992-06-29 Japanese speech recognizing method Pending JPH0612091A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP4170898A JPH0612091A (en) 1992-06-29 1992-06-29 Japanese speech recognizing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP4170898A JPH0612091A (en) 1992-06-29 1992-06-29 Japanese speech recognizing method

Publications (1)

Publication Number Publication Date
JPH0612091A true JPH0612091A (en) 1994-01-21

Family

ID=15913373

Family Applications (1)

Application Number Title Priority Date Filing Date
JP4170898A Pending JPH0612091A (en) 1992-06-29 1992-06-29 Japanese speech recognizing method

Country Status (1)

Country Link
JP (1) JPH0612091A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015094848A (en) * 2013-11-12 2015-05-18 インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation Information processor, information processing method and program
CN112927676A (en) * 2021-02-07 2021-06-08 北京有竹居网络技术有限公司 Method, device, equipment and storage medium for acquiring voice information
GB2610807A (en) * 2021-09-09 2023-03-22 Richard Curtis Matthew Improvements in child training rulers

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015094848A (en) * 2013-11-12 2015-05-18 インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation Information processor, information processing method and program
CN112927676A (en) * 2021-02-07 2021-06-08 北京有竹居网络技术有限公司 Method, device, equipment and storage medium for acquiring voice information
GB2610807A (en) * 2021-09-09 2023-03-22 Richard Curtis Matthew Improvements in child training rulers

Similar Documents

Publication Publication Date Title
CN109410914B (en) Method for identifying Jiangxi dialect speech and dialect point
CN108305634B (en) Decoding method, decoder and storage medium
US5878390A (en) Speech recognition apparatus equipped with means for removing erroneous candidate of speech recognition
US9978364B2 (en) Pronunciation accuracy in speech recognition
US20070219777A1 (en) Identifying language origin of words
US20070112569A1 (en) Method for text-to-pronunciation conversion
US20150073796A1 (en) Apparatus and method of generating language model for speech recognition
US20050197838A1 (en) Method for text-to-pronunciation conversion capable of increasing the accuracy by re-scoring graphemes likely to be tagged erroneously
Bianne-Bernard et al. Variable length and context-dependent HMM letter form models for Arabic handwritten word recognition
CN111933116A (en) Speech recognition model training method, system, mobile terminal and storage medium
CN111898342A (en) Chinese pronunciation verification method based on edit distance
JP3364631B2 (en) Statistical language model generation apparatus and speech recognition apparatus
Azim et al. Large vocabulary Arabic continuous speech recognition using tied states acoustic models
JPH0612091A (en) Japanese speech recognizing method
JP2938865B1 (en) Voice recognition device
JPH09134192A (en) Statistical language model forming device and speech recognition device
US20050203742A1 (en) System and method for computer recognition and interpretation of arbitrary spoken-characters
JP3240691B2 (en) Voice recognition method
JP3430265B2 (en) Japanese speech recognition method
JP3009709B2 (en) Japanese speech recognition method
Nakagawa et al. Spoken language identification by ergodic HMMs and its state sequences
JPH04291399A (en) Voice recognizing method
JP2968792B1 (en) Statistical language model generation device and speech recognition device
JP2001188556A (en) Method and device for voice recognition
JPH1185183A (en) Speech recognition system and its apparatus as well as storage medium recording speech recognition processing program

Legal Events

Date Code Title Description
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20040113

RD01 Notification of change of attorney

Effective date: 20040123

Free format text: JAPANESE INTERMEDIATE CODE: A7426

RD03 Notification of appointment of power of attorney

Effective date: 20040123

Free format text: JAPANESE INTERMEDIATE CODE: A7423

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20040123

FPAY Renewal fee payment (prs date is renewal date of database)

Year of fee payment: 4

Free format text: PAYMENT UNTIL: 20080206

FPAY Renewal fee payment (prs date is renewal date of database)

Year of fee payment: 5

Free format text: PAYMENT UNTIL: 20090206

FPAY Renewal fee payment (prs date is renewal date of database)

Free format text: PAYMENT UNTIL: 20090206

Year of fee payment: 5

FPAY Renewal fee payment (prs date is renewal date of database)

Free format text: PAYMENT UNTIL: 20100206

Year of fee payment: 6

LAPS Cancellation because of no payment of annual fees