JP2012022069A - Speech recognition method, and device and program for the same - Google Patents

Speech recognition method, and device and program for the same Download PDF

Info

Publication number
JP2012022069A
JP2012022069A JP2010158472A JP2010158472A JP2012022069A JP 2012022069 A JP2012022069 A JP 2012022069A JP 2010158472 A JP2010158472 A JP 2010158472A JP 2010158472 A JP2010158472 A JP 2010158472A JP 2012022069 A JP2012022069 A JP 2012022069A
Authority
JP
Japan
Prior art keywords
word
reliability
recognition
speech
document
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2010158472A
Other languages
Japanese (ja)
Other versions
JP5406797B2 (en
Inventor
Taichi Asami
太一 浅見
Satoru Kobashigawa
哲 小橋川
Yoshikazu Yamaguchi
義和 山口
Hirokazu Masataki
浩和 政瀧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nippon Telegraph and Telephone Corp
Original Assignee
Nippon Telegraph and Telephone Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nippon Telegraph and Telephone Corp filed Critical Nippon Telegraph and Telephone Corp
Priority to JP2010158472A priority Critical patent/JP5406797B2/en
Publication of JP2012022069A publication Critical patent/JP2012022069A/en
Application granted granted Critical
Publication of JP5406797B2 publication Critical patent/JP5406797B2/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

PROBLEM TO BE SOLVED: To provide a speech recognition method, for instance, appropriate for generation of a text to be used for data mining.SOLUTION: The speech recognition method comprises: a speech recognition process; a speech document recognition reliability calculation process, a speech document removal process; and a word removal process. The speech recognition process outputs a speech recognition result, in which the reliability of word recognition is added to each word after speech recognition processing is applied to an input speech document, and the speech document recognition reliability calculation process accepts the speech recognition result as an input, and calculates and outputs speech document recognition reliability which is the recognition reliability for the whole speech document. Then the speech document removal process accepts the speech recognition result and the speech document recognition reliability as an input and removes a speech document that falls short of a prescribed threshold of speech document recognition reliability, and the word removal process removes a word that falls short of a prescribed threshold of word recognition reliability, from the speech recognition results of the speech documents that are not removed in the speech document removal process.

Description

この発明は、例えばデータマイニングに用いるテキストを生成するのに好適な音声認識方法とその装置と、プログラムに関する。   The present invention relates to a speech recognition method and apparatus suitable for generating text used for data mining, for example, and a program.

テキストデータとして収集されたデータを統計的に分析する手法を、一般的にテキストマイニングと称する。そのテキストデータを得る目的で音声認識を用いる場合がある。音声認識には、誤認識が付き物である。そこで、従来から、この音声認識誤りを減らす工夫が試みられている。   A technique for statistically analyzing data collected as text data is generally referred to as text mining. Voice recognition may be used for the purpose of obtaining the text data. Misrecognition is a natural part of speech recognition. Thus, conventionally, attempts have been made to reduce this speech recognition error.

例えば、音声認識結果に、その認識結果の確からしさを表す認識信頼度を付与する方法がある。特許文献1に、音声認識処理の探索の結果の上位N位までのNベスト候補のスコア1位の単語w1の認識信頼度を、単語w1のスコアとスコア2位以下でw1と異なる単語w2とのスコア差を単語w1の継続時間長で正規化した値とする考えが示されている。 For example, there is a method of giving a recognition reliability indicating the certainty of the recognition result to the voice recognition result. In Patent Document 1, the recognition reliability of the word w 1 with the highest score of the N best candidates up to the top N as a result of the search of the speech recognition process is different from the score of the word w 1 with the score of 2 or lower and w 1. The idea that the score difference from the word w 2 is normalized by the duration of the word w 1 is shown.

また、別の方法として、音声認識結果中の各単語間の関連の強さを計測して周囲の単語と関連の強い単語に、高い認識信頼度を付与し、関連の弱い単語に低い認識信頼度を付与する方法がある(非特許文献1)。この方法は、単語w及び単語wの直前k個の単語と直後の1単語とのn個の単語集合N(w)を音声認識結果から取得する。そして、その単語集合N(w)に含まれる全ての2単語の組み合わせ(wi,wj)に対して、事前に学習コーパス上で算出した相互情報量MI(wi,wj)を用いて単語間の強さS(wi,wj)を計算する。また、単語集合N(w)中の全ての単語tについての関連の強さS(t,wi)の平均値を文脈一貫性尺度SC(t)として計算する。 Another method is to measure the strength of association between each word in the speech recognition result, and give high recognition confidence to words that are strongly related to surrounding words, and low recognition confidence to weakly related words. There is a method of providing a degree (Non-Patent Document 1). In this method, n word sets N (w) of the word w and k words immediately before the word w and one word immediately after the word w are acquired from the speech recognition result. Then, the mutual information MI (w i , w j ) calculated in advance on the learning corpus is used for all combinations (w i , w j ) of two words included in the word set N (w). Then, the strength S (w i , w j ) between words is calculated. Further, the average value of the relation strength S (t, w i ) for all the words t in the word set N (w) is calculated as the context consistency measure SC (t).

大量に蓄積された音声文書をデータマイニング処理する際は、上記したような認識信頼度が付与されたテキストが用いられる。   When data mining processing is performed on a large amount of stored voice documents, text with the above-described recognition reliability is used.

特開2005−148342号公報JP 2005-148342 A

D. Inkpen, A. Desilets, “Semantic Similarity for Detecting Recognition Errors in Automatic Speech Transcripts,”Proceedings of HLT/EMNLP, pp.49-56, October 2005.D. Inkpen, A. Desilets, “Semantic Similarity for Detecting Recognition Errors in Automatic Speech Transcripts,” Proceedings of HLT / EMNLP, pp.49-56, October 2005.

大量に蓄積された音声文書を音声認識によってテキスト化し、全文検索・文書分類・パターン抽出などのテキストマイニング処理を行う際、実際には発声されていない単語が音声認識誤りによってテキスト中に現れることにより、誤った情報がテキストマイニング結果に含まれてしまう問題がある。テキストマイニング結果に誤った情報が多数含まれる場合、テキストマイニング利用者が得られた結果を見て有用な知見を得ることは難しい。   When a large amount of speech documents are converted into text by speech recognition and text mining processing such as full-text search, document classification, and pattern extraction is performed, words that are not actually spoken appear in the text due to speech recognition errors. There is a problem that incorrect information is included in the text mining result. When a lot of incorrect information is included in the text mining result, it is difficult for a text mining user to obtain useful knowledge by looking at the obtained result.

上記したように、音声認識結果の各単語に認識信頼度を付与して認識信頼度の低い単語をテキストマイニング処理の対象から除外することで、誤った情報の抽出量を削減することが可能である。しかし、音声認識結果全体の認識精度が低い場合、単語に対する認識信頼度は認識の確からしさを正しく表さない傾向がある。例えば、周囲雑音の大きな環境で収録された音声文書の認識誤り単語に適切に認識信頼度を付与することは難しい。また、話者の話し方の差異を原因とする場合も、音声文書全体の認識信頼度が低くなり各単語に適切な認識信頼度を付与できない。   As mentioned above, it is possible to reduce the amount of erroneous information extracted by giving recognition confidence to each word of the speech recognition result and excluding words with low recognition confidence from the target of text mining processing. is there. However, when the recognition accuracy of the entire speech recognition result is low, the recognition reliability for a word tends not to accurately represent the probability of recognition. For example, it is difficult to appropriately assign a recognition reliability to a recognition error word of an audio document recorded in an environment with a large ambient noise. In addition, even when the cause is a difference in the speaker's way of speaking, the recognition reliability of the entire voice document becomes low, and an appropriate recognition reliability cannot be given to each word.

この発明は、このような課題に鑑みてなされたものであり、音声認識によってテキストデータに認識誤り単語が含まれ難いようにした音声認識方法とその装置と、プログラムを提供することを目的とする。   The present invention has been made in view of the above problems, and an object of the present invention is to provide a speech recognition method, a device thereof, and a program that make it difficult for a recognition error word to be included in text data by speech recognition. .

この発明の音声認識方法は、音声認識過程と、音声文書認識信頼度計算過程と、音声文書除去過程と、単語除去過程と、を含む。音声認識過程は、入力される音声文書を音声認識処理した単語毎に単語認識信頼度を付与した音声認識結果を出力する。音声文書認識信頼度計算過程は、音声認識結果を入力として音声文書全体の認識信頼度である音声文書認識信頼度を計算して出力する。音声文書除去過程は、音声認識結果と音声文書認識信頼度とを入力として所定の音声文書認識信頼度閾値未満の音声文書を除去する。単語除去過程は、音声文書除去過程で除去されなかった音声文書の音声認識結果から所定の単語認識信頼度閾値未満の単語認識信頼度の単語を除去する。   The speech recognition method of the present invention includes a speech recognition process, a speech document recognition reliability calculation process, a speech document removal process, and a word removal process. The speech recognition process outputs a speech recognition result with word recognition reliability assigned to each word obtained by performing speech recognition processing on the input speech document. In the voice document recognition reliability calculation process, the voice recognition result, which is the recognition reliability of the entire voice document, is calculated and output using the voice recognition result as an input. In the voice document removal process, a voice document less than a predetermined voice document recognition reliability threshold is removed by inputting the voice recognition result and the voice document recognition reliability. In the word removal process, words having a word recognition reliability less than a predetermined word recognition reliability threshold are removed from the voice recognition result of the voice document that has not been removed in the voice document removal process.

この発明の音声認識方法は、音声文書単位での除去を行った後に、残った音声文書に対して単語単位の除去を行うので、認識信頼度が全体的に低い単語単位での除去が難しい音声認識結果を適切に除去することができる。したがって、例えば、テキストマイニング処理対象となるテキストデータに含まれる認識誤り単語の数を、効果的に削減することができる。その結果、テキストマイニングの利用者が有用な知見を得ることが可能となる。   In the speech recognition method according to the present invention, since the removal is performed in units of words after the removal in units of speech documents, it is difficult to remove in units of words whose recognition reliability is generally low. The recognition result can be appropriately removed. Therefore, for example, it is possible to effectively reduce the number of recognition error words included in text data to be subjected to text mining processing. As a result, users of text mining can obtain useful knowledge.

また、この発明の音声認識方法を、音声認識に用いる確率モデルの教師なし適応に利用しても効果的である。つまり、認識誤り単語の少ない音声認識結果を収集することが可能であるので、収集した認識誤りの少ない音声認識結果を教師なし適応に用いれば音声認識精度を向上させることができる。   It is also effective to use the speech recognition method of the present invention for unsupervised adaptation of a probability model used for speech recognition. That is, since speech recognition results with few recognition error words can be collected, the speech recognition accuracy can be improved by using the collected speech recognition results with few recognition errors for unsupervised adaptation.

この発明の音声認識装置100の機能構成例を示す図。The figure which shows the function structural example of the speech recognition apparatus 100 of this invention. 音声認識装置100の動作フローを示す図。The figure which shows the operation | movement flow of the speech recognition apparatus 100. Nベスト候補と単語認識信頼度について説明する図。The figure explaining N best candidate and word recognition reliability. 音声文書認識信頼度計算部20の機能構成例を示す図。The figure which shows the function structural example of the audio | voice document recognition reliability calculation part 20. FIG. 音声文書認識信頼度計算部20の動作フローを示す図。The figure which shows the operation | movement flow of the voice document recognition reliability calculation part 20. FIG. 音声認識部10が出力する単語wn、単語認識信頼度D(wn)の例を示す図。Word w n output by the speech recognition unit 10, shows an example of a word recognition reliability D (w n). この発明の音声認識装置200の機能構成例を示す図。The figure which shows the function structural example of the speech recognition apparatus 200 of this invention. 音声認識装置200の動作フローを示す図。The figure which shows the operation | movement flow of the speech recognition apparatus 200. 単語関連度テーブル作成装置150の機能構成例を示す図。The figure which shows the function structural example of the word related degree table creation apparatus 150. FIG. 単語集合を概念的に示す図。The figure which shows a word set notionally. 単語関連度テーブルの一例を示す図。The figure which shows an example of a word related degree table. 音声文書認識信頼度計算部70の機能構成例を示す図。The figure which shows the function structural example of the audio | voice document recognition reliability calculation part. この発明の音声認識装置300の機能構成例を示す図。The figure which shows the function structural example of the speech recognition apparatus 300 of this invention. 音声文書認識信頼度高速計算部90の機能構成例を示す図。The figure which shows the function structural example of the audio | voice document recognition reliability high speed calculation part 90. FIG. 音声文書認識信頼度高速計算部90の動作フローを示す図。The figure which shows the operation | movement flow of the voice document recognition reliability high speed calculation part. 単語集合と直前重複フラグと直後重複フラグの例を示す図。The figure which shows the example of a word set, an immediately preceding duplication flag, and an immediately following duplication flag. 単語集合取得ステップの動作フローの例を示す図。The figure which shows the example of the operation | movement flow of a word set acquisition step. 単語集合音響信頼度高速計算ステップの動作フローの例を示す図。The figure which shows the example of the operation | movement flow of a word set acoustic reliability high-speed calculation step.

以下、この発明の実施の形態を図面を参照して説明する。複数の図面中同一のものには同じ参照符号を付し、説明は繰り返さない。   Embodiments of the present invention will be described below with reference to the drawings. The same reference numerals are given to the same components in a plurality of drawings, and the description will not be repeated.

図1にこの発明の音声認識装置100の機能構成例を示す。その動作フローを図2に示す。音声認識装置100は、音声認識部10と、音声文書認識信頼度計算部20と、音声文書除去部30と、単語除去部40と、制御部50と、を具備する。音声認識装置100の各部の機能は、例えばROM、RAM、CPU等で構成されるコンピュータに所定のプログラムが読み込まれて、CPUがそのプログラムを実行することで実現されるものである。   FIG. 1 shows a functional configuration example of the speech recognition apparatus 100 of the present invention. The operation flow is shown in FIG. The speech recognition apparatus 100 includes a speech recognition unit 10, a speech document recognition reliability calculation unit 20, a speech document removal unit 30, a word removal unit 40, and a control unit 50. The function of each part of the speech recognition apparatus 100 is realized by reading a predetermined program into a computer composed of, for example, a ROM, a RAM, a CPU, etc., and executing the program by the CPU.

音声認識部10は、入力される音声文書を音声認識処理した単語毎に単語認識信頼度を付与した音声認識結果を出力する(ステップS10)。音声認識部10は、図示しない内部の音響分析部によって、音声文書を、数十msecのフレームと呼ばれる単位でLPCケプストラム、MFCC、その他の音響特徴パラメータ系列に分析する。そして、辞書と言語モデルを用いて入力音声に対する認識結果候補の探索を、音響特徴パラメータ系列について行う。探索の結果、上位N位までのNベスト候補が、単語認識信頼度と共に音声認識結果として出力される。なお、音声文書とは、例えば、コールセンターにおける顧客とオペレータとの間で交わされる会話の集合のことであり、一要件についてまとめられた音声データである。また、例えば一つの講義が1個の音声ファイルにまとめられた様なものである。   The speech recognition unit 10 outputs a speech recognition result in which word recognition reliability is given to each word obtained by performing speech recognition processing on the input speech document (step S10). The voice recognition unit 10 analyzes the voice document into an LPC cepstrum, MFCC, and other acoustic feature parameter series in units called frames of several tens of msec by an internal acoustic analysis unit (not shown). Then, a recognition result candidate for the input speech is searched for the acoustic feature parameter series using a dictionary and a language model. As a result of the search, N best candidates up to the top N are output as a speech recognition result together with the word recognition reliability. Note that the voice document is, for example, a set of conversations exchanged between a customer and an operator at a call center, and is voice data compiled for one requirement. For example, one lecture is organized into one audio file.

ここで、図3を参照してNベスト候補と単語認識信頼度について説明する。なお、Nベスト候補と単語認識信頼度については従来技術である。単語認識信頼度については、例えば特許文献1に記載されている。   Here, N best candidates and word recognition reliability will be described with reference to FIG. Note that N best candidates and word recognition reliability are conventional techniques. The word recognition reliability is described in Patent Document 1, for example.

図3の横軸は、経過時間でありフレームで表す。縦軸は、フレーム単位で探索された単語列候補を、スコアの高い順番に並べたNベスト候補である。スコアとは探索時の尤度のことである。   The horizontal axis in FIG. 3 represents elapsed time and is represented by a frame. The vertical axis represents N best candidates in which word string candidates searched for in units of frames are arranged in descending order of scores. The score is the likelihood at the time of search.

単語認識信頼度は、フレームt*において単語w**(*は任意の整数)と異なる単語がNベスト候補中に存在する場合、単語w**のフレームt*におけるスコアと対立候補単語のフレームt*における次の順位のスコアとの間のスコア差によって与えられる。図3に示す例では、フレームt1〜t4の音響特徴パラメータ系列で探索された第1位候補の単語w11(11は第1候補の1番目の単語であることを表す)の単語認識信頼度は、対向する単語が第3位候補の単語w31と第2位候補の単語w21であるので、それぞれのスコア差(●)の合計をフレーム数で除した値が単語認識信頼度となる。対立候補が存在しない単語w13については、予め決められた固定値(○)が用いられて単語認識信頼度となる。この単語認識信頼度が候補毎に累積されて単語列の認識信頼度となる。 Word recognition reliability, if the word w ** in frame t * (* is an arbitrary integer) is different word present in the N-best candidate words w ** frame t score paired candidacy word frames in * Given by the score difference between the next ranking score at t * . In the example shown in FIG. 3, the word recognition of the first candidate word w 11 (11 represents the first word of the first candidate) searched for in the acoustic feature parameter series of frames t 1 to t 4. Since the opposite words are the third candidate word w 31 and the second candidate word w 21 , the value obtained by dividing the total score difference (●) by the number of frames is the word recognition reliability. It becomes. For the word w 13 the conflict candidate does not exist, the word recognition reliability by predetermined fixed value (○) is used. This word recognition reliability is accumulated for each candidate and becomes the word string recognition reliability.

音声文書認識信頼度計算部20は、単語毎の単語認識信頼度及び単語列の認識信頼度から音声文書全体の認識信頼度である音声文書認識信頼度を計算して出力する(ステップS20)。   The voice document recognition reliability calculation unit 20 calculates and outputs a voice document recognition reliability that is the recognition reliability of the entire voice document from the word recognition reliability for each word and the word string reliability (step S20).

音声文書除去部30は、音声認識部10が出力する音声認識結果と音声文書認識信頼度計算部20が出力する音声文書認識信頼度とを入力として、所定の音声文書認識信頼度閾値θd未満の音声文書を除去する(ステップS30)。 The voice document removal unit 30 receives the voice recognition result output from the voice recognition unit 10 and the voice document recognition reliability output from the voice document recognition reliability calculation unit 20 as input, and is less than a predetermined voice document recognition reliability threshold θ d. Are deleted (step S30).

単語除去部40は、音声文書除去部30で除去されなかった音声文書の音声認識結果から所定の単語認識信頼度閾値θw未満の単語認識信頼度の単語を除去する(ステップS40)。音声文書認識信頼度閾値θdと単語認識信頼度閾値θwは、予め定数として各部に備えておいても良いし、外部から与えても良い。 Word removal unit 40 removes the word of a predetermined word recognition confidence threshold θ word recognition reliability of less than w results speech recognition of the speech documents which were not removed in the spoken document removal unit 30 (step S40). The voice document recognition reliability threshold value θ d and the word recognition reliability threshold value θ w may be previously provided as constants in each unit, or may be given from the outside.

このように、音声認識装置100は、音声文書単位での除去を行った後に、残った音声文書に対して単語単位の除去した音声認識結果を出力するので、音声認識結果に含まれる誤認識単語を削減することができる。   As described above, the speech recognition apparatus 100 outputs the speech recognition result in which the word unit is removed from the remaining speech document after performing the removal in the speech document unit, so that the erroneous recognition word included in the speech recognition result. Can be reduced.

図4に音声文書認識信頼度計算部20の機能構成を示して更に詳しく動作を説明する。その動作フローを図5に示す。音声文書認識信頼度計算部20は、単語時間長取得手段21と、正規化解除手段22と、音声文書総時間長計算手段23と、信頼度累積手段24と、音声文書認識信頼度算出手段25と、を備える。   FIG. 4 shows the functional configuration of the voice document recognition reliability calculation unit 20, and the operation will be described in more detail. The operation flow is shown in FIG. The voice document recognition reliability calculation unit 20 includes a word time length acquisition unit 21, a denormalization unit 22, a voice document total time length calculation unit 23, a reliability accumulation unit 24, and a voice document recognition reliability calculation unit 25. And comprising.

単語時間長取得手段21は、音声認識部20が出力する単語の単語時間長を求める(ステップS21)。単語時間長取得手段21は、最初に音声文書の信頼度合計値D(W)と音声文書の総時間長WDを0に初期化する(ステップS50)。ステップS50〜S52は、図1に示した制御部50が処理する。   The word time length acquisition unit 21 obtains the word time length of the word output from the voice recognition unit 20 (step S21). The word time length acquisition unit 21 first initializes the total reliability D (W) of the voice document and the total time length WD of the voice document to 0 (step S50). Steps S50 to S52 are processed by the control unit 50 shown in FIG.

図6に、音声認識部20が出力する単語wn、単語認識信頼度D(wn)の例を示す。ここでは、音声ファイルが1個のNベスト候補から成る例で説明する。つまり添え字は一桁で説明する。例えば、単語w1は名詞「本日」であり、単語認識信頼度D(wn)は9891であり、その始端時刻(wdnsFn)と終端時刻(wdneFn)は0.00−0.98[秒]である。単語認識信頼度D(wn)の値は、マイナスの値もあり得る。図6の例では、「おいたわしい」がそれに該当する。第1位候補の単語のスコアが、下位の候補の単語のスコアよりも小さい場合に単語認識信頼度D(wn)はマイナスとなる。この場合、第1位候補のその単語の信頼度は相当低いことを意味する。この単語認識信頼度D(wn)の値は、音声認識結果の音響的な信頼度を表すものである。 FIG. 6 shows an example of the word w n and the word recognition reliability D (w n ) output from the speech recognition unit 20. Here, an example in which an audio file is composed of one N best candidate will be described. In other words, the subscript is described with one digit. For example, the word w 1 is the noun “today”, the word recognition reliability D (w n ) is 9891, and its start time (wd n sFn) and end time (wd n eFn) are 0.00-0. It is 98 [seconds]. The value of the word recognition reliability D (w n ) may be a negative value. In the example of FIG. 6, “Nice” corresponds to this. When the score of the first candidate word is smaller than the score of the lower candidate word, the word recognition reliability D (w n ) is negative. In this case, it means that the reliability of the word of the first candidate is considerably low. The value of the word recognition reliability D (w n ) represents the acoustic reliability of the speech recognition result.

単語時間長取得手段21は、単語w1の終端時刻から始端時刻を引いたその単語時間長wd1を、0.98[秒]若しくは、フレーム時間を例えば10msecとした場合、98フレームとして取得する(ステップS21)。 The word time length acquisition means 21 acquires the word time length wd 1 obtained by subtracting the start time from the end time of the word w 1 as 0.98 [sec] or 98 frames when the frame time is set to 10 msec, for example. (Step S21).

正規化解除手段22は、単語認識信頼度D(w1)=9891に単語時間長wd1を乗じた単語信頼度wcを計算する(ステップS22)。正規化解除手段22は、フレーム数で正規化されていた単語認識信頼度の正規化を解除する働きをする。 The denormalization means 22 calculates a word reliability wc obtained by multiplying the word recognition reliability D (w 1 ) = 9891 by the word time length wd 1 (step S22). The normalization cancellation means 22 functions to cancel the normalization of the word recognition reliability that has been normalized by the number of frames.

信頼度累積手段24は、正規化が解除された単語信頼度wcを累積した単語信頼度累積値D(W)を求める(ステップS24)音声文書総時間長計算手段23は、単語時間長取得手段21で取得した単語時間長wd*を音声ファイル全体で累積した音声文書総時間長WDを計算する(ステップS23)。上記したステップS21〜S23の処理は、音声ファイルの全ての単語wnについて処理されるまで、単語を更新(ステップS52)しながら繰り返される(ステップS51のno)。 The reliability accumulating means 24 obtains a word reliability accumulated value D (W) obtained by accumulating the normalized word reliability wc (step S24). The voice document total time length calculating means 23 is a word time length obtaining means. The voice document total time length WD obtained by accumulating the word time length wd * acquired in 21 in the entire voice file is calculated (step S23). Processing in steps S21~S23 described above is repeated until the processing for all words w n of the audio file, while updating the word (step S52) (no in step S51).

音声文書認識信頼度算出手段25は、単語信頼度累積値D(W)を音声文書総時間長WDで除した音声文書の音声文書認識信頼度docCを算出する(ステップS25)。音声文書認識信頼度算出手段25は、音声文書のフレーム当たりの認識信頼度を求めることになる。この音声文書認識信頼度docCは、音声文書の音声認識結果の音響的な品質を表す指標となる。   The voice document recognition reliability calculation means 25 calculates the voice document recognition reliability docC of the voice document obtained by dividing the word reliability cumulative value D (W) by the voice document total time length WD (step S25). The voice document recognition reliability calculation means 25 obtains the recognition reliability per frame of the voice document. The voice document recognition reliability docC is an index representing the acoustic quality of the voice recognition result of the voice document.

音声文書除去部30は、音声文書の音響的な信頼度を表す音声文書認識信頼度docCと、事前に決められた音声文書認識信頼度閾値θdとを比較して音声文書認識信頼度docCがθd以上の場合、音声文書をそのまま単語除去部40に出力する。音声文書認識信頼度docCがθd未満の場合は、その音声文書を単語除去部40に出力しない。 The voice document removal unit 30 compares the voice document recognition reliability docC representing the acoustic reliability of the voice document with a predetermined voice document recognition reliability threshold θ d to obtain a voice document recognition reliability docC. If it is equal to or greater than θ d , the voice document is output as it is to the word removal unit 40. When the voice document recognition reliability docC is less than θ d , the voice document is not output to the word removal unit 40.

音声文書認識信頼度閾値θdは実数値であり、大きな値(例えば30000程度)に設定すると出力される音声文書の音声認識精度は高くなる。小さな値(例えば-30000程度)に設定すると出力される音声文書の音声認識精度は低くなるが、出力される音声文書の数は増加する。 The voice document recognition reliability threshold θ d is a real value, and when set to a large value (for example, about 30000), the voice recognition accuracy of the output voice document is increased. When a small value (for example, about −30000) is set, the voice recognition accuracy of the output voice document is lowered, but the number of output voice documents is increased.

単語除去部40は、音声文書除去部30が出力した音声文書を構成する単語とその単語に付与された単語認識信頼度D(wn)を入力として、単語認識信頼度D(wn)の値が所定の単語認識信頼度閾値θw未満であれば当該単語を、除去されたことを表す所定の記号、例えば“<rejected>”に置換して音声文書を出力する。 The word removal unit 40 receives the words constituting the voice document output from the voice document removal unit 30 and the word recognition reliability D (w n ) assigned to the words, and inputs the word recognition reliability D (w n ). If the value is less than a predetermined word recognition reliability threshold value θw, the word is replaced with a predetermined symbol indicating that it has been removed, for example, “<rejected>”, and a speech document is output.

以上の処理によって最終的に出力される音声認識結果は、音声認識精度の比較的高い音声文書から更に単語認識信頼度が低い単語が除去されたものになり、音声認識結果に含まれる誤認識単語を削減することができる。この音声認識装置100を用いてデータマイニング用のテキストデータを取得すると、テキストデータに含まれる誤認識単語の数を減らすことができるので、テキストマイニングの利用者が有用な知見を得ることが可能になる。   The speech recognition result that is finally output by the above processing is obtained by removing words with lower word recognition reliability from a speech document having a relatively high speech recognition accuracy, and is a misrecognized word included in the speech recognition result. Can be reduced. When text data for data mining is acquired using the speech recognition apparatus 100, the number of misrecognized words included in the text data can be reduced, so that a user of text mining can obtain useful knowledge. Become.

図7にこの発明の音声認識装置200の機能構成例を示す。その動作フローを図8に示す。音声認識装置200は、上記した音声文書認識信頼度docCを、音響信頼度と文脈信頼度を組み合わせた値にしたものである。よって、音声文書認識信頼度計算部70のみが、音声認識装置200と異なる。音響信頼度とは、音声認識装置100の音声文書認識信頼度docCと同じものである。以降、実施例1で述べた音声文書認識信頼度docCを音響信頼度と称する。   FIG. 7 shows a functional configuration example of the speech recognition apparatus 200 of the present invention. The operation flow is shown in FIG. The speech recognition apparatus 200 is a combination of the above-described speech document recognition reliability docC and a combination of acoustic reliability and context reliability. Therefore, only the speech document recognition reliability calculation unit 70 is different from the speech recognition apparatus 200. The acoustic reliability is the same as the voice document recognition reliability docC of the voice recognition apparatus 100. Hereinafter, the voice document recognition reliability docC described in the first embodiment is referred to as acoustic reliability.

音声認識装置200の音声文書認識信頼度計算部70は、音声認識結果から文脈信頼度を得るために、音声認識結果を構成する単語間の関連度を示す単語関連度テーブル60を参照する。単語関連度テーブル60を作成する単語関連度テーブル作成装置150の動作を説明する。   The speech document recognition reliability calculation unit 70 of the speech recognition apparatus 200 refers to the word association degree table 60 indicating the association degree between words constituting the speech recognition result in order to obtain context reliability from the speech recognition result. The operation of the word association degree table creation device 150 that creates the word association degree table 60 will be described.

〔単語関連度テーブル作成装置〕
図9に単語関連度テーブル作成装置150の機能構成例を示す。単語関連度テーブル作成装置150は、学習コーパス81と、形態素解析部82と、学習コーパス単語集合取得部83と、単語リスト84と、単語カウント部85と、単語関連度計算部86と、テーブル配列部87と、を具備する。
[Word relevance table creation device]
FIG. 9 shows a functional configuration example of the word association degree table creation device 150. The word association degree table creation device 150 includes a learning corpus 81, a morpheme analysis unit 82, a learning corpus word set acquisition unit 83, a word list 84, a word count unit 85, a word association degree calculation unit 86, and a table array. Part 87.

学習コーパス81は音声文書を大規模に集積したものである。形態素解析部82は、学習コーパス81から音声文書を読み出して単語に分割する周知の形態素解析処理を行い、各単語の前後に単語境界を表す記号、例えば「\n」を付与した単語境界付き学習コーパスを出力する。形態素解析処理は周知であり、例えば参考文献「特許第3379643号」に記載されている。   The learning corpus 81 is a collection of voice documents on a large scale. The morpheme analysis unit 82 performs a well-known morpheme analysis process that reads a speech document from the learning corpus 81 and divides it into words, and learns with word boundaries provided with symbols representing word boundaries before and after each word, for example, “\ n” Output corpus. The morphological analysis processing is well known and is described in, for example, a reference document “Japanese Patent No. 3379643”.

学習コーパス単語集合取得部83は、形態素解析部82が出力する単語境界付き学習コーパスの先頭から末尾まで、窓幅n単語、窓シフト量m単語で窓かけを行い、各窓に含まれる単語リスト84に記載された単語をまとめて単語集合とし、窓ごとの単語集合を出力する。単語リスト84は、音声認識結果に出現し得る全ての単語が記載されたものであり、事前に作成しておく。図10に、単語集合を概念的に示す。横方向は時間経過であり、単語集合をN1〜Nで示す。mは窓シフト量であり、nは窓幅である。隣り合う単語集合は、n-m個の単語を共有する関係にある。 The learning corpus word set acquisition unit 83 performs windowing with a window width n words and a window shift amount m words from the beginning to the end of the word-boundary learning corpus output by the morpheme analysis unit 82, and a word list included in each window The words described in 84 are combined into a word set, and a word set for each window is output. The word list 84 describes all the words that can appear in the speech recognition result, and is created in advance. FIG. 10 conceptually shows the word set. The horizontal direction is the passage of time, and the word set is indicated by N 1 to N h . m is the window shift amount, and n is the window width. Adjacent word sets share a relationship of nm words.

単語カウント部85は、学習コーパス単語集合取得部83が出力する単語集合を入力として、単語集合内の各単語の単独生起回数C(w)、各単語ペアの生起回数C(wi,wj)、単語集合の総数をカウントして出力する。単語wの生起回数C(w)とは、単語wを含む単語集合の個数である。単語ペア(wi,wj)の生起回数C(wi,wj)とは、wiとwjを共に含む単語集合の個数である。 The word count unit 85 receives the word set output from the learning corpus word set acquisition unit 83 as an input, and the number of occurrences C (w) of each word in the word set and the number of occurrences C (w i , w j of each word pair) ), Count and output the total number of word sets. The number of occurrences C (w) of the word w is the number of word sets including the word w. The number of occurrences C (w i , w j ) of the word pair (w i , w j ) is the number of word sets including both w i and w j .

単語関連度計算部86は、各単語ペア(wi,wj)の関連度S(wi,wj)を例えば式(1)で計算する。 The word relevance calculation unit 86 calculates the relevance S (w i , w j ) of each word pair (w i , w j ) using, for example, equation (1).

Figure 2012022069
Figure 2012022069

Nは単語集合の総数、C(w)は単語wの単独生起回数、C(wi,wj)は単語wiとwjの共起回数である。関連度S(wi,wj)の値が大きいと、それらの単語同士の関連性が高いことを意味する。 N is the total number of word sets, C (w) is the number of single occurrences of word w, and C (w i , w j ) is the number of co-occurrence of words w i and w j . When the value of the relevance S (w i , w j ) is large, it means that the relevance between these words is high.

テーブル配列部87は、単語wiとwjから計算した関連度S(wi,wj)を参照できるテーブルの形式に配列する。図11に単語関連度テーブル60の一例を示す。最上列と最左列は単語w1〜wNであり、各行と各列の交差する欄にそれぞれの単語の関連度S(wi,wj)が配列される。 The table arrangement unit 87 arranges the relevance S (w i , w j ) calculated from the words w i and w j in a table format that can be referred to. FIG. 11 shows an example of the word association degree table 60. The uppermost column and the leftmost column are words w 1 to w N , and the relevance S (w i , w j ) of each word is arranged in a column where each row and each column intersect.

図12に音声文書認識信頼度計算部70の機能構成例を示す。音声文書認識信頼度計算部70は、認識結果単語集合取得手段71と、単語集合音響信頼度計算手段72と、単語集合信頼度計算手段73と、信頼度統合手段74と、を具備する。   FIG. 12 shows a functional configuration example of the voice document recognition reliability calculation unit 70. The voice document recognition reliability calculation unit 70 includes a recognition result word set acquisition unit 71, a word set acoustic reliability calculation unit 72, a word set reliability calculation unit 73, and a reliability integration unit 74.

認識結果単語集合取得手段71は、音声認識結果の単語wkと単語認識信頼度C(wk)を、その先頭から所定の数n個ずつnより小さい数のm個移動させながらh個の単語集合に分割する。つまり、音声文書の音声認識結果の先頭からn単語を取得し、得られた単語集合をN1とする。そして、音声認識結果の先頭からm単語目からn単語を取得し、得られた単語集合をN2とする。次に、音声認識結果の先頭から2m単語目からn単語を取得し、得られた単語集合をN3とする。以上の処理を音声文書の末尾に到達するまで繰り返し、h個の単語集合Nk(Nk:N1〜Nh)を取得する。kは、実施例1ではnを用いた着目している単語集合及び単語を識別する変数である。 The recognition result word set acquisition unit 71 moves the words w k and the word recognition reliability C (w k ) of the speech recognition result by moving a predetermined number n from the head by m, which is smaller than n. Divide into word sets. That is, n words are acquired from the head of the speech recognition result of the speech document, and the obtained word set is set to N 1 . Then, to get the n words of m words from the beginning of the speech recognition results, resulting word set to N 2. Next, n words are acquired from the 2m-th word from the beginning of the speech recognition result, and the obtained word set is defined as N 3 . The above processing is repeated until the end of the speech document is reached, and h word sets N k (N k : N 1 to N h ) are acquired. k is a variable that identifies a focused word set and word using n in the first embodiment.

単語集合音響信頼度計算手段72は、各単語集合に含まれる全ての単語の単語認識信頼度D(w)と単語時間長wdkを取得して実施例1で説明済みの音声文書認識信頼度計算部20と同じ処理をして単語集合Nk毎に音響信頼度CA(Nk)を計算する。 The word set acoustic reliability calculation means 72 acquires the word recognition reliability D (w) and the word time length wd k of all the words included in each word set, and the speech document recognition reliability described in the first embodiment. The acoustic reliability CA (N k ) is calculated for each word set N k by performing the same processing as the calculation unit 20.

単語集合文脈信頼度計算手段73は、各単語集合に含まれる全ての2単語の組み合わせ(wi,wj)に対して単語の関連の強さS(wi,wj)を単語関連度テーブル60を参照して求め、その平均値を各単語集合の文脈信頼度CL(Nk)として計算する。 The word set context reliability calculation means 73 calculates the word association strength S (w i , w j ) for all combinations of two words (w i , w j ) included in each word set. The average value is calculated as the context reliability CL (N k ) of each word set.

信頼度統合手段74は、それぞれh個の音響信頼度CA(Nk)と文脈信頼度CL(Nk)の平均値を音声文書認識信頼度として計算して出力する。このように、音響信頼度CA(Nk)と文脈信頼度CL(Nk)を用いて音声文書の認識信頼度を算出することで、実施例1の音声文書認識信頼度よりも精度の高い音声文書認識信頼度とすることができる。 The reliability integration unit 74 calculates and outputs an average value of h acoustic reliability CA (N k ) and context reliability CL (N k ) as a voice document recognition reliability. In this way, by calculating the recognition reliability of the voice document using the acoustic reliability CA (N k ) and the context reliability CL (N k ), the accuracy is higher than the voice document recognition reliability of the first embodiment. The voice document recognition reliability can be obtained.

図13に、音声文書認識信頼度を計算する処理を省力化したこの発明の音声認識装置300の機能構成例を示す。音声認識装置300は、音声文書認識信頼度高速計算部90を備える点で、音声認識装置200と異なる。音声文書認識信頼度高速計算部90は、単語集合Nk間で重複する信頼度加算処理を省くことで、音響信頼度の計算を高速に計算するようにしたものである。 FIG. 13 shows an example of the functional configuration of the speech recognition apparatus 300 of the present invention in which the processing for calculating the speech document recognition reliability is saved. The speech recognition apparatus 300 is different from the speech recognition apparatus 200 in that it includes a speech document recognition reliability high speed calculation unit 90. The voice document recognition reliability high-speed calculation unit 90 calculates the acoustic reliability at high speed by omitting the reliability addition processing overlapping between the word sets Nk .

図14に音声文書認識信頼度高速計算部90の機能構成例を示す。その動作フローを図15に示す。音声文書認識信頼度高速計算部90は、単語集合取得手段91と単語集合音響信頼度高速計算手段92を備える点で、音声認識装置200の単語集合音響信頼度計算手段72と異なる。   FIG. 14 shows a functional configuration example of the voice document recognition reliability high-speed calculation unit 90. The operation flow is shown in FIG. The speech document recognition reliability high-speed calculation unit 90 is different from the word set acoustic reliability calculation unit 72 of the speech recognition apparatus 200 in that it includes a word set acquisition unit 91 and a word set acoustic reliability high-speed calculation unit 92.

単語集合取得手段91は、音響信頼度の計算量を減らす目的で使用する直前重複フラグと直後重複フラグの2つのフラグを各単語に付与して単語集合を取得する。直前重複フラグとは、当該単語が直前の単語集合にも含まれるか否かを示す真偽値である。直後重複フラグとは、当該単語が直後の単語集合にも含まれるか否かを示す真偽値である。   The word set obtaining unit 91 obtains a word set by assigning each word with two flags, the immediately preceding duplicate flag and the immediately following duplicate flag used for the purpose of reducing the calculation amount of the acoustic reliability. The immediately preceding duplicate flag is a true / false value indicating whether or not the word is also included in the immediately preceding word set. The immediately following duplication flag is a true / false value indicating whether or not the word is also included in the immediately following word set.

単語集合取得手段91は、音声認識部10から入力される音声認識結果の単語wkを、その先頭から所定の数n個ずつnより小さい数のm個移動させながらh個の単語集合に分割する(ステップS91)。単語集合に分割する際、単語集合取得手段91は、1番目の単語集合に追加する1番目からm番目の単語の直前重複フラグBFと直後重複フラグAFを偽、m+1番目からn番目の単語の直前重複フラグBFを偽及び直後重複フラグAFを真とし、N番目の単語集合に追加する1番目から(N-1)・m番目の単語の直前重複フラグを真及び直後重複フラグを偽、(N-1)・m+1番目からn+(N-1)・m番目の単語の直前重複フラグを偽及び直後重複フラグを真とする。 The word set acquisition unit 91 divides the word w k of the speech recognition result input from the speech recognition unit 10 into h word sets while moving a predetermined number n from the beginning, m less than n. (Step S91). When the word set is divided into word sets, the word set acquisition unit 91 sets the immediately preceding duplicate flag BF and the immediately following duplicate flag AF of the first to mth words to be added to the first word set to false, and the (m + 1) th to nth The word immediately preceding duplicate flag BF is false and the immediately following duplicate flag AF is true, and the first to (N-1) -mth word immediately preceding duplicate flags to be added to the Nth word set are true and the immediately following duplicate flag is false. , (N−1) · m + 1 to n + (N−1) · m-th word are set to false, and the next duplicate flag is set to true.

単語集合音響信頼度高速計算手段92は、重複区間記憶部920を備え、直後重複フラグBFが真の単語の単語認識信頼度にその単語の継続時間長を乗算して時間正規化を解除した値とその継続時間長を、重複区間記憶部920に記憶する。そして、直前重複フラグBFが偽で直後重複フラグAFが真の単語の単語認識信頼度にその単語の継続時間長を乗算して時間正規化を解除した値とその継続時間長と、重複区間記憶部920に記憶された値とから単語集合の音響信頼度を計算する(ステップS92)。   The word set acoustic reliability high-speed calculation unit 92 includes an overlapping section storage unit 920, and a value obtained by canceling time normalization by multiplying the word recognition reliability of a true word by the word recognition reliability of the true word and the duration of the word. And the duration length thereof are stored in the overlapping section storage unit 920. Then, the value obtained by multiplying the word recognition reliability of the true word by the duration length of the word and canceling the time normalization, the duration length, and the overlap interval memory The acoustic reliability of the word set is calculated from the value stored in the unit 920 (step S92).

図16に、単語集合N1〜Nhと、各単語集合に付与された直前重複フラグBFと直後重複フラグAFの例を示す。図16の横方向は経過時間である。横方向に重なり合う単語集合を縦方向にずらして表記している。 FIG. 16 shows an example of the word sets N 1 to N h and the immediately preceding duplicate flag BF and the immediately following duplicate flag AF given to each word set. The horizontal direction in FIG. 16 is the elapsed time. The word sets that overlap in the horizontal direction are shown shifted in the vertical direction.

単語集合N1の先頭からm個の単語は、前後の単語集合内の単語と重ならないので直前重複フラグBFと直後重複フラグAFは(BF:0,AF:0)である。単語集合N1の先頭からm+1番目からn番目の単語は、直後の単語集合(N2)の単語と重複するので直前重複フラグBFと直後重複フラグAFは(BF:0,AF:1)である。以降、BF:とAF:の表記を省略する場合もある。 The m words from the beginning of the word set N 1, the front and rear because the word does not overlap with the words in the set immediately before overlap flag BF and after overlap flag AF is (BF: 0: 0, AF ). Since the words m + 1 to n from the beginning of the word set N 1 overlap with the words in the immediately following word set (N 2 ), the immediately preceding duplicate flag BF and the immediately following duplicate flag AF are (BF: 0, AF: 1 ). Hereinafter, the notation of BF: and AF: may be omitted.

2番目の単語集合N2の1番目からm番目の単語は直前の単語集合(N1)の単語とだけ重複するので(1,0)、m+1番目からn番目の単語は直前と直後の両方の単語集合の単語と重複するので(1,1)、n+1番目からn+m番目の単語は直後の単語集合の単語のみと重複するので(0,1)である。3番目以降の単語集合も2番目の単語集合と同じ関係である。 Since the 1st to m-th words in the second word set N 2 overlap only with the words in the previous word set (N 1 ) (1,0), the m + 1-th to n-th words are immediately before and after (1, 1), the (n + 1) th to n + mth words overlap with only the words of the immediately following word set (0, 1). The third and subsequent word sets have the same relationship as the second word set.

図15の例では、mをm<n/2としたので(1,1)の状態が存在するが、m=n/2とすると直前重複フラグBFと直後重複フラグAFは、N1が((0,0),(0,1))、N2以降が((1,0),(0,1))となる。つまり、単語集合取得手段91は1番目から(N-1)・m番目の単語の直前重複フラグBFを真及び直後重複フラグAFを偽、(N-1)・m+1番目からn+(N-1)・m番目の単語の直前重複フラグBFを偽及び直後重複フラグAFを真とする。 In the example of FIG. 15, since m is set to m <n / 2, there is a state (1, 1). However, when m = n / 2, the immediately preceding duplicate flag BF and the immediately following duplicate flag AF have N 1 ( (0, 0), (0, 1)), and after N 2 are ((1, 0), (0, 1)). That is, the word set acquisition unit 91 sets the immediately preceding duplicate flag BF of the first to (N−1) · mth words to be true and the immediately following duplicate flag AF to be false, and (N−1) · m + 1 to n + (N -1)-The preceding duplicate flag BF of the mth word is false and the immediately following duplicate flag AF is true.

図16に、図15に示した単語集合を取得する単語集合取得手段91の動作フローを示す。図16は窓シフト量mをm<n/2とした例である。   FIG. 16 shows an operation flow of the word set acquisition unit 91 for acquiring the word set shown in FIG. FIG. 16 shows an example in which the window shift amount m is m <n / 2.

音声文書認識信頼度計算部90が処理を開始すると、単語集合取得手段91は単語集合を特定する変数Nkと単語wnを特定する変数wk、及びカウント値iをそれぞれNk=N1,wk=w1,i=1に初期化する(ステップS91a)。そして、音声認識結果の先頭からm番目までの単語w1〜wmまでの単語wkの直前重複フラグBFを0(偽)、直後重複フラグAFを0(偽)に設定する(ステップS91b〜S91d)。更に、単語wm+1〜wnまでの単語wkの直前重複フラグBFを0(偽)、直後重複フラグAFを1(真)に設定(ステップS91e〜S91g)して単語w1〜wnまでを1番目の単語集合N1として取得する。 When spoken document recognition reliability calculating unit 90 starts the process, the word set obtaining unit 91 variable w k, and the count value i of each N k = N 1 specifying the variable N k and words w n for specifying a word set , W k = w 1 , i = 1 is initialized (step S91a). Then, 0 last overlap flag BF word w k to word w 1 to w m from the beginning of the speech recognition result to m-th (false), the immediately overlap flag AF is set to 0 (false) (step S91b~ S91d). Further, the immediately preceding duplication flag BF of words w k + 1 to w n is set to 0 (false) and the immediately following duplication flag AF is set to 1 (true) (steps S91e to S91g), and the words w 1 to w are set. Up to n are acquired as the first word set N 1 .

次に、単語集合NkをNk=Nk+1(ステップS91h)として2番目の単語集合N2を取得する。ここで、着目する単語を表す変数wkを、先頭の単語w1からmi個移動させたwk=wmiとする(ステップS91i)。ここではi=1なのでwk=wmである。 Next, the second word set N 2 is obtained with the word set N k as N k = N k + 1 (step S91h). Here, it is assumed that the variable w k representing the word of interest is w k = w mi that is moved mi from the first word w 1 (step S91i). Here, since i = 1, w k = w m .

そこで、単語wm+m、つまり、先頭の単語から2m個の単語までの直前重複フラグBFを1(真)、直後重複フラグAFを0(偽)に設定する(ステップS91j〜S91m)。そして、先頭の単語から2m+1個からn+m(i-1)個までの単語の直前重複フラグBFを1(真)、直後重複フラグAFを1(真)に設定する(ステップS91n〜S91p)。更に、先頭の単語からn+m(i-1)+1個目からn+mi個までの単語の直前重複フラグBFを0(偽)、直後重複フラグAFを1(真)に設定する(ステップS91q〜S91s)。このような処理によって、単語wm〜wn+mは、その単語の直前重複フラグBFと直後重複フラグAFが図15に示すように設定され、単語集合N2として取得される。 Thus, the word w m + m , that is, the immediately preceding overlap flag BF from the first word to 2m words is set to 1 (true), and the immediately after overlap flag AF is set to 0 (false) (steps S91j to S91m). Then, the immediately preceding duplication flag BF is set to 1 (true) and the immediately following duplication flag AF is set to 1 (true) for 2m + 1 to n + m (i-1) words from the first word (steps S91n to S91n). S91p). Further, the immediately preceding overlap flag BF is set to 0 (false) and the immediately following overlap flag AF is set to 1 (true) for the n + m (i-1) +1 to n + mi words from the first word ( Steps S91q to S91s). By such processing, the words w m to w n + m are acquired as the word set N 2 with the immediately preceding overlapping flag BF and the immediately following overlapping flag AF of the word set as shown in FIG.

そして、先頭からn+mi+1個目の単語が存在すれば、カウント値iをインクリメント(ステップS91u)、単語集合を現す変数NkもNk=Nk+1とインクリメント(ステップS91v)され、ステップS91iの処理に戻る。 If the n + mi + 1-th word from the head exists, the count value i is incremented (step S91u), and the variable N k representing the word set is also incremented to N k = N k + 1 (step S91v). The process returns to step S91i.

以上の処理(ステップS91i〜S91v)は、音声認識結果の最後の単語になるまで繰り返される(ステップS91tのno)。その結果、音声認識結果の単語は、その先頭から所定の数n個ずつnより小さい数のm個移動させながらh個の単語集合に分割される。そして、各単語集合には、図15に示すように直前重複フラグBFと直後重複フラグAFとが付与される。   The above processing (steps S91i to S91v) is repeated until the last word of the speech recognition result is reached (no in step S91t). As a result, the words of the speech recognition result are divided into h word sets while moving a predetermined number n from the beginning by m, which is smaller than n. Then, as shown in FIG. 15, the immediately preceding overlap flag BF and the immediately following overlap flag AF are assigned to each word set.

単語集合音響信頼度高速計算手段92は、重複区間記憶部920を備え、直後重複フラグAFが真の単語の単語信頼度を2度計算しないようにしたものである。図18に示す単語集合音響信頼度高速計算手段92の動作フローを参照して説明する。   The word set acoustic reliability high-speed calculation means 92 includes an overlapping section storage unit 920, and does not calculate the word reliability of a word whose true overlap flag AF is true twice. The operation will be described with reference to the operation flow of the word set acoustic reliability high speed calculation means 92 shown in FIG.

単語集合音響信頼度高速計算手段92は、最初に単語集合を特定する変数Nkと単語wnを特定する変数wkを、それぞれNk=N1,wk=w1に初期化する(ステップS92a)。そして、単語集合音響信頼度高速計算手段92は、単語(N1,w1)〜(N1,wn)の単語認識信頼度にその単語の継続時間長を乗算して時間正規化を解除した単語信頼度を計算(ステップS92b)し、その単語信頼度wcnを累積する(ステップS92c)。と共に、単語の継続時間長wdnも累積する(ステップS92d)。この処理は、単語集合N1の最後の単語wnになるまで繰り返される(ステップS92i)。 Word set sound reliability fast computation means 92, first the variable w k for specifying the variable N k and words w n for specifying a word set is initialized to N k = N 1, w k = w 1 , respectively ( Step S92a). Then, the word set acoustic reliability fast calculation means 92 multiplies the word recognition reliability of the words (N 1 , w 1 ) to (N 1 , w n ) by the duration time of the word to cancel the time normalization. Get words reliability (step S92b) and, accumulating the word confidence wc n (step S92c). At the same time, the word duration time wd n is also accumulated (step S92d). This process is repeated until the last word w n of the word set N 1 is reached (step S92i).

そして、単語集合N1の先頭からm個目の単語wmから最後の単語wnまでの単語の直後重複フラグAFが1(真)となるので、その単語wm〜wnまでの単語信頼度wcnはBackDupDMに保存される(ステップS92f)。また、単語wm〜wnの継続時間長wdnはBackDupWDに保存される(ステップS92g)。BackDupDMとBackDupWDは、例えばキュー(queue)と呼ばれるデータ構造で単語信頼度wcと継続時間長wdを保存するものとする。 Then, since the duplication flag AF immediately after the word from the m-th word w m to the last word w n of the word set N 1 is 1 (true), the word trust for the words w m to w n degrees wc n is stored in BackDupDM (step S92f). Also, duration wd n words w m to w n are stored in BackDupWD (Step S92G). BackDupDM and BackDupWD store the word reliability wc k and the duration time wd k in a data structure called a queue, for example.

最後の単語wnまでの単語信頼度の計算が終わる(ステップS92iのyes)と、単語集合N1の音響信頼度を、単語信頼度wckの累積値S(wck)を継続時間長の累積値S(wdk)で除して計算する(ステップS92i)。 When the calculation of the word reliability up to the last word w n is finished (yes in step S92i), the acoustic reliability of the word set N 1 and the cumulative value S (wc k ) of the word reliability wc k are set as the duration length. Calculation is performed by dividing by the cumulative value S (wd k ) (step S92i).

2番目の単語集合N2の先頭からn-m番目、つまり先頭の単語からn個目の単語wm〜wnは、直前の単語集合N1と重複しているので既に計算済みである。よって、BackDupDMに保存されている計算済みの単語信頼度をPreDMにコピーする(ステップS92m)と共にBackDupWDに保存されている継続時間長をPreWDにコピーする(ステップS92n)。そして、単語集合N1で単語信頼度を計算していない直後重複フラグBFが0(偽)の単語wn+1〜wn+miの正規化を解除して単語信頼度と継続時間長の累積を計算する(ステップS92pとS92q)。なお、正規化解除のステップは作図の都合により表記を省略している。 N-m-th from the second head of the word set N 2, i.e. the top of n-th word wm~wn from the word is already computed so overlaps with the immediately preceding word set N 1. Therefore, the calculated word reliability stored in BackDupDM is copied to PreDM (step S92m), and the duration length stored in BackDupWD is copied to PreWD (step S92n). Then, duplicate flag BF immediately after the word set N 1 does not calculate the word reliability is 0 word w n + 1 ~w n + mi to release the normalized word reliability and duration of the (false) The accumulation is calculated (steps S92p and S92q). Note that the denormalization step is omitted for convenience of drawing.

新たに計算される直後重複フラグBFが0(偽)の単語wn+1〜wn+miの単語信頼度wcと継続時間長wdは、BackDupDMとBackDupWDに保存される(ステップS92r)。BackDupDMとBackDupWDは、古い順番にデータが消去される構造なので最新の単語w2m〜wn+mの単語信頼度wckと継続時間長wdkが保存される。音声認識結果の先頭の単語からn個目以降の単語(wn+1〜)では、直前重複フラグBF=0の単語についてのみ正規化を解除する計算と継続時間長の累積計算を行い、新たに計算した値をキュー(BackDupDM,BackDupWD)に保存する(図16参照)。 The word reliability wc k and the duration time wd k of the words wn + 1 to wn + mi whose duplicate flag BF is 0 (false) immediately after being newly calculated are stored in BackDupDM and BackDupWD (step S92r). . Since BackDupDM and BackDupWD are structures in which data is erased in the oldest order, the word reliability wc k and duration length wd k of the latest words w 2m to w n + m are stored. For the nth and subsequent words (w n + 1 ) from the first word in the speech recognition result, only the word with the previous duplication flag BF = 0 is denormalized and the cumulative duration is calculated. The calculated values are stored in queues (BackDupDM, BackDupWD) (see FIG. 16).

単語集合N2の最後の単語wn+miまで処理が終了する(ステップS92tのyes)と、単語集合N2の音響信頼度をBackDupDMに保存されている単語信頼度の和に新たに計算した単語信頼度を加えた累積値を、BackDupWDに保存されている継続時間長の和に新たに計算した継続時間長を加えた値で除して、単語集合N2の音響信頼度を計算する(ステップS92u)。 Last word w n + mi to process word set N 2 is completed (step yes in S92t), newly computed sum word confidence stored acoustic confidence word set N 2 to BackDupDM Calculate the acoustic reliability of the word set N 2 by dividing the cumulative value including the word reliability by the value obtained by adding the newly calculated duration to the sum of the durations stored in BackDupWD ( Step S92u).

そして、カウント値iがインクリメント(ステップS92x)され、着目する単語集合NkもインクリトされN3以降の単語集合に対して同じ処理が、音声認識結果の単語が無くなるまで繰り返される(ステップS92wのyes)。このように、直前の単語集合Nkで計算済みの単語集合と継続時間長は、コピーして計算されるので2度計算することが無い。 Then, the count value i is incremented (step S92x), the word set N k of interest is incremented, and the same processing is repeated for the word sets after N 3 until there are no words in the speech recognition result (yes in step S92w). ). In this way, the word set already calculated in the immediately preceding word set N k and the duration length are copied and calculated, and thus are not calculated twice.

つまり、単語集合音響信頼度高速計算手段92は、重複区間記憶部920に直後重複フラグBFが真の単語の単語認識信頼度D(wk)にその単語の継続時間長wdkを乗算して時間正規化を解除した値wckとその継続時間長wdkを記憶し、直前重複フラグAFが偽で直後重複フラグBFが真の単語の単語認識信頼度D(wk)にその単語wkの継続時間長を乗算して時間正規化を解除した値wckとその継続時間長wdkと、重複区間記憶部920に記憶された値とから単語集合wkの音響信頼度を計算する。よって、実施例1と2の音声認識装置よりも高速に音響信頼度を計算することができる。 That is, the word set acoustic reliability high-speed calculation unit 92 multiplies the word recognition reliability D (w k ) of a word whose true overlap flag BF is true by the overlap interval storage unit 920 by the duration duration wd k of the word. Stores the value wc k with the time normalization canceled and its duration wd k, and sets the word w k to the word recognition reliability D (w k ) of the word with the immediately preceding duplicate flag AF being false and the immediately following duplicate flag BF being true. The acoustic reliability of the word set w k is calculated from the value wc k obtained by canceling the time normalization by multiplying the continuous time length, the continuous time length wd k, and the value stored in the overlapping section storage unit 920. Therefore, the acoustic reliability can be calculated at a higher speed than the speech recognition apparatuses according to the first and second embodiments.

なお、直前重複フラグBFと直後重複フラグAFを用いて音響信頼度を高速に計算する例を説明したが、その処理方法は上記した例に限定されない。例えば、図18に示した例に加えて、重複区間の単語信頼度の合計と重複区間の継続時間長の合計を、それぞれ記憶して置くようにしても良い。そのようにすれば、音響信頼度を計算する度(ステップS92u)に行うキューに記憶された単語信頼度と継続時間長の累積値を計算する処理も削減することが可能である。   Although an example in which the acoustic reliability is calculated at high speed using the immediately preceding overlap flag BF and the immediately following overlap flag AF has been described, the processing method is not limited to the above-described example. For example, in addition to the example shown in FIG. 18, the total word reliability of overlapping sections and the total duration time of overlapping sections may be stored. By doing so, it is possible to reduce processing for calculating the cumulative value of the word reliability and the duration length stored in the queue every time the acoustic reliability is calculated (step S92u).

上記装置における処理手段をコンピュータによって実現する場合、各装置が有すべき機能の処理内容はプログラムによって記述される。そして、このプログラムをコンピュータで実行することにより、各装置における処理手段がコンピュータ上で実現される。   When the processing means in the above apparatus is realized by a computer, the processing contents of the functions that each apparatus should have are described by a program. Then, by executing this program on the computer, the processing means in each apparatus is realized on the computer.

この処理内容を記述したプログラムは、コンピュータで読み取り可能な記録媒体に記録しておくことができる。コンピュータで読み取り可能な記録媒体としては、例えば、磁気記録装置、光ディスク、光磁気記録媒体、半導体メモリ等どのようなものでもよい。具体的には、例えば、磁気記録装置として、ハードディスク装置、フレキシブルディスク、磁気テープ等を、光ディスクとして、DVD(Digital Versatile Disc)、DVD−RAM(Random Access Memory)、CD−ROM(Compact Disc Read Only Memory)、CD−R(Recordable)/RW(ReWritable)等を、光磁気記録媒体として、MO(Magneto Optical disc)等を、半導体メモリとしてEEP−ROM(Electronically Erasable and Programmable-Read Only Memory)等を用いることができる。   The program describing the processing contents can be recorded on a computer-readable recording medium. As the computer-readable recording medium, for example, any recording medium such as a magnetic recording device, an optical disk, a magneto-optical recording medium, and a semiconductor memory may be used. Specifically, for example, as a magnetic recording device, a hard disk device, a flexible disk, a magnetic tape or the like, and as an optical disk, a DVD (Digital Versatile Disc), a DVD-RAM (Random Access Memory), a CD-ROM (Compact Disc Read Only). Memory), CD-R (Recordable) / RW (ReWritable), etc., magneto-optical recording medium, MO (Magneto Optical disc), etc., semiconductor memory, EEP-ROM (Electronically Erasable and Programmable-Read Only Memory), etc. Can be used.

また、このプログラムの流通は、例えば、そのプログラムを記録したDVD、CD−ROM等の可搬型記録媒体を販売、譲渡、貸与等することによって行う。さらに、このプログラムをサーバコンピュータの記録装置に格納しておき、ネットワークを介して、サーバコンピュータから他のコンピュータにそのプログラムを転送することにより、このプログラムを流通させる構成としてもよい。   The program is distributed by selling, transferring, or lending a portable recording medium such as a DVD or CD-ROM in which the program is recorded. Further, the program may be distributed by storing the program in a recording device of a server computer and transferring the program from the server computer to another computer via a network.

また、各手段は、コンピュータ上で所定のプログラムを実行させることにより構成することにしてもよいし、これらの処理内容の少なくとも一部をハードウェア的に実現することとしてもよい。   Each means may be configured by executing a predetermined program on a computer, or at least a part of these processing contents may be realized by hardware.

Claims (5)

入力される音声文書を音声認識処理した単語毎に単語認識信頼度を付与した音声認識結果を出力する音声認識過程と、
上記音声認識結果を入力として音声文書全体の認識信頼度である音声文書認識信頼度を計算して出力する音声認識信頼度計算過程と、
上記音声認識結果と上記音声文書認識信頼度とを入力として所定の音声文書認識信頼度閾値未満の音声文書を除去する音声文書除去過程と、
上記音声文書除去過程で除去されなかった音声文書の音声認識結果から所定の単語認識信頼度閾値未満の単語認識信頼度の単語を除去する単語除去過程と、
を含む音声認識方法。
A speech recognition process for outputting a speech recognition result to which word recognition reliability is given for each word obtained by performing speech recognition processing on an input speech document;
A speech recognition reliability calculation process for calculating and outputting a speech document recognition reliability, which is a recognition reliability of the entire speech document, using the speech recognition result as an input;
An audio document removal process of removing an audio document having a threshold value less than a predetermined audio document recognition reliability threshold by inputting the audio recognition result and the audio document recognition reliability;
A word removal process for removing words having a word recognition reliability less than a predetermined word recognition reliability threshold from the voice recognition result of the voice document that has not been removed in the voice document removal process;
A speech recognition method including:
請求項1に記載した音声認識方法において、
上記音声認識信頼度計算過程は、
上記音声認識結果の単語を、その先頭から所定の数n個ずつnより小さい数のm個移動させながら単語集合に分割する単語集合取得ステップと、
上記単語認識信頼度にその単語の継続時間長を乗算して時間正規化を解除した値を足し合わせ、その足し合わせた値を上記単語集合の全単語の継続時間長の総和で除算して単語集合音響信頼度として求める単語集合音響信頼度計算ステップと、
上記音声認識結果に含まれる全ての単語間の組み合わせの2単語間の関連度を表した単語関連度テーブルを参照して上記単語集合に含まれる2単語の組み合わせの関連の強さの平均値を単語集合文脈信頼度として求める単語集合文脈信頼度計算ステップと、
上記音声文書全体の上記単語集合音響信頼度と上記単語集合文脈信頼度とを平均した値を、音声文書認識信頼度として求める信頼度統合ステップと、
を含むことを特徴とする音声認識方法。
The speech recognition method according to claim 1,
The above speech recognition reliability calculation process is as follows:
A word set acquisition step of dividing the word of the speech recognition result into a word set while moving a predetermined number n from the beginning by m, which is a number smaller than n;
Multiply the word recognition reliability by the duration of the word and cancel the time normalization, add the sum, and divide the sum by the sum of the durations of all words in the word set. A word set sound reliability calculation step to be obtained as a set sound reliability;
By referring to a word association degree table representing the degree of association between two words among all the combinations between words included in the speech recognition result, an average value of association strengths of the combinations of two words included in the word set is obtained. A word set context reliability calculation step to obtain as word set context reliability,
A reliability integration step of obtaining a value obtained by averaging the word set acoustic reliability and the word set context reliability of the entire voice document as a voice document recognition reliability;
A speech recognition method comprising:
請求項1に記載した音声認識方法において、
上記音声認識信頼度計算過程は、
上記音声認識結果の単語を、その先頭から所定の数n個ずつnより小さい数のm個移動させながら単語集合に分割し、1番目の単語集合に追加する1番目からm番目の単語の直前重複フラグと直後重複フラグを偽とし、N番目の単語集合に追加する1番目からN・m番目の単語の直前重複フラグを真及び直後重複フラグを偽、N・m+1番目からn+N・m番目の単語の直前重複フラグを偽及び直後重複フラグを真とする単語集合取得ステップと、
上記直後重複フラグが真の単語の上記単語認識信頼度にその単語の継続時間長を乗算して時間正規化を解除した値とその継続時間長を記憶する重複区間記憶ステップと、
上記直前重複フラグが偽で直後重複フラグが真の単語の上記単語認識信頼度にその単語の継続時間長を乗算して時間正規化を解除した値とその継続時間長と、上記重複区間記憶ステップで記憶された値とから当該単語集合の音響信頼度を計算する単語集合音響信頼度高速計算ステップと、
を含むことを特徴とする音声認識方法。
The speech recognition method according to claim 1,
The above speech recognition reliability calculation process is as follows:
The word of the speech recognition result is divided into a word set while moving a predetermined number n from the beginning by m less than n, and immediately before the first to mth words to be added to the first word set The duplicate flag and the immediately following duplicate flag are set to false, the immediately preceding duplicate flag of the first to N · mth words to be added to the Nth word set is true and the immediately following duplicate flag is false, and the N · m + 1 to n + N A word set acquisition step in which the immediately preceding duplicate flag of the mth word is false and the immediately following duplicate flag is true;
The overlapping section storage step for storing the value obtained by multiplying the word recognition reliability of the word for which the duplication flag is true by the duration length of the word and canceling the time normalization, and the duration length, and
A value obtained by multiplying the word recognition reliability of the word whose previous duplicate flag is false and the true duplicate flag is true by the duration length of the word and canceling the time normalization, the duration duration, and the overlapping section storing step A word set acoustic reliability fast calculation step for calculating the acoustic reliability of the word set from the value stored in
A speech recognition method comprising:
入力される音声文書を音声認識処理した単語毎に単語認識信頼度を付与した音声認識結果を出力する音声認識部と、
上記音声認識結果を入力として音声文書全体の認識信頼度である音声文書認識信頼度を計算して出力する音声認識信頼度計算部と、
上記音声認識結果と上記音声文書認識信頼度とを入力として所定の音声文書認識信頼度閾値未満の音声文書を除去するする音声文書除去部と、
上記音声文書除去部で除去されなかった音声文書の音声認識結果から所定の単語認識信頼度閾値未満の単語認識信頼度の単語を除去する単語除去部と、
を具備する音声認識装置。
A speech recognition unit that outputs a speech recognition result to which word recognition reliability is given for each word obtained by performing speech recognition processing on the input speech document;
A speech recognition reliability calculation unit that calculates and outputs a speech document recognition reliability that is a recognition reliability of the entire speech document by using the speech recognition result as an input;
A voice document removal unit that removes a voice document that is less than a predetermined voice document recognition reliability threshold by inputting the voice recognition result and the voice document recognition reliability;
A word removal unit that removes words having a word recognition reliability less than a predetermined word recognition reliability threshold from the voice recognition result of the voice document that has not been removed by the voice document removal unit;
A speech recognition apparatus comprising:
請求項1乃至3の何れかに記載した音声認識方法を、コンピュータに実行させるための音声認識方法プログラム。   A speech recognition method program for causing a computer to execute the speech recognition method according to claim 1.
JP2010158472A 2010-07-13 2010-07-13 Speech recognition method, apparatus and program thereof Expired - Fee Related JP5406797B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2010158472A JP5406797B2 (en) 2010-07-13 2010-07-13 Speech recognition method, apparatus and program thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2010158472A JP5406797B2 (en) 2010-07-13 2010-07-13 Speech recognition method, apparatus and program thereof

Publications (2)

Publication Number Publication Date
JP2012022069A true JP2012022069A (en) 2012-02-02
JP5406797B2 JP5406797B2 (en) 2014-02-05

Family

ID=45776413

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2010158472A Expired - Fee Related JP5406797B2 (en) 2010-07-13 2010-07-13 Speech recognition method, apparatus and program thereof

Country Status (1)

Country Link
JP (1) JP5406797B2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018134916A1 (en) * 2017-01-18 2018-07-26 三菱電機株式会社 Speech recognition device
WO2019030810A1 (en) * 2017-08-08 2019-02-14 三菱電機株式会社 Speech recognition device and speech recognition method

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002202797A (en) * 2000-11-16 2002-07-19 Sony Internatl Europ Gmbh For recognizing method speech
JP2005148342A (en) * 2003-11-14 2005-06-09 Nippon Telegr & Teleph Corp <Ntt> Method for speech recognition, device, and program and recording medium for implementing the same method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002202797A (en) * 2000-11-16 2002-07-19 Sony Internatl Europ Gmbh For recognizing method speech
JP2005148342A (en) * 2003-11-14 2005-06-09 Nippon Telegr & Teleph Corp <Ntt> Method for speech recognition, device, and program and recording medium for implementing the same method

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018134916A1 (en) * 2017-01-18 2018-07-26 三菱電機株式会社 Speech recognition device
WO2019030810A1 (en) * 2017-08-08 2019-02-14 三菱電機株式会社 Speech recognition device and speech recognition method
JPWO2019030810A1 (en) * 2017-08-08 2019-11-14 三菱電機株式会社 Speech recognition apparatus and speech recognition method

Also Published As

Publication number Publication date
JP5406797B2 (en) 2014-02-05

Similar Documents

Publication Publication Date Title
JP2007256342A (en) Clustering system, clustering method, clustering program and attribute estimation system using clustering program and clustering system
JP6622681B2 (en) Phoneme Breakdown Detection Model Learning Device, Phoneme Breakdown Interval Detection Device, Phoneme Breakdown Detection Model Learning Method, Phoneme Breakdown Interval Detection Method, Program
JP6585022B2 (en) Speech recognition apparatus, speech recognition method and program
KR20100130263A (en) Apparatus and method for extension of articulation dictionary by speech recognition
Khan et al. An intelligent system for spoken term detection that uses belief combination
Sadeghian et al. Towards an automatic speech-based diagnostic test for Alzheimer’s disease
Moyal et al. Phonetic search methods for large speech databases
Granell et al. A multimodal crowdsourcing framework for transcribing historical handwritten documents
Das et al. Optimal prosodic feature extraction and classification in parametric excitation source information for Indian language identification using neural network based Q-learning algorithm
Mary et al. Searching speech databases: features, techniques and evaluation measures
JP5406797B2 (en) Speech recognition method, apparatus and program thereof
US8639510B1 (en) Acoustic scoring unit implemented on a single FPGA or ASIC
Rahmawati et al. Java and Sunda dialect recognition from Indonesian speech using GMM and I-Vector
JP5149941B2 (en) Speech recognition method, apparatus and program thereof
KR20130126570A (en) Apparatus for discriminative training acoustic model considering error of phonemes in keyword and computer recordable medium storing the method thereof
JP2011248107A (en) Voice recognition result search method, apparatus and program for the same
US20050246172A1 (en) Acoustic model training method and system
Tejedor et al. Search on speech from spoken queries: the multi-domain International ALBAYZIN 2018 query-by-example spoken term detection evaluation
Chen et al. Topic segmentation on spoken documents using self-validated acoustic cuts
JP2001312293A (en) Method and device for voice recognition, and computer- readable storage medium
JP4478925B2 (en) Speech recognition result reliability verification apparatus, computer program, and computer
Ravi et al. Phoneme segmentation-based unsupervised pattern discovery and clustering of speech signals
JP5325176B2 (en) 2-channel speech recognition method, apparatus and program thereof
Dandeniya Profanity filtering in speech contents using deep learning algorithms
JP2005173008A (en) Voice analysis processing, voice processor using same, and medium

Legal Events

Date Code Title Description
A621 Written request for application examination

Free format text: JAPANESE INTERMEDIATE CODE: A621

Effective date: 20121101

A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20130610

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20130618

A521 Request for written amendment filed

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20130718

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20130827

A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20130828

A521 Request for written amendment filed

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20130920

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20131022

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20131101

R150 Certificate of patent or registration of utility model

Ref document number: 5406797

Country of ref document: JP

Free format text: JAPANESE INTERMEDIATE CODE: R150

S531 Written request for registration of change of domicile

Free format text: JAPANESE INTERMEDIATE CODE: R313531

R350 Written notification of registration of transfer

Free format text: JAPANESE INTERMEDIATE CODE: R350

LAPS Cancellation because of no payment of annual fees