WO2010113396A1 - Device, method, program for reading determination, computer readable medium therefore, and voice synthesis device - Google Patents

Device, method, program for reading determination, computer readable medium therefore, and voice synthesis device Download PDF

Info

Publication number
WO2010113396A1
WO2010113396A1 PCT/JP2010/001753 JP2010001753W WO2010113396A1 WO 2010113396 A1 WO2010113396 A1 WO 2010113396A1 JP 2010001753 W JP2010001753 W JP 2010001753W WO 2010113396 A1 WO2010113396 A1 WO 2010113396A1
Authority
WO
WIPO (PCT)
Prior art keywords
reading
word
information
reading determination
word set
Prior art date
Application number
PCT/JP2010/001753
Other languages
French (fr)
Japanese (ja)
Inventor
近藤玲史
安藤真一
Original Assignee
日本電気株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電気株式会社 filed Critical 日本電気株式会社
Priority to JP2011506983A priority Critical patent/JP5533853B2/en
Publication of WO2010113396A1 publication Critical patent/WO2010113396A1/en

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/08Text analysis or generation of parameters for speech synthesis out of text, e.g. grapheme to phoneme translation, prosody generation or stress or intonation determination

Definitions

  • the present invention relates to, for example, a reading determination device, a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium for determining a reading of a word having a plurality of reading candidates.
  • the present invention relates to a reading determination device, a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium thereof that can be easily and appropriately determined.
  • a method of determining how to read a given sentence for example, “reading” (reading kana, accent information, etc.) such as letters and words is defined in advance in the dictionary, and the grammatical connection relation of each word in the sentence is determined.
  • a method for determining how to read the entire sentence based on the reading defined in the dictionary while checking is widely known (for example, Non-Patent Documents 1 and 2).
  • a method of giving a more appropriate reading as a sentence by considering phonological rules such as rendaku and devoicing is also known.
  • Non-Patent Document 2 the part-of-speech relationship is used as the rule.
  • “market” has two types of readings, “Ichiba” and “shijo”, and “kuroko” (noun) means “mole” and “black”.
  • “ ⁇ ” is read as “iso”, but when used in general nouns, it becomes a flat accent, and when used in personal names, it becomes a head-high accent. Therefore, for example, such differences are also important when performing speech synthesis or the like. In order to give a correct reading to a sentence, it is desirable to appropriately select these multiple readings (including accents).
  • the present invention has been made to solve such problems, and is a reading determination device, a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium that can easily and appropriately determine how to read a word.
  • the main purpose is to provide
  • One aspect of the present invention for achieving the above object is a reading determination apparatus for determining how to read a word, having a plurality of reading candidates, and a word set comprising a plurality of element words similar to the reading candidate Respectively, a corpus database for storing corpus information including a plurality of example sentences, and the word generated by the word set generating means based on the corpus information stored in the corpus database.
  • It is a reading judgment apparatus characterized by comprising reading judgment information generating means for generating judgment information.
  • Another aspect of the present invention for achieving the above object is a reading determination method for determining how to read a word having a plurality of reading candidates, and includes a plurality of element words similar to the reading candidate.
  • Each of the word sets is generated, and based on the corpus information including a plurality of example sentences, the feature amounts for the plurality of element words of the generated word set are calculated, respectively, and the feature amounts for the plurality of element words of the calculated word set And reading method determination information in which the reading candidates are associated with each other.
  • one aspect of the present invention for achieving the above object is a non-transitory computer-readable medium that stores a reading determination program for determining how to read a word, having a plurality of reading candidates. Based on the processing for generating each word set composed of a plurality of element words similar to the candidate and corpus information including a plurality of example sentences, the feature amounts for the plurality of element words of the generated word set are calculated.
  • a non-temporary storing a reading judgment program that causes a computer to execute processing, and processing to generate reading judgment information in which the calculated feature quantities for a plurality of element words of the word set and the reading candidates are associated with each other Computer readable medium.
  • a reading determination device a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium thereof that can easily and appropriately determine how to read a word.
  • FIG. 1 is a functional block diagram of a reading determination apparatus according to an embodiment of the present invention.
  • the reading determination apparatus 10 is an apparatus for determining how to read a word having a plurality of reading candidates.
  • the reading determination device 10 generates a word set generation unit 14 that generates a word set, a corpus database 15 that stores corpus information, a feature amount calculation unit 16 that calculates feature amounts of element words, and generates reading determination information.
  • Reading information determination means 17 for reading is an apparatus for determining how to read a word having a plurality of reading candidates.
  • the reading determination device 10 generates a word set generation unit 14 that generates a word set, a corpus database 15 that stores corpus information, a feature amount calculation unit 16 that calculates feature amounts of element words, and generates reading determination information.
  • Reading information determination means 17 for reading.
  • the word set generation unit 14 generates a word set composed of a plurality of element words similar to reading candidates.
  • the corpus database 15 stores corpus information including a plurality of example sentences.
  • the feature amount calculation means 16 calculates the feature amounts for a plurality of element words of the word set generated by the word set generation means 14 based on the corpus information stored in the corpus database 15.
  • the reading determination information generation unit 17 generates reading determination information in which the feature amounts for a plurality of element words in the word set calculated by the feature amount calculation unit 16 are associated with the reading candidates.
  • FIG. 2 is a block diagram showing an example of a schematic system configuration of the reading determination device according to the first embodiment of the present invention.
  • the reading determination apparatus 10 includes a same pronunciation different pronunciation word set generation unit 11, a reading candidate DB (database) 12, a thesaurus DB (database) 13, a word set generation unit 14, and a corpus DB (database). ) 15, a context vector generation unit 16, and a reading determination information generation unit 17.
  • the same notation different pronunciation word set generation unit 11 reads from the reading candidate DB 12 a plurality of reading candidates (reading candidate 1, reading candidate 2,..., Reading candidates for the input word. M) and word meanings corresponding to each (meaning 1, meaning 2,..., Meaning M) are acquired.
  • a phoneme string used when uttering a given text is referred to as “reading”.
  • a Japanese syllable sequence is used as a phoneme sequence, but an arbitrary phonetic symbol sequence such as an international phonetic symbol (IPA) can be used without depending on a language.
  • IPA international phonetic symbol
  • “how to read” includes accent information (accent position, separation, etc.) indicating how the corresponding phoneme string is uttered, and supplementary reading information such as vowel devoicing, accompanying the phoneme string. Also good.
  • the same notation different pronunciation word set generation unit 11 outputs a plurality of reading candidates for the acquired input word to the word set generation unit 14.
  • the reading candidate DB 12 stores, for example, a plurality of entries in which words, parts of speech, readings, and categories on the thesaurus are associated as a set as shown in FIG.
  • the notation different pronunciation word set generation unit 11 searches and acquires a corresponding entry from a plurality of entries stored in the reading candidate DB 12 based on the input word “Kuroko”.
  • the thesaurus DB 13 stores systematic thesaurus dictionary information by classifying each word according to the upper / lower relationship, partial / whole relationship, synonym relationship, synonym relationship, etc. of a plurality of words as shown in FIG. is doing.
  • the word set generation unit (word set generation means) 14 selects a word set composed of a plurality of element words similar to a plurality of reading candidates from the notation different pronunciation word set generation unit 11 based on thesaurus dictionary information of the thesaurus DB 13. , Respectively.
  • the plurality of element words similar to the reading candidate include, for example, words belonging to the same category as the reading candidate category on the thesaurus, and include words having the same meaning in a broad sense.
  • word set 1 ⁇ element word 1-1: stain
  • element word 1-6 lentigo ⁇
  • category ⁇ body stain ⁇
  • the number of element words N1 6.
  • word set 2 ⁇ element word 2-1: Kuroko
  • element word 2-2 black
  • element word 2-3 black
  • element word 2-4 guardianship
  • element word 2-5 black tool ⁇
  • Category ⁇ kabuki assistant ⁇
  • the number of element words N2 5.
  • the word set generation unit 14 uses the thesaurus dictionary information in the thesaurus DB 13 to generate a word set of reading candidates.
  • the word set generation unit 14 is not limited to this, and may use synonym dictionary information, for example.
  • the dictionary information can be used.
  • the word set generation unit 14 can determine whether each word belongs to the same category by calculating a case frame.
  • the reading candidate DB 12 can have corresponding frames (each frame 1,..., Each frame N) instead of the word meaning (meaning 1,..., Word N).
  • the word set generation unit 14 outputs the generated word set to the context vector generation unit 16.
  • the corpus DB 15 stores corpus information including a plurality of example sentences (text data).
  • the example of the corpus information may not be given a reading.
  • the context vector generation unit 16 calculates context vectors for a plurality of element words of the word set generated by the word set generation unit 14 based on the corpus information stored in the corpus DB 15.
  • the context vector generation unit 16 first extracts an example sentence using the element word from the corpus information of the corpus DB 15 for each element word of the word set generated by the word set generation unit 14. .
  • the context vector generation unit 16 calculates the context vector of the element word using each extracted example sentence.
  • the context vector is used as, for example, a feature amount indicating the situation of words around the corresponding word.
  • Non-Patent Document 2 (Iwanami Lecture, Software Science 15, Natural Language Processing, pp. 421-424), 1 does not appear when the index word T (i) appears in the entire document.
  • a document vector having 0 as a coefficient is known.
  • the context vector in this embodiment focuses on the independent words around the corresponding element word in the vector space around all the independent words included in the example sentence of the corpus information, and calculates the coefficient. You may choose.
  • the description will be given focusing on the context vector as the feature amount expressing the situation of the words around the corresponding element word, but the present invention is not limited to this.
  • the corpus information of the words around the corresponding element word is used.
  • a feature quantity other than the vector expression may be used, such as using the sum of the appearance probabilities.
  • it will be described as a context vector including feature quantities other than these vector expressions.
  • the context vector when calculating the context vector, it is possible to extract words that represent more features using grammatical knowledge, compress the context vector by a general method such as synonym compression or dimension compression of each element word, The use efficiency of the context vector space may be improved. Regardless of the use or non-use of these methods, it can be said that the storage of the context vector can be realized at a low additional cost because it requires a smaller storage capacity than the storage of the word bigram.
  • the context vector generation unit 16 selects all of the example sentences or a predetermined number of example sentences. A similar process is performed to calculate a plurality of context vectors. On the other hand, when there is no example sentence using the corresponding element word in the corpus information, the context vector generation unit 16 does not calculate the context vector of the element word.
  • the context vector generation unit 16 outputs to the reading determination information generation unit 17 the context vector calculated as described above and used as a plurality of element words of the word set.
  • the reading determination information generation unit (reading determination information generation means) 17 generates reading determination information in which context vectors for a plurality of element words of the word set calculated by the context vector generation unit 16 are associated with reading candidates. For example, for each reading candidate, the reading determination information generation unit 17 calculates a representative average context vector (representative context vector) obtained by arithmetically averaging a plurality of corresponding context vectors in the context vector space. Then, the reading determination information generation unit 17 generates reading determination information in which each reading candidate is associated with the calculated representative average context vector (arithmetic mean value).
  • the reading determination device 10 includes, for example, a CPU (Central Processing Unit) that performs control processing, arithmetic processing, and the like, a ROM (Read (OnlymMemory) that stores a control program executed by the CPU, an arithmetic program, processing data, and the like Is composed of a microcomputer mainly including a RAM (Random Access Memory) or the like that temporarily stores. Also, the same notation different pronunciation word set generation unit 11, word set generation unit 14, context vector generation unit 16, reading determination information generation unit 17, and later-described reading determination unit 21 are stored in the ROM, for example, It can be realized by a program executed by the CPU.
  • a CPU Central Processing Unit
  • ROM Read (OnlymMemory) that stores a control program executed by the CPU
  • an arithmetic program, processing data, and the like Is composed of a microcomputer mainly including a RAM (Random Access Memory) or the like that temporarily stores.
  • FIG. 5 is a flowchart showing an example of a processing flow of the reading determination apparatus according to the first embodiment of the present invention.
  • the same notation different pronunciation word set generation unit 11 acquires a plurality of reading candidates for the input word from the reading candidate DB 12 based on the input word input by the user (step S101), and acquires the plurality of acquired readings. Candidates are output to the word set generation unit 14.
  • the word set generation unit 14 generates a word set composed of a plurality of element words similar to a plurality of reading candidates from the notation different pronunciation word set generation unit 11 based on thesaurus dictionary information in the thesaurus DB 13.
  • the generated word set is output to the context vector generation unit 16.
  • the context vector generation unit 16 calculates context vectors for a plurality of element words of the word set generated by the word set generation unit 14 based on the corpus information stored in the corpus DB 15 (step S103). The calculated context vector is output to the reading determination information generation unit 17.
  • the reading determination information generation unit 17 calculates a representative average context vector obtained by arithmetically averaging a plurality of corresponding context vectors in the context vector space for each reading candidate (step S104). Then, the reading determination information generation unit 17 generates reading determination information in which each reading candidate is associated with the calculated average context vector (step S105).
  • a word set composed of element words similar to a plurality of reading candidates is generated, and an average context vector for each reading candidate is calculated using corpus information. Then, reading determination information in which each reading candidate is associated with the average context vector is generated.
  • the information amount of the reading determination information can be effectively increased and the accuracy thereof can be improved. Therefore, more appropriate and highly accurate reading determination information can be acquired, and further, how to read a word can be determined easily and appropriately using this reading determination information.
  • the present embodiment is more effective than the amount of information related to an arbitrary synonym pronoun word set obtained from the corpus information.
  • the information amount of the reading determination information generated using the word set and the corpus information is larger and more accurate.
  • it is not necessary to prepare a large amount of learning corpus information to which correct reading is given, or to write rules for a large number of homophones with the same notation, and information on reading determination information This is superior in that the amount can be increased efficiently and the accuracy of reading can be improved.
  • since reading is estimated based on information such as word similarity, improvement in reading accuracy can be expected by improving the estimation accuracy.
  • FIG. 6 is a block diagram showing a schematic system configuration of the reading determination apparatus according to the second embodiment of the present invention.
  • the reading determination device 20 reads the input word based on the reading determination information generated by the reading determination information generation unit 17 in addition to the configuration of the reading determination device 10 according to the first embodiment. It further includes a reading determination unit 21 for determining, and an output device 22 for outputting the determination of the reading.
  • the output device 22 for example, a display device, a printer device, an audio output device, or the like can be used.
  • the reading determination device 20 according to the present embodiment can determine how to read an input word online, for example.
  • FIG. 7 is a flowchart showing an example of a processing flow of the reading determination apparatus according to the second embodiment of the present invention. For example, information specifying an input sentence and an input word in the input sentence is input to the context vector generation unit 16 (step S201).
  • the context vector generation unit 16 calculates a context vector for the specified input word in the input sentence as in the first embodiment (step S202), and outputs it to the reading determination unit 21.
  • the reading determination unit 21 inputs the input word in the input sentence based on the context vector calculated by the context vector generation unit 16 and the average context vector of the reading determination information generated by the reading determination information generation unit 17. Is read (step S203).
  • the reading determination information is a plurality of sets of information in which an average context vector and a reading candidate are associated with each other.
  • the reading determination unit 21 selects a reading candidate corresponding to the average context vector having the smallest cosine distance (high similarity) from the context vector of the input word among the plurality of average context vectors of the reading determination information. It is determined that the input word is read.
  • the reading determination unit 21 outputs the determined input word reading to the output device 22.
  • the output device 22 outputs how to read the input word output from the reading determination unit 21 by, for example, screen display, print display, voice, or the like (step S204).
  • the most similar reading is selected from the features of each reading candidate represented by the average context vector, and thus a more appropriate reading can be determined.
  • FIG. 8 is a block diagram showing a schematic system configuration of a reading determination apparatus according to the third embodiment of the present invention.
  • the reading determination device 30 according to the third embodiment includes an example sentence word acquisition unit 31 that acquires example sentence information of input words from the corpus DB 15, and a reading determination unit 21. And a reading DB 32 for storing the reading of the word determined by the above.
  • the reading determination device 30 according to the present embodiment can determine how to read an input word offline, for example.
  • the example sentence word acquisition unit 31 acquires a plurality of example sentences including the input word from the corpus DB 15, extracts information specifying the input word from each example sentence, and outputs the information to the context vector generation unit 16.
  • the context vector generation unit 16 calculates the context vector for the input word of each example sentence using the information of each example sentence from the example sentence word acquisition unit 31, and outputs it to the reading determination unit 21.
  • the reading determination unit 21 selects reading candidates corresponding to the average context vector having the smallest cosine distance from the context vector for each example sentence among the plurality of average context vectors of the reading determination information generated by the reading determination information generation unit 17. And how to read the input word in the example sentence. Then, the reading determination unit 21 outputs the determined reading of the input word in each example sentence to the reading DB 32 and stores it in the reading DB 32. Further, the reading determination unit 21 selects one of the statistically highest readings from the plurality of readings stored in the reading DB 32, determines the reading as the corresponding input word, and outputs the determined reading as the output device 22. Output for.
  • a plurality of context vectors for the input word are generated using the corpus information, each reading is determined, and stored in the reading DB 32. Then, based on the frequency of the plurality of readings stored in the reading DB 32, it is possible to statistically determine how to read the input word.
  • the reading determination information generation unit 17 stores reading determination information in which each reading candidate is associated with one representative average context vector. Although generated, in the reading determination device 40 according to the fourth embodiment, the reading determination information generation unit 47 generates reading determination information in which each reading candidate and a plurality of context vectors are associated with each other.
  • the reading determination information is a set of combinations of each context vector calculated by the context vector generation unit 16 and corresponding reading candidates corresponding to the number of context vectors.
  • the reading determination information of the first to third embodiments is a collection of the average context vector and the reading candidate corresponding to the average context vector as many as the number of reading candidates. Therefore, the reading determination information according to the present embodiment has a larger information amount and higher accuracy.
  • the reading determination information generation unit 47 sets the reading determination information obtained by combining all the context vectors generated by the context vector generation unit 16 and the reading candidates corresponding to the respective context vectors. Output for.
  • the reading determination device 40 In the reading determination device 40 according to the fourth embodiment, other configurations are substantially the same as the reading determination devices 10, 20, and 30 according to the first to third embodiments. Therefore, the same parts are denoted by the same reference numerals, and detailed description thereof is omitted.
  • the reading determination unit 21 resembles the cosine distance or the like between the context vector obtained from the input sentence and the input word and all context vectors of the reading determination information. Each degree may be calculated. Then, the reading determination unit 21 determines the reading candidate corresponding to the context vector having the highest similarity among all the context vectors of the reading determination information as the reading of the input word.
  • the reading determination unit 21 calculates the similarity between the context vector obtained from each example sentence in the corpus DB 15 and all the context vectors of the reading determination information. Also good. Then, the reading determination unit 21 determines the reading candidate corresponding to the context vector having the highest similarity to the context vector of each example sentence among all the context vectors of the reading determination information as the reading of the input word in the example sentence. Output to the reading DB 32.
  • the word set generation unit 14 includes a word set including element words belonging to the same category on the thesaurus as a word set including a plurality of element words similar to a plurality of reading candidates.
  • the word set generation unit 54 generates a word set including element words belonging to adjacent categories in the same hierarchy on the thesaurus.
  • the other configuration is substantially the same as that of the reading determination device 10 according to the first embodiment. Therefore, the same parts are denoted by the same reference numerals, and detailed description thereof is omitted.
  • the element words having the similar relationship include, in addition to the element words belonging to the same category on the thesaurus, the element words belonging to the adjacent category of the same hierarchy on the thesaurus.
  • the adjacent category refers to the closest category using, for example, the degree of relationship between categories defined on the thesaurus.
  • the word set generation unit 54 extracts, for example, element words that belong to the same category as the category to which the reading candidate 1 “mole” belongs and a close category on the thesaurus shown in FIG. Then, the word set generation unit 54 follows the similarity between categories defined in advance in the thesaurus from the “body surface state” category and the “body color” category in the same hierarchy as the “body stain” category. , Select the “body surface” category. Further, the word set generation unit 54 extracts element words belonging to the “body stain” category and the “body surface state” category, and generates a word set 1.
  • the word set including the element words belonging to the adjacent categories of the same hierarchy on the thesaurus is generated.
  • a word set including a broader synonym can be generated.
  • the word set generation unit 14 includes a word set including element words belonging to the same category on the thesaurus as a word set including a plurality of element words similar to a plurality of reading candidates.
  • the word set generation unit 64 generates a word set including element words that belong to the upper hierarchy and / or lower hierarchy categories on the thesaurus. Also good. Thus, by generating a word set including broader synonyms for the upper and lower relations of the concept, the information amount of the reading determination information can be effectively increased and the accuracy thereof can be improved.
  • the other configuration is substantially the same as that of the reading determination device 10 according to the first embodiment. Therefore, the same parts are denoted by the same reference numerals, and detailed description thereof is omitted.
  • the word set generation unit 64 can control the degree of similarity depending on how many upper and / or lower hierarchies are targeted in the thesaurus.
  • generation part 64 produces
  • the element words having the similar relationship belong to the category of the upper hierarchy and / or the lower hierarchy of the number of hierarchies set in advance on the thesaurus. Element words are also included.
  • the word set generation unit 64 includes element words that belong to the same category as the category to which the reading candidate 1 “mole” belongs, and elements that belong to categories of upper and lower hierarchies having a preset number of hierarchies. Each word is extracted.
  • the word set generation unit 64 extracts element words that belong to the categories of the upper one layer and the lower one layer, and generates the word set 1.
  • the reading determination device 60 in addition to the element words belonging to the same category on the thesaurus, it belongs to the category of the upper hierarchy and / or the lower hierarchy of the number of hierarchies set in advance on the thesaurus.
  • a word set including element words By generating a word set including element words, a word set including a broader synonym can be generated.
  • FIG. 9 is a block diagram showing a schematic system configuration of a reading determination apparatus according to the seventh embodiment of the present invention.
  • the reading determination apparatus 70 according to the seventh embodiment of the present invention overlaps with an element word deletion unit 71 that detects and deletes overlapping element words.
  • a context vector deletion unit 72 that detects and deletes the context vector.
  • the element word deletion unit 71 detects overlapping element words among the word sets for the plurality of reading candidates 1 to M generated by the word set generation unit 14.
  • the above-described overlapping element word indicates, for example, a case where at least one set of element words overlaps. Then, the element word deletion unit 71 deletes one of the overlapping element words from the word set, and outputs the deleted word set to the context vector generation unit 16. On the other hand, the element word deletion unit 71 outputs a word set that does not include duplicated element words to the context vector generation unit 16 as it is.
  • overlapping element words have the same context vector. For this reason, if there are overlapping element words between word sets of a plurality of reading candidates, the degree of duplication of reading determination information generated based on the element words also increases. Therefore, by removing previously overlapping element words, the degree of separation in the reading determination information can be increased, and the accuracy of the reading determination information can be increased.
  • the context vector deletion unit 72 selects the same context vector among the word sets 1 to M from among the context vectors corresponding to the element words of the plurality of word sets 1 to M generated by the context vector generation unit 16. Then, one of the context vectors is deleted and output to the reading determination information generation unit 17.
  • the same context vector refers to a case where at least one set of context vectors is the same, for example.
  • the degree of duplication of the reading determination information generated based on the context vectors also increases. Therefore, by removing overlapping context vectors in advance, the degree of separation in the reading determination information can be increased, and the accuracy of the reading determination information can be increased.
  • the context vector deletion unit 72 is closer than a predetermined distance from among the context vectors for the element words of the plurality of word sets 1 to M generated by the context vector generation unit 16, and is similar to each other.
  • the detected context vector may be detected, and one of the context vectors may be deleted.
  • the context vector deletion unit 72 determines that the distance is closer than the predetermined distance.
  • the context vector deletion unit 72 deletes one of the detected sets of context vectors close to each other and outputs the result to the reading determination information generation unit 17.
  • the degree of duplication of the reading determination information generated based on the context vectors also increases. Therefore, by removing context vectors that are close in advance, the degree of separation in the reading determination information can be increased, and the accuracy of the reading determination information can be increased.
  • the context vector deletion unit 72 may multiply the context vector generated by the context vector generation unit 16 by a weighting factor for enhancing the feature.
  • the context vector detection unit 72 may detect the above-described adjacent context vector using the context vector multiplied by the weighting factor, and delete the detected context vector.
  • the importance of each element word V (i) is b (i).
  • the importance b (i) for example, a value of a scale tf-idf indicating whether or not a corresponding word is characteristic when it appears in the corpus information may be used.
  • the value of tf-idf is a value calculated from two indices, tf (word appearance frequency) and idf (reverse appearance frequency).
  • a weighting coefficient is set according to the importance b (i) of each word V (i).
  • the context vector D is multiplied by the weighting coefficient, whereby the difference regarding the word with high characteristic is emphasized and the difference regarding the word with low characteristic is reduced. Therefore, it is possible to perform similarity calculation more reflecting the characteristics of the corpus information.
  • FIG. 10 is a block diagram showing a schematic system configuration of a speech synthesizer according to the eighth embodiment of the present invention.
  • a speech synthesizer 80 according to the fifth embodiment includes a morpheme analyzer 81 that performs morphological analysis of an input sentence, a reading determination device 20 according to the second embodiment, and a speech generator 82 that generates synthesized speech. ing.
  • the morpheme analysis unit 81 performs morpheme analysis on the input sentence, divides the input sentence into morphemes, extracts independent words from the plurality of morphemes, and outputs them to the reading determination device 20.
  • the voice generation unit 82 Based on the information on how to read the input sentence output from the reading determination device 20, the voice generation unit 82 generates a waveform of a synthesized voice for the input sentence using, for example, a waveform connection type speech synthesis method.
  • the reading information used in the speech synthesis includes, for example, not only a phoneme string but also information on an accent position.
  • the noun “Tani” can be uttered separately in a head-high shape when used as a person's name, and in a flat shape when used as the opposite of a mountain.
  • FIG. 11 is a flowchart showing an example of the processing flow of the speech synthesizer according to the eighth embodiment of the present invention.
  • the morpheme analysis unit 81 performs morpheme analysis on the input sentence (step S302), divides the input sentence into a plurality of morphemes, and becomes independent. Extract words. Then, the morpheme analysis unit 81 outputs the extracted independent word as an input word together with the input sentence to the reading determination device 20.
  • the reading determination device 20 performs the above-described reading determination process based on the input sentence and the input word from the morphological analysis unit 81 (step S303), and determines the reading for all independent words (step S304).
  • step S305 information on how to read the input sentence is generated.
  • the reading determination device 20 outputs information on how to read the generated input sentence to the voice generation unit 82.
  • the voice generation unit 82 generates a synthesized voice waveform based on the information on how to read the input sentence from the reading judgment device 20 (step S306), and outputs the voice of the generated synthesized voice waveform (step S307).
  • the present invention is not limited to the above-described embodiment, and can be appropriately changed without departing from the spirit of the present invention.
  • the present invention has been described as a hardware configuration, but the present invention is not limited to this.
  • the present invention can also realize arbitrary processing by causing a CPU to execute a computer program.
  • Non-transitory computer readable media include various types of tangible storage media (tangible storage medium). Examples of non-transitory computer-readable media include magnetic recording media (eg flexible disks, magnetic tapes, hard disk drives), magneto-optical recording media (eg magneto-optical discs), CD-ROMs (Read Only Memory), CD-Rs, CD-R / W, semiconductor memory (for example, mask ROM, PROM (Programmable ROM), EPROM (Erasable ROM), flash ROM, RAM (random access memory)) are included.
  • the program may also be supplied to the computer by various types of temporary computer-readable media. Examples of transitory computer readable media include electrical signals, optical signals, and electromagnetic waves.
  • the temporary computer-readable medium can supply the program to the computer via a wired communication path such as an electric wire and an optical fiber, or a wireless communication path.
  • the present invention can be applied to, for example, a reading determination device that determines an appropriate reading for a word or a sentence.
  • Reading Judgment Device 11 Same Pronunciation Different Pronunciation Group Generation Unit 12 Reading DB 13 Thesaurus DB 14 word set generator 15 corpus DB 16 Context vector generation unit 17 Reading determination information generation unit 21 Reading determination unit 22 Output device 31 Example word acquisition unit 32 Reading DB 71 Element word deletion output unit 72 Context vector deletion unit 80 Speech synthesizer

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Machine Translation (AREA)
  • Document Processing Apparatus (AREA)

Abstract

A reading determination device (10) determines a reading of a word which has a plurality of reading candidates. The reading determination device (10) is also provided with a word set generation means (14) which generates word sets comprising a plurality of element words similar to the reading candidates, respectively; a corpus database (15) which stores corpus information containing a plurality of example sentences; a feature amount calculation means for calculating feature amounts for the respective plurality of element words of the word sets generated by the word set generation means (14) on the basis of the corpus information stored in the corpus database (15); and a reading determination information generation means (17) which generates reading determination information wherein the feature amounts for the plurality of element words of the word sets calculated by the feature amount calculation means are correlated to the respective reading candidates.

Description

[規則37.2に基づきISAが決定した発明の名称] 読み方判断装置、方法、プログラム、及びそのコンピュータ可読媒体、並びに音声合成装置[Name of Invention Determined by ISA Based on Rule 37.2] Reading Judgment Device, Method, Program, Computer-Readable Medium, and Speech Synthesizer
 本発明は、例えば、複数の読み方候補を有する、単語の読み方を判断するための読み方判断装置、音声合成装置、読み方判断方法、読み方判断プログラム、及びそのコンピュータ可読媒体に関し、特に、単語の読み方を容易かつ適切に判断できる読み方判断装置、音声合成装置、読み方判断方法、読み方判断プログラム、及びそのコンピュータ可読媒体に関するものである。 The present invention relates to, for example, a reading determination device, a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium for determining a reading of a word having a plurality of reading candidates. The present invention relates to a reading determination device, a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium thereof that can be easily and appropriately determined.
 与えられた文章の読み方を判断する方法として、例えば、予め辞書に文字や単語などの「読み方」(読み仮名やアクセント情報等)を定義しておき、文章中の各単語の文法的接続関係をチェックしつつ、辞書に定義された読み方に基づいて、文章全体の読み方を判断する方法が広く知られている(例えば、非特許文献1及び2)。さらに、連濁や無声化などの音韻規則を考慮することで、文章としてより適切な読み方を付与する方法も知られている。 As a method of determining how to read a given sentence, for example, “reading” (reading kana, accent information, etc.) such as letters and words is defined in advance in the dictionary, and the grammatical connection relation of each word in the sentence is determined. A method for determining how to read the entire sentence based on the reading defined in the dictionary while checking is widely known (for example, Non-Patent Documents 1 and 2). Furthermore, a method of giving a more appropriate reading as a sentence by considering phonological rules such as rendaku and devoicing is also known.
 上記非特許文献2に示す形態素解析によれば、その規則として品詞関係が使用されている。しかしながら、例えば、日本語においては、同一表記および同一品詞でありながら、利用分野や意味によって複数の読み方を有する単語(同表記異発音語組)が多数存在する。具体的には、「市場」(名詞)は、「いちば」及び「しじょう」の二種類の読み方を有しており、「黒子」(名詞)は、「ほくろ」及び「くろご」の二種類の読み方を有している。また、「磯」(名詞)の読み方は、「いそ」であるが、一般名詞で使用される場合、平板型のアクセントになり、人名で使用される場合、頭高型のアクセントになる。したがって、例えば、音声合成等を行う場合には、こうした違いも重要となる。文章に正しい読み方を付与するためには、これら複数の読み方(アクセント等を含む)を、適切に選択することが望ましい。 According to the morphological analysis shown in Non-Patent Document 2 above, the part-of-speech relationship is used as the rule. However, for example, in Japanese, there are many words (same notation different pronunciation word sets) having the same notation and the same part of speech but having a plurality of readings depending on the field of use and meaning. Specifically, “market” (noun) has two types of readings, “Ichiba” and “shijo”, and “kuroko” (noun) means “mole” and “black”. There are two types of reading. In addition, “磯” (noun) is read as “iso”, but when used in general nouns, it becomes a flat accent, and when used in personal names, it becomes a head-high accent. Therefore, for example, such differences are also important when performing speech synthesis or the like. In order to give a correct reading to a sentence, it is desirable to appropriately select these multiple readings (including accents).
 上記選択を行うために、例えば、「市場」の前に「魚」が連接する場合は、その読み方を「いちば」とし、同様に前に「株式」が連接する場合は、その読み方を「しじょう」として、読み分ける方法が考えられる。すなわち、「魚」及び「市場」の連鎖が存在する文章の読み方を統計的に調べて、「さかないちば」「さかなしじょう」「うおいちば」「うおしじょう」などの複数の読み方の中から、一番多く読まれた読み方「うおいちば」を抽出し、その読み方を正解とする方法が考えられる。この方法は、例えば、単語bigramの学習頻度に応じて読み方を決定するとの考え方に基づくものであり、適切なbigramセットを定義することで読み方の精度を向上することが可能となる。 In order to make the above selection, for example, when “fish” is connected before “market”, the reading is “first”, and when “stock” is connected before, the reading is “ "Shijo" can be considered as a method of reading differently. In other words, we will statistically investigate how to read texts that have a chain of “fish” and “market”, and use multiple readings such as “Saizen Chiba”, “Sakanashi Jyo”, “Uoichiba”, and “Uoshijo”. A method of extracting the most frequently read “Uoichiba” from among them and making the reading correct is conceivable. This method is based on the idea of determining how to read according to the learning frequency of the word bigram, for example, and it is possible to improve the accuracy of reading by defining an appropriate bigram set.
 しかしながら、上述したような同表記異発音語組は、多数組存在しており、また、これらの同表記異発音語組に連接する可能性のある単語数も非常に多い。このため、適切なbigramセットを作成するためには、正解の読み方を含む学習コーパスが多量に必要となるが、そのような多量の学習コーパスを得ることはあまり実用的とは言えない。また、単語bigramの学習を行う代わりに、同表記異発音語組ごとに連接する特徴のある付属語などを、予め規則として記述し、解析時に利用する方法も考えられる。しかしながら、全ての同表記異発音語組に対して規則を記述することは、実質的に困難と言える。 However, there are a large number of the same notation different pronunciation groups as described above, and the number of words that can be connected to these same notation different pronunciation groups is very large. For this reason, in order to create an appropriate bigram set, a large amount of learning corpora including how to read correct answers is required, but it is not very practical to obtain such a large amount of learning corpora. In addition, instead of learning the word bigram, there may be a method in which, for example, an attached word having a characteristic connected to each notation differently pronounced word set is described as a rule in advance and used at the time of analysis. However, it can be said that it is substantially difficult to describe the rules for all the homophones having the same notation.
 本発明は、このような問題点を解決するためになされたものであり、単語の読み方を容易かつ適切に判断できる読み方判断装置、音声合成装置、読み方判断方法、読み方判断プログラム、及びコンピュータ可読媒体を提供することを主たる目的とする。 The present invention has been made to solve such problems, and is a reading determination device, a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium that can easily and appropriately determine how to read a word. The main purpose is to provide
 上記目的を達成するための本発明の一態様は、複数の読み方候補を有する、単語の読み方を判断するための読み方判断装置であって、前記読み方候補に類似する複数の要素単語からなる単語集合を、夫々生成する単語集合生成手段と、複数の例文を含むコーパス情報を記憶するコーパスデータベースと、前記コーパスデータベースに記憶された前記コーパス情報に基づいて、前記単語集合生成手段により生成された前記単語集合の複数の要素単語に対する特徴量を夫々算出する特徴量算出手段と、前記特徴量算出手段により算出された前記単語集合の複数の要素単語に対する特徴量と、前記読み方候補とを夫々関連付けた読み方判断情報を生成する読み方判断情報生成手段と、を備える、ことを特徴とする読み方判断装置である。 One aspect of the present invention for achieving the above object is a reading determination apparatus for determining how to read a word, having a plurality of reading candidates, and a word set comprising a plurality of element words similar to the reading candidate Respectively, a corpus database for storing corpus information including a plurality of example sentences, and the word generated by the word set generating means based on the corpus information stored in the corpus database. A feature amount calculating means for calculating feature amounts for a plurality of element words in the set, a reading method in which the feature amounts for the plurality of element words in the word set calculated by the feature amount calculating means are associated with the reading candidates, respectively. It is a reading judgment apparatus characterized by comprising reading judgment information generating means for generating judgment information.
 また、上記目的を達成するための本発明の一態様は、複数の読み方候補を有する、単語の読み方を判断するための読み方判断方法であって、前記読み方候補に類似する複数の要素単語からなる単語集合を、夫々生成し、複数の例文を含むコーパス情報に基づいて、前記生成した単語集合の複数の要素単語に対する特徴量を夫々算出し、前記算出した単語集合の複数の要素単語に対する特徴量と、前記読み方候補とを夫々関連付けた読み方判断情報を生成する、ことを特徴とする読み方判断方法である。 Another aspect of the present invention for achieving the above object is a reading determination method for determining how to read a word having a plurality of reading candidates, and includes a plurality of element words similar to the reading candidate. Each of the word sets is generated, and based on the corpus information including a plurality of example sentences, the feature amounts for the plurality of element words of the generated word set are calculated, respectively, and the feature amounts for the plurality of element words of the calculated word set And reading method determination information in which the reading candidates are associated with each other.
 さらに、上記目的を達成するための本発明の一態様は、複数の読み方候補を有する、単語の読み方を判断するための読み方判断プログラムを格納する非一時的なコンピュータ可読媒体であって、前記読み方候補に類似する複数の要素単語からなる単語集合を、夫々生成する処理と、複数の例文を含むコーパス情報に基づいて、前記生成された前記単語集合の複数の要素単語に対する特徴量を夫々算出する処理と、前記算出した前記単語集合の複数の要素単語に対する特徴量と、前記読み方候補とを夫々関連付けた読み方判断情報を生成する処理と、をコンピュータに実行させる読み方判断プログラムを格納する非一時的なコンピュータ可読媒体である。 Furthermore, one aspect of the present invention for achieving the above object is a non-transitory computer-readable medium that stores a reading determination program for determining how to read a word, having a plurality of reading candidates. Based on the processing for generating each word set composed of a plurality of element words similar to the candidate and corpus information including a plurality of example sentences, the feature amounts for the plurality of element words of the generated word set are calculated. A non-temporary storing a reading judgment program that causes a computer to execute processing, and processing to generate reading judgment information in which the calculated feature quantities for a plurality of element words of the word set and the reading candidates are associated with each other Computer readable medium.
 本発明によれば、単語の読み方を容易かつ適切に判断できる読み方判断装置、音声合成装置、読み方判断方法、読み方判断プログラム、及びそのコンピュータ可読媒体を提供することができる。 According to the present invention, it is possible to provide a reading determination device, a speech synthesizer, a reading determination method, a reading determination program, and a computer readable medium thereof that can easily and appropriately determine how to read a word.
本発明の実施形態に係る読み方判断装置の機能ブロック図である。It is a functional block diagram of the reading judgment apparatus concerning the embodiment of the present invention. 本発明の第1実施形態に係る読み方判断装置の概略的なシステム構成の一例を示すブロック図である。It is a block diagram which shows an example of the schematic system configuration | structure of the reading judgment apparatus which concerns on 1st Embodiment of this invention. 読み方候補DBに記憶された、単語、品詞、読み方、及び、シソーラス上のカテゴリーを、一組にして関連付けた複数のエントリの一例を示す図である。It is a figure which shows an example of the some entry which linked | related the word, the part of speech, the reading method, and the category on a thesaurus memorize | stored in reading candidate DB. シソーラスDBに記憶されたシソーラス辞書情報の一例を示す図である。It is a figure which shows an example of thesaurus dictionary information memorize | stored in the thesaurus DB. 本発明の第1実施形態に係る読み方判断装置の処理フローの一例を示すフローチャートである。It is a flowchart which shows an example of the processing flow of the reading judgment apparatus which concerns on 1st Embodiment of this invention. 本発明の第2実施形態に係る読み方判断装置の概略的なシステム構成を示すブロック図である。It is a block diagram which shows the schematic system configuration | structure of the reading judgment apparatus which concerns on 2nd Embodiment of this invention. 本発明の第2実施形態に係る読み方判断装置の処理フローの一例を示すフローチャートである。It is a flowchart which shows an example of the processing flow of the reading judgment apparatus which concerns on 2nd Embodiment of this invention. 本発明の第3実施形態に係る読み方判断装置の概略的なシステム構成を示すブロック図である。It is a block diagram which shows the schematic system configuration | structure of the reading judgment apparatus which concerns on 3rd Embodiment of this invention. 本発明の第7実施形態に係る読み方判断装置の概略的なシステム構成を示すブロック図である。It is a block diagram which shows the schematic system configuration | structure of the reading judgment apparatus which concerns on 7th Embodiment of this invention. 本発明の第8実施形態に係る音声合成装置の概略的なシステム構成を示すブロック図である。It is a block diagram which shows the schematic system configuration | structure of the speech synthesizer concerning 8th Embodiment of this invention. 本発明の第8実施形態に係る音声合成装置の処理フローの一例を示すフローチャートである。It is a flowchart which shows an example of the processing flow of the speech synthesizer concerning 8th Embodiment of this invention.
 以下、図面を参照して本発明の実施の形態について説明する。図1は、本発明の実施形態に係る読み方判断装置の機能ブロック図である。本実施形態に係る読み方判断装置10は、複数の読み方候補を有する、単語の読み方を判断するための装置である。また、読み方判断装置10は、単語集合を生成する単語集合生成手段14と、コーパス情報を記憶するコーパスデータベース15と、要素単語の特徴量を算出する特徴量算出手段16と、読み方判断情報を生成する読み方判断情報生成手段17と、を備えている。 Hereinafter, embodiments of the present invention will be described with reference to the drawings. FIG. 1 is a functional block diagram of a reading determination apparatus according to an embodiment of the present invention. The reading determination apparatus 10 according to the present embodiment is an apparatus for determining how to read a word having a plurality of reading candidates. The reading determination device 10 generates a word set generation unit 14 that generates a word set, a corpus database 15 that stores corpus information, a feature amount calculation unit 16 that calculates feature amounts of element words, and generates reading determination information. Reading information determination means 17 for reading.
 単語集合生成手段14は、読み方候補に類似する複数の要素単語からなる単語集合を、夫々生成する。また、コーパスデータベース15は、複数の例文を含むコーパス情報を記憶する。さらに、特徴量算出手段16は、コーパスデータベース15に記憶されたコーパス情報に基づいて、単語集合生成手段14により生成された単語集合の複数の要素単語に対する特徴量を夫々算出する。読み方判断情報生成手段17は、特徴量算出手段16により算出された単語集合の複数の要素単語に対する特徴量と、読み方候補とを夫々関連付けた読み方判断情報を生成する。このように、読み方候補に類似する単語集合と、コーパス情報とを用いることで、読み方判断情報の情報量を効果的に増加させ、その精度を向上させることができる。したがって、より適切かつ高精度な読み方判断情報を取得でき、さらに、この読み方判断情報を用いて、単語の読み方を容易かつ適切に判断できる。
(第1実施形態)
The word set generation unit 14 generates a word set composed of a plurality of element words similar to reading candidates. The corpus database 15 stores corpus information including a plurality of example sentences. Further, the feature amount calculation means 16 calculates the feature amounts for a plurality of element words of the word set generated by the word set generation means 14 based on the corpus information stored in the corpus database 15. The reading determination information generation unit 17 generates reading determination information in which the feature amounts for a plurality of element words in the word set calculated by the feature amount calculation unit 16 are associated with the reading candidates. Thus, by using the word set similar to the reading candidate and the corpus information, the information amount of the reading determination information can be effectively increased and the accuracy thereof can be improved. Therefore, more appropriate and highly accurate reading determination information can be acquired, and further, how to read a word can be determined easily and appropriately using this reading determination information.
(First embodiment)
 図2は、本発明の第1実施形態に係る読み方判断装置の概略的なシステム構成の一例を示すブロック図である。本実施形態に係る読み方判断装置10は、同表記異発音語組生成部11と、読み方候補DB(データベース)12と、シソーラスDB(データベース)13と、単語集合生成部14と、コーパスDB(データベース)15と、文脈ベクトル生成部16と、読み方判断情報生成部17と、を備えている。 FIG. 2 is a block diagram showing an example of a schematic system configuration of the reading determination device according to the first embodiment of the present invention. The reading determination apparatus 10 according to the present embodiment includes a same pronunciation different pronunciation word set generation unit 11, a reading candidate DB (database) 12, a thesaurus DB (database) 13, a word set generation unit 14, and a corpus DB (database). ) 15, a context vector generation unit 16, and a reading determination information generation unit 17.
 同表記異発音語組生成部11は、ユーザにより入力された入力単語に基づいて、読み方候補DB12から、その入力単語に対する複数の読み方候補(読み方候補1、読み方候補2、・・・、読み方候補M)と、夫々に対応する単語意味(語意1、語意2、・・・、語意M)と、からなる同表記異発音語組を取得する。 Based on the input word input by the user, the same notation different pronunciation word set generation unit 11 reads from the reading candidate DB 12 a plurality of reading candidates (reading candidate 1, reading candidate 2,..., Reading candidates for the input word. M) and word meanings corresponding to each (meaning 1, meaning 2,..., Meaning M) are acquired.
 ここで、例えば、与えられたテキストを発声する際に用いる音韻列を「読み方」と称するものとする。また、以下、音韻列として、日本語の音節の並びを用いて説明するが、言語に依存せずに、国際音声記号(IPA)などの任意の音声記号列を用いることもできる。さらに、「読み方」には、音韻列に付随して、該当音韻列をどのように発声するかを示すアクセント情報(アクセント位置、区切り等)や、母音無声化などの読み方補助情報を含んでいても良い。 Here, for example, a phoneme string used when uttering a given text is referred to as “reading”. In the following description, a Japanese syllable sequence is used as a phoneme sequence, but an arbitrary phonetic symbol sequence such as an international phonetic symbol (IPA) can be used without depending on a language. Furthermore, “how to read” includes accent information (accent position, separation, etc.) indicating how the corresponding phoneme string is uttered, and supplementary reading information such as vowel devoicing, accompanying the phoneme string. Also good.
 同表記異発音語組生成部11は、取得した入力単語に対する複数の読み方候補を、単語集合生成部14に対して出力する。 The same notation different pronunciation word set generation unit 11 outputs a plurality of reading candidates for the acquired input word to the word set generation unit 14.
 読み方候補DB12は、例えば、図3に示すような、単語、品詞、読み方、及び、シソーラス上のカテゴリーを、一組にして関連付けた複数のエントリを記憶している。例えば、表記異発音語組生成部11は、入力単語「黒子」に基づいて、読み方候補DB12に記憶された複数のエントリ中から該当するエントリを検索し、取得する。この場合、表記異発音語組生成部11は、読み方候補DB12から、例えば、2つ(読み方の個数M=2)のエントリ、読み方候補1=「ほくろ」(カテゴリーは、身体の染み)、及び、読み方候補2=「くろご」(カテゴリーは、歌舞伎の補助員)を取得する。 The reading candidate DB 12 stores, for example, a plurality of entries in which words, parts of speech, readings, and categories on the thesaurus are associated as a set as shown in FIG. For example, the notation different pronunciation word set generation unit 11 searches and acquires a corresponding entry from a plurality of entries stored in the reading candidate DB 12 based on the input word “Kuroko”. In this case, the notation different pronunciation word set generation unit 11 reads from the reading candidate DB 12, for example, two entries (number of readings M = 2), reading candidate 1 = “mole” (category is body stain), and , Reading candidate 2 = “Kurogo” (category is Kabuki assistant).
 シソーラスDB13は、例えば、図4に示すような、複数の単語の上位/下位関係、部分/全体関係、同義関係、類義関係などによって、各単語を分類し、体系化したシソーラス辞書情報を記憶している。 The thesaurus DB 13 stores systematic thesaurus dictionary information by classifying each word according to the upper / lower relationship, partial / whole relationship, synonym relationship, synonym relationship, etc. of a plurality of words as shown in FIG. is doing.
 単語集合生成部(単語集合生成手段)14は、表記異発音語組生成部11からの、複数の読み方候補に類似する複数の要素単語からなる単語集合を、シソーラスDB13のシソーラス辞書情報に基づいて、夫々生成する。ここで、上記読み方候補に類似する複数の要素単語には、例えば、シソーラス上で読み方候補のカテゴリーと同一カテゴリーに属する単語が含まれており、広い意味で同義の単語が含まれる。 The word set generation unit (word set generation means) 14 selects a word set composed of a plurality of element words similar to a plurality of reading candidates from the notation different pronunciation word set generation unit 11 based on thesaurus dictionary information of the thesaurus DB 13. , Respectively. Here, the plurality of element words similar to the reading candidate include, for example, words belonging to the same category as the reading candidate category on the thesaurus, and include words having the same meaning in a broad sense.
 単語集合生成部14は、シソーラスDB13のシソーラス辞書情報に基づいて、例えば、図4に示すシソーラス上において、読み方候補1=「ほくろ」と同一カテゴリーに属する単語を抽出し、読み方候補1=「ほくろ」に対する単語集合1を生成する。ここで、単語集合1={要素単語1-1:染み、要素単語1-2:しみ、要素単語1-3:黒子、要素単語1-4:ほくろ、要素単語1-5:ホクロ、要素単語1-6:lentigo}となり、カテゴリー={身体の染み}となり、要素単語の数N1=6となる。 The word set generation unit 14 extracts words belonging to the same category as reading candidate 1 = “mole” on the thesaurus shown in FIG. 4 based on thesaurus dictionary information in the thesaurus DB 13, and reading candidate 1 = “mole” 1 is generated. Here, word set 1 = {element word 1-1: stain, element word 1-2: blot, element word 1-3: black, element word 1-4: mole, element word 1-5: mole, element word 1-6: lentigo}, category = {body stain}, and the number of element words N1 = 6.
 同様に、単語集合生成部14は、読み方候補2=「くろご」に対する、単語集合2を生成する。ここで、単語集合2={要素単語2-1:黒子、要素単語2-2:くろこ、要素単語2-3:くろご、要素単語2-4:後見、要素単語2-5:黒具}となり、カテゴリー={歌舞伎の補助員}となり、要素単語の数N2=5となる。 Similarly, the word set generation unit 14 generates the word set 2 for the reading candidate 2 = “Kurogo”. Here, word set 2 = {element word 2-1: Kuroko, element word 2-2: black, element word 2-3: black, element word 2-4: guardianship, element word 2-5: black tool }, Category = {kabuki assistant}, and the number of element words N2 = 5.
 なお、単語集合生成部14は、シソーラスDB13のシソーラス辞書情報を用いて、読み方候補の単語集合を生成しているが、これに限らず、例えば、同義語辞書情報などを用いてもよく、任意の辞書情報を用いることができる。また、単語集合生成部14は、非特許文献2に開示されているように、格フレームを計算することで、各単語が上記同一カテゴリーに属するかどうかを判定することもできる。さらに、読み方候補DB12には、単語意味(語意1、・・・、語意N)の代わりに、対応各フレーム(各フレーム1、・・・、各フレームN)を持つこともできる。単語集合生成部14は、生成した単語集合を、文脈ベクトル生成部16に対して出力する。 The word set generation unit 14 uses the thesaurus dictionary information in the thesaurus DB 13 to generate a word set of reading candidates. However, the word set generation unit 14 is not limited to this, and may use synonym dictionary information, for example. The dictionary information can be used. Further, as disclosed in Non-Patent Document 2, the word set generation unit 14 can determine whether each word belongs to the same category by calculating a case frame. Further, the reading candidate DB 12 can have corresponding frames (each frame 1,..., Each frame N) instead of the word meaning (meaning 1,..., Word N). The word set generation unit 14 outputs the generated word set to the context vector generation unit 16.
 コーパスDB15は、複数の例文(テキストデータ)を含むコーパス情報を記憶している。ここで、コーパス情報の例文には、読み方が付与されてなくてもよい。 The corpus DB 15 stores corpus information including a plurality of example sentences (text data). Here, the example of the corpus information may not be given a reading.
 文脈ベクトル生成部16は、コーパスDB15に記憶されたコーパス情報に基づいて、単語集合生成部14により生成された単語集合の複数の要素単語に対して文脈ベクトルを夫々算出する。文脈ベクトル生成部16は、まず、単語集合生成部14により生成された単語集合の要素単語の夫々に対して、コーパスDB15のコーパス情報の中から、その要素単語が使用されている例文を抽出する。そして、文脈ベクトル生成部16は、抽出した各例文を用いて、その要素単語の文脈ベクトルを夫々算出する。ここで、文脈ベクトルは、例えば、該当単語の周辺の単語の状況等を表す特徴量として用いられている。 The context vector generation unit 16 calculates context vectors for a plurality of element words of the word set generated by the word set generation unit 14 based on the corpus information stored in the corpus DB 15. The context vector generation unit 16 first extracts an example sentence using the element word from the corpus information of the corpus DB 15 for each element word of the word set generated by the word set generation unit 14. . Then, the context vector generation unit 16 calculates the context vector of the element word using each extracted example sentence. Here, the context vector is used as, for example, a feature amount indicating the situation of words around the corresponding word.
 また、非特許文献2(岩波講座 ソフトウェア科学15 自然言語処理、pp.421-424)において説明されるベクトル空間法において、文書全体に索引語T(i)が出現する場合は1を、出現しない場合は0を、係数として持つ文書ベクトルが知られている。一方、本実施形態における文脈ベクトルも上記文書ベクトルと同様に、コーパス情報の例文に含まれる全ての自立語を軸とするベクトル空間において、該当要素単語の周辺の自立語に着目して、係数を選択してもよい。 In addition, in the vector space method described in Non-Patent Document 2 (Iwanami Lecture, Software Science 15, Natural Language Processing, pp. 421-424), 1 does not appear when the index word T (i) appears in the entire document. In this case, a document vector having 0 as a coefficient is known. On the other hand, in the same manner as the document vector, the context vector in this embodiment focuses on the independent words around the corresponding element word in the vector space around all the independent words included in the example sentence of the corpus information, and calculates the coefficient. You may choose.
 例えば、文脈ベクトル生成部16が、コーパスDB15のコーパス情報から、その要素単語T7が使用されている例文S{T21 T32 T52 T7 T42 T64 T73 T12}を抽出したとする。そして、文脈ベクトル生成部16は、要素単語T7に対して、前後2単語(wl=2)づつ抽出した文脈ベクトルを、下記式で算出することができる。
  D(T7;S)=Σ(i=1,t)a(i)*V(i)
=[0…0 V(32) 0…0 V(42) 0….0 V(52)0….0 V(64) 0….0]
For example, it is assumed that the context vector generation unit 16 extracts an example sentence S {T21 T32 T52 T7 T42 T64 T73 T12} in which the element word T7 is used from the corpus information of the corpus DB 15. Then, the context vector generation unit 16 can calculate a context vector extracted by two words before and after (wl = 2) for the element word T7 by the following formula.
D (T7; S) = Σ (i = 1, t) a (i) * V (i)
= [0 ... 0 V (32) 0 ... 0 V (42) 0 ... 0.0 V (52) 0 .... 0.0 V (64) 0 .... 0.0]
 なお、本実施形態において、該当要素単語の周辺の単語の状況を表現する特徴量として、文脈ベクトルに着目して説明するが、これに限らず、例えば、該当要素単語の周辺の単語のコーパス情報における出現確率の和を用いるなど、ベクトル表現以外の特徴量を用いてもよい。以降の説明では、これらベクトル表現以外の特徴量も含めて、文脈ベクトルとして説明するものとする。 In the present embodiment, the description will be given focusing on the context vector as the feature amount expressing the situation of the words around the corresponding element word, but the present invention is not limited to this. For example, the corpus information of the words around the corresponding element word is used. A feature quantity other than the vector expression may be used, such as using the sum of the appearance probabilities. In the following description, it will be described as a context vector including feature quantities other than these vector expressions.
 また、文脈ベクトルを算出する際、各単語が同一層で複数の意味を有するような曖昧性を低減するために、シソーラス中に複数存在する同一単語のうち、どの単語に該当するかを決め、カテゴリーを割り振ることも有用である。この方法は、例えば、非特許文献2(岩波講座 ソフトウェア科学15 自然言語処理、pp.235-240)に示す格フレームを用いて同定することができる。 Further, when calculating the context vector, in order to reduce ambiguity such that each word has a plurality of meanings in the same layer, determine which word corresponds to the same word existing in the thesaurus, It is also useful to assign categories. This method can be identified using, for example, a case frame shown in Non-Patent Document 2 (Iwanami Lecture, Software Science 15, Natural Language Processing, pp. 235-240).
 さらに、文脈ベクトルを算出する際、文法的知識を用いてより特徴を表す単語を抽出してもよく、各要素単語の同義語圧縮や次元圧縮などの一般的な手法により文脈ベクトルを圧縮し、文脈ベクトル空間の利用効率を向上させても良い。なお、これらの方法の使用及び不使用を問わず、文脈ベクトルの格納は、単語bigramなどの格納と比較して、所要の記憶容量が小さくて済むため、追加コストが少なく実現可能と言える。 Furthermore, when calculating the context vector, it is possible to extract words that represent more features using grammatical knowledge, compress the context vector by a general method such as synonym compression or dimension compression of each element word, The use efficiency of the context vector space may be improved. Regardless of the use or non-use of these methods, it can be said that the storage of the context vector can be realized at a low additional cost because it requires a smaller storage capacity than the storage of the word bigram.
 さらにまた、該当要素単語が使用されている例文がコーパスDB15のコーパス情報中に複数存在する場合には、文脈ベクトル生成部16は、その全ての例文、あるいは予め定めた数の例文に対して、同様の処理を行い、複数の文脈ベクトルを算出する。一方で、該当要素単語が使われている例文がコーパス情報中に存在しない場合には、文脈ベクトル生成部16は、その要素単語の文脈ベクトルを算出しないこととなる。文脈ベクトル生成部16は、上述のようにして算出した、単語集合の複数の要素単語にする文脈ベクトルを、読み方判断情報生成部17に対して出力する。 Furthermore, when there are a plurality of example sentences in which the corresponding element word is used in the corpus information of the corpus DB 15, the context vector generation unit 16 selects all of the example sentences or a predetermined number of example sentences. A similar process is performed to calculate a plurality of context vectors. On the other hand, when there is no example sentence using the corresponding element word in the corpus information, the context vector generation unit 16 does not calculate the context vector of the element word. The context vector generation unit 16 outputs to the reading determination information generation unit 17 the context vector calculated as described above and used as a plurality of element words of the word set.
 読み方判断情報生成部(読み方判断情報生成手段)17は、文脈ベクトル生成部16により算出された単語集合の複数の要素単語に対する文脈ベクトルと、読み方候補とを夫々関連付けた読み方判断情報を生成する。読み方判断情報生成部17は、例えば、読み方候補毎に、対応する複数の文脈ベクトルを文脈ベクトル空間上で相加平均した、代表の平均文脈ベクトル(代表文脈ベクトル)を算出する。そして、読み方判断情報生成部17は、各読み方候補と、算出した代表の平均文脈ベクトル(相加平均値)と、を夫々関連付けた読み方判断情報を生成する。 The reading determination information generation unit (reading determination information generation means) 17 generates reading determination information in which context vectors for a plurality of element words of the word set calculated by the context vector generation unit 16 are associated with reading candidates. For example, for each reading candidate, the reading determination information generation unit 17 calculates a representative average context vector (representative context vector) obtained by arithmetically averaging a plurality of corresponding context vectors in the context vector space. Then, the reading determination information generation unit 17 generates reading determination information in which each reading candidate is associated with the calculated representative average context vector (arithmetic mean value).
 なお、読み方判断装置10は、例えば、制御処理、演算処理等と行うCPU(Central Processing Unit)、CPUによって実行される制御プログラム、演算プログラム等が記憶されたROM(Read Only Memory)、処理データ等を一時的に記憶するRAM(Random Access Memory)等からなるマイクロコンピュータを中心にしてハードウェア構成されている。また、同表記異発音語組生成部11、単語集合生成部14、文脈ベクトル生成部16、読み方判断情報生成部17、及び、後述の読み方判断部21は、例えば、上記ROMに記憶され、上記CPUによって実行されるプログラムによって実現することができる。 The reading determination device 10 includes, for example, a CPU (Central Processing Unit) that performs control processing, arithmetic processing, and the like, a ROM (Read (OnlymMemory) that stores a control program executed by the CPU, an arithmetic program, processing data, and the like Is composed of a microcomputer mainly including a RAM (Random Access Memory) or the like that temporarily stores. Also, the same notation different pronunciation word set generation unit 11, word set generation unit 14, context vector generation unit 16, reading determination information generation unit 17, and later-described reading determination unit 21 are stored in the ROM, for example, It can be realized by a program executed by the CPU.
 図5は、本発明の第1実施形態に係る読み方判断装置の処理フローの一例を示すフローチャートである。まず、同表記異発音語組生成部11は、ユーザにより入力された入力単語に基づいて、読み方候補DB12から、その入力単語に対する複数の読み方候補を取得し(ステップS101)、取得した複数の読み方候補を単語集合生成部14に対して出力する。 FIG. 5 is a flowchart showing an example of a processing flow of the reading determination apparatus according to the first embodiment of the present invention. First, the same notation different pronunciation word set generation unit 11 acquires a plurality of reading candidates for the input word from the reading candidate DB 12 based on the input word input by the user (step S101), and acquires the plurality of acquired readings. Candidates are output to the word set generation unit 14.
 次に、単語集合生成部14は、表記異発音語組生成部11からの、複数の読み方候補に類似する複数の要素単語からなる単語集合を、シソーラスDB13のシソーラス辞書情報に基づいて、夫々生成し(ステップS102)、生成した単語集合を文脈ベクトル生成部16に対して出力する。 Next, the word set generation unit 14 generates a word set composed of a plurality of element words similar to a plurality of reading candidates from the notation different pronunciation word set generation unit 11 based on thesaurus dictionary information in the thesaurus DB 13. In step S102, the generated word set is output to the context vector generation unit 16.
 その後、文脈ベクトル生成部16は、コーパスDB15に記憶されたコーパス情報に基づいて、単語集合生成部14により生成された単語集合の複数の要素単語に対して文脈ベクトルを夫々算出し(ステップS103)、算出した文脈ベクトルを、読み方判断情報生成部17に出力する。 Thereafter, the context vector generation unit 16 calculates context vectors for a plurality of element words of the word set generated by the word set generation unit 14 based on the corpus information stored in the corpus DB 15 (step S103). The calculated context vector is output to the reading determination information generation unit 17.
 さらに、読み方判断情報生成部17は、読み方候補毎に、対応する複数の文脈ベクトルを文脈ベクトル空間上で相加平均した、代表の平均文脈ベクトルを算出する(ステップS104)。そして、読み方判断情報生成部17は、各読み方候補と、算出した平均文脈ベクトルと、を夫々関連付けた読み方判断情報を生成する(ステップS105)。 Further, the reading determination information generation unit 17 calculates a representative average context vector obtained by arithmetically averaging a plurality of corresponding context vectors in the context vector space for each reading candidate (step S104). Then, the reading determination information generation unit 17 generates reading determination information in which each reading candidate is associated with the calculated average context vector (step S105).
 以上、第1実施形態に係る読み方判断装置10によれば、複数の読み方候補に類似する要素単語からなる単語集合を夫々生成し、コーパス情報を用いて、各読み方候補に対する平均文脈ベクトルを算出し、各読み方候補と平均文脈ベクトルとを関連付けた読み方判断情報を生成する。このように、読み方候補に類似する単語集合と、コーパス情報とを用いることで、読み方判断情報の情報量を効果的に増加させ、その精度を向上させることができる。したがって、より適切かつ高精度な読み方判断情報を取得でき、さらに、この読み方判断情報を用いて、単語の読み方を容易かつ適切に判断できる。 As described above, according to the reading determination device 10 according to the first embodiment, a word set composed of element words similar to a plurality of reading candidates is generated, and an average context vector for each reading candidate is calculated using corpus information. Then, reading determination information in which each reading candidate is associated with the average context vector is generated. Thus, by using the word set similar to the reading candidate and the corpus information, the information amount of the reading determination information can be effectively increased and the accuracy thereof can be improved. Therefore, more appropriate and highly accurate reading determination information can be acquired, and further, how to read a word can be determined easily and appropriately using this reading determination information.
 なお、例えば、入力単語の正解の読み方を含むコーパス情報(学習コーパス)を大量に用意できたとしても、そのコーパス情報から得られる任意の同表記異発音語組に関する情報量よりも、本実施形態のように、単語集合とコーパス情報とを用いて生成した読み方判断情報の情報量の方が、より多く、高精度であることは言うまでもない。しかも、本実施形態によれば、正解の読み方が付与された学習コーパス情報を多量に用意したり、多数の同表記異発音語組に対して規則を記述する必要がなく、読み方判断情報の情報量を効率的に増加させ、読み方判断の精度を向上させることができる点でより優れている。また、本実施形態によれば、単語の類似性などの情報に基づいて、読み方を推定するため、それらの推定精度が向上することにより、読み方の精度向上も期待できる。
(第2実施形態)
Note that, for example, even if a large amount of corpus information (learning corpus) including how to read the correct answer of the input word can be prepared, the present embodiment is more effective than the amount of information related to an arbitrary synonym pronoun word set obtained from the corpus information. Needless to say, the information amount of the reading determination information generated using the word set and the corpus information is larger and more accurate. Moreover, according to the present embodiment, it is not necessary to prepare a large amount of learning corpus information to which correct reading is given, or to write rules for a large number of homophones with the same notation, and information on reading determination information This is superior in that the amount can be increased efficiently and the accuracy of reading can be improved. Further, according to the present embodiment, since reading is estimated based on information such as word similarity, improvement in reading accuracy can be expected by improving the estimation accuracy.
(Second Embodiment)
 図6は、本発明の第2実施形態に係る読み方判断装置の概略的なシステム構成を示すブロック図である。第2実施形態に係る読み方判断装置20は、第1実施形態に係る読み方判断装置10の構成に加えて、読み方判断情報生成部17により生成された読み方判断情報に基づいて、入力単語の読み方を判断する読み方判断部21と、その読み方の判断を出力する出力装置22と、を、更に備えている。なお、出力装置22として、例えば、ディスプレイ装置、プリンタ装置、音声出力装置などを用いることができる。また、本実施形態に係る読み方判断装置20は、例えば、オンラインで入力単語の読み方を判断できる。 FIG. 6 is a block diagram showing a schematic system configuration of the reading determination apparatus according to the second embodiment of the present invention. The reading determination device 20 according to the second embodiment reads the input word based on the reading determination information generated by the reading determination information generation unit 17 in addition to the configuration of the reading determination device 10 according to the first embodiment. It further includes a reading determination unit 21 for determining, and an output device 22 for outputting the determination of the reading. As the output device 22, for example, a display device, a printer device, an audio output device, or the like can be used. The reading determination device 20 according to the present embodiment can determine how to read an input word online, for example.
 第2実施形態に係る読み方判断装置20において、他の構成は、第1実施形態に係る読み方判断装置10と略同一である。したがって、同一部分に同一符号を付して、詳細な説明は省略する。 In the reading determination apparatus 20 according to the second embodiment, other configurations are substantially the same as those of the reading determination apparatus 10 according to the first embodiment. Therefore, the same reference numerals are given to the same parts, and detailed description is omitted.
 図7は、本発明の第2実施形態に係る読み方判断装置の処理フローの一例を示すフローチャートである。例えば、文脈ベクトル生成部16には、入力文章と、その入力文章中の入力単語を特定する情報が入力される(ステップS201)。 FIG. 7 is a flowchart showing an example of a processing flow of the reading determination apparatus according to the second embodiment of the present invention. For example, information specifying an input sentence and an input word in the input sentence is input to the context vector generation unit 16 (step S201).
 次に、文脈ベクトル生成部16は、上記第1実施形態と同様に、入力文章中の、特定された入力単語に対する文脈ベクトルを算出し(ステップS202)、読み方判断部21に対して出力する。 Next, the context vector generation unit 16 calculates a context vector for the specified input word in the input sentence as in the first embodiment (step S202), and outputs it to the reading determination unit 21.
 その後、読み方判断部21は、文脈ベクトル生成部16により算出された文脈ベクトルと、読み方判断情報生成部17により生成された読み方判断情報の平均文脈ベクトルと、に基づいて、入力文章中における入力単語の読み方を判断する(ステップS203)。 Thereafter, the reading determination unit 21 inputs the input word in the input sentence based on the context vector calculated by the context vector generation unit 16 and the average context vector of the reading determination information generated by the reading determination information generation unit 17. Is read (step S203).
 ここで、上記第1実施形態によれば、読み方判断情報は、平均文脈ベクトルと、読み候補とを夫々関連付けた複数組の情報となっている。読み方判断部21は、例えば、読み方判断情報の複数の平均文脈ベクトルのうち、入力単語の文脈ベクトルとのコサイン距離が一番小さい(類似度が高い)平均文脈ベクトルに対応する読み方候補を、その入力単語の読み方と判断する。読み方判断部21は、判断した入力単語の読み方を出力装置22に対して出力する。出力装置22は、読み方判断部21から出力された入力単語の読み方を、例えば、画面表示、プリント表示、音声等により出力する(ステップS204)。 Here, according to the first embodiment, the reading determination information is a plurality of sets of information in which an average context vector and a reading candidate are associated with each other. For example, the reading determination unit 21 selects a reading candidate corresponding to the average context vector having the smallest cosine distance (high similarity) from the context vector of the input word among the plurality of average context vectors of the reading determination information. It is determined that the input word is read. The reading determination unit 21 outputs the determined input word reading to the output device 22. The output device 22 outputs how to read the input word output from the reading determination unit 21 by, for example, screen display, print display, voice, or the like (step S204).
 以上、第2実施形態に係る読み方判断装置20によれば、平均文脈ベクトルにより表される各読み方候補の特徴のうち、最も類似した読み方が選択されるため、より適切な読み方を判断できる。
(第3実施形態)
As described above, according to the reading determination device 20 according to the second embodiment, the most similar reading is selected from the features of each reading candidate represented by the average context vector, and thus a more appropriate reading can be determined.
(Third embodiment)
 図8は、本発明の第3実施形態に係る読み方判断装置の概略的なシステム構成を示すブロック図である。第3実施形態に係る読み方判断装置30は、第2実施形態に係る読み方判断装置20の構成に加えて、コーパスDB15から入力単語の例文情報を取得する例文単語取得部31と、読み方判断部21により判断された単語の読み方を記憶する読み方DB32と、を更に備えている。本実施形態に係る読み方判断装置30は、例えば、オフラインで入力単語の読み方を判断できる。 FIG. 8 is a block diagram showing a schematic system configuration of a reading determination apparatus according to the third embodiment of the present invention. In addition to the configuration of the reading determination device 20 according to the second embodiment, the reading determination device 30 according to the third embodiment includes an example sentence word acquisition unit 31 that acquires example sentence information of input words from the corpus DB 15, and a reading determination unit 21. And a reading DB 32 for storing the reading of the word determined by the above. The reading determination device 30 according to the present embodiment can determine how to read an input word offline, for example.
 第3実施形態に係る読み方判断装置30において、他の構成は、第2実施形態に係る読み方判断装置20と略同一である。したがって、同一部分に同一符号を付して、詳細な説明は省略する。 In the reading determination apparatus 30 according to the third embodiment, other configurations are substantially the same as those of the reading determination apparatus 20 according to the second embodiment. Therefore, the same reference numerals are given to the same parts, and detailed description is omitted.
 例えば、例文単語取得部31は、入力単語を含む複数の例文をコーパスDB15から取得し、各例文中から入力単語を特定する情報を抽出し、文脈ベクトル生成部16に対して出力する。文脈ベクトル生成部16は、例文単語取得部31からの各例文の情報を用いて、各例文の入力単語に対する文脈ベクトルを夫々算出し、読み方判断部21に対して出力する。 For example, the example sentence word acquisition unit 31 acquires a plurality of example sentences including the input word from the corpus DB 15, extracts information specifying the input word from each example sentence, and outputs the information to the context vector generation unit 16. The context vector generation unit 16 calculates the context vector for the input word of each example sentence using the information of each example sentence from the example sentence word acquisition unit 31, and outputs it to the reading determination unit 21.
 読み方判断部21は、読み方判断情報生成部17により生成された読み方判断情報の複数の平均文脈ベクトルのうち、各例文に対する文脈ベクトルとのコサイン距離が一番小さい平均文脈ベクトルに対応する読み方候補を、その例文における入力単語の読み方と夫々判断する。そして、読み方判断部21は、判断した各例文における入力単語の読み方を、読み方DB32に夫々出力し、読み方DB32に記憶させる。さらに、読み方判断部21は、読み方DB32に記憶された複数の読み方から、統計的に最も頻度が高い読み方を一つ選択し、該当入力単語の読み方として判断し、その判断した読み方を出力装置22に対して出力する。 The reading determination unit 21 selects reading candidates corresponding to the average context vector having the smallest cosine distance from the context vector for each example sentence among the plurality of average context vectors of the reading determination information generated by the reading determination information generation unit 17. And how to read the input word in the example sentence. Then, the reading determination unit 21 outputs the determined reading of the input word in each example sentence to the reading DB 32 and stores it in the reading DB 32. Further, the reading determination unit 21 selects one of the statistically highest readings from the plurality of readings stored in the reading DB 32, determines the reading as the corresponding input word, and outputs the determined reading as the output device 22. Output for.
 以上、第3実施形態に係る読み方判断装置30によれば、コーパス情報を用いて入力単語に対する複数の文脈ベクトルを生成し、夫々の読み方を判断し、読み方DB32に蓄積する。そして、読み方DB32に蓄積された複数の読み方の頻度に基づいて、統計的に、その入力単語に対する読み方を判断することができる。
(第4実施形態)
As described above, according to the reading determination device 30 according to the third embodiment, a plurality of context vectors for the input word are generated using the corpus information, each reading is determined, and stored in the reading DB 32. Then, based on the frequency of the plurality of readings stored in the reading DB 32, it is possible to statistically determine how to read the input word.
(Fourth embodiment)
 上記第1乃至3実施形態に係る読み方判断装置10、20、30において、読み方判断情報生成部17は、各読み方候補と、代表となる1つの平均文脈ベクトルと、を夫々関連付けた読み方判断情報を生成しているが、第4実施形態に係る読み方判断装置40において、読み方判断情報生成部47は、各読み方候補と、複数の文脈ベクトルと、を夫々関連付けた読み方判断情報を生成する。 In the reading determination devices 10, 20, and 30 according to the first to third embodiments, the reading determination information generation unit 17 stores reading determination information in which each reading candidate is associated with one representative average context vector. Although generated, in the reading determination device 40 according to the fourth embodiment, the reading determination information generation unit 47 generates reading determination information in which each reading candidate and a plurality of context vectors are associated with each other.
 この場合、読み方判断情報は、文脈ベクトル生成部16により算出された各文脈ベクトルと、それに対応する読み方候補との組を、文脈ベクトルの数だけ集めたものとなる。一方で、上記第1乃至3実施形態の読み方判断情報は、平均文脈ベクトルと、それに対応する読み方候補との組を、読み方候補の数だけ集めたものとなっている。したがって、本実施形態に係る読み方判断情報は、より情報量が大きくなりかつ高精度になる。 In this case, the reading determination information is a set of combinations of each context vector calculated by the context vector generation unit 16 and corresponding reading candidates corresponding to the number of context vectors. On the other hand, the reading determination information of the first to third embodiments is a collection of the average context vector and the reading candidate corresponding to the average context vector as many as the number of reading candidates. Therefore, the reading determination information according to the present embodiment has a larger information amount and higher accuracy.
 読み方判断情報生成部47は、上述のように、文脈ベクトル生成部16により生成された全文脈ベクトルと、各文脈ベクトルに対応する読み方候補とを夫々組にした読み方判断情報を、読み方判断部21に対して出力する。 As described above, the reading determination information generation unit 47 sets the reading determination information obtained by combining all the context vectors generated by the context vector generation unit 16 and the reading candidates corresponding to the respective context vectors. Output for.
 第4実施形態に係る読み方判断装置40において、他の構成は、第1乃至第3実施形態に係る読み方判断装置10、20、30と略同一である。したがって、同一部分には同一符号を付して詳細な説明は省略する。 In the reading determination device 40 according to the fourth embodiment, other configurations are substantially the same as the reading determination devices 10, 20, and 30 according to the first to third embodiments. Therefore, the same parts are denoted by the same reference numerals, and detailed description thereof is omitted.
 なお、本実施形態において、上記第2実施形態と同様に、読み方判断部21は、入力文章及び入力単語から得られた文脈ベクトルと、読み方判断情報の全文脈ベクトルとの、コサイン距離等の類似度を夫々算出してもよい。そして、読み方判断部21は、読み方判断情報の全文脈ベクトルのうち、最も類似度が高い文脈ベクトルに対応する読み方候補を、その入力単語の読み方と判断する。 In this embodiment, as in the second embodiment, the reading determination unit 21 resembles the cosine distance or the like between the context vector obtained from the input sentence and the input word and all context vectors of the reading determination information. Each degree may be calculated. Then, the reading determination unit 21 determines the reading candidate corresponding to the context vector having the highest similarity among all the context vectors of the reading determination information as the reading of the input word.
 また、本実施形態において、上記第3実施形態と同様に、読み方判断部21は、コーパスDB15の各例文から得られた文脈ベクトルと、読み方判断情報の全文脈ベクトルとの類似度を算出してもよい。そして、読み方判断部21は、読み方判断情報の全文脈ベクトルのうち、各例文の文脈ベクトルとの類似度が最も高い文脈ベクトルに対応する読み方候補を、その例文における入力単語の読み方と判断し、読み方DB32に対して出力する。 In the present embodiment, as in the third embodiment, the reading determination unit 21 calculates the similarity between the context vector obtained from each example sentence in the corpus DB 15 and all the context vectors of the reading determination information. Also good. Then, the reading determination unit 21 determines the reading candidate corresponding to the context vector having the highest similarity to the context vector of each example sentence among all the context vectors of the reading determination information as the reading of the input word in the example sentence. Output to the reading DB 32.
 以上、第4実施形態に係る読み方判断装置40によれば、上記第1乃至第3実施形態のように、単一の代表文脈ベクトルを求め、各読み方間において十分な分離が困難となる場合でも、各要素単語の文脈ベクトルの類似性を用いて、適切な読み方判断が可能となる。
(第5実施形態)
As described above, according to the reading determination device 40 according to the fourth embodiment, even when a single representative context vector is obtained and sufficient separation between readings is difficult as in the first to third embodiments. Using the similarity of the context vectors of each element word, an appropriate reading method can be determined.
(Fifth embodiment)
 上記第1実施形態に係る読み方判断装置10において、単語集合生成部14は、複数の読み方候補に類似する複数の要素単語からなる単語集合として、シソーラス上で同一カテゴリーに属する要素単語を含む単語集合を生成しているが、第5実施形態に係る読み方判断装置50において、単語集合生成部54は、シソーラス上で同一階層の近接したカテゴリーに属する要素単語も含む単語集合を生成する。これにより、より広い類義語を含む単語集合を生成することで、読み方判断情報の情報量を効果的に増加させ、その精度を向上させることができる。 In the reading determination device 10 according to the first embodiment, the word set generation unit 14 includes a word set including element words belonging to the same category on the thesaurus as a word set including a plurality of element words similar to a plurality of reading candidates. However, in the reading determination apparatus 50 according to the fifth embodiment, the word set generation unit 54 generates a word set including element words belonging to adjacent categories in the same hierarchy on the thesaurus. Thus, by generating a word set including a broader synonym, it is possible to effectively increase the information amount of the reading determination information and improve the accuracy thereof.
 第5実施形態に係る読み方判断装置50において、他の構成は、第1実施形態に係る読み方判断装置10と略同一である。したがって、同一部分には同一符号を付して詳細な説明は省略する。 In the reading determination device 50 according to the fifth embodiment, the other configuration is substantially the same as that of the reading determination device 10 according to the first embodiment. Therefore, the same parts are denoted by the same reference numerals, and detailed description thereof is omitted.
 なお、単語集合生成部54は、近接したカテゴリーの選択方法を変更することにより、類似度を制御してもよい。例えば、上記第1実施形態と同様に、同表記異発音語組生成部11は、入力単語「黒子」に対して、読み方候補DB12から、例えば、2つのエントリ、読み方候補1=「ほくろ」(カテゴリーは身体の染み)、及び、読み方候補2=「くろご」(カテゴリーは歌舞伎の補助員)を取得する。そして、単語集合生成部54は、各読み方候補1、2に対応して、その読み方候補と類似関係にある、複数の要素単語から構成される単語集合を生成する。 Note that the word set generation unit 54 may control the similarity by changing the method of selecting adjacent categories. For example, in the same way as in the first embodiment, the same notation different pronunciation word set generation unit 11 receives, for example, two entries from the reading candidate DB 12 for the input word “Kuroko”, reading candidate 1 = “mole” ( Category is body stain), and reading candidate 2 = "Kurogo" (category is Kabuki assistant). And the word set production | generation part 54 produces | generates the word set comprised from the some element word which has a similar relationship with the reading candidate corresponding to each reading candidate 1 and 2. FIG.
 ここで、上記類似関係の要素単語には、上述の如く、シソーラス上で同一カテゴリーに属する要素単語に加えて、シソーラス上で同一階層の近接したカテゴリーに属する要素単語も含まれる。また、上記近接したカテゴリーとは、例えば、シソーラス上で定義されたカテゴリー間の関係の度合を用いて、最も近い1つのカテゴリーを指す。なお、図4において、読み方候補1=「ほくろ」を含むカテゴリーに対して、上位方向に1階層、及び下位方向に1階層を示している。 Here, as described above, the element words having the similar relationship include, in addition to the element words belonging to the same category on the thesaurus, the element words belonging to the adjacent category of the same hierarchy on the thesaurus. The adjacent category refers to the closest category using, for example, the degree of relationship between categories defined on the thesaurus. In FIG. 4, for the category including the reading candidate 1 = “mole”, one level is shown in the upper direction and one level is shown in the lower direction.
 単語集合生成部54は、例えば、図4に示すシソーラス上において、読み方候補1「ほくろ」の属するカテゴリーと同一カテゴリー及び近接したカテゴリーに属する要素単語を、夫々抽出する。そして、単語集合生成部54は、「身体の染み」カテゴリーと同一階層にある「身体の表面様態」カテゴリー、および、「身体の色」カテゴリーから、予めシソーラスに定義されたカテゴリー間の類似性に従って、「身体の表面様態」カテゴリーを選択する。さらに、単語集合生成部54は、これら「身体の染み」カテゴリーおよび「身体の表面様態」カテゴリーに属する要素単語を抽出し、単語集合1を生成する。 The word set generation unit 54 extracts, for example, element words that belong to the same category as the category to which the reading candidate 1 “mole” belongs and a close category on the thesaurus shown in FIG. Then, the word set generation unit 54 follows the similarity between categories defined in advance in the thesaurus from the “body surface state” category and the “body color” category in the same hierarchy as the “body stain” category. , Select the “body surface” category. Further, the word set generation unit 54 extracts element words belonging to the “body stain” category and the “body surface state” category, and generates a word set 1.
 なお、単語集合1={要素単語1-1:染み、要素単語1-2:しみ、要素単語1-3:黒子、要素単語1-4:ほくろ、要素単語1-5:ホクロ、要素単語1-6:lentigo、要素単語1-7:にきび、要素単語1-8:吹出物、要素単語1-9:毛孔}となり、カテゴリー={身体の染み}となり、要素単語の数N1=9となる。 Note that word set 1 = {element word 1-1: stain, element word 1-2: blot, element word 1-3: black child, element word 1-4: mole, element word 1-5: mole, element word 1 -6: lentigo, element word 1-7: acne, element word 1-8: pimple, element word 1-9: pores, category = {body stain}, and number of element words N1 = 9.
 以上、第5実施形態に係る読み方判断装置50によれば、シソーラス上で同一カテゴリーに属する要素単語に加えて、シソーラス上で同一階層の近接したカテゴリーに属する要素単語も含む単語集合を生成することで、より広い類義語を含む単語集合を生成することができる。
(第6実施形態)
As described above, according to the reading determination device 50 according to the fifth embodiment, in addition to the element words belonging to the same category on the thesaurus, the word set including the element words belonging to the adjacent categories of the same hierarchy on the thesaurus is generated. Thus, a word set including a broader synonym can be generated.
(Sixth embodiment)
 上記第1実施形態に係る読み方判断装置10において、単語集合生成部14は、複数の読み方候補に類似する複数の要素単語からなる単語集合として、シソーラス上で同一カテゴリーに属する要素単語を含む単語集合を生成しているが、第6実施形態に係る読み方判断装置60において、単語集合生成部64は、シソーラス上で上位階層及び/又は下位階層のカテゴリーに属する要素単語も含む単語集合を生成してもよい。これにより、概念の上位下位関係を対象とした、より広い類義語を含む単語集合を生成することで、読み方判断情報の情報量を効果的に増加させ、その精度を向上させることができる。 In the reading determination device 10 according to the first embodiment, the word set generation unit 14 includes a word set including element words belonging to the same category on the thesaurus as a word set including a plurality of element words similar to a plurality of reading candidates. In the reading determination device 60 according to the sixth embodiment, the word set generation unit 64 generates a word set including element words that belong to the upper hierarchy and / or lower hierarchy categories on the thesaurus. Also good. Thus, by generating a word set including broader synonyms for the upper and lower relations of the concept, the information amount of the reading determination information can be effectively increased and the accuracy thereof can be improved.
 第6実施形態に係る読み方判断装置60において、他の構成は、第1実施形態に係る読み方判断装置10と略同一である。したがって、同一部分には同一符号を付して詳細な説明は省略する。 In the reading determination device 60 according to the sixth embodiment, the other configuration is substantially the same as that of the reading determination device 10 according to the first embodiment. Therefore, the same parts are denoted by the same reference numerals, and detailed description thereof is omitted.
 なお、単語集合生成部64は、シソーラス上において、何階層の上位階層及び/又は下位階層までを対象範囲にするかによって、類似度を制御することができる。 Note that the word set generation unit 64 can control the degree of similarity depending on how many upper and / or lower hierarchies are targeted in the thesaurus.
 例えば、上記第1実施形態と同様に、同表記異発音語組生成部11は、入力単語「黒子」に対して、読み方候補DB12から、例えば、2つのエントリ、読み方候補1=「ほくろ」(カテゴリーは身体の染み)、及び、読み方候補2=「くろご」(カテゴリーは歌舞伎の補助員)を取得する。そして、単語集合生成部64は、各読み方候補1、2に対応して、その読み方候補1、2と類似関係にある、複数の要素単語から構成される単語集合を生成する。 For example, in the same way as in the first embodiment, the same notation different pronunciation word set generation unit 11 receives, for example, two entries from the reading candidate DB 12 for the input word “Kuroko”, reading candidate 1 = “mole” ( Category is body stain), and reading candidate 2 = "Kurogo" (category is Kabuki assistant). And the word set production | generation part 64 produces | generates the word set comprised from the some element word which has the similar relationship with the reading candidate 1 and 2 corresponding to each reading candidate 1 and 2. FIG.
 ここで、上記類似関係の要素単語には、上述の如く、シソーラス上で同一カテゴリーに属する要素単語に加えて、シソーラス上で、予め設定した階層数の上位階層及び/又は下位階層のカテゴリーに属する要素単語も含まれる。 Here, as described above, in addition to the element words belonging to the same category on the thesaurus, the element words having the similar relationship belong to the category of the upper hierarchy and / or the lower hierarchy of the number of hierarchies set in advance on the thesaurus. Element words are also included.
 単語集合生成部64は、例えば、図4に示すシソーラス上において、読み方候補1「ほくろ」の属するカテゴリーと同一カテゴリーの要素単語、及び予め設定した階層数の上位階層及び下位階層のカテゴリーに属する要素単語を、夫々抽出する。 For example, on the thesaurus shown in FIG. 4, the word set generation unit 64 includes element words that belong to the same category as the category to which the reading candidate 1 “mole” belongs, and elements that belong to categories of upper and lower hierarchies having a preset number of hierarchies. Each word is extracted.
 ここで、「身体の染み」カテゴリーの上位1階層に存在するのは、「身体の表面」カテゴリー1つであり、同じく下位1階層に存在するのは、「身体の染みの色」カテゴリーおよび「身体の染みの形状」カテゴリーの2つである。そこで、単語集合生成部64は、これら上位1階層および下位1階層のカテゴリーに属する要素単語を抽出し、単語集合1を生成する。 Here, there is one “body surface” category in the upper level of the “body stain” category, and there are also “color of body stain” categories and “ It is in the category “Body stain shape”. Therefore, the word set generation unit 64 extracts element words that belong to the categories of the upper one layer and the lower one layer, and generates the word set 1.
 なお、単語集合1={要素単語1-1:染み、要素単語1-2:しみ、要素単語1-3:黒子、要素単語1-4:ほくろ、要素単語1-5:ホクロ、要素単語1-6:lentigo、要素単語1-7:色、要素単語1-8:染み、要素単語1-9:皺、要素単語1-10:赤、要素単語1-11:黒、要素単語1-12:灰色、要素単語1-13:丸、要素単語1-14:点、要素単語1-15:三角}となり、カテゴリー={身体の染み}となり、要素単語の数N1=15となる。 Note that word set 1 = {element word 1-1: stain, element word 1-2: blot, element word 1-3: black child, element word 1-4: mole, element word 1-5: mole, element word 1 -6: Lentigo, Element word 1-7: Color, Element word 1-8: Stain, Element word 1-9: Amber, Element word 1-10: Red, Element word 1-11: Black, Element word 1-12 : Gray, element word 1-13: circle, element word 1-14: dot, element word 1-15: triangle}, category = {body stain}, and the number of element words N1 = 15.
 以上、第6実施形態に係る読み方判断装置60によれば、シソーラス上で同一カテゴリーに属する要素単語に加えて、シソーラス上で、予め設定した階層数の上位階層及び/又は下位階層のカテゴリーに属する要素単語も含む単語集合を生成することで、より広い類義語を含む単語集合を生成することができる。
(第7実施形態)
As described above, according to the reading determination device 60 according to the sixth embodiment, in addition to the element words belonging to the same category on the thesaurus, it belongs to the category of the upper hierarchy and / or the lower hierarchy of the number of hierarchies set in advance on the thesaurus. By generating a word set including element words, a word set including a broader synonym can be generated.
(Seventh embodiment)
 図9は、本発明の第7実施形態に係る読み方判断装置の概略的なシステム構成を示すブロック図である。本発明の第7実施形態に係る読み方判断装置70は、第1実施形態に係る読み方判断装置10の構成に加えて、重複する要素単語を検出し、削除する要素単語削除部71と、重複する文脈ベクトルを検出し、削除する文脈ベクトル削除部72と、を更に備えている。 FIG. 9 is a block diagram showing a schematic system configuration of a reading determination apparatus according to the seventh embodiment of the present invention. In addition to the configuration of the reading determination apparatus 10 according to the first embodiment, the reading determination apparatus 70 according to the seventh embodiment of the present invention overlaps with an element word deletion unit 71 that detects and deletes overlapping element words. And a context vector deletion unit 72 that detects and deletes the context vector.
 要素単語削除部71は、単語集合生成部14により生成された複数の読み方候補1~Mに対する単語集合間で、重複する要素単語を検出する。ここで、上記重複する要素単語とは、例えば、少なくとも一組の要素単語が重複する場合を指す。そして、要素単語削除部71は、重複する要素単語の一方を単語集合の中から削除し、削除した単語集合を文脈ベクトル生成部16に出力する。一方、要素単語削除部71は、重複した要素単語を含まない単語集合を、そのまま、文脈ベクトル生成部16に出力する。 The element word deletion unit 71 detects overlapping element words among the word sets for the plurality of reading candidates 1 to M generated by the word set generation unit 14. Here, the above-described overlapping element word indicates, for example, a case where at least one set of element words overlaps. Then, the element word deletion unit 71 deletes one of the overlapping element words from the word set, and outputs the deleted word set to the context vector generation unit 16. On the other hand, the element word deletion unit 71 outputs a word set that does not include duplicated element words to the context vector generation unit 16 as it is.
 ここで、重複する要素単語は、対応する文脈ベクトルも同一となる。このため、複数の読み方候補の単語集合間において重複する要素単語が存在すると、その要素単語に基づいて生成される読み方判断情報の重複度も大きくなる。したがって、予め重複する要素単語を除去することによって、読み方判断情報における分離度を高くし、読み方判断情報の精度を高めることができる。 Here, overlapping element words have the same context vector. For this reason, if there are overlapping element words between word sets of a plurality of reading candidates, the degree of duplication of reading determination information generated based on the element words also increases. Therefore, by removing previously overlapping element words, the degree of separation in the reading determination information can be increased, and the accuracy of the reading determination information can be increased.
 文脈ベクトル削除部72は、文脈ベクトル生成部16により生成される複数の単語集合1~Mの要素単語に夫々対応する文脈ベクトルの中から、各単語集合1~M間で同一となる文脈ベクトルを検出し、その一方の文脈ベクトルを削除し、読み方判断情報生成部17に出力する。なお、上記同一となる文脈ベクトルとは、例えば、少なくとも一組の文脈ベクトルが同一となる場合を指す。 The context vector deletion unit 72 selects the same context vector among the word sets 1 to M from among the context vectors corresponding to the element words of the plurality of word sets 1 to M generated by the context vector generation unit 16. Then, one of the context vectors is deleted and output to the reading determination information generation unit 17. The same context vector refers to a case where at least one set of context vectors is the same, for example.
 ここで、複数の単語集合1~M間において、重複する文脈ベクトルが存在すると、文脈ベクトルに基づいて生成される読み方判断情報の重複度も大きくなる。したがって、予め重複する文脈ベクトルを除去することによって、読み方判断情報における分離度を高くし、読み方判断情報の精度を高めることができる。 Here, if there are overlapping context vectors among the plurality of word sets 1 to M, the degree of duplication of the reading determination information generated based on the context vectors also increases. Therefore, by removing overlapping context vectors in advance, the degree of separation in the reading determination information can be increased, and the accuracy of the reading determination information can be increased.
 なお、文脈ベクトル削除部72は、文脈ベクトル生成部16により生成された複数の単語集合1~Mの要素単語の夫々に対する文脈ベクトルの中から、予め定められた距離よりも近接し、相互に類似した文脈ベクトルを検出し、その一方の文脈ベクトルを削除してもよい。 The context vector deletion unit 72 is closer than a predetermined distance from among the context vectors for the element words of the plurality of word sets 1 to M generated by the context vector generation unit 16, and is similar to each other. The detected context vector may be detected, and one of the context vectors may be deleted.
 例えば、文脈ベクトル削除部72は、文脈ベクトル空間上でのコサイン距離が、予め定めた閾値εよりも小さいとき、予め定められた距離よりも近接していると判断する。文脈ベクトル削除部72は、検出された相互に近接する文脈ベクトルの組のうち一方を削除し、読み方判断情報生成部17に出力する。 For example, when the cosine distance on the context vector space is smaller than the predetermined threshold ε, the context vector deletion unit 72 determines that the distance is closer than the predetermined distance. The context vector deletion unit 72 deletes one of the detected sets of context vectors close to each other and outputs the result to the reading determination information generation unit 17.
 ここで、複数の単語集合1~M間において、近接する文脈ベクトルが存在すると、文脈ベクトルに基づいて生成される読み方判断情報の重複度も大きくなる。したがって、予め近接する文脈ベクトルを除去することによって、読み方判断情報における分離度を高くし、読み方判断情報の精度を高めることができる。 Here, when there are adjacent context vectors among the plurality of word sets 1 to M, the degree of duplication of the reading determination information generated based on the context vectors also increases. Therefore, by removing context vectors that are close in advance, the degree of separation in the reading determination information can be increased, and the accuracy of the reading determination information can be increased.
 また、文脈ベクトル削除部72は、文脈ベクトル生成部16により生成された文脈ベクトルに対して、特徴性を強調するための重み係数を乗算してもよい。文脈ベクトル検出部72は、重み係数を乗算した文脈ベクトルを用いて、上述の近接する文脈ベクトルを検出し、検出した文脈ベクトルを削除してもよい。 Further, the context vector deletion unit 72 may multiply the context vector generated by the context vector generation unit 16 by a weighting factor for enhancing the feature. The context vector detection unit 72 may detect the above-described adjacent context vector using the context vector multiplied by the weighting factor, and delete the detected context vector.
 例えば、文脈ベクトルD=Σ(i=1、t)a(i)*V(i)について、各要素単語V(i)の重要度をb(i)とする。ここで、重要度b(i)は、例えば、該当単語がコーパス情報中に出現する際に特徴的であるか否かを示す尺度tf-idfの値を用いてもよい。なお、このtf-idfの値は、tf(単語の出現頻度)とidf(逆出現頻度)との二つの指標から算出される値である。この各単語V(i)の重要度b(i)に応じて、重み係数を設定する。これにより、2つの文脈ベクトルの類似度を求める際に、文脈ベクトルDに重み係数を乗算することにより、特徴性の高い単語に関する差異は強調され、特徴性の低い単語に関する差異は縮小される。したがって、コーパス情報の特徴をより反映した類似度計算を行うことができる。
(第8実施形態)
For example, for the context vector D = Σ (i = 1, t) a (i) * V (i), the importance of each element word V (i) is b (i). Here, as the importance b (i), for example, a value of a scale tf-idf indicating whether or not a corresponding word is characteristic when it appears in the corpus information may be used. The value of tf-idf is a value calculated from two indices, tf (word appearance frequency) and idf (reverse appearance frequency). A weighting coefficient is set according to the importance b (i) of each word V (i). As a result, when obtaining the similarity between two context vectors, the context vector D is multiplied by the weighting coefficient, whereby the difference regarding the word with high characteristic is emphasized and the difference regarding the word with low characteristic is reduced. Therefore, it is possible to perform similarity calculation more reflecting the characteristics of the corpus information.
(Eighth embodiment)
 図10は、本発明の第8実施形態に係る音声合成装置の概略的なシステム構成を示すブロック図である。第5実施形態に係る音声合成装置80は、入力文章の形態素解析を行う形態素解析部81と、第2実施形態に係る読み方判断装置20と、合成音声を生成する音声生成部82と、を備えている。 FIG. 10 is a block diagram showing a schematic system configuration of a speech synthesizer according to the eighth embodiment of the present invention. A speech synthesizer 80 according to the fifth embodiment includes a morpheme analyzer 81 that performs morphological analysis of an input sentence, a reading determination device 20 according to the second embodiment, and a speech generator 82 that generates synthesized speech. ing.
 形態素解析部81は、入力文章に対して形態素解析を行うことで、入力文章を形態素に分割し、複数の形態素のうち自立語を抽出し、読み方判断装置20に出力する。音声生成部82は、読み方判断装置20から出力された入力文章の読み方の情報に基づいて、例えば、波形接続型音声合成方式などを用いて、入力文章に対する合成音声の波形を生成する。なお、上記音声合成で用いる読み方の情報には、例えば、単に音韻列だけでなく、アクセント位置の情報を含むものとする。これにより、例えば名詞「谷」という単語を、人名としての用法では頭高型に、山の反対を表す用法では平板型に発話し分けることも可能となる。 The morpheme analysis unit 81 performs morpheme analysis on the input sentence, divides the input sentence into morphemes, extracts independent words from the plurality of morphemes, and outputs them to the reading determination device 20. Based on the information on how to read the input sentence output from the reading determination device 20, the voice generation unit 82 generates a waveform of a synthesized voice for the input sentence using, for example, a waveform connection type speech synthesis method. Note that the reading information used in the speech synthesis includes, for example, not only a phoneme string but also information on an accent position. As a result, for example, the noun “Tani” can be uttered separately in a head-high shape when used as a person's name, and in a flat shape when used as the opposite of a mountain.
 図11は、本発明の第8実施形態に係る音声合成装置の処理フローの一例を示すフローチャートである。形態素解析部81に入力文章が入力されると(ステップS301)、形態素解析部81は、その入力文章に対して形態素解析を行って(ステップS302)、入力文章を複数の形態素に分割し、自立語を抽出する。そして、形態素解析部81は、入力文章と共に、抽出した自立語を入力単語として、読み方判断装置20に出力する。次に、読み方判断装置20は、形態素解析部81からの入力文章及び入力単語に基づいて、上述の読み方判断処理を行い(ステップS303)、全ての自立語について、読み方を確定し(ステップS304)、入力文章の読み方の情報を生成する(ステップS305)。読み方判断装置20は、生成した入力文章の読み方の情報を、音声生成部82に対して出力する。音声生成部82は、読み方判断装置20からの入力文章の読み方の情報に基づいて、合成音声波形を生成し(ステップS306)、生成した合成音声波形の音声を出力する(ステップS307)。 FIG. 11 is a flowchart showing an example of the processing flow of the speech synthesizer according to the eighth embodiment of the present invention. When the input sentence is input to the morpheme analysis unit 81 (step S301), the morpheme analysis unit 81 performs morpheme analysis on the input sentence (step S302), divides the input sentence into a plurality of morphemes, and becomes independent. Extract words. Then, the morpheme analysis unit 81 outputs the extracted independent word as an input word together with the input sentence to the reading determination device 20. Next, the reading determination device 20 performs the above-described reading determination process based on the input sentence and the input word from the morphological analysis unit 81 (step S303), and determines the reading for all independent words (step S304). Then, information on how to read the input sentence is generated (step S305). The reading determination device 20 outputs information on how to read the generated input sentence to the voice generation unit 82. The voice generation unit 82 generates a synthesized voice waveform based on the information on how to read the input sentence from the reading judgment device 20 (step S306), and outputs the voice of the generated synthesized voice waveform (step S307).
 なお、本発明は上記実施の形態に限られたものではなく、趣旨を逸脱しない範囲で適宜変更することが可能である。また、上述の実施の形態では、本発明をハードウェアの構成として説明したが、本発明は、これに限定されるものではない。本発明は、任意の処理を、CPUにコンピュータプログラムを実行させることにより実現することも可能である。) Note that the present invention is not limited to the above-described embodiment, and can be appropriately changed without departing from the spirit of the present invention. In the above-described embodiments, the present invention has been described as a hardware configuration, but the present invention is not limited to this. The present invention can also realize arbitrary processing by causing a CPU to execute a computer program. )
 プログラムは、様々なタイプの非一時的なコンピュータ可読媒体(non-transitory computer readable medium)を用いて格納され、コンピュータに供給することができる。非一時的なコンピュータ可読媒体は、様々なタイプの実体のある記録媒体(tangible storage medium)を含む。非一時的なコンピュータ可読媒体の例は、磁気記録媒体(例えばフレキシブルディスク、磁気テープ、ハードディスクドライブ)、光磁気記録媒体(例えば光磁気ディスク)、CD-ROM(Read Only Memory)、CD-R、CD-R/W、半導体メモリ(例えば、マスクROM、PROM(Programmable ROM)、EPROM(Erasable PROM)、フラッシュROM、RAM(random access memory))を含む。また、プログラムは、様々なタイプの一時的なコンピュータ可読媒体(transitory computer readable medium)によってコンピュータに供給されてもよい。一時的なコンピュータ可読媒体の例は、電気信号、光信号、及び電磁波を含む。一時的なコンピュータ可読媒体は、電線及び光ファイバ等の有線通信路、又は無線通信路を介して、プログラムをコンピュータに供給できる。 The program can be stored and supplied to a computer using various types of non-transitory computer readable media. Non-transitory computer readable media include various types of tangible storage media (tangible storage medium). Examples of non-transitory computer-readable media include magnetic recording media (eg flexible disks, magnetic tapes, hard disk drives), magneto-optical recording media (eg magneto-optical discs), CD-ROMs (Read Only Memory), CD-Rs, CD-R / W, semiconductor memory (for example, mask ROM, PROM (Programmable ROM), EPROM (Erasable ROM), flash ROM, RAM (random access memory)) are included. The program may also be supplied to the computer by various types of temporary computer-readable media. Examples of transitory computer readable media include electrical signals, optical signals, and electromagnetic waves. The temporary computer-readable medium can supply the program to the computer via a wired communication path such as an electric wire and an optical fiber, or a wireless communication path.
 本発明は、例えば、単語や文章などに対する適切な読み方を判断する読み方判断装置に適用可能である。 The present invention can be applied to, for example, a reading determination device that determines an appropriate reading for a word or a sentence.
 この出願は、2009年3月31日に出願された日本出願特願2009-084920を基礎とする優先権を主張し、その開示の全てをここに取り込む。 This application claims priority based on Japanese Patent Application No. 2009-084920 filed on Mar. 31, 2009, the entire disclosure of which is incorporated herein.
  10  読み方判断装置
  11  同表記異発音語組生成部
  12  読み方DB
  13  シソーラスDB
  14  単語集合生成部
  15  コーパスDB
  16  文脈ベクトル生成部
  17  読み方判断情報生成部
  21  読み方判断部
  22  出力装置
  31  例文単語取得部
  32  読み方DB
  71  要素単語削除出部
  72  文脈ベクトル削除部
  80  音声合成装置
10 Reading Judgment Device 11 Same Pronunciation Different Pronunciation Group Generation Unit 12 Reading DB
13 Thesaurus DB
14 word set generator 15 corpus DB
16 Context vector generation unit 17 Reading determination information generation unit 21 Reading determination unit 22 Output device 31 Example word acquisition unit 32 Reading DB
71 Element word deletion output unit 72 Context vector deletion unit 80 Speech synthesizer

Claims (23)

  1.  複数の読み方候補を有する、単語の読み方を判断するための読み方判断装置であって、
     前記読み方候補に類似する複数の要素単語からなる単語集合を、夫々生成する単語集合生成手段と、
     複数の例文を含むコーパス情報を記憶するコーパスデータベースと、
     前記コーパスデータベースに記憶された前記コーパス情報に基づいて、前記単語集合生成手段により生成された前記単語集合の複数の要素単語に対する特徴量を夫々算出する特徴量算出手段と、
     前記特徴量算出手段により算出された前記単語集合の複数の要素単語に対する特徴量と、前記読み方候補とを夫々関連付けた読み方判断情報を生成する読み方判断情報生成手段と、を備える、ことを特徴とする読み方判断装置。
    A reading determination device for determining how to read a word having a plurality of reading candidates,
    A word set generation means for generating a word set composed of a plurality of element words similar to the reading candidate,
    A corpus database that stores corpus information including multiple example sentences;
    Based on the corpus information stored in the corpus database, feature amount calculation means for calculating feature amounts for a plurality of element words of the word set generated by the word set generation means;
    A reading determination information generating means for generating reading determination information in which the characteristic amounts for a plurality of element words of the word set calculated by the feature amount calculation means are associated with the reading candidates, respectively. Reading judgment device.
  2.  請求項1記載の読み方判断装置であって、
     前記特徴量算出手段は、前記コーパスデータベースに記憶された前記コーパス情報に基づいて、前記単語集合生成手段により生成された前記単語集合の複数の要素単語に対する文脈ベクトルを、前記特徴量として夫々算出する文脈ベクトル生成部を有する、ことを特徴とする読み方判断装置。
    The reading judgment device according to claim 1,
    The feature amount calculating means calculates, as the feature amounts, context vectors for a plurality of element words of the word set generated by the word set generation means based on the corpus information stored in the corpus database. A reading judgment device characterized by having a context vector generation unit.
  3.  請求項2記載の読み方判断装置であって、
     前記読み方判断情報生成手段は、前記文脈ベクトル生成部により算出された前記単語集合の複数の要素単語に対する文脈ベクトルに基づいて、代表となる文脈ベクトルを算出し、該代表の文脈ベクトルと、前記読み方候補とを夫々関連付けた読み方判断情報を生成する、ことを特徴とする読み方判断装置。
    A reading judgment device according to claim 2,
    The reading determination information generating means calculates a representative context vector based on the context vectors for a plurality of element words of the word set calculated by the context vector generation unit, the representative context vector, and the reading A reading determination device characterized by generating reading determination information associated with each candidate.
  4.  請求項3記載の読み方判断装置であって、
     前記読み方判断情報生成手段は、前記文脈ベクトル生成部により算出された前記単語集合の複数の要素単語に対する文脈ベクトルの平均値を、前記代表の文脈ベクトルとして算出する、ことを特徴とする読み方判断装置。
    A reading judgment device according to claim 3,
    The reading determination information generating means calculates, as the representative context vector, an average value of context vectors for a plurality of element words of the word set calculated by the context vector generation unit. .
  5.  請求項1乃至4のうちいずれか1項記載の読み方判断装置であって、
     前記単語集合の要素単語は、前記読み方候補の同義語を含む、ことを特徴とする読み方判断装置。
    A reading judgment device according to any one of claims 1 to 4,
    An element word of the word set includes a synonym of the reading candidate.
  6.  請求項1乃至5のうちいずれか1項記載の読み方判断装置であって、
     前記単語集合の要素単語は、シソーラス上において、前記読み方候補の属するカテゴリーに対して上位及び下位関係にあるカテゴリーに属する単語を含む、ことを特徴とする読み方判断装置。
    The reading judgment device according to any one of claims 1 to 5,
    An element word of the word set includes a word belonging to a category that has a higher and lower relationship with respect to a category to which the reading candidate belongs on a thesaurus.
  7.  請求項1乃至6のうちいずれか1項記載の読み方判断装置であって、
     前記読み方候補間の分離度を高める処理を行う分離処理手段を更に備える、ことを特徴とする読み方判断装置。
    The reading judgment device according to any one of claims 1 to 6,
    A reading judgment device further comprising a separation processing means for performing a process for increasing the degree of separation between the reading candidates.
  8.  請求項7記載の読み方判断装置であって、
     前記分離処理手段は、前記単語集合間の要素単語の中から、重複する前記要素単語を検出し、その一方を削除する単語削除部を有する、ことを特徴とする読み方判断装置。
    The reading determination device according to claim 7,
    The reading determination apparatus according to claim 1, wherein the separation processing unit includes a word deletion unit that detects the overlapping element word from element words between the word sets and deletes one of the element words.
  9.  請求項7又は8記載の読み方判断装置であって、
     前記分離処理手段は、前記単語集合間の要素単語に対応する文脈ベクトルの中から、同一となる文脈ベクトルを検出し、その一方を削除する文脈ベクトル削除部を有する、ことを特徴とする読み方判断装置。
    The reading judgment device according to claim 7 or 8,
    The separation processing means includes a context vector deletion unit that detects the same context vector from context vectors corresponding to element words between the word sets and deletes one of the context vectors. apparatus.
  10.  請求項7乃至9のうちいずれか1項記載の読み方判断装置であって、
     前記分離処理手段は、前記単語集合間の要素単語に対応する文脈ベクトルの中から、相互に類似する文脈ベクトルを検出し、その一方を削除する文脈ベクトル削除部を有する、ことを特徴とする読み方判断装置。
    The reading judgment device according to any one of claims 7 to 9,
    The method according to claim 1, wherein the separation processing unit includes a context vector deletion unit that detects mutually similar context vectors from context vectors corresponding to element words between the word sets and deletes one of the context vectors. Judgment device.
  11.  請求項9記載の読み方判断装置であって、
     前記文脈ベクトル削除部は、前記要素単語の重要度に応じた重み係数を夫々設定し、該重み係数を乗じた前記文脈ベクトルに基づいて、前記相互に類似する文脈ベクトルを検出する、ことを特徴とする読み方判断装置。
    The reading judgment device according to claim 9, wherein
    The context vector deletion unit sets a weighting factor corresponding to the importance of the element word, and detects the similar context vectors based on the context vector multiplied by the weighting factor. Reading judgment device.
  12.  請求項1乃至10のうちいずれか1項記載の読み方判断装置であって、
     前記読み方判断情報生成手段により生成された前記読み方判断情報に基づいて、単語の読み方を判断する読み方判断手段を更に備える、ことを特徴とする読み方判断装置。
    The reading judgment device according to any one of claims 1 to 10,
    A reading judgment apparatus, further comprising reading judgment means for judging how to read a word based on the reading judgment information generated by the reading judgment information generation means.
  13.  請求項11記載の読み方判断装置であって、
     読み方判断手段は、前記読み方判断情報の複数の代表の文脈ベクトルのうち、入力単語の文脈ベクトルとの類似度が高い前記代表の文脈ベクトルに対応する読み方候補を、その入力単語の読み方と判断する、ことを特徴とする読み方判断装置。
    The reading judgment device according to claim 11,
    The reading determination means determines a reading candidate corresponding to the representative context vector having a high similarity to the context vector of the input word among the plurality of representative context vectors of the reading determination information as the reading of the input word. The reading judgment apparatus characterized by the above.
  14.  請求項12記載の読み方判断装置であって、
     前記単語を含む複数の例文情報をコーパスデータベースから取得する例文単語取得手段を更に備え、
     文脈ベクトル生成手段は、前記例文単語取得手段により取得された各例文情報に基づいて、文脈ベクトルを夫々生成し、
     前記読み方判断手段は、前記文脈ベクトルに対応した例文情報における前記単語の読み方を夫々判断し、該判断された複数の読み方のうち、最も頻度の高い読み方を、前記単語の読み方として、決定する、ことを特徴とする読み方判断装置。
    The reading judgment device according to claim 12, wherein
    Further comprising example sentence word acquisition means for acquiring a plurality of example sentence information including the word from a corpus database;
    The context vector generation means generates a context vector based on each example sentence information acquired by the example sentence word acquisition means,
    The reading determination means determines how to read the word in the example sentence information corresponding to the context vector, and determines the most frequent reading among the determined reading as the reading of the word; A device for judging how to read.
  15.  請求項1乃至14のうちいずれか1項記載の読み方判断装置であって、
     シソーラス辞書情報を記憶するシソーラスデータベースを更に備え、
     前記単語集合生成手段は、前記読み方候補に類似する複数の要素単語からなる単語集合を、前記シソーラスデータベースのシソーラス辞書情報に基づいて、生成する、ことを特徴とする読み方判断装置。
    The reading judgment device according to any one of claims 1 to 14,
    A thesaurus database for storing thesaurus dictionary information;
    The reading method determining apparatus according to claim 1, wherein the word set generation unit generates a word set including a plurality of element words similar to the reading candidate based on thesaurus dictionary information of the thesaurus database.
  16.  請求項1乃至15のうちいずれか1項記載の読み方判断装置を備え、
     該読み方判断装置により判断された前記単語の読み方に基づいて、音声を合成する、ことを特徴とする音声合成装置。
    The reading judgment device according to any one of claims 1 to 15,
    A speech synthesizer characterized by synthesizing speech based on how to read the word determined by the reading determination device.
  17.  請求項16記載の音声合成装置であって、
     入力文章に対して形態素解析を行い、前記入力文章を形態素に分割する形態素解析手段と、
     前記形態素解析手段により分割された前記形態素の読み方を判断する前記読み方判断装置と、
     前記読み方判断装置により判断された前記入力文章の読み方に基づいて、音声を合成する音声生成部と、を備える、ことを特徴とする音声合成装置。
    The speech synthesizer according to claim 16, wherein
    Morphological analysis is performed on the input text, and the morphological analysis means for dividing the input text into morphemes,
    The reading determination device for determining how to read the morpheme divided by the morpheme analysis means;
    A speech synthesizer comprising: a speech generation unit that synthesizes speech based on how to read the input sentence determined by the reading determination device.
  18.  複数の読み方候補を有する、単語の読み方を判断するための読み方判断方法であって、
     前記読み方候補に類似する複数の要素単語からなる単語集合を、夫々生成し、
     複数の例文を含むコーパス情報に基づいて、前記生成した単語集合の複数の要素単語に対する特徴量を夫々算出し、
     前記算出した単語集合の複数の要素単語に対する特徴量と、前記読み方候補とを夫々関連付けた読み方判断情報を生成する、ことを特徴とする読み方判断方法。
    A reading determination method for determining how to read a word having a plurality of reading candidates,
    A word set composed of a plurality of element words similar to the reading candidate is generated respectively.
    Based on the corpus information including a plurality of example sentences, the feature amounts for the plurality of element words of the generated word set are calculated, respectively.
    A reading determination method characterized by generating reading determination information in which feature quantities for a plurality of element words in the calculated word set are associated with the reading candidates.
  19.  請求項18記載の読み方判断方法であって、
     前記コーパス情報に基づいて、前記生成した単語集合の複数の要素単語に対する文脈ベクトルを、前記特徴量として夫々算出する、ことを特徴とする読み方判断方法。
    A method for judging how to read according to claim 18,
    A reading method determination method, characterized in that, based on the corpus information, context vectors for a plurality of element words of the generated word set are respectively calculated as the feature amounts.
  20.  請求項19記載の読み方判断方法であって、
     前記算出した単語集合の複数の要素単語に対する文脈ベクトルに基づいて、代表となる文脈ベクトルを算出し、該代表の文脈ベクトルと、前記読み方候補とを夫々関連付けた読み方判断情報を生成する、ことを特徴とする読み方判断方法。
    A method for judging how to read according to claim 19,
    Calculating a representative context vector based on context vectors for a plurality of element words of the calculated word set, and generating reading determination information in which the representative context vector and the reading candidate are associated with each other. Characteristic reading method.
  21.  請求項18乃至20のうちいずれか1項記載の読み方判断方法であって、
     前記読み方候補間の分離度を高める処理を更に行う、ことを特徴とする読み方判断方法。
    A method of judging how to read according to any one of claims 18 to 20,
    A method for judging how to read, further comprising a process of increasing the degree of separation between the reading candidates.
  22.  請求項18乃至21のうちいずれか1項記載の読み方判断方法であって、
     前記生成した前記読み方判断情報に基づいて、単語の読み方を更に判断する、ことを特徴とする読み方判断方法。
    A method of judging how to read according to any one of claims 18 to 21,
    A reading determination method, further comprising: determining how to read a word based on the generated reading determination information.
  23.  複数の読み方候補を有する、単語の読み方を判断するための読み方判断プログラムを格納した非一時的なコンピュータ可読媒体であって、
     前記読み方候補に類似する複数の要素単語からなる単語集合を、夫々生成する処理と、
     複数の例文を含むコーパス情報に基づいて、前記生成された前記単語集合の複数の要素単語に対する特徴量を夫々算出する処理と、
     前記算出した前記単語集合の複数の要素単語に対する特徴量と、前記読み方候補とを夫々関連付けた読み方判断情報を生成する処理と、をコンピュータに実行させる読み方判断プログラムを格納した非一時的なコンピュータ可読媒体。
    A non-transitory computer-readable medium having a plurality of reading candidates and storing a reading determination program for determining how to read a word,
    A process of generating a word set composed of a plurality of element words similar to the reading candidate,
    Based on corpus information including a plurality of example sentences, a process of calculating feature amounts for a plurality of element words of the generated word set, respectively,
    A non-transitory computer-readable program that stores a reading determination program that causes a computer to execute a process for generating reading determination information that associates the calculated feature quantities for a plurality of element words of the word set with the reading candidates. Medium.
PCT/JP2010/001753 2009-03-31 2010-03-11 Device, method, program for reading determination, computer readable medium therefore, and voice synthesis device WO2010113396A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2011506983A JP5533853B2 (en) 2009-03-31 2010-03-11 Reading judgment device, method, program, and speech synthesizer

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2009-084920 2009-03-31
JP2009084920 2009-03-31

Publications (1)

Publication Number Publication Date
WO2010113396A1 true WO2010113396A1 (en) 2010-10-07

Family

ID=42827715

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2010/001753 WO2010113396A1 (en) 2009-03-31 2010-03-11 Device, method, program for reading determination, computer readable medium therefore, and voice synthesis device

Country Status (2)

Country Link
JP (1) JP5533853B2 (en)
WO (1) WO2010113396A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019008477A (en) * 2017-06-22 2019-01-17 富士通株式会社 Discrimination program, discrimination device and discrimination method
JP2020052819A (en) * 2018-09-27 2020-04-02 大日本印刷株式会社 Information processing apparatus, information processing method, and program

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06119144A (en) * 1992-10-02 1994-04-28 Toshiba Corp Document read-alout device
JPH07114572A (en) * 1993-10-18 1995-05-02 Sharp Corp Document classifying device
JPH1115497A (en) * 1997-06-19 1999-01-22 Fujitsu Ltd Name reading-out speech synthesis device
JP2008134750A (en) * 2006-11-28 2008-06-12 Nippon Telegr & Teleph Corp <Ntt> Data classifier, data classification method, data classification program, and recording medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5146979B2 (en) * 2006-06-02 2013-02-20 株式会社国際電気通信基礎技術研究所 Ambiguity resolution device and computer program in natural language

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06119144A (en) * 1992-10-02 1994-04-28 Toshiba Corp Document read-alout device
JPH07114572A (en) * 1993-10-18 1995-05-02 Sharp Corp Document classifying device
JPH1115497A (en) * 1997-06-19 1999-01-22 Fujitsu Ltd Name reading-out speech synthesis device
JP2008134750A (en) * 2006-11-28 2008-06-12 Nippon Telegr & Teleph Corp <Ntt> Data classifier, data classification method, data classification program, and recording medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KIYOTAKA UCHIMOTO ET AL.: "Doshi no Goiteki Chishiki Kakutoku ni Okeru Ruigigo no Yorei o Mochiita Tagisei no Ruibetsu", INFORMATION PROCESSING SOCIETY OF JAPAN KENKYU HOKOKU, INFORMATION PROCESSING SOCIETY OF JAPAN, vol. 94, no. 47, 28 May 1994 (1994-05-28), pages 105 - 112 *
YOSHIKI TANBA ET AL.: "Tango Vector o Mochiita Tagigo no Imi Suitei -Kyoki Vector to Teigi Kyori Vector no Hikaku", INFORMATION PROCESSING SOCIETY OF JAPAN KENKYU HOKOKU, INFORMATION PROCESSING SOCIETY OF JAPAN, vol. 94, no. 63, 22 July 1994 (1994-07-22), pages 49 - 56 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2019008477A (en) * 2017-06-22 2019-01-17 富士通株式会社 Discrimination program, discrimination device and discrimination method
JP2020052819A (en) * 2018-09-27 2020-04-02 大日本印刷株式会社 Information processing apparatus, information processing method, and program
JP7115187B2 (en) 2018-09-27 2022-08-09 大日本印刷株式会社 Information processing device, information processing method and program

Also Published As

Publication number Publication date
JPWO2010113396A1 (en) 2012-10-04
JP5533853B2 (en) 2014-06-25

Similar Documents

Publication Publication Date Title
TWI539441B (en) Speech recognition method and electronic apparatus
Sridhar et al. Exploiting acoustic and syntactic features for automatic prosody labeling in a maximum entropy framework
Qian et al. Automatic prosody prediction and detection with conditional random field (crf) models
Tachbelie et al. Using different acoustic, lexical and language modeling units for ASR of an under-resourced language–Amharic
KR101735195B1 (en) Method, system and recording medium for converting grapheme to phoneme based on prosodic information
JP5524138B2 (en) Synonym dictionary generating apparatus, method and program thereof
Gelas et al. Quality assessment of crowdsourcing transcriptions for African languages
Sangeetha et al. Speech translation system for english to dravidian languages
Wu et al. Automatic generation of synthesis units and prosodic information for Chinese concatenative synthesis
Cucu et al. SMT-based ASR domain adaptation methods for under-resourced languages: Application to Romanian
Suzuki et al. Accent sandhi estimation of Tokyo dialect of Japanese using conditional random fields
KR20080045413A (en) Method for predicting phrase break using static/dynamic feature and text-to-speech system and method based on the same
Cole et al. Corpus phonology with speech resources
Zhang et al. A study on functional loads of phonetic contrasts under context based on mutual information of Chinese text and phonemes
Kayte et al. A text-to-speech synthesis for Marathi language using festival and Festvox
Wutiwiwatchai et al. Thai text-to-speech synthesis: a review
Wang et al. RNN-based prosodic modeling for mandarin speech and its application to speech-to-text conversion
JP5533853B2 (en) Reading judgment device, method, program, and speech synthesizer
Chu et al. A concatenative Mandarin TTS system without prosody model and prosody modification.
Sitaram et al. Text to speech in new languages without a standardized orthography
Pellegrini et al. Automatic word decompounding for asr in a morphologically rich language: Application to amharic
Prahallad Automatic building of synthetic voices from audio books
Anto et al. Text to speech synthesis system for English to Malayalam translation
Akinwonmi Development of a prosodic read speech syllabic corpus of the Yoruba language
JPH0962286A (en) Voice synthesizer and the method thereof

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10758182

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2011506983

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10758182

Country of ref document: EP

Kind code of ref document: A1