WO2013012093A1 - Information processing system, method of learning recognition dictionary, and information processing program - Google Patents

Information processing system, method of learning recognition dictionary, and information processing program Download PDF

Info

Publication number
WO2013012093A1
WO2013012093A1 PCT/JP2012/068747 JP2012068747W WO2013012093A1 WO 2013012093 A1 WO2013012093 A1 WO 2013012093A1 JP 2012068747 W JP2012068747 W JP 2012068747W WO 2013012093 A1 WO2013012093 A1 WO 2013012093A1
Authority
WO
WIPO (PCT)
Prior art keywords
probability
instance
reference vector
bag
correct
Prior art date
Application number
PCT/JP2012/068747
Other languages
French (fr)
Japanese (ja)
Inventor
利憲 細井
博義 宮野
Original Assignee
株式会社Nec情報システムズ
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社Nec情報システムズ filed Critical 株式会社Nec情報システムズ
Publication of WO2013012093A1 publication Critical patent/WO2013012093A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks

Definitions

  • the present invention relates to an information processing system, a recognition dictionary learning method, and an information processing program, and more particularly to an information processing system, a recognition dictionary learning method, and an information processing program used in pattern recognition.
  • Recognition of an input vector may be performed by setting a category to which a reference vector closest to the input vector belongs as a recognition result of the input vector based on a distance calculation between a reference vector group called a recognition dictionary and the input vector. is there.
  • the recognition accuracy varies depending on the value of the reference vector. Therefore, how to set the value of the reference vector is important for improving the recognition accuracy.
  • Patent Document 3 and Non-Patent Document 1 disclose Learning Vector Quantization (LVQ) as a recognition dictionary learning method using a reference vector. According to LVQ, the learning process is completed in a very short time compared to a statistical pattern recognition method that does not use other reference vectors such as a perceptron type neural network or a support vector machine.
  • LVQ Learning Vector Quantization
  • Patent Document 1 discloses an example of a technique improved from this LVQ.
  • Non-Patent Document 2 discloses specific effects for one of the methods described in Patent Document 1.
  • Non-Patent Document 3 discloses an improved method of these methods.
  • the reference vector value is automatically set using a plurality of input vectors as learning data.
  • Patent Document 1 Patent Document 3, Non-Patent Document 1, Non-Patent Document 2, and Non-Patent Document 3
  • it is necessary to correctly assign the category to which all input vectors used for learning belong.
  • it takes time and effort to assign the correct category in advance.
  • patent document 3 non-patent document 1, non-patent document 2, non-patent document 3, when the information prepared in advance is incomplete for the category to which the input vector used for learning belongs, We could't make a recognition dictionary from that information.
  • the objective of this invention is providing the information processing system which solves the above-mentioned subject.
  • an information processing system includes a reference vector storage unit that holds a group of reference vectors, an instance selection unit that selects one instance from a bag including a plurality of instances, and the reference vector.
  • Reference vector specifying means for specifying the related reference vector most relevant to the selected instance from the reference vector group stored in the storage means, and instance probability calculating means for calculating the instance correct probability that the category of the instance is correct
  • Bag probability calculation means for calculating a bag correct probability that the bag category is correct using the probability that the instance category included in the bag is correct, and the related reference using the bag correct probability
  • a reference vector correcting means for correcting the vector It is characterized in.
  • a recognition dictionary learning method selects one instance from a bag including a plurality of instances, and selects the instance selected from a reference vector group stored in a reference vector storage means.
  • a related reference vector that is most relevant to the category calculates an instance correct probability that the selected category of the instance is correct, and uses the probability that the category of the instance included in the bag is correct,
  • a bag correct probability that is a correct answer is calculated, and the related reference vector is corrected using the bag correct probability.
  • an information processing program includes a computer that includes an instance selection unit that selects one instance from a bag including a plurality of instances, and a reference vector group stored in a reference vector storage unit.
  • a reference vector specifying means for specifying a related reference vector most relevant to the selected instance
  • an instance probability calculating means for calculating an instance correct probability that the category of the selected instance is correct
  • Bag probability calculation means for calculating a bag correct probability that the bag category is correct using the probability that the category of the instance is correct, and a reference vector correction that corrects the related reference vector using the bag correct probability It is characterized by operating as a means.
  • a recognition dictionary can be created even when the category to which the input vector used for learning belongs is not completely known.
  • the information processing system 100 is a device for learning a recognition dictionary used in pattern recognition.
  • the information processing system 100 includes a reference vector storage unit 101, an instance selection unit 102, a reference vector identification unit 103, an instance probability calculation unit 104, a bag probability calculation 105, and a reference vector correction unit 106.
  • the reference vector storage unit 101 holds a reference vector group.
  • the instance selection unit 102 selects one instance from the bag including a plurality of instances. Further, the reference vector specifying unit 103 specifies a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage unit 101.
  • the instance probability calculation unit 104 calculates an instance correct probability that the instance category is correct. Then, the bag probability calculation unit 105 calculates the bag correct probability that the category of the bag is correct using the probability that the category of the instance included in the bag is correct.
  • the reference vector correction unit 106 corrects the related reference vector using the bag correct answer probability.
  • FIG. 2 illustrates the concept of bags and instances.
  • the category of the bag is defined as “positive”. If only a negative instance exists in the bag, the category of the bag is defined as “negative”. Based on the definition of this category, let pij be the probability that the j-th instance in the i-th bag is positive. At this time, the probability that one or more instances in the bag are positive, that is, the probability that the bag is positive, pi, is calculated by the following equation. For example, when classifying into two categories of “positive” and “negative (non-positive)”, even if there is incomplete information that any one of n sets of instances is “positive”, the information The processing system 400 cannot learn well.
  • the information processing system 400 cannot apply information of a category that is not complete to the learning process. That is, before the information processing system 400 performs the learning process, the information regarding the category needs to be completely given to all the input vectors used for learning. However, this work takes time and effort. In addition, when it is difficult to accurately define the category of the input vector, it is difficult to completely give information on the category in the first place.
  • the information processing system 400 calculates the probability that the instance is the correct category, and further calculates the probability that the bag is the correct category.
  • FIG. 3 schematically represents a recognition method using a recognition dictionary by the information processing system 400 of the present embodiment.
  • the recognition dictionary includes reference vectors 311 to 31n belonging to a certain category A and reference vectors 321 to 32n belonging to another category B.
  • the information processing system 400 identifies the instance 303 and the closest reference vector 312 among the reference vectors 311 to 31n belonging to the category A.
  • the information processing system 400 calculates the distance d1. Further, the information processing system 400 specifies the instance 303 and the closest reference vector 322 among the reference vectors 321 to 32n belonging to the category B. Then, the information processing system 400 calculates the distance d2. Then, the information processing system 400 determines that the instance 303 also belongs to the category to which the closest reference vector 312 belongs. In the case of FIG. 3, since d1 ⁇ d2, the information processing system 400 determines that the instance 303 belongs to category A.
  • FIG. 4 is a diagram for explaining a functional configuration of the information processing system 400 according to the present embodiment.
  • the information processing system 400 calculates a distance between a group of reference vectors (also referred to as templates, prototypes, and representative vectors) called a recognition dictionary and an input vector. Based on the calculation result, the information processing system 400 sets the category (also referred to as a class or label) to which the reference vector closest to the input vector belongs as the recognition result of the input vector.
  • a group of reference vectors also referred to as templates, prototypes, and representative vectors
  • the information processing system 400 sets the category (also referred to as a class or label) to which the reference vector closest to the input vector belongs as the recognition result of the input vector.
  • x a group of reference vectors
  • wk k is 1 to K
  • each is a vector, but the bar is omitted here for simplification.
  • the present embodiment includes a data processing device 401 and a storage device 402.
  • the data processing device 401 includes a bag selection unit 411, an instance selection unit 412, a reference vector identification unit 413, an instance probability calculation unit 414, a bag probability calculation unit 415, a probability base correction coefficient calculation unit 416, and a reference vector.
  • a correction unit 417 and an end determination unit 418 are included.
  • the storage device 402 includes a learning data storage unit 421 and a reference vector storage unit 422.
  • the learning data storage unit 421 stores information related to learning instance groups (bags).
  • the reference vector storage unit 422 stores information on a reference vector group that is a recognition dictionary. Specifically, the reference vector storage unit 422 stores individual reference vectors and category information to which the reference vectors belong.
  • the bag selection unit 411 selects one bag i from all bags used for learning.
  • the instance selection unit 412 selects one instance (one input vector) j from the bag i selected by the bag selection unit 411.
  • the reference vector specifying unit 413 determines a reference vector pair related to the instance j selected by the instance selecting unit 412. Specifically, the reference vector specifying unit 413 is closest to the reference vector group in the same category as the instance j from the inter-vector distance between the instance j and each reference vector held in the reference vector storage unit 422.
  • the first reference vector W1 and its first distance d1 are calculated.
  • the reference vector specifying unit 413 further calculates the second reference vector W2 closest to the reference vector group in a category different from the instance j and the second distance d2. Note that it is sufficient for the reference vector specifying unit 413 to search for the closest reference vector. Therefore, the reference vector specifying unit 413 does not necessarily calculate the distances from all the reference vectors.
  • the instance probability calculation unit 414 estimates a probability pij that the instance is a correct category.
  • the correct category is a category assigned to the bag i to which the instance j belongs.
  • any method may be used as a method of estimating the probability pij.
  • the probability pij can be expressed as follows.
  • the probability pij is also written as the following equation.
  • ⁇ ij can be obtained based on the following equation.
  • ⁇ ij is written by Hiroyoshi Miyano and Nagaki Ishidera, “Learning Vector Quantization Method Using Heavy Distribution Function of Support”, IEICE, IEICE Technical Report, vol. 110, no. 187, PRMU 2010-81, pp. 185-192, September 2010, ⁇ k may be used.
  • the monotone increasing function R ( ⁇ ) may be, for example, the following expression.
  • the monotonically increasing function R ( ⁇ ) may be expressed by the following equation using the reference vector learning frequency t. Furthermore, the monotonically increasing function R ( ⁇ ) may be the following equation using the reference vector learning count t and arbitrary constants ⁇ 0 and ⁇ 1.
  • the bag probability calculation unit 415 calculates a probability qi that the bag is likely to be the correct category from the probability pij that each instance in the bag i is likely to be the correct category. This calculation formula is expressed by the following formula.
  • the probability-based correction coefficient calculation unit 416 calculates, for each instance, a probability-based correction coefficient uij for updating the reference vector based on the probability pij that the instance is a correct category and the probability qi that the bag to which the instance belongs is the correct category. calculate.
  • the probability-based correction coefficient calculation unit 416 may obtain the correction coefficient uij so as to maximize the likelihood L, which means the likelihood of the recognition result.
  • the likelihood L may be defined based on the probability qi. For example, assume that L is defined by the following equation. In this case, the probability-based correction coefficient calculation unit 416 may calculate the probability-based correction coefficient uij using the following formula.
  • the reference vector correction unit 417 corrects the first reference vector w1 and the second reference vector w2 stored in the reference vector storage unit 422 based on the probability base correction coefficient uij calculated by the probability base correction coefficient calculation unit 416. To do. More specifically, the reference vector correction unit 417 calculates a value calculated from the function D2 ( ⁇ ) that converts the second distance d2, a probability-based correction coefficient uij, an arbitrary coefficient ⁇ 1, and an instance x and w1. A vector obtained by multiplying the difference vector is added (or subtracted) to w1 as a vector. By this process, the reference vector correction unit 417 corrects the first reference vector w1.
  • the reference vector correction unit 417 calculates a value calculated from the function D1 (•) that converts the first distance d1, a probability-based correction coefficient uij, and an arbitrary coefficient ⁇ 2, and the difference between the instances x and w2. A vector obtained by multiplying the vector is added (or subtracted) as a vector to w2. By this process, the reference vector correction unit 417 corrects the second reference vector w2.
  • the function D2 (•) includes a power calculation of d2 and a calculation that further divides by the power of the sum of d1 and d2.
  • the function D1 (•) is composed of a power calculation of d1 and a calculation for further dividing by the power of the sum of d1 and d2.
  • the power here refers to the calculation of the y power using an arbitrary real number y of 0 or more.
  • D1 (•) and D2 (•) may be functions represented by the following expressions.
  • the end determination unit 418 determines the end of learning (correction) of the reference vector. (Operation) Next, the overall operation of the present embodiment will be described in detail with reference to the flowchart of FIG. In the following description, the number of bags used for learning is M. The number of instances in the bag i is Ni.
  • an instance (input vector) used for learning is prepared, and a bag composed of one or more instances is prepared.
  • a category is assigned to all bags.
  • the bag selection unit 411 selects one bag i from the learning bag group stored in the learning data storage unit 421 (S501).
  • the instance selection unit 412 selects one instance in the selected bag (S503).
  • the reference vector specifying unit 413 searches for the reference vector closest to the selected instance (S505).
  • the instance probability calculation unit 414 calculates a probability Pij that the instance is correct (S507).
  • the instance selection unit 412 determines whether or not the processing in steps S503 to S507 has been completed for all instances in the bag (S509), and if not, the processing returns to step S503.
  • the information processing apparatus 400 includes the instance selection process (S503) by the instance selection unit 412, the reference vector specification process (S505) by the reference vector specification unit 413, and the instance probability calculation process (S507) by the instance probability calculation unit 414. Is repeatedly executed.
  • the bag probability calculation unit 415 calculates a probability qi that the selected bag i is correct (step S511). The instance selection unit 412 selects one instance again from the bag.
  • the probability-based correction coefficient calculation unit 416 calculates a probability-based correction coefficient for the selected instance (S513). Further, the reference vector correction unit 417 uses the first reference vector w1 and the second reference vector w2 stored in the reference vector storage unit 422 based on the probability base correction coefficient uij calculated by the probability base correction coefficient calculation unit 415. Is corrected (S515). Next, the instance selection unit 412 determines whether or not the processing of steps S512 to S515 has been completed for all instances in the bag (S516), and if not, returns to step S503. The information processing apparatus 400 repeatedly executes the processes in steps S511 to S515 for all instances in the selected bag i.
  • the information processing apparatus 400 corrects all reference vectors stored in the reference vector storage unit 422.
  • the bag selection unit 411 determines whether or not the processing of steps S501 to S516 has been completed for all bags (S517), and if not, the processing returns to step S501.
  • the information processing apparatus 400 repeatedly executes the processes in steps S501 to S517 for all bags.
  • the information processing apparatus 400 corrects all reference vectors stored in the reference vector storage unit 422.
  • the end determination unit 418 determines whether or not to end the learning of the reference vector. If the end determination unit 418 determines not to end, the process returns to step S501.
  • the information processing apparatus 400 repeatedly executes the processing from step S501 to step S517 for all learning bags.
  • the end determination unit 418 determines to end the learning of the reference vector
  • the information processing apparatus 400 ends the process.
  • the end determination unit 418 may determine to end the learning of the reference vector when the information processing apparatus 400 repeatedly executes the above-described process a predetermined number of times.
  • the bag probability calculation unit 415 calculates the probability that the bag is the correct category.
  • the reference vector correction unit 417 corrects the reference vector so as to maximize the likelihood L derived therefrom. Therefore, the information processing apparatus 400 can perform learning only with the information of the category assigned to the bag, not the instance.
  • the likelihood L derived from the probability that the bag is the correct category can be maximized by using the probability base correction coefficient uij calculated by the probability base correction coefficient calculation unit 416 used in the present embodiment.
  • the function S to be minimized can be defined as follows using a logarithmic function that is a monotone function.
  • the update formula of the reference vector wk is expressed as follows using a constant ⁇ .
  • the following expression is an expression obtained by expanding the second term on the right side of the above expression. Further expansion of the above expression leads to the following expression 4 and expression 5 using the constant ⁇ .
  • Equation 4 corresponds to a probability-based correction coefficient.
  • Equation 3 is derived from Equation 4 and Equation 5. Since the following expression described above can be interpreted as an expression for normalizing the magnitude of the distance value, the same effect can be obtained even if the numerator and the denominator are raised.
  • the operation of the present embodiment is described using “probability of being correct”. In the case of two categories of “positive” and “negative (non-positive)”, the “probability of incorrect answer” or “probability of positive” or “probability of negative” is obvious. . Therefore, it can be said that this embodiment uses these probabilities.
  • the present embodiment since the reference vector is corrected based on the probability that the bag is correct, a learning process for maximizing the correctness (evaluation scale) of the entire bag to be learned is performed. As a result, even if there is no category information corresponding to all the input vectors used for learning, the present embodiment learns only from the category information corresponding to the set (bag) of input vectors (instances) used for learning. it can. Therefore, this embodiment can create a recognition dictionary. In the present embodiment, the preparation for the learning process can be completed only by adding coarser information before the learning process. In addition, the present embodiment can perform learning even when it is difficult to accurately define the category of each input vector in the first place or when the category can be defined only for a set of input vectors.
  • the present embodiment can be applied to a use for detecting a specific object from a video or a use for identifying and authenticating a person or object from a video.
  • a use for detecting a specific object from a video or a use for identifying and authenticating a person or object from a video.
  • the probability base correction coefficient calculation unit 416 according to the second embodiment may calculate the probability base correction coefficient uij by the following formula. As long as a category is assigned to a bag that is a set of instances, this embodiment can be used.
  • a vector generated by minutely changing a value of a vector component of an instance to be learned may be created.
  • input vector a vector generated by minutely changing a value of a vector component of an instance to be learned
  • a set in which a plurality of instances that are originally generated from a single instance with a slight variation may be prepared as one bag.
  • the instance is an instance including noise, for example, if one of the instances generated from the instance is reduced in noise, the present embodiment It is possible to execute a learning process that is hardly affected.
  • one bag only needs to be configured by that one instance. In the present embodiment, the same effect as in the second embodiment can be obtained.
  • the present invention may be applied to a single device. Furthermore, the present invention can also be applied to a case where an information processing program that implements the functions of the embodiments is supplied directly or remotely to a system or apparatus. Therefore, in order to realize the functions of the present invention with a computer, a program installed in the computer, a medium storing the program, and a WWW (World Wide Web) server that downloads the program are also included in the scope of the present invention. .
  • the information processing system 100, the information processing system 400, the data processing device 401, and the storage device 402 are respectively a computer and a program that controls the computer, dedicated hardware, or a program that controls the computer and the computer and dedicated hardware. It can be realized by a combination.
  • the bag probability calculation unit 415, the probability base correction count calculation unit 416, the reference vector correction unit 417, and the end determination unit 418 are, for example, for realizing the function of each unit read into the memory from the recording medium storing the program. It can be realized by a dedicated program and a processor that executes the program.
  • the reference vector storage unit 101, the learning data storage unit 421, and the reference vector storage unit 422 can be realized by a memory or a hard disk device included in the computer.
  • the reference vector storage unit 101, the instance selection unit 102, the reference vector specification unit 103, the instance probability calculation unit 104, the bag probability calculation unit 105, the reference vector correction unit 106, the bag selection unit 411, the instance selection unit 412, the reference vector specification Unit 413, instance probability calculation unit 414, bag probability calculation unit 415, probability-based correction count calculation unit 416, reference vector correction unit 417, end determination unit 418, learning data storage unit 421, part of reference vector storage unit 422 or All can be realized by a dedicated circuit for realizing the function of each unit.
  • (Appendix 1) Reference vector storage means for holding a reference vector group; An instance selection means for selecting one instance from a bag including a plurality of instances; Reference vector specifying means for specifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means; Instance probability calculating means for calculating an instance correct probability that the category of the instance is correct; Bag probability calculation means for calculating a bag correct probability that the category of the bag is correct using a probability that the category of the instance included in the bag is correct; Reference vector correcting means for correcting the related reference vector using the bag correct answer probability; An information processing system comprising: (Appendix 2) The reference vector specifying means has a first reference vector that is the closest to the instance in the same category as the bag to which the selected instance belongs, and a distance from the instance that is the closest to a category different from the bag to which the instance belongs.
  • the information processing system according to appendix 1, wherein a near second reference vector is specified as the related reference vector.
  • the reference vector specifying means calculates a first distance indicating a distance between the instance and the first reference vector, and a second distance indicating a distance between the instance and the second reference vector;
  • the information processing system according to supplementary note 2, wherein the reference vector correcting unit corrects the related reference vector using the first distance and the second distance.
  • the reference vector correcting means includes Supplementary note 3 wherein the related reference vector is corrected using a conversion distance value obtained by conversion by a power calculation of the first distance and a calculation by dividing by a power of the sum of the first distance and the second distance. Information processing system described in 1.
  • the instance probability calculating means calculates the instance correct probability by dividing a difference between the first distance and the second distance by a sum of the first distance and the second distance.
  • the information processing system according to appendix 3 or 4.
  • Probability-based correction coefficient calculating means for calculating a probability-based correction coefficient for correcting the related reference vector using the bag correct answer probability, The information processing system according to any one of appendices 1 to 5, wherein the reference vector correcting unit corrects the reference vector based on a probability-based correction coefficient.
  • the probability-based correction coefficient calculating means is configured to multiply the instance correct answer probability by the bag incorrect answer probability as the probability that the bag to which the instance belongs is an incorrect answer category, and the instance incorrect answer probability as the probability that the instance is an incorrect answer category. 7. The information processing system according to claim 6, wherein a probability-based correction coefficient is calculated by multiplying the bag correct answer probability.
  • the probability-based correction coefficient calculating means calculates a probability-based correction coefficient based on a value obtained by dividing the product of the instance correct answer probability and the bag incorrect answer probability by the bag correct answer probability, and the instance incorrect answer probability. The information processing system according to 6 or 7.
  • the supplementary note 1 according to any one of supplementary notes 1 to 8, wherein the bag probability calculating means calculates a probability that a category of at least one instance among the instances included in the bag is correct.
  • the information processing system in any one of thru
  • a recognition dictionary learning method characterized by that.
  • An instance selection means for selecting one instance from a bag including a plurality of instances; Reference vector specifying means for specifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means; An instance probability calculating means for calculating an instance correct probability that the category of the selected instance is correct; Bag probability calculation means for calculating a bag correct probability that the category of the bag is correct using a probability that the category of the instance included in the bag is correct; Reference vector correcting means for correcting the related reference vector using the bag correct answer probability; Information processing program. While the present invention has been described with reference to the embodiments, the present invention is not limited to the above embodiments.

Abstract

[Problem] To create a recognition dictionary, even when categories to which input vectors to be used for learning belong is not completely known. [Solution] An information processing system is characterized in being provided with: a reference vector storage means for retaining reference vectors; an instance selecting means for selecting one instance from a bag containing a plurality of instances; a reference vector specifying means for specifying a related reference vector that is most related to the selected instance, from among the reference vectors stored in the reference vector storage means; an instance probability calculating means for calculating an instance-correct probability that is the probability that the category of the instance is correct; a bag probability calculating means for calculating a bag-correct probability that is the probability that the category of the bag is correct; and a reference vector modifying means for modifying the related reference vector using the bag correct probability.

Description

情報処理システム、認識辞書学習方法および情報処理プログラムInformation processing system, recognition dictionary learning method, and information processing program
 本発明は、情報処理システム、認識辞書学習方法および情報処理プログラムに関し、特に、パターン認識で用いられる、情報処理システム、認識辞書学習方法および情報処理プログラムに関する。 The present invention relates to an information processing system, a recognition dictionary learning method, and an information processing program, and more particularly to an information processing system, a recognition dictionary learning method, and an information processing program used in pattern recognition.
 入力ベクトルの認識は、認識辞書と呼ばれる参照ベクトル群と入力ベクトルとの距離計算に基づいて、入力ベクトルに最も近い参照ベクトルが属するカテゴリを、その入力ベクトルの認識結果とすることにより行われることがある。この場合、参照ベクトルの値によって認識精度が変化する。そのため、どのように参照ベクトルの値を設定するかが認識精度を上げる上で重要である。
 特許文献3及び非特許文献1には、参照ベクトルを用いた認識辞書学習手法として、Learning Vector Quantization(LVQ)が開示されている。LVQによれば、パーセプトロン型ニューラルネットワークやサポートベクトルマシンといった他の参照ベクトルを使わない統計的パターン認識手法と比較して、非常に短時間で学習処理が完了する。また、LVQによれば、参照ベクトルの数を自由に設計できるため、その数を削減するだけで学習処理と認識処理の双方を高速化できる。さらに、認識処理が距離計算処理だけで実現できかつ並列型の計算にすることも容易である。そのため、ソフトウェアのプログラミングやICチップ上で実現するのが容易である。
 特許文献1は、このLVQを改良した手法の一例を開示している。また非特許文献2は、特許文献1に記載された方式の1種について具体的な効果を開示している。
 また、非特許文献3は、これらの手法の改良方式を開示している。ここでは、複数の入力ベクトルを学習データとして用いて自動的に参照ベクトルの値が設定される。
Recognition of an input vector may be performed by setting a category to which a reference vector closest to the input vector belongs as a recognition result of the input vector based on a distance calculation between a reference vector group called a recognition dictionary and the input vector. is there. In this case, the recognition accuracy varies depending on the value of the reference vector. Therefore, how to set the value of the reference vector is important for improving the recognition accuracy.
Patent Document 3 and Non-Patent Document 1 disclose Learning Vector Quantization (LVQ) as a recognition dictionary learning method using a reference vector. According to LVQ, the learning process is completed in a very short time compared to a statistical pattern recognition method that does not use other reference vectors such as a perceptron type neural network or a support vector machine. In addition, according to LVQ, the number of reference vectors can be freely designed, so that both the learning process and the recognition process can be accelerated only by reducing the number of reference vectors. Furthermore, the recognition process can be realized only by the distance calculation process, and it is easy to make a parallel type calculation. Therefore, it is easy to implement software programming or an IC chip.
Patent Document 1 discloses an example of a technique improved from this LVQ. Non-Patent Document 2 discloses specific effects for one of the methods described in Patent Document 1.
Non-Patent Document 3 discloses an improved method of these methods. Here, the reference vector value is automatically set using a plurality of input vectors as learning data.
特許第3452160号公報Japanese Patent No. 3452160 特開平5−124550号公報JP-A-5-124550 特開平6−333052号公報JP-A-6-333052
 特許文献1、特許文献3、非特許文献1、非特許文献2、非特許文献3に記載の方式では、学習に用いるすべての入力ベクトルについて、それが属するカテゴリを正しく付与しておく必要がある。しかし、このように事前に正しいカテゴリを付与するには手間や時間がかかる。また、そもそも事前に正しいカテゴリを付与できないこともある。
 そして、特許文献1、特許文献3、非特許文献1、非特許文献2、非特許文献3の技術では、学習に用いる入力ベクトルが属するカテゴリについて、事前に用意された情報が不完全な場合、その情報から認識辞書を作ることができなかった。
 本発明の目的は、上述の課題を解決する情報処理システムを提供することにある。
In the methods described in Patent Document 1, Patent Document 3, Non-Patent Document 1, Non-Patent Document 2, and Non-Patent Document 3, it is necessary to correctly assign the category to which all input vectors used for learning belong. . However, it takes time and effort to assign the correct category in advance. Also, there are cases where correct categories cannot be assigned in advance.
And in the technique of patent document 1, patent document 3, non-patent document 1, non-patent document 2, non-patent document 3, when the information prepared in advance is incomplete for the category to which the input vector used for learning belongs, We couldn't make a recognition dictionary from that information.
The objective of this invention is providing the information processing system which solves the above-mentioned subject.
 上記目的を達成するため、本発明に係る情報処理システムは、参照ベクトル群を保持する参照ベクトル記憶手段と、複数のインスタンスを含むバッグ内から1つのインスタンスを選択するインスタンス選択手段と、前記参照ベクトル記憶手段に記憶された参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定する参照ベクトル特定手段と、前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算するインスタンス確率算出手段と、前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算するバッグ確率算出手段と、前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する参照ベクトル修正手段と、を備えたことを特徴とする。
 上記目的を達成するため、本発明に係る認識辞書学習方法は、複数のインスタンスを含むバッグ内から1つのインスタンスを選択し、参照ベクトル記憶手段に記憶された参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定し、選択された前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算し、前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算し、前記バッグ正解確率を用いて、前記関連参照ベクトルを修正することを特徴とする。
 上記目的を達成するため、本発明に係る情報処理プログラムは、コンピュータを、複数のインスタンスを含むバッグ内から1つのインスタンスを選択するインスタンス選択手段と、参照ベクトル記憶手段に記憶された参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定する参照ベクトル特定手段と、選択された前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算するインスタンス確率算出手段と、前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算するバッグ確率算出手段と、前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する参照ベクトル修正手段として動作させることを特徴とする。
In order to achieve the above object, an information processing system according to the present invention includes a reference vector storage unit that holds a group of reference vectors, an instance selection unit that selects one instance from a bag including a plurality of instances, and the reference vector. Reference vector specifying means for specifying the related reference vector most relevant to the selected instance from the reference vector group stored in the storage means, and instance probability calculating means for calculating the instance correct probability that the category of the instance is correct Bag probability calculation means for calculating a bag correct probability that the bag category is correct using the probability that the instance category included in the bag is correct, and the related reference using the bag correct probability A reference vector correcting means for correcting the vector, It is characterized in.
To achieve the above object, a recognition dictionary learning method according to the present invention selects one instance from a bag including a plurality of instances, and selects the instance selected from a reference vector group stored in a reference vector storage means. A related reference vector that is most relevant to the category, calculates an instance correct probability that the selected category of the instance is correct, and uses the probability that the category of the instance included in the bag is correct, A bag correct probability that is a correct answer is calculated, and the related reference vector is corrected using the bag correct probability.
In order to achieve the above object, an information processing program according to the present invention includes a computer that includes an instance selection unit that selects one instance from a bag including a plurality of instances, and a reference vector group stored in a reference vector storage unit. Included in the bag; a reference vector specifying means for specifying a related reference vector most relevant to the selected instance; an instance probability calculating means for calculating an instance correct probability that the category of the selected instance is correct; Bag probability calculation means for calculating a bag correct probability that the bag category is correct using the probability that the category of the instance is correct, and a reference vector correction that corrects the related reference vector using the bag correct probability It is characterized by operating as a means.
 本発明によれば、学習に用いる入力ベクトルが属するカテゴリが完全にはわかっていない場合にも、認識辞書を作ることができる。 According to the present invention, a recognition dictionary can be created even when the category to which the input vector used for learning belongs is not completely known.
本発明の第1実施形態の構成を示すブロック図である。It is a block diagram which shows the structure of 1st Embodiment of this invention. バッグと、バッグに付与するカテゴリの考え方を説明するための図である。It is a figure for demonstrating the view of the category given to a bag and a bag. バッグ内のインスタンスのカテゴリ判定について説明するための図である。It is a figure for demonstrating the category determination of the instance in a bag. 本発明の第2実施形態の構成を示すブロック図である。It is a block diagram which shows the structure of 2nd Embodiment of this invention. 本発明の第2実施形態の動作を示すフローチャートである。It is a flowchart which shows operation | movement of 2nd Embodiment of this invention.
 以下に、図面を参照して、本発明の実施の形態について例示的に詳しく説明する。ただし、以下の実施の形態に記載されている構成要素は、あくまで例示であり、本発明の技術範囲をそれらのみに限定する趣旨のものではない。
 [第1実施形態]
 本発明の第1実施形態としての情報処理システム100について、図1を用いて説明する。情報処理システム100は、パターン認識で用いられる認識辞書の学習のための装置である。
 図1に示すように、情報処理システム100は、参照ベクトル記憶部101と、インスタンス選択部102と、参照ベクトル特定部103、インスタンス確率算出部104、バッグ確率算出105、参照ベクトル修正部106とを含む。
 参照ベクトル記憶部101は、参照ベクトル群を保持する。インスタンス選択部102は、複数のインスタンスを含むバッグ内から1つのインスタンスを選択する。
 さらに、参照ベクトル特定部103は、参照ベクトル記憶部101に記憶された参照ベクトル群から、選択されたインスタンスに最も関連する関連参照ベクトルを特定する。インスタンス確率算出部104は、インスタンスのカテゴリが正解であるインスタンス正解確率を計算する。そしてバッグ確率算出部105は、バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算する。参照ベクトル修正部106は、バッグ正解確率を用いて、関連参照ベクトルを修正する。
 以上の構成により、学習に用いる入力ベクトルが属するカテゴリが完全にはわかっていない場合にも、認識辞書を作ることができる。
 [第2実施形態]
 (前提技術)
 本発明の第2実施形態として、参照ベクトルを用いた認識辞書学習手法(LVQ)を前提とした情報処理システム400について説明する。まず、具体的な実施形態の説明に入る前に、統計的パターン認識の分野で、カテゴリの付与の仕方のルールを規定するMultiple Instance Learningと呼ばれる概念について説明する。この概念では、入力ベクトルは「インスタンス」と呼ばれる。また、インスタンスの集合は「バッグ」と呼ばれる。さらに、すべてのインスタンス1つずつにカテゴリが付与されるのではなく、バッグにのみカテゴリが付与される。
 図2は、バッグとインスタンスの概念を図示したものである。カテゴリとして「ポジティブ」と「ネガティブ」の2種類のみがある場合、バッグの中にポジティブインスタンスが1つでも存在すれば、そのバッグのカテゴリは「ポジティブ」定義される。また、バッグの中にネガティブインスタンスのみが存在すれば、そのバッグのカテゴリは「ネガティブ」と定義される。
 このカテゴリの定義に基づき、i番目のバッグ内のj番目のインスタンスがポジティブである確率をpijとする。このとき、バッグ内の1つ以上のインスタンスがポジティブである確率つまりバッグがポジティブである確率であるpiは、以下の数式によって計算される。
Figure JPOXMLDOC01-appb-M000001
 たとえば、「ポジティブ」「ネガティブ(非ポジティブ)」という2種類のカテゴリに分類する場合、インスタンスn個の集合のうちどれか1個が「ポジティブ」であるという不完全な情報があっても、情報処理システム400はうまく学習できない。参照ベクトルを更新する係数を求めるには、学習に用いるすべてのインスタンスとそれぞれのインスタンスが属するカテゴリが、1対1に対応付けられている必要がある。そのため、情報処理システム400は、完全でないカテゴリの情報があっても、それを学習処理に適用できない。
 つまり、情報処理システム400が学習処理をする前に、学習に用いる入力ベクトルすべてに対して、カテゴリに関する情報が完全に付与されている必要がある。しかし、このための作業には手間や時間がかかってしまう。また、入力ベクトルのカテゴリを正確に定義することが難しい場合、そもそもカテゴリに関する情報を完全に付与しておくことは困難である。たとえば、入力ベクトルが画像パターンであり、カテゴリが「顔」と「顔でない」という2種類のカテゴリである場合、顔の目・鼻・口だけを含むパターンや、顔の輪郭だけでなく首や肩まで含まれているパターンのカテゴリを「顔」と定義すべきか「顔でない」と定義すべきか判断が難しい。顔の目・鼻・口だけを含むパターンと、顔の輪郭まで含まれているパターンと、首や肩まで含まれているパターンのまとまりに対して、「少なくともどれか1つが顔」であると定義するのが自然な考え方といえる。
 このため、本実施形態では、情報処理システム400は、インスタンスが正解カテゴリである確率を計算し、さらにそのバッグが正解カテゴリである確率を計算する。そして、情報処理システム400は、インスタンスが属するバッグが正解カテゴリである確率を使って、参照ベクトルを修正する。このようにすることで、インスタンスが属するカテゴリが完全に明確ではなくても、情報処理システム400は、学習処理(つまり参照ベクトルの修正)を行なうことができる。
 (構成)
 図3は、本実施形態の情報処理システム400による認識辞書を用いた認識手法を模式的に表現したものである。認識辞書は、あるカテゴリAに属する参照ベクトル311~31nと別のカテゴリBに属する参照ベクトル321~32nを含んでいる。カテゴリを認識処理したいインスタンス303がある場合、情報処理システム400は、カテゴリAに属する参照ベクトル311~31nのなかで、インスタンス303と最近接の参照ベクトル312を特定する。そして、情報処理システム400は、その距離d1を計算する。また、情報処理システム400は、カテゴリBに属する参照ベクトル321~32nのなかで、インスタンス303と最近接の参照ベクトル322を特定する。そして、情報処理システム400は、その距離d2を計算する。そして、情報処理システム400は、最も近い参照ベクトル312が属するカテゴリに、インスタンス303も属すると判断する。図3の場合、d1<d2であるため、情報処理システム400は、インスタンス303はカテゴリAに属すると判断する。
 次に本実施形態に係る情報処理システム400について、図4を用いて説明する。図4は、本実施形態に係る情報処理システム400の機能構成を説明するための図である。情報処理システム400は、認識辞書と呼ばれる参照ベクトル(テンプレート、プロトタイプ、代表ベクトルとも呼ばれる)群と入力ベクトルとの距離を計算する。そして、情報処理システム400は、その計算結果に基づいて、入力ベクトルに最も近い参照ベクトルが属するカテゴリ(クラス、ラベルとも呼ばれる)を、その入力ベクトルの認識結果とする。
 以下では、説明のために、1つのインスタンスをx、参照ベクトルをwk(kは1~K)と表記する(それぞれベクトルではあるが、ここでは簡略化のためバーを省略する)。
 図4を参照すると、本実施形態は、データ処理装置401と、記憶装置402とを備える。データ処理装置401は、バッグ選択部411と、インスタンス選択部412と、参照ベクトル特定部413と、インスタンス確率算出部414と、バッグ確率算出部415と、確率ベース修正係数算出部416と、参照ベクトル修正部417と、終了判定部418とを有する。
 記憶装置402は、学習用データ記憶部421と参照ベクトル記憶部422を有する。学習用データ記憶部421は、学習用のインスタンス群(バッグ)に関する情報を記憶している。参照ベクトル記憶部422は、認識辞書である参照ベクトル群の情報を記憶している。具体的には、参照ベクトル記憶部422は、個々の参照ベクトルと、それが属するカテゴリ情報を記憶する。なお、本発明の動作中に参照ベクトルが修正され、その結果、学習された参照ベクトルが得られる。
 バッグ選択部411は、学習に用いる全バッグから1つのバッグiを選択する。インスタンス選択部412は、バッグ選択部411で選択されたバッグiの中から1つのインスタンス(1つの入力ベクトル)jを選択する。
 参照ベクトル特定部413は、インスタンス選択部412で選択されたインスタンスjに関連する参照ベクトル対を決定する。具体的には、参照ベクトル特定部413は、インスタンスjと、参照ベクトル記憶部422に保持される個々の参照ベクトルとのベクトル間距離から、インスタンスjと同じカテゴリの参照ベクトル群の中で最も近い第1の参照ベクトルW1とその第1の距離d1を算出する。また、参照ベクトル特定部413は、さらに、インスタンスjと異なるカテゴリの参照ベクトル群の中で最も近い第2の参照ベクトルW2とその第2の距離d2を算出する。なお、参照ベクトル特定部413は、もっとも近い参照ベクトルを検索できれば十分である。そのため、参照ベクトル特定部413は、必ずしもすべての参照ベクトルとの距離を計算する必要はない。
 インスタンス確率算出部414は、インスタンスが正解のカテゴリらしい確率pijを推定する。ここで、正解のカテゴリは、インスタンスjが属するバッグiに付与されたカテゴリのことである。なお、この確率pijを推定する方式は任意の方式で構わない。以下、確率pijを推定する方法の1例を説明する。確率pijは、以下のように表現できる。
Figure JPOXMLDOC01-appb-M000002
 一方、インスタンスの正解カテゴリらしさをμij、値域が[0,1]の任意の単調増加関数をR(−)とすると、確率pijは、以下の式のようにも書かれる。
Figure JPOXMLDOC01-appb-M000003
 μijについては様々な導出方式が考えられる。たとえば以下の式に基づいてμijを求めることができる。
Figure JPOXMLDOC01-appb-M000004
 他にも、μijは、宮野博義、石寺永記著「裾野の重たい分布関数を用いた学習ベクトル量子化手法」電子情報通信学会、信学技報、vol.110、no.187、PRMU2010−81、pp.185−192、2010年9月に記載されている、ηkであってもよい。
 また、単調増加関数R(−)は、たとえば、以下の式であればよい。
Figure JPOXMLDOC01-appb-M000005
 また、参照ベクトルの学習を繰り返すほど認識の信頼性があがることを考慮すると、単調増加関数R(−)は、参照ベクトルの学習回数tを用いた以下の式であってもよい。
Figure JPOXMLDOC01-appb-M000006
 さらには、単調増加関数R(−)は、参照ベクトルの学習回数tと任意の定数γ0とγ1とを用いた以下の式であってもよい。
Figure JPOXMLDOC01-appb-M000007
 バッグ確率算出部415は、バッグiの中の個々のインスタンスが正解カテゴリらしい確率pijから、バッグが正解カテゴリらしい確率qiを算出する。この計算式は以下の式で表現される。
Figure JPOXMLDOC01-appb-M000008
 確率ベース修正係数算出部416は、インスタンスが正解カテゴリらしい確率pijとインスタンスが属するバッグが正解カテゴリらしい確率qiとに基づいて、参照ベクトルを更新するための確率ベースの修正係数uijを、インスタンスごとに算出する。
 確率ベース修正係数算出部416は、認識結果の尤もらしさを意味する尤度Lを最大化するように、修正係数uijを求めればよい。尤度Lは、確率qiに基づいて定義されていればよい。
 たとえば、Lが以下の式で定義されているとする。
Figure JPOXMLDOC01-appb-M000009
 この場合、確率ベース修正係数算出部416は、確率ベース修正係数uijを以下の式によって計算すればよい。
Figure JPOXMLDOC01-appb-M000010
 参照ベクトル修正部417は、確率ベース修正係数算出部416が算出した確率ベース修正係数uijに基づいて、参照ベクトル記憶部422に記憶された第1の参照ベクトルw1と第2の参照ベクトルw2を修正する。詳しく説明すると、参照ベクトル修正部417は、第2の距離d2を変換する関数D2(・)から計算される値と、確率ベース修正係数uijと、任意の係数α1と、をインスタンスxとw1との差分ベクトルに乗じて得られるベクトルを、w1にベクトルとして加算(または減算)する。この処理によって、参照ベクトル修正部417は、第1の参照ベクトルw1を修正する。
 一方、参照ベクトル修正部417は、第1の距離d1を変換する関数D1(・)から計算される値と、確率ベース修正係数uijと、任意の係数α2と、をインスタンスxとw2との差分ベクトルに乗じて得られるベクトルを、w2にベクトルとして加算(または減算)する。この処理によって、参照ベクトル修正部417は、第2の参照ベクトルw2を修正する。この修正の計算は、以下の数式3のように表現できる。
Figure JPOXMLDOC01-appb-M000011
 なお、関数D2(・)は、d2のべき乗計算、およびさらにd1とd2の和のべき乗で除算する計算で構成される。関数D1(・)は、d1のべき乗計算、およびさらにd1とd2の和のべき乗で除算する計算で構成される。ここでべき乗とは、0以上の任意の実数yを用いたy乗の計算を指す。たとえば、D1(・)とD2(・)は、以下の式で表される関数であればよい。
Figure JPOXMLDOC01-appb-M000012
 終了判定部418は、参照ベクトルの学習(修正)の終了を判定する。
 (動作)
 次に、図5のフローチャートを参照して本実施の形態の全体の動作について詳細に説明する。以下の説明では、学習に用いるバッグの個数はMである。そして、バッグi内のインスタンスの個数はNiである。なお、動作の前に、学習に用いるインスタンス(入力ベクトル)を用意し、1個以上のインスタンスで構成されるバッグを用意しておく。実際のデータ構造としては、インスタンスにバッグの番号を対応付けておけば十分である。また、すべてのバッグにはカテゴリを付与しておく。
 はじめに、バッグ選択部411が、学習用データ記憶部421に記憶されている学習用のバッグ群から1つのバッグiを選択する(S501)。次に、インスタンス選択部412が、選択されたバッグ内にあるインスタンスを1つ選択する(S503)。そして、参照ベクトル特定部413が、選択されたインスタンスに最も近い参照ベクトルを検索する(S505)。さらに、インスタンス確率算出部414が、そのインスタンスが正解である確率Pijを計算する(S507)。次に、インスタンス選択部412は、バッグ内の全インスタンスについて、ステップS503~S507の処理が終了したかを判定し(S509)、終了していなければ、処理はステップS503に戻る。
 このように、情報処理装置400は、インスタンス選択部412によるインスタンス選択処理(S503)と参照ベクトル特定部413による参照ベクトル特定処理(S505)とインスタンス確率算出部414によるインスタンス確率算出処理(S507)とを繰返し実行する。
 次にバッグ確率算出部415は、選択されたバッグiが正解である確率qiを計算する(ステップS511)。インスタンス選択部412は、バッグ内から再度インスタンスを1つ選択する。そして、確率ベース修正係数算出部416が、選択されたインスタンスに関して、確率ベースの修正係数を算出する(S513)。さらに参照ベクトル修正部417は、確率ベース修正係数算出部415が算出した確率ベース修正係数uijに基づいて、参照ベクトル記憶部422に記憶されている第1の参照ベクトルw1と第2の参照ベクトルw2を修正する(S515)。次に、インスタンス選択部412は、バッグ内の全インスタンスについて、ステップS512~S515の処理が終了したかを判定し(S516)、終了していなければ、ステップS503に戻る。
 情報処理装置400は、これらのステップS511~S515の処理を、選択されたバッグi内にあるすべてのインスタンスについて、繰返し実行する。このことによって、情報処理装置400は、参照ベクトル記憶部422に記憶されているすべての参照ベクトルを修正する。これにより、バッグ単位での学習処理が終了する。
 次に、バッグ選択部411は、全バッグについて、ステップS501~S516の処理が終了したかを判定し(S517)、終了していなければ、処理はステップS501に戻る。
 情報処理装置400は、これらのステップS501~S517の処理を、全てのバッグについて繰返し実行する。このことによって、情報処理装置400は、参照ベクトル記憶部422に記憶されているすべての参照ベクトルを修正する。
 ステップS518では、終了判定部418が、参照ベクトルの学習を終了するか否かを判定する。終了判定部418が終了しないと判定した場合は、処理はステップS501に戻る。そして、情報処理装置400は、ステップS501からステップS517までの処理をすべての学習用のバッグについて繰返し実行する。
 一方、終了判定部418が参照ベクトルの学習を終了すると判定した場合は、情報処理装置400は、処理を終える。終了判定部418は、たとえば、あらかじめ決められた所定の回数だけ情報処理装置400が前述の処理を繰返し実行したら、参照ベクトルの学習を終了するように判定してもよい。
 以上のように本実施の形態では、バッグ確率算出部415が、バッグが正解カテゴリである確率を算出する。そして、参照ベクトル修正部417が、そこから導出される尤もらしさLを最大化するように参照ベクトルを修正する。そのため、情報処理装置400は、インスタンスではなくバッグに付与されたカテゴリの情報のみでも学習を実行することができる。
 次に、本実施形態で用いる確率ベース修正係数算出部416で計算される確率ベース修正係数uijを用いることで、バッグが正解カテゴリである確率から導出される尤もらしさLを最大化できることを説明する。Pijが上述の数式1のように定義され、Lが数式2のように定義される場合に、最急降下法でLを最大化することを考える。
 この場合、最小化すべき関数Sは、単調関数である対数関数を用いて、以下の式のように定義されうる。
Figure JPOXMLDOC01-appb-M000013
 最急降下法では、参照ベクトルwkの更新式は定数αを用いて以下のように表される。
Figure JPOXMLDOC01-appb-M000014
 以下の式は、上の式の右辺第2項を展開した式である。
Figure JPOXMLDOC01-appb-M000015
 上の式をさらに展開することにより、以下の、数式4と、定数βを用いた数式5とが導かれる。
Figure JPOXMLDOC01-appb-M000016
Figure JPOXMLDOC01-appb-M000017
 数式4は確率ベース修正係数に相当する。数式4と数式5とから数式3が導かれる。
 上述した以下の式は、距離値の大きさを正規化するための式であると解釈されうるため、この分子および分母をべき乗しても同様の効果を得ることができる。
Figure JPOXMLDOC01-appb-M000018
 なお、本実施形態の動作は「正解である確率」を用いて記載されている。「ポジティブ」と「ネガティブ(非ポジティブ)」という2カテゴリであれば、「不正解である確率」または「ポジティブである確率」または「ネガティブである確率」から「正解である確率」は自明である。そのため、本実施形態は、これらの確率を用いているともいえる。
 本実施形態は、バッグが正解である確率に基づいて参照ベクトルを修正するので、学習するバッグ全体についての正解らしさ(評価尺度)を最大化する学習処理を行う。これにより、学習に用いるすべての入力ベクトルそれぞれに対応するカテゴリの情報がなくても、本実施形態は、学習に用いる入力ベクトル(インスタンス)の集合(バッグ)に対応するカテゴリの情報だけからでも学習できる。従って、本実施形態は、認識辞書を作ることができる。本実施形態は、学習処理をする前に、より粗い情報を付与するだけで学習処理の準備を完了できる。また、そもそも入力ベクトル個々のカテゴリを正確に定義することが困難な場合や、入力ベクトルの集合のみにしかカテゴリを定義できない場合にも、本実施形態は、学習を行うことができる。
 本実施形態は、映像から特定の物体を検出する用途や、映像から人物や物体を同定し認証する用途に適用できる。特に膨大なデータを学習して認識に利用しようとする場合、物体の正解カテゴリを完全に付与することが困難である。しかし、データの集合に対してのみ正解カテゴリが付与されればよいため、本実施形態は、実用的に適用できる。
 [第3実施形態]
 上記第2実施形態に係る確率ベース修正係数算出部416は、以下の式によって確率ベース修正係数uijを計算してもよい。
Figure JPOXMLDOC01-appb-M000019
 インスタンスの集合であるバッグに対してカテゴリが付与されていさえすれば、本実施形態を利用することが可能である。たとえば、学習する対象のインスタンス(入力ベクトル)のベクトル成分について、値を微小変動させて生成したベクトルが作られればよい。そして、このように、もともと1つのインスタンスから、微小変動させて生成した複数のインスタンスをまとめた集合が、1つのバッグとして用意されてもよい。
 こうすることにより、インスタンスが、たとえばノイズが含まれるようなインスタンスであっても、それから生成されるインスタンスのうち1つでもノイズが低減されるようなインスタンスであれば、本実施形態は、ノイズの影響を受け難い学習処理を実行することができる。一方、インスタンスが、たとえば、カテゴリが正確に付与されているインスタンスがあるならば、その1つのインスタンスのみで1つのバッグが構成されればよい。
 本実施形態では、第2実施形態と同様の効果を得ることができる。バッグが正解カテゴリである確率から導出される尤もらしさLを、以下の数式により定義するものと考える。
Figure JPOXMLDOC01-appb-M000020
 すると、最急降下法で最小化すべき関数Sを、以下の数式のように求めることができる。
Figure JPOXMLDOC01-appb-M000021
 この場合、第2実施形態の効果の説明に用いた数式に相当する、以下の式を導くことができる。
Figure JPOXMLDOC01-appb-M000022
 したがって、確率ベース修正係数uijを計算することで、Lを最大化することができる。これにより、本実施形態は、インスタンスではなくバッグに付与されたカテゴリの情報のみでも、学習を実行することができる。
 [他の実施形態]
 以上、本発明の実施形態について詳述したが、それぞれの実施形態に含まれる別々の特徴を如何様に組み合わせたシステムまたは装置も、本発明の範疇に含まれる。
 また、本発明は、複数の機器から構成されるシステムに適用されてもよい。本発明は、単体の装置に適用されてもよい。さらに、本発明は、実施形態の機能を実現する情報処理プログラムが、システムあるいは装置に直接あるいは遠隔から供給される場合にも適用可能である。したがって、本発明の機能をコンピュータで実現するために、コンピュータにインストールされるプログラム、あるいはそのプログラムを格納した媒体、そのプログラムをダウンロードさせるWWW(World Wide Web)サーバも、本発明の範疇に含まれる。
 情報処理システム100、情報処理システム400、データ処理装置401、記憶装置402は、それぞれ、コンピュータ及びコンピュータを制御するプログラム、専用のハードウェア、又は、コンピュータ及びコンピュータを制御するプログラムと専用のハードウェアの組合せにより実現することができる。
 インスタンス選択部102、参照ベクトル特定部103、インスタンス確率算出部104、バッグ確率算出部105、参照ベクトル修正部106、バッグ選択部411、インスタンス選択部412、参照ベクトル特定部413、インスタンス確率算出部414、バッグ確率算出部415、確率ベース修正計数算出部416、参照ベクトル修正部417、終了判定部418は、例えば、プログラムを記憶する記録媒体からメモリに読み込まれた、各部の機能を実現するための専用のプログラムと、そのプログラムを実行するプロセッサにより実現することができる。また、参照ベクトル記憶部101、学習用データ記憶部421、参照ベクトル記憶部422は、コンピュータが含むメモリやハードディスク装置により実現することができる。あるいは、参照ベクトル記憶部101、インスタンス選択部102、参照ベクトル特定部103、インスタンス確率算出部104、バッグ確率算出部105、参照ベクトル修正部106、バッグ選択部411、インスタンス選択部412、参照ベクトル特定部413、インスタンス確率算出部414、バッグ確率算出部415、確率ベース修正計数算出部416、参照ベクトル修正部417、終了判定部418、学習用データ記憶部421、参照ベクトル記憶部422の一部又は全部を、各部の機能を実現する専用の回路によって実現することもできる。
 [実施形態の他の表現]
 上記の実施形態の一部または全部は、以下の付記のようにも記載されうるが、以下には限られない。
 (付記1)
 参照ベクトル群を保持する参照ベクトル記憶手段と、
 複数のインスタンスを含むバッグ内から1つのインスタンスを選択するインスタンス選択手段と、
 前記参照ベクトル記憶手段に記憶されている参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定する参照ベクトル特定手段と、
 前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算するインスタンス確率算出手段と、
 前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算するバッグ確率算出手段と、
 前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する参照ベクトル修正手段と、
 を備えたことを特徴とする情報処理システム。
 (付記2)
 前記参照ベクトル特定手段は、選択された前記インスタンスが属するバッグと同じカテゴリで最も前記インスタンスからの距離が近い第1参照ベクトルと、前記インスタンスが属するバッグとは異なるカテゴリで最も前記インスタンスからの距離が近い第2参照ベクトルとを、前記関連参照ベクトルとして特定することを特徴とする付記1に記載の情報処理システム。
 (付記3)
 前記参照ベクトル特定手段は、前記インスタンスと前記第1参照ベクトルとの距離を示す第1距離と、前記インスタンスと前記第2参照ベクトルとの距離を示す第2距離とを算出し、
 前記参照ベクトル修正手段は、前記第1距離および前記第2距離を用いて前記関連参照ベクトルを修正することを特徴とする付記2に記載の情報処理システム。
 (付記4)
 前記参照ベクトル修正手段は、
 第1距離のべき乗計算、および第1距離と前記第2距離の和のべき乗で除算する計算によって変換して得られる変換距離値を用いて前記関連参照ベクトルを修正することを特徴とする付記3に記載の情報処理システム。
 (付記5)
 前記インスタンス確率算出手段は、前記第1距離と前記第2距離との差を前記第1距離と前記第2距離との和で除算することにより、前記インスタンス正解確率を算出することを特徴とする付記3または4に記載の情報処理システム。
 (付記6)
 前記バッグ正解確率を用いて、前記関連参照ベクトルを修正するための確率ベース修正係数を計算する確率ベース修正係数算出手段をさらに備え、
 前記参照ベクトル修正手段は、確率ベース修正係数に基づいて参照ベクトルを修正することを特徴とする付記1乃至5のいずれか1項に記載の情報処理システム。
 (付記7)
 前記確率ベース修正係数算出手段は、インスタンス正解確率とインスタンスが属するバッグが不正解カテゴリである確率としてのバッグ不正解確率との乗算、および、インスタンスが不正解カテゴリである確率としてのインスタンス不正解確率と前記バッグ正解確率との乗算を行なうことにより確率ベース修正係数を算出することを特徴とする付記6に記載の情報処理システム。
 (付記8)
 前記確率ベース修正係数算出手段は、インスタンス正解確率とバッグ不正解確率との積をバッグ正解確率で除算した値、および、前記インスタンス不正解確率によって確率ベース修正係数を算出することを特徴とする付記6または7に記載の情報処理システム。
 (付記9)
 前記バッグ確率算出手段は、前記バッグに含まれるインスタンスのうち少なくとも1つ以上のインスタンスのカテゴリが正解である確率を算出することを付記1乃至8のいずれか1行に記載の特徴とする付記1乃至8のいずれかに記載の情報処理システム。
 (付記10)
 複数のインスタンスを含むバッグ内から1つのインスタンスを選択し、
 参照ベクトル記憶手段に記憶されている参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定し、
 選択された前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算し、
 前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算し、
 前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する
 ことを特徴とする認識辞書学習方法。
 (付記11)
 コンピュータを、
 複数のインスタンスを含むバッグ内から1つのインスタンスを選択するインスタンス選択手段と、
 参照ベクトル記憶手段に記憶されている参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定する参照ベクトル特定手段と、
 選択された前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算するインスタンス確率算出手段と、
 前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算するバッグ確率算出手段と、
 前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する参照ベクトル修正手段と
 して実行させることを特徴とする情報処理プログラム。
 以上、実施形態を参照して本願発明を説明したが、本願発明は上記実施形態に限定されものではない。本願発明の構成や詳細には、本願発明のスコープ内で当業者が理解し得る様々な変更をすることができる。
 この出願は、2011年7月19日に出願された日本出願特願2011−158339を基礎とする優先権を主張し、その開示の全てをここに取り込む。
Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the drawings. However, the constituent elements described in the following embodiments are merely examples, and are not intended to limit the technical scope of the present invention only to them.
[First Embodiment]
An information processing system 100 as a first embodiment of the present invention will be described with reference to FIG. The information processing system 100 is a device for learning a recognition dictionary used in pattern recognition.
As illustrated in FIG. 1, the information processing system 100 includes a reference vector storage unit 101, an instance selection unit 102, a reference vector identification unit 103, an instance probability calculation unit 104, a bag probability calculation 105, and a reference vector correction unit 106. Including.
The reference vector storage unit 101 holds a reference vector group. The instance selection unit 102 selects one instance from the bag including a plurality of instances.
Further, the reference vector specifying unit 103 specifies a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage unit 101. The instance probability calculation unit 104 calculates an instance correct probability that the instance category is correct. Then, the bag probability calculation unit 105 calculates the bag correct probability that the category of the bag is correct using the probability that the category of the instance included in the bag is correct. The reference vector correction unit 106 corrects the related reference vector using the bag correct answer probability.
With the above configuration, a recognition dictionary can be created even when the category to which the input vector used for learning belongs is not completely known.
[Second Embodiment]
(Prerequisite technology)
As a second embodiment of the present invention, an information processing system 400 based on a recognition dictionary learning method (LVQ) using reference vectors will be described. First, before entering a description of a specific embodiment, a concept called “Multiple Instance Learning” that defines a rule of how to assign a category in the field of statistical pattern recognition will be described. In this concept, an input vector is called an “instance”. A set of instances is called a “bag”. Furthermore, a category is not assigned to every instance, but only to a bag.
FIG. 2 illustrates the concept of bags and instances. When there are only two types of categories, “positive” and “negative”, if there is at least one positive instance in the bag, the category of the bag is defined as “positive”. If only a negative instance exists in the bag, the category of the bag is defined as “negative”.
Based on the definition of this category, let pij be the probability that the j-th instance in the i-th bag is positive. At this time, the probability that one or more instances in the bag are positive, that is, the probability that the bag is positive, pi, is calculated by the following equation.
Figure JPOXMLDOC01-appb-M000001
For example, when classifying into two categories of “positive” and “negative (non-positive)”, even if there is incomplete information that any one of n sets of instances is “positive”, the information The processing system 400 cannot learn well. In order to obtain the coefficient for updating the reference vector, it is necessary that all instances used for learning and the category to which each instance belongs have a one-to-one correspondence. For this reason, the information processing system 400 cannot apply information of a category that is not complete to the learning process.
That is, before the information processing system 400 performs the learning process, the information regarding the category needs to be completely given to all the input vectors used for learning. However, this work takes time and effort. In addition, when it is difficult to accurately define the category of the input vector, it is difficult to completely give information on the category in the first place. For example, if the input vector is an image pattern and the categories are two types of categories, “Face” and “Non-face”, a pattern that includes only the eyes, nose, and mouth of the face, It is difficult to determine whether the category of the pattern including the shoulder should be defined as “face” or “non-face”. For a pattern that includes only the eyes, nose, and mouth of the face, a pattern that includes the outline of the face, and a group of patterns that includes the neck and shoulders, at least one is a face. It is a natural way of thinking to define.
For this reason, in this embodiment, the information processing system 400 calculates the probability that the instance is the correct category, and further calculates the probability that the bag is the correct category. Then, the information processing system 400 corrects the reference vector using the probability that the bag to which the instance belongs is in the correct category. In this way, the information processing system 400 can perform learning processing (that is, correction of a reference vector) even if the category to which the instance belongs is not completely clear.
(Constitution)
FIG. 3 schematically represents a recognition method using a recognition dictionary by the information processing system 400 of the present embodiment. The recognition dictionary includes reference vectors 311 to 31n belonging to a certain category A and reference vectors 321 to 32n belonging to another category B. When there is an instance 303 for which the category is to be recognized, the information processing system 400 identifies the instance 303 and the closest reference vector 312 among the reference vectors 311 to 31n belonging to the category A. Then, the information processing system 400 calculates the distance d1. Further, the information processing system 400 specifies the instance 303 and the closest reference vector 322 among the reference vectors 321 to 32n belonging to the category B. Then, the information processing system 400 calculates the distance d2. Then, the information processing system 400 determines that the instance 303 also belongs to the category to which the closest reference vector 312 belongs. In the case of FIG. 3, since d1 <d2, the information processing system 400 determines that the instance 303 belongs to category A.
Next, an information processing system 400 according to the present embodiment will be described with reference to FIG. FIG. 4 is a diagram for explaining a functional configuration of the information processing system 400 according to the present embodiment. The information processing system 400 calculates a distance between a group of reference vectors (also referred to as templates, prototypes, and representative vectors) called a recognition dictionary and an input vector. Based on the calculation result, the information processing system 400 sets the category (also referred to as a class or label) to which the reference vector closest to the input vector belongs as the recognition result of the input vector.
Hereinafter, for the sake of explanation, one instance is represented by x and the reference vector is represented by wk (k is 1 to K) (each is a vector, but the bar is omitted here for simplification).
Referring to FIG. 4, the present embodiment includes a data processing device 401 and a storage device 402. The data processing device 401 includes a bag selection unit 411, an instance selection unit 412, a reference vector identification unit 413, an instance probability calculation unit 414, a bag probability calculation unit 415, a probability base correction coefficient calculation unit 416, and a reference vector. A correction unit 417 and an end determination unit 418 are included.
The storage device 402 includes a learning data storage unit 421 and a reference vector storage unit 422. The learning data storage unit 421 stores information related to learning instance groups (bags). The reference vector storage unit 422 stores information on a reference vector group that is a recognition dictionary. Specifically, the reference vector storage unit 422 stores individual reference vectors and category information to which the reference vectors belong. Note that the reference vector is modified during the operation of the present invention, resulting in a learned reference vector.
The bag selection unit 411 selects one bag i from all bags used for learning. The instance selection unit 412 selects one instance (one input vector) j from the bag i selected by the bag selection unit 411.
The reference vector specifying unit 413 determines a reference vector pair related to the instance j selected by the instance selecting unit 412. Specifically, the reference vector specifying unit 413 is closest to the reference vector group in the same category as the instance j from the inter-vector distance between the instance j and each reference vector held in the reference vector storage unit 422. The first reference vector W1 and its first distance d1 are calculated. Further, the reference vector specifying unit 413 further calculates the second reference vector W2 closest to the reference vector group in a category different from the instance j and the second distance d2. Note that it is sufficient for the reference vector specifying unit 413 to search for the closest reference vector. Therefore, the reference vector specifying unit 413 does not necessarily calculate the distances from all the reference vectors.
The instance probability calculation unit 414 estimates a probability pij that the instance is a correct category. Here, the correct category is a category assigned to the bag i to which the instance j belongs. Note that any method may be used as a method of estimating the probability pij. Hereinafter, an example of a method for estimating the probability pij will be described. The probability pij can be expressed as follows.
Figure JPOXMLDOC01-appb-M000002
On the other hand, if the correct answer category likelihood of an instance is μij, and an arbitrary monotonically increasing function with a value range of [0, 1] is R (−), the probability pij is also written as the following equation.
Figure JPOXMLDOC01-appb-M000003
There are various derivation methods for μij. For example, μij can be obtained based on the following equation.
Figure JPOXMLDOC01-appb-M000004
In addition, μij is written by Hiroyoshi Miyano and Nagaki Ishidera, “Learning Vector Quantization Method Using Heavy Distribution Function of Support”, IEICE, IEICE Technical Report, vol. 110, no. 187, PRMU 2010-81, pp. 185-192, September 2010, ηk may be used.
Further, the monotone increasing function R (−) may be, for example, the following expression.
Figure JPOXMLDOC01-appb-M000005
In consideration of the fact that the recognition reliability increases as the reference vector learning is repeated, the monotonically increasing function R (−) may be expressed by the following equation using the reference vector learning frequency t.
Figure JPOXMLDOC01-appb-M000006
Furthermore, the monotonically increasing function R (−) may be the following equation using the reference vector learning count t and arbitrary constants γ0 and γ1.
Figure JPOXMLDOC01-appb-M000007
The bag probability calculation unit 415 calculates a probability qi that the bag is likely to be the correct category from the probability pij that each instance in the bag i is likely to be the correct category. This calculation formula is expressed by the following formula.
Figure JPOXMLDOC01-appb-M000008
The probability-based correction coefficient calculation unit 416 calculates, for each instance, a probability-based correction coefficient uij for updating the reference vector based on the probability pij that the instance is a correct category and the probability qi that the bag to which the instance belongs is the correct category. calculate.
The probability-based correction coefficient calculation unit 416 may obtain the correction coefficient uij so as to maximize the likelihood L, which means the likelihood of the recognition result. The likelihood L may be defined based on the probability qi.
For example, assume that L is defined by the following equation.
Figure JPOXMLDOC01-appb-M000009
In this case, the probability-based correction coefficient calculation unit 416 may calculate the probability-based correction coefficient uij using the following formula.
Figure JPOXMLDOC01-appb-M000010
The reference vector correction unit 417 corrects the first reference vector w1 and the second reference vector w2 stored in the reference vector storage unit 422 based on the probability base correction coefficient uij calculated by the probability base correction coefficient calculation unit 416. To do. More specifically, the reference vector correction unit 417 calculates a value calculated from the function D2 (·) that converts the second distance d2, a probability-based correction coefficient uij, an arbitrary coefficient α1, and an instance x and w1. A vector obtained by multiplying the difference vector is added (or subtracted) to w1 as a vector. By this process, the reference vector correction unit 417 corrects the first reference vector w1.
On the other hand, the reference vector correction unit 417 calculates a value calculated from the function D1 (•) that converts the first distance d1, a probability-based correction coefficient uij, and an arbitrary coefficient α2, and the difference between the instances x and w2. A vector obtained by multiplying the vector is added (or subtracted) as a vector to w2. By this process, the reference vector correction unit 417 corrects the second reference vector w2. This correction calculation can be expressed as in Equation 3 below.
Figure JPOXMLDOC01-appb-M000011
The function D2 (•) includes a power calculation of d2 and a calculation that further divides by the power of the sum of d1 and d2. The function D1 (•) is composed of a power calculation of d1 and a calculation for further dividing by the power of the sum of d1 and d2. The power here refers to the calculation of the y power using an arbitrary real number y of 0 or more. For example, D1 (•) and D2 (•) may be functions represented by the following expressions.
Figure JPOXMLDOC01-appb-M000012
The end determination unit 418 determines the end of learning (correction) of the reference vector.
(Operation)
Next, the overall operation of the present embodiment will be described in detail with reference to the flowchart of FIG. In the following description, the number of bags used for learning is M. The number of instances in the bag i is Ni. Before operation, an instance (input vector) used for learning is prepared, and a bag composed of one or more instances is prepared. As an actual data structure, it is sufficient to associate a bag number with an instance. A category is assigned to all bags.
First, the bag selection unit 411 selects one bag i from the learning bag group stored in the learning data storage unit 421 (S501). Next, the instance selection unit 412 selects one instance in the selected bag (S503). Then, the reference vector specifying unit 413 searches for the reference vector closest to the selected instance (S505). Further, the instance probability calculation unit 414 calculates a probability Pij that the instance is correct (S507). Next, the instance selection unit 412 determines whether or not the processing in steps S503 to S507 has been completed for all instances in the bag (S509), and if not, the processing returns to step S503.
As described above, the information processing apparatus 400 includes the instance selection process (S503) by the instance selection unit 412, the reference vector specification process (S505) by the reference vector specification unit 413, and the instance probability calculation process (S507) by the instance probability calculation unit 414. Is repeatedly executed.
Next, the bag probability calculation unit 415 calculates a probability qi that the selected bag i is correct (step S511). The instance selection unit 412 selects one instance again from the bag. Then, the probability-based correction coefficient calculation unit 416 calculates a probability-based correction coefficient for the selected instance (S513). Further, the reference vector correction unit 417 uses the first reference vector w1 and the second reference vector w2 stored in the reference vector storage unit 422 based on the probability base correction coefficient uij calculated by the probability base correction coefficient calculation unit 415. Is corrected (S515). Next, the instance selection unit 412 determines whether or not the processing of steps S512 to S515 has been completed for all instances in the bag (S516), and if not, returns to step S503.
The information processing apparatus 400 repeatedly executes the processes in steps S511 to S515 for all instances in the selected bag i. Thus, the information processing apparatus 400 corrects all reference vectors stored in the reference vector storage unit 422. Thus, the learning process for each bag is completed.
Next, the bag selection unit 411 determines whether or not the processing of steps S501 to S516 has been completed for all bags (S517), and if not, the processing returns to step S501.
The information processing apparatus 400 repeatedly executes the processes in steps S501 to S517 for all bags. Thus, the information processing apparatus 400 corrects all reference vectors stored in the reference vector storage unit 422.
In step S518, the end determination unit 418 determines whether or not to end the learning of the reference vector. If the end determination unit 418 determines not to end, the process returns to step S501. Then, the information processing apparatus 400 repeatedly executes the processing from step S501 to step S517 for all learning bags.
On the other hand, when the end determination unit 418 determines to end the learning of the reference vector, the information processing apparatus 400 ends the process. For example, the end determination unit 418 may determine to end the learning of the reference vector when the information processing apparatus 400 repeatedly executes the above-described process a predetermined number of times.
As described above, in the present embodiment, the bag probability calculation unit 415 calculates the probability that the bag is the correct category. Then, the reference vector correction unit 417 corrects the reference vector so as to maximize the likelihood L derived therefrom. Therefore, the information processing apparatus 400 can perform learning only with the information of the category assigned to the bag, not the instance.
Next, it will be described that the likelihood L derived from the probability that the bag is the correct category can be maximized by using the probability base correction coefficient uij calculated by the probability base correction coefficient calculation unit 416 used in the present embodiment. . Let us consider maximizing L by the steepest descent method when Pij is defined as Equation 1 and L is defined as Equation 2.
In this case, the function S to be minimized can be defined as follows using a logarithmic function that is a monotone function.
Figure JPOXMLDOC01-appb-M000013
In the steepest descent method, the update formula of the reference vector wk is expressed as follows using a constant α.
Figure JPOXMLDOC01-appb-M000014
The following expression is an expression obtained by expanding the second term on the right side of the above expression.
Figure JPOXMLDOC01-appb-M000015
Further expansion of the above expression leads to the following expression 4 and expression 5 using the constant β.
Figure JPOXMLDOC01-appb-M000016
Figure JPOXMLDOC01-appb-M000017
Equation 4 corresponds to a probability-based correction coefficient. Equation 3 is derived from Equation 4 and Equation 5.
Since the following expression described above can be interpreted as an expression for normalizing the magnitude of the distance value, the same effect can be obtained even if the numerator and the denominator are raised.
Figure JPOXMLDOC01-appb-M000018
The operation of the present embodiment is described using “probability of being correct”. In the case of two categories of “positive” and “negative (non-positive)”, the “probability of incorrect answer” or “probability of positive” or “probability of negative” is obvious. . Therefore, it can be said that this embodiment uses these probabilities.
In the present embodiment, since the reference vector is corrected based on the probability that the bag is correct, a learning process for maximizing the correctness (evaluation scale) of the entire bag to be learned is performed. As a result, even if there is no category information corresponding to all the input vectors used for learning, the present embodiment learns only from the category information corresponding to the set (bag) of input vectors (instances) used for learning. it can. Therefore, this embodiment can create a recognition dictionary. In the present embodiment, the preparation for the learning process can be completed only by adding coarser information before the learning process. In addition, the present embodiment can perform learning even when it is difficult to accurately define the category of each input vector in the first place or when the category can be defined only for a set of input vectors.
The present embodiment can be applied to a use for detecting a specific object from a video or a use for identifying and authenticating a person or object from a video. In particular, when enormous amounts of data are learned and used for recognition, it is difficult to completely assign the correct category of the object. However, since the correct category only needs to be assigned to a set of data, this embodiment can be applied practically.
[Third Embodiment]
The probability base correction coefficient calculation unit 416 according to the second embodiment may calculate the probability base correction coefficient uij by the following formula.
Figure JPOXMLDOC01-appb-M000019
As long as a category is assigned to a bag that is a set of instances, this embodiment can be used. For example, a vector generated by minutely changing a value of a vector component of an instance to be learned (input vector) may be created. In this way, a set in which a plurality of instances that are originally generated from a single instance with a slight variation may be prepared as one bag.
In this way, if the instance is an instance including noise, for example, if one of the instances generated from the instance is reduced in noise, the present embodiment It is possible to execute a learning process that is hardly affected. On the other hand, if there is an instance in which, for example, a category is correctly assigned, one bag only needs to be configured by that one instance.
In the present embodiment, the same effect as in the second embodiment can be obtained. It is assumed that the likelihood L derived from the probability that the bag is the correct category is defined by the following mathematical formula.
Figure JPOXMLDOC01-appb-M000020
Then, the function S to be minimized by the steepest descent method can be obtained as in the following equation.
Figure JPOXMLDOC01-appb-M000021
In this case, the following equation corresponding to the equation used to explain the effects of the second embodiment can be derived.
Figure JPOXMLDOC01-appb-M000022
Therefore, L can be maximized by calculating the probability-based correction coefficient uij. Thereby, this embodiment can perform learning only with the information of the category provided to the bag instead of the instance.
[Other Embodiments]
As mentioned above, although embodiment of this invention was explained in full detail, the system or apparatus which combined the separate characteristic contained in each embodiment how was included in the category of this invention.
Further, the present invention may be applied to a system composed of a plurality of devices. The present invention may be applied to a single device. Furthermore, the present invention can also be applied to a case where an information processing program that implements the functions of the embodiments is supplied directly or remotely to a system or apparatus. Therefore, in order to realize the functions of the present invention with a computer, a program installed in the computer, a medium storing the program, and a WWW (World Wide Web) server that downloads the program are also included in the scope of the present invention. .
The information processing system 100, the information processing system 400, the data processing device 401, and the storage device 402 are respectively a computer and a program that controls the computer, dedicated hardware, or a program that controls the computer and the computer and dedicated hardware. It can be realized by a combination.
Instance selection unit 102, reference vector specification unit 103, instance probability calculation unit 104, bag probability calculation unit 105, reference vector correction unit 106, bag selection unit 411, instance selection unit 412, reference vector specification unit 413, instance probability calculation unit 414 The bag probability calculation unit 415, the probability base correction count calculation unit 416, the reference vector correction unit 417, and the end determination unit 418 are, for example, for realizing the function of each unit read into the memory from the recording medium storing the program. It can be realized by a dedicated program and a processor that executes the program. The reference vector storage unit 101, the learning data storage unit 421, and the reference vector storage unit 422 can be realized by a memory or a hard disk device included in the computer. Alternatively, the reference vector storage unit 101, the instance selection unit 102, the reference vector specification unit 103, the instance probability calculation unit 104, the bag probability calculation unit 105, the reference vector correction unit 106, the bag selection unit 411, the instance selection unit 412, the reference vector specification Unit 413, instance probability calculation unit 414, bag probability calculation unit 415, probability-based correction count calculation unit 416, reference vector correction unit 417, end determination unit 418, learning data storage unit 421, part of reference vector storage unit 422 or All can be realized by a dedicated circuit for realizing the function of each unit.
[Other expressions of embodiment]
A part or all of the above-described embodiment can be described as in the following supplementary notes, but is not limited thereto.
(Appendix 1)
Reference vector storage means for holding a reference vector group;
An instance selection means for selecting one instance from a bag including a plurality of instances;
Reference vector specifying means for specifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means;
Instance probability calculating means for calculating an instance correct probability that the category of the instance is correct;
Bag probability calculation means for calculating a bag correct probability that the category of the bag is correct using a probability that the category of the instance included in the bag is correct;
Reference vector correcting means for correcting the related reference vector using the bag correct answer probability;
An information processing system comprising:
(Appendix 2)
The reference vector specifying means has a first reference vector that is the closest to the instance in the same category as the bag to which the selected instance belongs, and a distance from the instance that is the closest to a category different from the bag to which the instance belongs. The information processing system according to appendix 1, wherein a near second reference vector is specified as the related reference vector.
(Appendix 3)
The reference vector specifying means calculates a first distance indicating a distance between the instance and the first reference vector, and a second distance indicating a distance between the instance and the second reference vector;
The information processing system according to supplementary note 2, wherein the reference vector correcting unit corrects the related reference vector using the first distance and the second distance.
(Appendix 4)
The reference vector correcting means includes
Supplementary note 3 wherein the related reference vector is corrected using a conversion distance value obtained by conversion by a power calculation of the first distance and a calculation by dividing by a power of the sum of the first distance and the second distance. Information processing system described in 1.
(Appendix 5)
The instance probability calculating means calculates the instance correct probability by dividing a difference between the first distance and the second distance by a sum of the first distance and the second distance. The information processing system according to appendix 3 or 4.
(Appendix 6)
Probability-based correction coefficient calculating means for calculating a probability-based correction coefficient for correcting the related reference vector using the bag correct answer probability,
The information processing system according to any one of appendices 1 to 5, wherein the reference vector correcting unit corrects the reference vector based on a probability-based correction coefficient.
(Appendix 7)
The probability-based correction coefficient calculating means is configured to multiply the instance correct answer probability by the bag incorrect answer probability as the probability that the bag to which the instance belongs is an incorrect answer category, and the instance incorrect answer probability as the probability that the instance is an incorrect answer category. 7. The information processing system according to claim 6, wherein a probability-based correction coefficient is calculated by multiplying the bag correct answer probability.
(Appendix 8)
The probability-based correction coefficient calculating means calculates a probability-based correction coefficient based on a value obtained by dividing the product of the instance correct answer probability and the bag incorrect answer probability by the bag correct answer probability, and the instance incorrect answer probability. The information processing system according to 6 or 7.
(Appendix 9)
The supplementary note 1 according to any one of supplementary notes 1 to 8, wherein the bag probability calculating means calculates a probability that a category of at least one instance among the instances included in the bag is correct. The information processing system in any one of thru | or 8.
(Appendix 10)
Select an instance from within a bag containing multiple instances,
Identifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means;
Calculating an instance correct probability that the category of the selected instance is correct;
Using the probability that the category of the instance included in the bag is correct, the bag correct probability that the category of the bag is correct is calculated,
The related reference vector is corrected using the bag correct probability.
A recognition dictionary learning method characterized by that.
(Appendix 11)
Computer
An instance selection means for selecting one instance from a bag including a plurality of instances;
Reference vector specifying means for specifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means;
An instance probability calculating means for calculating an instance correct probability that the category of the selected instance is correct;
Bag probability calculation means for calculating a bag correct probability that the category of the bag is correct using a probability that the category of the instance included in the bag is correct;
Reference vector correcting means for correcting the related reference vector using the bag correct answer probability;
Information processing program.
While the present invention has been described with reference to the embodiments, the present invention is not limited to the above embodiments. Various changes that can be understood by those skilled in the art can be made to the configuration and details of the present invention within the scope of the present invention.
This application claims the priority on the basis of Japanese application Japanese Patent Application No. 2011-158339 for which it applied on July 19, 2011, and takes in those the indications of all here.

Claims (10)

  1.  参照ベクトル群を保持する参照ベクトル記憶手段と、
     複数のインスタンスを含むバッグ内から1つのインスタンスを選択するインスタンス選択手段と、
     前記参照ベクトル記憶手段に記憶された参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定する参照ベクトル特定手段と、
     前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算するインスタンス確率算出手段と、
     前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算するバッグ確率算出手段と、
     前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する参照ベクトル修正手段と、
     を備えたことを特徴とする情報処理システム。
    Reference vector storage means for holding a reference vector group;
    An instance selection means for selecting one instance from a bag including a plurality of instances;
    Reference vector specifying means for specifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means;
    Instance probability calculating means for calculating an instance correct probability that the category of the instance is correct;
    Bag probability calculation means for calculating a bag correct probability that the category of the bag is correct using a probability that the category of the instance included in the bag is correct;
    Reference vector correcting means for correcting the related reference vector using the bag correct answer probability;
    An information processing system comprising:
  2.  前記参照ベクトル特定手段は、選択された前記インスタンスが属するバッグと同じカテゴリで最も前記インスタンスからの距離が近い第1参照ベクトルと、前記インスタンスが属するバッグとは異なるカテゴリで最も前記インスタンスからの距離が近い第2参照ベクトルとを、前記関連参照ベクトルとして特定することを特徴とする請求項1に記載の情報処理システム。 The reference vector specifying means has a first reference vector that is the closest to the instance in the same category as the bag to which the selected instance belongs, and a distance from the instance that is closest to a category different from the bag to which the instance belongs. The information processing system according to claim 1, wherein a second reference vector that is close is specified as the related reference vector.
  3.  前記参照ベクトル特定手段は、前記インスタンスと前記第1参照ベクトルとの距離を示す第1距離と、前記インスタンスと前記第2参照ベクトルとの距離を示す第2距離とを算出し、
     前記参照ベクトル修正手段は、前記第1距離および前記第2距離を用いて前記関連参照ベクトルを修正することを特徴とする請求項2に記載の情報処理システム。
    The reference vector specifying means calculates a first distance indicating a distance between the instance and the first reference vector, and a second distance indicating a distance between the instance and the second reference vector;
    The information processing system according to claim 2, wherein the reference vector correcting unit corrects the related reference vector using the first distance and the second distance.
  4.  前記参照ベクトル修正手段は、
     第1距離のべき乗計算、および第1距離と前記第2距離の和のべき乗で除算する計算によって変換して得られる変換距離値を用いて前記関連参照ベクトルを修正することを特徴とする請求項3に記載の情報処理システム。
    The reference vector correcting means includes
    The related reference vector is corrected using a conversion distance value obtained by conversion by a power calculation of a first distance and a calculation of dividing by a power of a sum of the first distance and the second distance. 3. The information processing system according to 3.
  5.  前記インスタンス確率算出手段は、前記第1距離と前記第2距離との差を前記第1距離と前記第2距離との和で除算することにより、前記インスタンス正解確率を算出することを特徴とする請求項3または4に記載の情報処理システム。 The instance probability calculating means calculates the instance correct probability by dividing a difference between the first distance and the second distance by a sum of the first distance and the second distance. The information processing system according to claim 3 or 4.
  6.  前記バッグ正解確率を用いて、前記関連参照ベクトルを修正するための確率ベース修正係数を計算する確率ベース修正係数算出手段をさらに備え、
     前記参照ベクトル修正手段は、確率ベース修正係数に基づいて参照ベクトルを修正することを特徴とする請求項1乃至5のいずれか1項に記載の情報処理システム。
    Probability-based correction coefficient calculating means for calculating a probability-based correction coefficient for correcting the related reference vector using the bag correct answer probability,
    The information processing system according to claim 1, wherein the reference vector correcting unit corrects the reference vector based on a probability-based correction coefficient.
  7.  前記確率ベース修正係数算出手段は、インスタンス正解確率とインスタンスが属するバッグが不正解カテゴリである確率としてのバッグ不正解確率との乗算、および、インスタンスが不正解カテゴリである確率としてのインスタンス不正解確率と前記バッグ正解確率との乗算を行なうことにより確率ベース修正係数を算出することを特徴とする請求項6に記載の情報処理システム。 The probability-based correction coefficient calculating means includes multiplying the instance correct answer probability by the bag incorrect answer probability as the probability that the bag to which the instance belongs is an incorrect answer category, and the instance incorrect answer probability as the probability that the instance is an incorrect answer category. The information processing system according to claim 6, wherein a probability-based correction coefficient is calculated by performing multiplication of the bag correct answer probability.
  8.  前記確率ベース修正係数算出手段は、インスタンス正解確率とバッグ不正解確率との積をバッグ正解確率で除算した値、および、前記インスタンス不正解確率によって確率ベース修正係数を算出することを特徴とする請求項6または7に記載の情報処理システム。 The probability-based correction coefficient calculation means calculates a probability-based correction coefficient based on a value obtained by dividing a product of an instance correct answer probability and a bag incorrect answer probability by a bag correct answer probability, and the instance incorrect answer probability. Item 8. The information processing system according to Item 6 or 7.
  9.  複数のインスタンスを含むバッグ内から1つのインスタンスを選択し、
     参照ベクトル記憶手段に記憶された参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定し、
     選択された前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算し、
     前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算し、
     前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する
     ことを特徴とする認識辞書学習方法。
    Select an instance from within a bag containing multiple instances,
    Identifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means;
    Calculating an instance correct probability that the category of the selected instance is correct;
    Using the probability that the category of the instance included in the bag is correct, the bag correct probability that the category of the bag is correct is calculated,
    The recognition dictionary learning method, wherein the related reference vector is corrected using the bag correct answer probability.
  10.  コンピュータを、
     複数のインスタンスを含むバッグ内から1つのインスタンスを選択するインスタンス選択手段と、
     参照ベクトル記憶手段に記憶された参照ベクトル群から、選択された前記インスタンスに最も関連する関連参照ベクトルを特定する参照ベクトル特定手段と、
     選択された前記インスタンスのカテゴリが正解であるインスタンス正解確率を計算するインスタンス確率算出手段と、
     前記バッグに含まれるインスタンスのカテゴリが正解である確率を用いて、該バッグのカテゴリが正解であるバッグ正解確率を計算するバッグ確率算出手段と、
     前記バッグ正解確率を用いて、前記関連参照ベクトルを修正する参照ベクトル修正手段と
     して動作させることを特徴とする情報処理プログラムを記憶する記録媒体。
    Computer
    An instance selection means for selecting one instance from a bag including a plurality of instances;
    Reference vector specifying means for specifying a related reference vector most relevant to the selected instance from the reference vector group stored in the reference vector storage means;
    An instance probability calculating means for calculating an instance correct probability that the category of the selected instance is correct;
    Bag probability calculation means for calculating a bag correct probability that the category of the bag is correct using a probability that the category of the instance included in the bag is correct;
    A recording medium for storing an information processing program which is operated as reference vector correcting means for correcting the related reference vector using the bag correct answer probability.
PCT/JP2012/068747 2011-07-19 2012-07-18 Information processing system, method of learning recognition dictionary, and information processing program WO2013012093A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2011158339 2011-07-19
JP2011-158339 2011-07-19

Publications (1)

Publication Number Publication Date
WO2013012093A1 true WO2013012093A1 (en) 2013-01-24

Family

ID=47558265

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2012/068747 WO2013012093A1 (en) 2011-07-19 2012-07-18 Information processing system, method of learning recognition dictionary, and information processing program

Country Status (2)

Country Link
JP (1) JPWO2013012093A1 (en)
WO (1) WO2013012093A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104718553A (en) * 2013-06-04 2015-06-17 松下电器(美国)知识产权公司 Information processing system for identifying used commodities in domestic electrical appliances, and security system
CN105069506A (en) * 2015-07-24 2015-11-18 中国地质大学(武汉) Blast furnace hanging diagnosis method capable of processing conflict information effectively

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002251592A (en) * 2001-02-22 2002-09-06 Toshiba Corp Learning method for pattern recognition dictionary

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2002251592A (en) * 2001-02-22 2002-09-06 Toshiba Corp Learning method for pattern recognition dictionary

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YI ZHANG ET AL.: "Learning from multi-topic web documents for contextual advertisement", PROCEEDINGS OF THE 14TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 27 August 2008 (2008-08-27), pages 1051 - 1059 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104718553A (en) * 2013-06-04 2015-06-17 松下电器(美国)知识产权公司 Information processing system for identifying used commodities in domestic electrical appliances, and security system
CN104718553B (en) * 2013-06-04 2018-09-11 松下电器(美国)知识产权公司 Information processing system
CN105069506A (en) * 2015-07-24 2015-11-18 中国地质大学(武汉) Blast furnace hanging diagnosis method capable of processing conflict information effectively
CN105069506B (en) * 2015-07-24 2017-11-14 中国地质大学(武汉) A kind of blast furnace hanging diagnostic method for handling conflicting information

Also Published As

Publication number Publication date
JPWO2013012093A1 (en) 2015-02-23

Similar Documents

Publication Publication Date Title
Ciliberto et al. Convex learning of multiple tasks and their structure
JP6209879B2 (en) Convolutional neural network classifier system, training method, classification method and use thereof
Zhou et al. Learning from the wisdom of crowds by minimax entropy
US8886574B2 (en) Generalized pattern recognition for fault diagnosis in machine condition monitoring
JP6498107B2 (en) Classification apparatus, method, and program
US10318874B1 (en) Selecting forecasting models for time series using state space representations
WO2019055499A1 (en) Systems and methods for predicting chemical reactions
US20130151442A1 (en) Method for learning task skill and robot using thereof
JP6003492B2 (en) Character recognition device and program
Afsari et al. Group action induced distances for averaging and clustering linear dynamical systems with applications to the analysis of dynamic scenes
EP3745309A1 (en) Training a generative adversarial network
Horak et al. Classification of SURF image features by selected machine learning algorithms
JP6962123B2 (en) Label estimation device and label estimation program
WO2016084326A1 (en) Information processing system, information processing method, and recording medium
WO2013012093A1 (en) Information processing system, method of learning recognition dictionary, and information processing program
US20230222392A1 (en) Computer-readable recording medium storing detection program, detection method, and detection device
CN109389217B (en) Learning method based on Grassmann kernel
CN114422450B (en) Network traffic analysis method and device based on multi-source network traffic data
WO2014073366A1 (en) Information processing system, recognition dictionary learning method, and recognition dictionary learning program
EP2889724B1 (en) System and method for selecting features for identifying human activities in a human-computer interacting environment
Luebke et al. Linear dimension reduction in classification: adaptive procedure for optimum results
Giampouras et al. Online Bayesian low-rank subspace learning from partial observations
Ying et al. Network topology inference with sparsity and Laplacian constraints
CN116467102B (en) Fault detection method and device based on edge algorithm
US20140119641A1 (en) Character recognition apparatus, character recognition method, and computer-readable medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 12814908

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2013524761

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 12814908

Country of ref document: EP

Kind code of ref document: A1