WO2022249392A1 - 情報処理装置、情報処理方法、及びプログラム - Google Patents
情報処理装置、情報処理方法、及びプログラム Download PDFInfo
- Publication number
- WO2022249392A1 WO2022249392A1 PCT/JP2021/020174 JP2021020174W WO2022249392A1 WO 2022249392 A1 WO2022249392 A1 WO 2022249392A1 JP 2021020174 W JP2021020174 W JP 2021020174W WO 2022249392 A1 WO2022249392 A1 WO 2022249392A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- training
- examples
- machine learning
- artificial
- information processing
- Prior art date
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 87
- 238000003672 processing method Methods 0.000 title claims description 35
- 238000012549 training Methods 0.000 claims abstract description 440
- 238000010801 machine learning Methods 0.000 claims abstract description 204
- 238000000034 method Methods 0.000 claims description 117
- 230000002194 synthesizing effect Effects 0.000 claims description 42
- 230000006870 function Effects 0.000 claims description 15
- 238000004422 calculation algorithm Methods 0.000 claims description 14
- 238000003066 decision tree Methods 0.000 claims description 14
- 238000002372 labelling Methods 0.000 claims description 9
- 238000012545 processing Methods 0.000 description 30
- 238000010586 diagram Methods 0.000 description 25
- 238000011156 evaluation Methods 0.000 description 12
- 230000000694 effects Effects 0.000 description 9
- 230000015572 biosynthetic process Effects 0.000 description 5
- 230000010354 integration Effects 0.000 description 5
- 238000012706 support-vector machine Methods 0.000 description 5
- 238000003786 synthesis reaction Methods 0.000 description 5
- 239000000284 extract Substances 0.000 description 4
- 102220470087 Ribonucleoside-diphosphate reductase subunit M2_S20A_mutation Human genes 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 3
- 230000002950 deficient Effects 0.000 description 3
- 238000005070 sampling Methods 0.000 description 3
- 101000746134 Homo sapiens DNA endonuclease RBBP8 Proteins 0.000 description 2
- 101000969031 Homo sapiens Nuclear protein 1 Proteins 0.000 description 2
- 102100021133 Nuclear protein 1 Human genes 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 238000007637 random forest analysis Methods 0.000 description 2
- 240000007594 Oryza sativa Species 0.000 description 1
- 235000007164 Oryza sativa Nutrition 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000007667 floating Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 235000009566 rice Nutrition 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
Definitions
- the present invention relates to technology for generating examples used in machine learning.
- Non-Patent Document 1 the minority class instance (training example) closest to the decision boundary of the Support Vector Machine and the neighboring minority class instance are synthesized to create a minority class virtual instance. It is described to generate
- Non-Patent Document 1 has the problem that virtual instances (artificial examples) are generated near the decision boundary, and artificial examples are not generated in regions other than near the decision boundary where training examples are insufficient. rice field.
- One aspect of the present invention has been made in view of the above problems, and an example of its purpose is to provide a technique capable of generating artificial examples in areas where training examples used for machine learning training are lacking. That is.
- An information processing apparatus provides a machine learning model group including a plurality of acquisition means for acquiring a plurality of training examples, and a plurality of machine learning models for outputting a prediction result from input of the training examples.
- a selection means for selecting, from among the plurality of training examples, a training example having variations in a plurality of prediction results obtained using the machine learning model group after training; and the plurality of training generating means for generating artificial examples by synthesizing two or more of the training examples including the selected training example.
- An information processing method includes obtaining a plurality of training examples, and generating a machine learning model group including a plurality of machine learning models that take the training examples as input and output a prediction result using the plurality of training examples. training, selecting, from among the plurality of training examples, a training example in which a plurality of prediction results obtained using the group of machine learning models after training vary, and among the plurality of training examples, Combining two or more training examples containing the selected training examples to generate an artificial example.
- a program according to one aspect of the present invention is a program for causing a computer to function as an information processing apparatus, the computer comprising an acquisition unit that acquires a plurality of training examples, and a machine that receives the examples as input and outputs a prediction result.
- training means for training a machine learning model group including a plurality of learning models using the plurality of training examples; and a plurality of prediction results obtained by using the machine learning model group after training among the plurality of training examples.
- selection means for selecting a training example with variation in the training examples; and generation means for generating an artificial example by synthesizing two or more training examples including the selected training example from among the plurality of training examples. .
- artificial examples can be generated in areas lacking training examples used for machine learning training.
- FIG. 1 is a block diagram showing the configuration of an information processing device according to exemplary Embodiment 1 of the present invention
- FIG. FIG. 3 is a flow diagram showing the flow of an information processing method according to exemplary embodiment 1 of the present invention
- FIG. 4 is a diagram schematically showing a specific example of an information processing method according to exemplary embodiment 1 of the present invention
- FIG. 2 schematically illustrates an artificial example generated by exemplary embodiment 1 of the present invention
- FIG. 7 is a block diagram showing the configuration of an information processing apparatus according to exemplary Embodiment 2 of the present invention
- FIG. 7 is a flow diagram showing the flow of an information processing method according to exemplary embodiment 2 of the present invention
- FIG. 7 is a diagram schematically showing a specific example of an information processing method according to exemplary embodiment 2 of the present invention
- FIG. 10 is a flow diagram showing the flow of a first generation process according to exemplary embodiment 2 of the present invention
- FIG. 11 is a flow diagram showing the flow of a second generation process according to exemplary embodiment 2 of the present invention
- FIG. 11 is a flow diagram showing the flow of a third generation process according to exemplary embodiment 2 of the present invention
- FIG. 11 is a flow diagram showing the flow of an information processing method according to exemplary embodiment 3 of the present invention
- FIG. 11 is a flow diagram showing the flow of an information processing method according to exemplary embodiment 4 of the present invention
- FIG. 11 is a diagram schematically illustrating an information processing method according to exemplary embodiment 5 of the present invention
- FIG. 4 is a diagram schematically explaining an artificial example generated by the technique described in Non-Patent Document 1
- 1 is a block diagram showing the configuration of a computer functioning as an information processing device according to exemplary embodiments 1 to 5 of the present invention
- FIG. 4 is a diagram schematically explaining an artificial example generated by the technique described in Non-Patent Document 1
- 1 is a block diagram showing the configuration of a computer functioning as an information processing device according to exemplary embodiments 1 to 5 of the present invention
- FIG. 1 is a block diagram showing the configuration of an information processing device 10. As shown in FIG.
- the information processing device 10 is a device that generates artificial examples from a plurality of training examples using a machine learning model group.
- the information processing device 10 includes an acquisition unit 11, a training unit 12, a selection unit 13, and a generation unit 14, as shown in FIG.
- the acquisition unit 11 is an example of a configuration that implements the acquisition means described in the claims.
- the training unit 12 is an example of a configuration that implements training means described in the claims.
- the selection unit 13 is an example of a configuration that implements the selection means described in the claims.
- the generation unit 14 is an example of a configuration that implements the generation means described in the claims.
- the acquisition unit 11 acquires a plurality of training examples.
- the training unit 12 trains a machine learning model group including a plurality of machine learning models that output prediction results using examples as input, using a plurality of training examples.
- the selection unit 13 selects, from among the plurality of training examples, training examples in which there are variations in the plurality of prediction results obtained using the machine learning model group after training.
- the generating unit 14 generates an artificial example by synthesizing two or more training examples including the selected training example from among the plurality of training examples.
- a machine learning model group includes a plurality of machine learning models. Each machine learning model takes an example as an input and outputs a prediction result.
- a prediction result may include, for example, a predicted probability that each of a plurality of labels is predicted. In this case, the label with the highest prediction probability may be referred to as the prediction result.
- a machine learning model is, by way of example, a model generated using a machine learning algorithm such as a decision tree, neural network, random forest, or support vector machine. However, the machine learning algorithms used to generate each machine learning model are not limited to these.
- the multiple machine learning models may all be models generated using the same machine learning algorithm. At least two of the plurality of machine learning models may be models generated using different machine learning algorithms.
- the machine learning model group may be stored, for example, in the memory of the information processing device 10 or may be stored in another device communicably connected to the information processing device 10 .
- the machine learning model group does not necessarily have to be "training target machine learning models to be trained using generated artificial examples”.
- the group of machine learning models may include some or all of the machine learning models to be trained.
- the machine learning model group does not have to include the machine learning model to be trained.
- the number of machine learning models to be trained may be plural or singular.
- An example is information that is input to each machine learning model and includes features. In other words, examples exist in feature space.
- a training example is an example used to train a set of machine learning models. The training examples may be examples obtained by observation or may be artificial examples that are artificially generated.
- a training example in which a plurality of prediction results vary is a training example in which the evaluation result of variation indicates that the variation is large.
- evaluation of variability means evaluating whether or not the variability of a plurality of prediction results is large.
- the variability evaluation may be an evaluation based on the entropy of voting results. The entropy of voting results will be described in detail in exemplary embodiment 2 below.
- the evaluation of variation may be an evaluation based on the ratio of prediction results showing the same label among a plurality of prediction results.
- the evaluation of variation is not limited to the one described above.
- a training example evaluated as having a large variation in a plurality of prediction results is also referred to as a “training example with a variation in prediction results”.
- a training example evaluated as not having a large variation in a plurality of prediction results is also described as a “training example having a small variation in prediction results”.
- FIG. 2 is a flow diagram showing the flow of the information processing method S10. As shown in FIG. 2, the information processing method S10 includes steps S101 to S104.
- step S101 acquisition process
- the acquisition unit 11 acquires a plurality of training examples.
- the acquisition unit 11 may acquire a plurality of training examples by reading them from memory.
- the acquisition unit 11 may acquire a plurality of training examples from an input device or from a device connected via a network.
- the plurality of training examples obtained in this step includes one or both of observed examples and artificial examples.
- step S102 training process
- the training unit 12 trains a group of machine learning models using a plurality of training examples acquired in step S101.
- the training examples used for training each machine learning model group may be a part or all of the plurality of training examples acquired in step S101.
- step S103 In step S ⁇ b>103 (selection processing), the selection unit 13 selects training examples having variations in prediction results from among the plurality of training examples.
- the selection unit 13 may select one such training example, or may select a plurality of such training examples.
- the selection unit 13 inputs a training example to be evaluated among a plurality of training examples to each machine learning model after training, and acquires a prediction result output from each machine learning model. Thereby, the selection unit 13 obtains a plurality of prediction results for the training example to be evaluated. In addition, the selection unit 13 evaluates variations in the obtained prediction results. If the selection unit 13 evaluates that a plurality of prediction results have large variations, the selection unit 13 selects the training example as "a training example having variations in prediction results."
- the selection unit 13 makes some or all of the plurality of training examples subject to variation evaluation. For example, when each machine learning model group is trained using a part of a plurality of training examples, the selection unit 13 selects the other part of the plurality of training examples (i.e., the machine learning model group was not used for training training examples) may be evaluated.
- step S104 generation processing
- the generation unit 14 generates an artificial example by synthesizing two or more training examples including the training example selected in step S103 among the plurality of training examples. For example, the generation unit 14 may synthesize the selected training example with other training examples that exist nearby in the feature space. Further, for example, when a plurality of training examples are selected in step S103, the generation unit 14 may synthesize the selected training examples.
- the generating unit 14 may generate one artificial example by synthesizing two training examples, or may generate one artificial example by synthesizing three or more training examples. In this step, the generation unit 14 may generate one artificial example, or may generate a plurality of artificial examples.
- the synthesizing process performed by the generating unit 14 is represented by the following equation (1) as an example.
- Equation (1) ⁇ xv represents an artificial example, and x i represents a training example selected by the selection unit 13 .
- x j may be another training example selected by the selection unit 13, or may be another training example that is not selected. If it is another training example that has not been selected, then x j is a training example that is in the neighborhood of x i .
- ⁇ is a weighting coefficient that satisfies 0 ⁇ 1.
- the generation unit 14 determines the value of the coefficient ⁇ using random numbers generated by a random function. Note that the synthesizing process performed by the generating unit 14 is not limited to the method described above, and the generating unit 14 may synthesize a plurality of training examples by other methods.
- FIG. 3 is a diagram schematically showing a specific example of the information processing method S10.
- the plurality of training examples T acquired by the acquisition unit 11 in step S101 include training examples t1, t2, t3, .
- the machine learning model group trained by the training unit 12 in step S102 includes machine learning models m1, m2, m3, . Each of the machine learning models m1, m2, m3, . Each of the machine learning models m1, m2, m3, .
- the selection unit 13 evaluates variations in a plurality of prediction results for training examples t1 to t10 to be evaluated.
- solid-line circles indicate training examples with variability in prediction results
- dashed-line circles indicate training examples with small variability in prediction results.
- a plurality of prediction results obtained from machine learning models m1, m2, m3, . is evaluated.
- training examples t6, t9, and t10 a plurality of prediction results obtained from machine learning models m1, m2, m3, .
- training examples t3 and t4 two of a plurality of prediction results obtained from machine learning models m1, m2, m3, . . . are "A" and one is "B.” is evaluated as large.
- training examples t7 and t8 two of a plurality of prediction results obtained from machine learning models m1, m2, m3, . . . are "B" and one is "A.” is evaluated as large.
- the selection unit 13 selects training examples t3, t4, t7, and t8 with variations in prediction results.
- the generation unit 14 generates an artificial example t51 by synthesizing the training example t3 with variations in prediction results and the neighboring training example t5. Further, the generation unit 14 generates an artificial example t52 by synthesizing the training example t4 in which the prediction results vary and the neighboring training example t1. Further, the generation unit 14 generates an artificial example t53 by synthesizing a plurality of training examples t7 and t8 having different prediction results. In FIG. 3, the double-lined circle indicates an artificial example.
- the exemplary embodiment obtains a plurality of training examples, trains a machine learning model family including a plurality of machine learning models that take the examples as input and output a prediction result using the plurality of training examples, and trains the plurality of training examples. From among the training examples, select training examples with variations in the prediction results obtained using the machine learning model group after training, and synthesize two or more training examples including the selected training examples from among the plurality of training examples. and to generate an artificial example.
- training examples with multiple prediction results that vary are considered to be in areas where there are insufficient training examples in the feature space.
- Artificial examples obtained by synthesizing multiple training examples containing such training examples are likely to be generated in areas lacking training examples.
- the present exemplary embodiment can generate artificial examples in regions lacking training examples.
- FIG. 4 is a diagram schematically illustrating an artificial example generated by this exemplary embodiment.
- FIG. 14 is a diagram schematically explaining an artificial example generated by the technique described in Non-Patent Document 1.
- FIG. 4 and 14 solid-line circles indicate training examples with small variations in prediction results, dashed-line circles indicate training examples with variations in prediction results, and double-line circles indicate artificial examples. show.
- Regions R1, R2, and R3 indicate regions in the feature amount space. Regions R1, R2, and R3 contain artificial examples with varying prediction results, and are regions lacking artificial examples.
- Non-Patent Document 1 As shown in FIG. 14, the technique described in Non-Patent Document 1 generates an artificial example in a region R1 near the decision boundary B by the support vector machine. However, with the technique described in Non-Patent Document 1, it is difficult to generate artificial examples in regions R2 and R3 other than the vicinity of the decision boundary B, where training examples are insufficient.
- the exemplary embodiment generates artificial examples by synthesizing multiple training examples, including training examples with varying prediction results.
- the present exemplary embodiment can generate artificial examples in regions R1, R2, R3 lacking training examples.
- this exemplary embodiment can suppress the generation of artificial examples biased toward a partial region R1.
- this exemplary embodiment employs a configuration in which a group of machine learning models are used to select training examples with varying prediction results.
- this exemplary embodiment suppresses the generation of artificial examples in biased regions compared to the case of generating artificial examples in the vicinity of the decision boundary like the technique described in Non-Patent Document 1. be able to.
- This also allows the present exemplary embodiment to generate artificial examples in regions that are more deficient in training examples compared to using prediction probabilities to select training examples with varying prediction results. be able to. This is because, for example, when a decision tree is included in a group of machine learning models, the decision tree may make an erroneous prediction with a prediction probability of 1.
- FIG. 5 is a block diagram showing the configuration of the information processing device 20.
- the information processing device 20 is a device that generates an artificial example from a plurality of examples using the machine learning model group COM0.
- the machine learning model group COM0 is configured almost similarly to the machine learning model group in the first exemplary embodiment. However, in this exemplary embodiment, at least two of the plurality of machine learning models included in the machine learning model group COM0 were generated using different machine learning algorithms.
- the machine learning model group COM0 includes machine learning models that are training targets to be trained using artificial examples.
- At least one machine learning model included in the machine learning model group COM0 is a decision tree.
- the machine learning model being trained is a decision tree.
- the information processing device 20 includes, as shown in FIG. .
- the acquisition unit 21 is an example of a configuration that implements the acquisition means described in the claims.
- the training unit 22 is an example of a configuration that implements training means described in the claims.
- the selection unit 23 is an example of a configuration that implements the selection means described in the claims.
- the generation unit 24 is an example of a configuration that implements the generation means described in the claims.
- the labeling unit 25 is an example of a configuration that implements the labeling means described in the claims.
- the output unit 26 is an example of a configuration that implements the output means described in the claims.
- the acquisition unit 21 is configured in the same manner as the acquisition unit 11 in the first exemplary embodiment.
- the training unit 22 is configured almost similarly to the training unit 12 in exemplary embodiment 1, but differs in that the machine learning model group COM0 is divided into a plurality of groups for training. Details of the training process by the training unit 22 will be described later.
- the selection unit 23 is configured in substantially the same manner as the selection unit 23 in exemplary embodiment 1, but differs in the details of the training examples to be evaluated for variations in prediction results. The details of the training examples to be evaluated will be described later.
- the generating unit 24 generates artificial examples by synthesizing the training examples selected by the selecting unit 23 and examples existing in the vicinity of the selected training examples in the feature space. As an example, the generator 24 generates an artificial example by synthesizing two training examples according to Equation (1) above.
- the labeling unit 25 labels some or all of the training examples and artificial examples.
- the label assigning unit 25 may assign a label based on information output from an input device that accepts user operations. Further, as an example, the labeling unit 25 may give labels obtained by inputting training examples and artificial examples to a machine learning model trained to output labels with examples as input.
- the machine learning model that outputs the label is a machine learning model different from each machine learning model included in the machine learning model group COM0.
- the machine learning model that outputs the label is preferably a model with higher prediction accuracy than at least one machine learning model included in the machine learning model group. For example, if the machine learning model to be trained included in the machine learning model group is a decision tree, the machine learning model that outputs the label may be a random forest.
- the output unit 26 outputs the artificial example generated by the generation unit 24.
- the output unit 26 may store the artificial example generated by the generation unit 24 in a recording medium such as an external storage device.
- the output unit 26 may output the artificial example to an output device such as a display device, for example.
- the control unit 27 controls each unit of the information processing device 20 .
- the control unit 27 particularly adds the artificial examples generated by the generation unit 24 to a plurality of training examples, and causes the acquisition unit 21, the training unit 22, the selection unit 23, and the generation unit 24 to function again. .
- FIG. 6 is a flow diagram showing the flow of the information processing method S20.
- step S201 acquisition process
- the acquisition unit 11 acquires a plurality of training examples.
- the plurality of training examples to be acquired may include examples obtained by observation, or may include artificial examples.
- Step S202 In step S ⁇ b>202 , the label assigning unit 25 assigns a label to each of the plurality of training examples acquired by the acquiring unit 21 .
- step S203 training process
- the training unit 22 trains each of the multiple machine learning model groups using some or all of the multiple training examples acquired by the acquisition unit 21 .
- the details of the training process for training each machine learning model group will be described later.
- step S204 In step S ⁇ b>204 (selection processing), the selection unit 23 selects one or more training examples with variations in prediction results from among the plurality of training examples acquired by the acquisition unit 21 .
- the selection processing performed by the selection unit 23 will be described later.
- the details of the process of selecting training examples with varying prediction results will be described later.
- step S205 generation processing
- the generation unit 24 identifies a plurality of training examples including the training example selected by the selection unit 23 as synthesis targets.
- the generation unit 24 also generates an artificial example by synthesizing a plurality of training examples specified as synthesis targets. The details of the generation process performed by the generation unit 24 will be described later.
- step S206 the label assigning unit 25 assigns a label to each of the artificial examples generated by the generating unit 24.
- FIG. in step S207 the control unit 27 determines whether to end the training process. As an example, when the number of times the processes of steps S203 to S206 have been executed is equal to or greater than a predetermined threshold, the control unit 27 determines to end the training process. On the other hand, when the number of times the processes of steps S203 to S206 have been executed is less than the predetermined threshold, it is determined that the training process is not finished. If the training process is not to end (NO in step S207), the control unit 27 proceeds to the process of step S208. On the other hand, when ending the training process (YES in step S207), the control unit 27 proceeds to the process of step S209.
- Step S208 the control unit 27 adds one or more artificial examples generated in step S206 executed so far to a plurality of training examples. After completing the process of step S208, the control unit 27 returns to the process of step S203. In other words, the control unit 27 adds the artificial example to the plurality of training examples and causes the acquisition unit 21, the training unit 22, the selection unit 23, and the generation unit 24 to function again.
- step S209 the output unit 26 outputs one or more artificial examples generated in step S206 executed so far.
- One or more artificial examples thus generated using information processing method S20 are used to train a machine learning model to be trained.
- the training unit 22 may perform the process of training the machine learning model to be trained.
- FIG. 7 is a diagram schematically showing a specific example of the information processing method S20.
- each divided group will be referred to as a machine learning model group COMi.
- a plurality of machine learning models included in the machine learning model group COMi are hereinafter referred to as machine learning models mi-j.
- the plurality of machine learning models mi-j included in the machine learning model group COMi may all be models generated by the same machine learning algorithm, or at least two of them may be models generated by mutually different machine learning algorithms. It can be a model.
- the training unit 22 extracts the training example group Di from the training example group T acquired by the acquisition unit 21 in step S201.
- Training example group Di is part of training example group T.
- the training unit 22 may extract the training example group Di by random sampling.
- the training example groups Di1 and Di2 may contain all the same training examples, or some or all of them may be different.
- the selection unit 23 evaluates the variability of each training example to be evaluated, using an index of vote entropy in the QBC (query by committee) method.
- the following equation (2) is an equation that indicates the training example ⁇ x for which the entropy of the voting result is the maximum.
- C represents the total number of machine learning models mi-j in the machine learning model group COMi.
- V(y) indicates the number of machine learning models mi-j that have predicted the label y in the machine learning model group COMi.
- the selection unit 23 may select the training example ⁇ x indicated by the equation (2) as a training example having variations in prediction results. In this case, the number of training examples with variations in prediction results selected by the selection unit 23 for each machine learning model group COMi is one.
- the selection unit 23 uses the M machine learning model group COMi to select M training examples with varying prediction results. For each machine learning model group COMi, the selection unit 23 may select a predetermined number of training examples in descending order of entropy of the voting result, or may select training examples whose entropy of the voting result is equal to or greater than a threshold. good. In this case, the number of training examples with variations in prediction results selected by the selection unit 23 for each machine learning model group COMi can be plural. In other words, in this case, the selection unit 23 selects M or more training examples with variations in prediction results using the M machine learning model group COMi. Further, the selection unit 23 may randomly select one or a predetermined number from among the M or more training examples with variations in the prediction results selected in this way, or the entropy of the voting results may be One or a predetermined number may be selected in descending order.
- step S205 the generation unit 24 performs any one of the first generation processing S30, the second generation processing S40, and the third generation processing S50 using training examples having variations in the prediction results, thereby generating artificial training examples.
- the first generation process S30 is a process of generating an artificial example by synthesizing training examples having variations in prediction results and neighboring training examples.
- the second generation process is a process of generating an artificial example by synthesizing two or more training examples with varying prediction results.
- the third generation process is a process of selectively executing either the first generation process or the second generation process.
- the first generation processing S30 will be described with reference to FIG.
- FIG. 8 is a flowchart showing the flow of the first generation processing S30.
- the first generation process S30 includes steps S301 and S302.
- step S204 previously performed, one or a plurality of training examples with varying prediction results are selected.
- the generation unit 24 executes the following steps S301 and S302 for each of one or more training examples (hereinafter referred to as the training examples) having variations in prediction results.
- Step S301 the generating unit 24 selects training examples near the training example.
- Neighboring training examples may be training examples with varying prediction results or training examples with less variation in prediction results.
- the neighboring training example may be a training example in the training example group T that is closest to the training example in the feature space.
- the neighboring training example may be a training example whose distance in the feature space from the training example in the training example group T is equal to or less than a threshold.
- Step S302 the generating unit 24 generates an artificial example by synthesizing the training example and the neighboring training example selected in step S301.
- an artificial example tv1-1 is generated by synthesizing a training example t1 with variations in the prediction results and neighboring training examples.
- an artificial example tv1-2 is generated by synthesizing the training example t2 with variations in prediction results and its neighboring training examples.
- the generation unit 24 may use the above formula (1) as an example of synthesis processing.
- the generating unit 24 may use known techniques such as MUNGE (see reference 1) and SMOTE (see reference 2) as other examples of combining processing.
- FIG. 9 is a flowchart for explaining the flow of the second generation processing S40.
- the second generation process S40 includes steps S401 and S402.
- the second generation process can be executed when a plurality of training examples with different prediction results are selected in step S204.
- the generation unit 24 executes the following steps S401 and S402 for each of a plurality of training examples (hereinafter referred to as training examples) having variations in prediction results.
- Step S401 the selection unit 23 selects another training example different from the training example from among the plurality of training examples having variations in prediction results.
- the selection unit 23 selects two or more training examples having a plurality of variations in prediction results from among the plurality of training examples.
- the selection unit 23 may randomly select such other training examples from a plurality of training examples with varying prediction results.
- the selection unit 23 selects, as such another training example, one having the smallest distance in the feature amount space from the training example among a plurality of training examples having variations in prediction results, or a distance equal to or less than a threshold. can be selected. Note that if the training example has already been used for synthesis, the processing of steps S401 and S402 relating to the training example need not be executed.
- Step S402 the generation unit 24 generates an artificial example by synthesizing the training example and the other training example selected in step S401.
- the artificial example tv2-1 is generated by synthesizing the training examples t11 and t12 with different prediction results.
- the two or more training examples synthesized by the generation unit 24 may be selected using the same machine learning model group COMi as in this example, or at least one of them may be different from the others. It may be selected using the machine learning model group COMi.
- the generator 24 selects two or more training examples from training examples t1, t2, . . . , t11, t12, .
- step S402 It may be synthesized to produce artificial examples tv1-1, tv1-2, tv2-1, or tv2-2. Note that the technique used in the combining process in step S402 is the same as described in step S302, so detailed description will not be repeated.
- FIG. 10 is a flowchart for explaining the flow of the third generation process S50.
- the third generation process S50 includes steps S501 to S503.
- step S204 one or more training examples with varying prediction results are selected.
- the generation unit 24 executes the following steps S501 to S503 for each of one or more training examples (hereinafter referred to as the training examples) having variations in prediction results.
- Step S501 the generator 24 selects either the first generation process or the second generation process.
- the generator 24 may select the first generation process using a probability p determined by a random function, and select the second generation process if the first generation process is not selected.
- the method of selecting either the first generation process or the second generation process is not limited to the method using the probability p, and may be another method.
- Step S502 the generation unit 24 determines which one is selected. When the first generation process is selected, the generation unit 24 proceeds to the process of step S503 and executes the first generation process. On the other hand, when the second generation process is selected, the generation unit 24 proceeds to the process of step S504 and executes the second generation process. The details of the first generation process and the second generation process are as described above.
- This exemplary embodiment has a configuration of executing a first generation process for generating an artificial example by synthesizing a training example with a variation in prediction results and instruction examples in the vicinity thereof.
- the artificial examples generated by the first generation process are generated in the vicinity of the training examples in which the prediction results vary.
- Training examples with varying prediction results are considered to be regions in the feature space that are deficient in training examples.
- such artificial examples are generated in regions lacking training examples.
- this exemplary embodiment has a configuration of executing a second generation process of synthesizing two or more training examples having variations in prediction results to generate an artificial example.
- the artificial examples generated by the second generation process are obtained by synthesizing training examples in a region lacking training examples. Therefore, areas where such artificial examples exist are also likely to be deficient in training examples.
- this exemplary embodiment has a configuration in which either the first generation process or the second generation process is selected and executed to execute the third generation process for generating an artificial example.
- the artificial example generated by the third generation process is generated by the first generation process or the second generation process.
- the region in which the artificial example is generated by the first generation process and the region in which the artificial example is generated by the second generation process can be different. Therefore, when a plurality of artificial examples are generated by the third generation process, there is a high possibility that these artificial examples will be generated distributed over more diverse regions lacking training examples.
- the exemplary embodiment generates artificial examples that are biased toward regions where training examples are sufficient by performing any of the first generation process, the second generation process, and the third generation process. Instead, artificial examples can be generated in areas where training examples are more scarce.
- the group of machine learning models includes a machine learning model to be trained. This allows the exemplary embodiments to generate artificial examples that are more effective at improving the accuracy of the machine learning model under training.
- this exemplary embodiment employs a configuration in which at least two of the machine learning model groups are models generated by mutually different machine learning algorithms.
- the machine learning model to be trained is a decision tree and not a support vector machine.
- such training target machine learning models are included in the machine learning model group COM0. Therefore, compared to the technique described in Non-Patent Document 1, which generates artificial examples near the decision boundary of the support vector machine, the present exemplary embodiment provides artificial examples that are more effective in improving the accuracy of the machine learning model to be trained. can be generated.
- generator 24 generates a plurality of artificial examples.
- the generation unit 24 integrates two artificial examples that satisfy the similarity condition into one artificial example, among the plurality of generated artificial examples.
- the similarity condition is a condition indicating that examples are similar.
- the similarity condition may be, for example, that the cosine similarity is greater than or equal to a threshold, or that the distance in the feature amount space is less than or equal to the threshold.
- similar conditions are not limited to these. Details of the process of integration will be described later.
- FIG. 11 is a flow diagram illustrating the flow of the information processing method S20A according to the third exemplary embodiment.
- the information processing method S20A shown in FIG. 11 is configured substantially in the same manner as the information processing method S20 according to exemplary embodiment 2, but differs in that step S205A is further included.
- Step S205A the generation unit 24 integrates two similar artificial examples among the artificial examples generated in step S205. Specifically, the generation unit 24 determines whether or not any of the artificial example generated in step S205 this time and the artificial example generated in step S205 up to the previous time satisfies the similarity condition. If it is determined that the similarity condition is met, the generation unit 24 integrates two artificial examples that satisfy the similarity condition.
- An example of the integration process is the process of synthesizing two artificial examples.
- the generator 24 synthesizes two artificial examples to generate one artificial example, and deletes the original two artificial examples that satisfy the similarity condition.
- Another example of integration processing is processing for deleting one of two artificial examples. Note that the integration process is not limited to the process described above, as long as one artificial example generated by referring to the two artificial examples is used instead of the two artificial examples that satisfy the similarity condition.
- deleting an artificial example means deleting it from objects to be labeled in step S206 and objects to be added to training examples in step S208. As a result, the integrated artificial examples are labeled and added to the training examples.
- the generation unit generates a plurality of artificial examples, and integrates two artificial examples satisfying similarity conditions among the plurality of generated artificial examples into one artificial example. Adopted.
- the present exemplary embodiment generates artificial examples that can more efficiently improve the accuracy of machine learning models in areas lacking training examples by integrating artificial examples that satisfy similarity conditions. can do.
- the generation unit 24 outputs artificial examples having variations in a plurality of prediction results obtained using the post-training machine learning model group COM0 among the generated artificial examples.
- an artificial example with variation is an artificial example showing that the evaluation result of variation indicates "large variation”. The details of the variability evaluation have been described above and will not be repeated.
- the generation unit 24 performs post-evaluation of variations in the generated artificial examples using the post-training machine learning model group COM0, and adopts artificial examples having variations in prediction results from the post-evaluation.
- FIG. 12 is a flow diagram illustrating the flow of the information processing method S20B according to the fourth illustrative embodiment.
- the information processing method S20B shown in FIG. 12 is configured substantially in the same manner as the information processing method S20 according to exemplary embodiment 2, but differs in that step S205B is further included.
- Step S205B In step S205B, the generation unit 24 post-evaluates the artificial examples generated in step S205.
- the generating unit 24 uses the machine learning model group COM0 to evaluate the variation in prediction results for the artificial example. For example, in the example shown in FIG. 7, the generation unit 24 evaluates the variation of the prediction results for the artificial example tv1-1 using the machine learning model group COM1.
- the machine learning model group COM1 used to evaluate the variability is desirably the one used to evaluate the training example t1 referenced to generate the artificial example tv1-1.
- the details of the process of evaluating the variability of prediction results using the machine learning model group COM0 are as described in the second exemplary embodiment.
- the generation unit 24 determines that the prediction results for the artificial example generated in step S205 do not vary greatly, the generation unit 24 deletes the artificial example.
- deleting an artificial example means deleting it from objects to be labeled in step S206 and objects to be added to training examples in step S208. As a result, artificial examples with varying prediction results are labeled and added to the training examples.
- ⁇ Effects of this exemplary embodiment> a configuration is adopted in which the generation unit outputs artificial examples in which there are variations in a plurality of prediction results obtained using the machine learning model group after training, among the generated artificial examples. It is
- artificial examples obtained by synthesizing multiple training examples, including training examples with varying prediction results do not necessarily have varying prediction results.
- artificial examples generated in this way may have small variations in prediction results.
- Training a machine learning model using training examples with small variations in prediction results is not efficient in improving the accuracy of the machine learning model.
- the present exemplary embodiment generates artificial examples that can more efficiently improve the accuracy of machine learning models in areas lacking training examples by post-evaluating the generated artificial examples. be able to.
- This exemplary embodiment is a form in which the configuration of the machine learning model group COM0 and steps S203 to S204 in the information processing method S20 in exemplary embodiment 2 are modified as follows. This exemplary embodiment will be described with reference to FIG. FIG. 13 is a diagram schematically illustrating an information processing method according to this exemplary embodiment.
- Each machine learning model mj is a model generated by the same machine learning algorithm.
- each machine learning model mj may be a decision tree.
- Step S203 the training unit 22 extracts a training example group Dj from the training example group T acquired by the acquisition unit 21 in step S201.
- Training example group Dj is part of training example group T.
- the training unit 22 may extract the training example group Dj by random sampling.
- Machine learning model groups mj1 and mj2 are trained such that the parameters constituting each are different from each other by using training example groups Dj1 and Dj2, which are at least partially different.
- Step S204 of this exemplary embodiment the selection unit 23 evaluates the variation of prediction results for each training example included in the training example group T using the machine learning model group COM0. Further, the selection unit 23 selects training examples with variations in prediction results. In the example of FIG. 13, the selection unit 23 selects training examples t1, t3, .
- step S205 is as described in the second exemplary embodiment. That is, in the example of FIG. 13, any one of the first generation process, the second generation process, and the third generation process is executed for each of the training examples t1, t2, . As a result, artificial examples tv1, tv2, . . . are generated.
- This exemplary embodiment uses models generated by the same machine learning algorithm as machine learning models constituting a group of machine learning models, and selects training examples having variations in prediction results from the acquired training example group. Choose, and adopt the configuration.
- this exemplary embodiment can generate effective artificial examples by improving the accuracy of the machine learning model when all the machine learning included in the machine learning model group is a decision tree. I will explain why. Decision trees can vary greatly in term structure for small changes in the training examples. Therefore, by using a group of machine learning models including a plurality of decision trees, it is possible to more accurately select training examples with varying prediction results.
- Some or all of the functions of the information processing apparatuses 10 and 20 may be realized by hardware such as integrated circuits (IC chips), or may be realized by software.
- the information processing apparatuses 10 and 20 are implemented by computers that execute instructions of programs, which are software that implements each function, for example.
- An example of such a computer (hereinafter referred to as computer C) is shown in FIG.
- Computer C comprises at least one processor C1 and at least one memory C2.
- a program P for operating the computer C as the information processing apparatuses 10 and 20 is recorded in the memory C2.
- the processor C1 reads the program P from the memory C2 and executes it, thereby realizing each function of the information processing apparatuses 10 and 20 .
- processor C1 for example, CPU (Central Processing Unit), GPU (Graphic Processing Unit), DSP (Digital Signal Processor), MPU (Micro Processing Unit), FPU (Floating point number Processing Unit), PPU (Physics Processing Unit) , a microcontroller, or a combination thereof.
- memory C2 for example, a flash memory, HDD (Hard Disk Drive), SSD (Solid State Drive), or a combination thereof can be used.
- the computer C may further include a RAM (Random Access Memory) for expanding the program P during execution and temporarily storing various data.
- Computer C may further include a communication interface for sending and receiving data to and from other devices.
- Computer C may further include an input/output interface for connecting input/output devices such as a keyboard, mouse, display, and printer.
- the program P can be recorded on a non-temporary tangible recording medium M that is readable by the computer C.
- a recording medium M for example, a tape, disk, card, semiconductor memory, programmable logic circuit, or the like can be used.
- the computer C can acquire the program P via such a recording medium M.
- the program P can be transmitted via a transmission medium.
- a transmission medium for example, a communication network or broadcast waves can be used.
- Computer C can also obtain program P via such a transmission medium.
- Appendix 1 an acquisition means for acquiring a plurality of training examples; a training means for training a machine learning model group including a plurality of machine learning models that output prediction results using examples as input, using the plurality of training examples; selection means for selecting, from among the plurality of training examples, a training example in which a plurality of prediction results obtained using the machine learning model group after training varies; generating means for generating an artificial example by synthesizing two or more training examples including the selected training example from among the plurality of training examples; Information processing device with
- artificial examples are generated using training examples with variations in prediction results, so artificial examples are not generated in biased areas such as the vicinity of the decision boundary. Artificial examples can often be generated.
- Appendix 2 The information processing apparatus according to appendix 1, wherein the generating means generates the artificial example by synthesizing the selected training example and an example existing in the vicinity of the selected training example in a feature space.
- the selection means selects, from the plurality of training examples, two or more training examples in which the plurality of prediction results vary;
- the information processing apparatus according to appendix 1, wherein the generating means generates the artificial example by synthesizing two or more of the selected training examples.
- training examples with different prediction results are combined to generate artificial examples, so it is possible to generate artificial examples with high accuracy in regions where training examples are insufficient.
- the generating means is a first generation process of synthesizing the selected training example and examples existing in the vicinity of the selected training example in a feature space; a second generation process that combines two or more of the selected training examples to generate the artificial example;
- the information processing apparatus according to appendix 1, wherein the artificial example is generated by executing any one of:
- either the first generation process or the second generation process is selectively used to generate artificial examples. Artificial examples can be generated.
- the machine learning model group is repeatedly trained using the training examples to which the generated artificial examples are added, so it is possible to more accurately select training examples with varying prediction results.
- artificial examples can be generated in areas where artificial examples are scarce.
- the generating means is generating a plurality of said artificial examples; 5.
- the information processing apparatus according to any one of appendices 1 to 4, wherein two artificial examples satisfying a similarity condition among the plurality of artificial examples are integrated into one artificial example.
- Appendix 7 The method according to any one of appendices 1 to 6, wherein the generating means outputs, among the artificial examples, artificial examples in which a plurality of prediction results obtained using the group of machine learning models after training vary. Information processing equipment.
- Appendix 8 The information processing apparatus according to any one of appendices 1 to 7, wherein the machine learning model group includes a training target machine learning model trained using the artificial example.
- Appendix 9 The information processing apparatus according to any one of appendices 1 to 8, wherein at least two of the machine learning model groups use different machine learning algorithms.
- Appendix 11 11. The information processing device according to any one of appendices 1 to 10, wherein at least one of the machine learning model groups is a decision tree.
- Appendix 12 12. The information processing apparatus according to any one of appendices 1 to 11, further comprising labeling means for labeling some or all of the plurality of training examples and the artificial examples.
- (Appendix 13) obtaining multiple training examples; training, using the plurality of training examples, a machine learning model group including a plurality of machine learning models that output prediction results using examples as input; Selecting, from among the plurality of training examples, a training example in which a plurality of prediction results obtained using the machine learning model group after training varies; synthesizing two or more training examples including the selected training example from among the plurality of training examples to generate an artificial example; Information processing method including.
- a program for causing a computer to function as an information processing device comprising: an acquisition means for acquiring a plurality of training examples; a training means for training a machine learning model group including a plurality of machine learning models that output prediction results using examples as input, using the plurality of training examples; selection means for selecting, from among the plurality of training examples, a training example in which a plurality of prediction results obtained using the machine learning model group after training varies; generating means for generating an artificial example by synthesizing two or more training examples including the selected training example from among the plurality of training examples; A program that acts as a
- Appendix 15 A computer-readable recording medium on which the program according to appendix 14 is recorded.
- said processor comprising: an acquisition process for acquiring a plurality of training examples; a training process of training a machine learning model group including a plurality of machine learning models that output prediction results using examples as input, using the plurality of training examples; A selection process of selecting, from among the plurality of training examples, a training example in which a plurality of prediction results obtained using the machine learning model group after training varies; and a generation process of generating an artificial example by synthesizing two or more training examples including the selected training example from among the plurality of training examples.
- the information processing apparatus may further include a memory, and the memory stores a program for causing the processor to execute the -process, the -process, and the -process. good too.
- this program may be recorded in a computer-readable non-temporary tangible recording medium.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Description
本発明の第1の例示的実施形態について、図面を参照して詳細に説明する。本例示的実施形態は、後述する例示的実施形態の基本となる形態である。
本例示的実施形態に係る情報処理装置10の構成について、図1を参照して説明する。図1は、情報処理装置10の構成を示すブロック図である。情報処理装置10は、複数の訓練用例から、機械学習モデル群を用いて人工用例を生成する装置である。
機械学習モデル群は、複数の機械学習モデルを含む。各機械学習モデルは、用例を入力として予測結果を出力する。予測結果は、例えば、複数のラベルの各々が予測される予測確率を含むものであってもよい。この場合、最も予測確率が高いラベルを、予測結果と記載する場合もある。機械学習モデルは、一例として、決定木、ニューラルネットワーク、ランダムフォレスト、又はサポートベクタマシン等の機械学習アルゴリズムを用いて生成されたモデルである。ただし、各機械学習モデルの生成に用いられる機械学習アルゴリズムは、これらに限られない。複数の機械学習モデルは、全て同一の機械学習アルゴリズムを用いて生成されたモデルであってもよい。また、複数の機械学習モデルのうち少なくとも2つが、互いに異なる機械学習アルゴリズムを用いて生成されたモデルであってもよい。機械学習モデル群は、例えば情報処理装置10のメモリに記憶されていてもよいし、情報処理装置10と通信可能に接続された他の装置に記憶されていてもよい。
用例は、各機械学習モデルに入力される情報であり、特徴量を含む。換言すると、用例は、特徴量空間に存在する。訓練用例は、機械学習モデル群の訓練に用いる用例である。訓練用例は、観測により得られる用例であってもよいし、人工的に生成された人工用例であってもよい。
複数の予測結果にばらつきがある訓練用例とは、ばらつきの評価結果が「ばらつきが大きい」ことを示す訓練用例である。例えば、ばらつきの評価とは、複数の予測結果のばらつきが大きいか否かを評価することである。具体例として、ばらつきの評価は、投票結果のエントロピーに基づく評価であってもよい。投票結果のエントロピーについては、後述の例示的実施形態2で詳細を説明する。また、ばらつきの評価は、複数の予測結果のうち同一のラベルを示す予測結果の割合に基づく評価であってもよい。ただし、ばらつきの評価は、上述したものに限られない。以降、「複数の予測結果のばらつきが大きいと評価した訓練用例」を、「予測結果にばらつきがある訓練用例」とも記載する。また、「複数の予測結果のばらつきが大きくないと評価した訓練用例」を、「予測結果のばらつきが小さい訓練用例」とも記載する。
本例示的実施形態に係る情報処理方法S10の流れについて、図2を参照して説明する。図2は、情報処理方法S10の流れを示すフロー図である。図2に示すように、情報処理方法S10は、ステップS101~S104を含む。
ステップS101(取得処理)において、取得部11は、複数の訓練用例を取得する。例えば、取得部11は、複数の訓練用例をメモリから読み込むことにより取得してもよい。また、例えば、取得部11は、複数の訓練用例を、入力装置から取得してもよいし、ネットワークを介して接続された装置から取得してもよい。本ステップで取得する複数の訓練用例は、観測用例及び人工用例の一方又は両方を含んでいる。
ステップS102(訓練処理)において、訓練部12は、ステップS101で取得した複数の訓練用例を用いて、機械学習モデル群を訓練する。ここで、機械学習モデル群の各々の訓練に用いる訓練用例は、ステップS101で取得した複数の訓練用例の一部であってもよいし全部であってもよい。
ステップS103(選択処理)において、選択部13は、複数の訓練用例のうち、予測結果にばらつきがある訓練用例を選択する。選択部13は、そのような訓練用例を1つ選択してもよいし、複数選択してもよい。
ステップS104(生成処理)において、生成部14は、複数の訓練用例のうち、ステップS103で選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成する。例えば、生成部14は、選択した訓練用例と、特徴量空間においてその近傍に存在する他の訓練用例とを合成してもよい。また、例えば、生成部14は、ステップS103において複数の訓練用例を選択した場合、選択した複数の訓練用例同士を合成してもよい。また、生成部14は、2つの訓練用例を合成して1つの人工用例を生成してもよいし、3以上の訓練用例を合成して1つの人工用例を生成してもよい。また、生成部14は、本ステップにおいて、1つの人工用例を生成してもよいし、複数の人工用例を生成してもよい。
2つの訓練用例を合成して1つの人工用例を生成する場合、生成部14が行う合成処理は、一例として、以下の(1)式で表される。
情報処理方法S10の具体例について、図3を参照して説明する。図3は、情報処理方法S10の具体例を模式的に示す図である。
本例示的実施形態は、複数の訓練用例を取得し、用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、複数の訓練用例を用いて訓練し、複数の訓練用例のうち、訓練後の機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択し、複数の訓練用例のうち、選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成する、との構成を採用している。
本発明の第2の例示的実施形態について、図面を参照して詳細に説明する。なお、例示的実施形態1にて説明した構成要素と同じ機能を有する構成要素については、同じ符号を付し、その説明を適宜省略する。
本例示的実施形態に係る情報処理装置20の構成について、図5を参照して説明する。図5は、情報処理装置20の構成を示すブロック図である。情報処理装置20は、複数の用例から、機械学習モデル群COM0を用いて人工用例を生成する装置である。
機械学習モデル群COM0は、例示的実施形態1における機械学習モデル群とほぼ同様に構成される。ただし、本例示的実施形態では、機械学習モデル群COM0に含まれる複数の機械学習モデルは、そのうち少なくとも2つが、互いに異なる機械学習アルゴリズムを用いて生成されたものである。
本例示的実施形態に係る情報処理方法S20の流れについて、図6を参照して説明する。図6は、情報処理方法S20の流れを示すフロー図である。
ステップS201(取得処理)において、取得部11は、複数の訓練用例を取得する。取得する複数の訓練用例は、観測により得られた用例を含んでいてもよいし、人工用例を含んでいてもよい。
ステップS202において、ラベル付与部25は、取得部21が取得した複数の訓練用例の各々にラベルを付与する。
ステップS203(訓練処理)において、訓練部22は、取得部21が取得した複数の訓練用例の一部または全部を用いて、複数の機械学習モデル群の各々を訓練する。機械学習モデル群の各々を訓練する訓練処理の詳細については後述する。
ステップS204(選択処理)において、選択部23は、取得部21が取得した複数の訓練用例のうち、予測結果にばらつきがある訓練用例を1つ以上選択する。選択部23が行う選択処理については後述する。予測結果にばらつきがある訓練用例を選択する処理の詳細については後述する。
ステップS205(生成処理)において、生成部24は、選択部23が選択した訓練用例を含む複数の訓練用例を、合成対象として特定する。また、生成部24は、合成対象として特定した複数の訓練用例を合成して人工用例を生成する。生成部24が行う生成処理の詳細については後述する。
ステップS206において、ラベル付与部25は、生成部24が生成した人工用例の各々にラベルを付与する。ステップS207において、制御部27は、訓練処理を終了するかを判定する。制御部27は、一例として、ステップS203~S206の処理を実行した回数が所定の閾値以上である場合、訓練処理を終了すると判定する。一方、ステップS203~S206の処理を実行した回数が所定の閾値未満である場合、訓練処理を終了しないと判定する。訓練処理を終了しない場合(ステップS207にてNO)、制御部27はステップS208の処理に進む。一方、訓練処理を終了する場合(ステップS207にてYES)、制御部27はステップS209の処理に進む。
ステップS208において、制御部27は、これまでに実行したステップS206で生成された1以上の人工用例を複数の訓練用例に追加する。ステップS208の処理を終えると、制御部27は、ステップS203の処理に戻る。換言すると、制御部27は、人工用例を複数の訓練用例に追加して、取得部21、訓練部22、選択部23、および生成部24を再度機能させる。
ステップS209において、出力部26は、これまでに実行したステップS206で生成された1以上の人工用例を出力する。
このようにして情報処理方法S20を用いて生成された1つ以上の人工用例は、訓練対象の機械学習モデルを訓練するために用いられる。訓練対象の機械学習モデルを訓練する処理は、例えば、訓練部22が実行してもよい。
ステップS203~S204における訓練処理及び選択処理の具体例について、図7を参照して説明する。図7は、情報処理方法S20の具体例を模式的に示す図である。
ステップS205における生成処理の具体例について説明する。ステップS205において、生成部24は、予測結果にばらつきがある訓練用例を用いて、第1生成処理S30、第2生成処理S40、及び第3生成処理S50の何れかを実行することにより、人工用例を生成する。第1生成処理S30は、予測結果にばらつきがある訓練用例とその近傍の訓練用例とを合成して人工用例を生成する処理である。第2生成処理は、予測結果にばらつきがある2つ以上の訓練用例を合成して人工用例を生成する処理である。第3生成処理は、第1生成処理及び第2生成処理の何れかを選択的に実行する処理である。
第1生成処理S30について、図8を参照して説明する。図8は、第1生成処理S30の流れを示すフロー図である。図8において、第1生成処理S30は、ステップS301~S302を含む。ここで、先行して実施されたステップS204では、予測結果にばらつきがある1または複数の訓練用例が選択されている。生成部24は、予測結果にばらつきがある1または複数の訓練用例のそれぞれ(以下では、当該訓練用例と記載)について、以下のステップS301~S302を実行する。
ステップS301において、生成部24は、当該訓練用例の近傍の訓練用例を選択する。近傍の訓練用例は、予測結果にばらつきがある訓練用例であってもよいし、予測結果のばらつきが小さい訓練用例であってもよい。例えば、近傍の訓練用例は、訓練用例群Tのうち、当該訓練用例との特徴量空間における距離が最も近い訓練用例であってもよい。また、例えば、近傍の訓練用例は、訓練用例群Tのうち、当該訓練用例との特徴量空間における距離が閾値以下の訓練用例であってもよい。
ステップS302において、生成部24は、当該訓練用例と、ステップS301で選択した近傍の訓練用例とを合成して人工用例を生成する。例えば、図7の例では、予測結果にばらつきがある訓練用例t1とその近傍の訓練用例とを合成して人工用例tv1-1が生成される。また、予測結果にばらつきがある訓練用例t2とその近傍の訓練用例とを合成して人工用例tv1-2が生成される。
[参考文献2] Chawla, N. V., Bowyer, K. W., Hall, L. O., and Kegelmeyer, W. P., “SMOTE: Synthetic minority over-sampling technique”, Journal of Artificial Intelligent Research, 16, 321-357 (2002).
(第2生成処理)
第2生成処理S40について、図9を参照して説明する。図9は、第2生成処理S40の流れを説明するフロー図である。図9に示すように、第2生成処理S40は、ステップS401~S402を含む。なお、第2生成処理は、ステップS204において、予測結果にばらつきがある訓練用例が複数選択されている場合に実行可能である。生成部24は、生成部24は、予測結果にばらつきがある複数の訓練用例のそれぞれ(以下、当該訓練用例と記載)について、以下のステップS401~S402を実行する。
ステップS401において、選択部23は、予測結果にばらつきがある複数の訓練用例のうち、当該訓練用例とは異なる他の訓練用例を選択する。換言すると、選択部23は、複数の訓練用例のうち、複数の予測結果にばらつきがある2つ以上の訓練用例を選択する。例えば、選択部23は、そのような他の訓練用例を、予測結果にばらつきがある複数の訓練用例からランダムに選択してもよい。また、例えば、選択部23は、そのような他の訓練用例として、予測結果にばらつきがある複数の訓練用例のうち当該訓練用例との特徴量空間における距離が最も小さいもの、又は距離が閾値以下のものを選択してもよい。なお、当該訓練用例が、既に合成に用いられている場合、当該訓練用例に関するステップS401~S402の処理は、実行されなくてもよい。
ステップS402において、生成部24は、当該訓練用例と、ステップS401で選択した他の訓練用例とを合成して人工用例を生成する。例えば、図7の例では、予測結果にばらつきがある訓練用例t11及びt12を合成して人工用例tv2-1が生成される。ここで、生成部24が合成する2以上の訓練用例は、この例のように、同一の機械学習モデル群COMiを用いて選択したものであってもよいし、そのうち少なくとも1つが他とは異なる機械学習モデル群COMiを用いて選択したものであってもよい。例えば、図7の例では、生成部24は、予測結果にばらつきのある訓練用例t1、t2、…、t11、t12、…の中から2つ以上の訓練用例を選択し、選択した訓練用例を合成して人工用例tv1-1、tv1-2、tv2-1、又はtv2-2を生成してもよい。なお、ステップS402における合成処理に用いる手法については、ステップS302で説明した通りであるため、詳細な説明を繰り返さない。
第3生成処理について、図10を参照して説明する。図10は、第3生成処理S50の流れを説明するフロー図である。図10において、第3生成処理S50は、ステップS501~S503を含む。ここで、ステップS204では、予測結果にばらつきがある1または複数の訓練用例が選択されている。生成部24は、予測結果にばらつきがある1または複数の訓練用例のそれぞれ(以下では、当該訓練用例と記載)について、以下のステップS501~S503を実行する。
ステップS501において、生成部24は、第1生成処理及び第2生成処理の何れかを選択する。例えば、生成部24は、ランダム関数により決定した確率pを用いて第1生成処理を選択し、第1生成処理を選択しなかった場合に第2生成処理を選択してもよい。なお、第1生成処理及び第2生成処理の何れかを選択する手法は、確率pを用いる手法に限らず、他の手法であってもよい。
ステップS502において、生成部24は、いずれを選択したかを判定する。第1生成処理を選択した場合、生成部24はステップS503の処理に進み、第1生成処理を実行する。一方、第2生成処理を選択した場合、生成部24はステップS504の処理に進み、第2生成処理を実行する。第1生成処理及び第2生成処理の詳細については、上述した通りである。
本例示的実施形態は、予測結果にばらつきがある訓練用例と、その近傍にある訓令用例とを合成して人工用例を生成する第1生成処理を実行する、との構成を有する。
本発明の例示的実施形態3について、図面を参照して詳細に説明する。なお、例示的実施形態2にて説明した構成要素と同じ機能を有する構成要素については、同じ符号を付記してその説明を繰り返さない。本例示的実施形態は、例示的実施形態2における生成部24を次のように変形した形態である。
本例示的実施形態において、生成部24は、複数の人工用例を生成する。また、生成部24は、生成した複数の人工用例のうち、類似条件を満たす2つの人工用例を1つの人工用例に統合する。ここで、類似条件とは、用例が類似することを示す条件である。類似条件は、例えばコサイン類似度が閾値以上であることであってもよいし、特徴量空間における距離が閾値以下であることであってもよい。ただし、類似条件はこれらに限られない。統合する処理の詳細については後述する。
本例示的実施形態における情報処理方法S20Aについて、図11を参照して説明する。図11は、例示的実施形態3に係る情報処理方法S20Aの流れを説明するフロー図である。図11に示す情報処理方法S20Aは、例示的実施形態2に係る情報処理方法S20とほぼ同様に構成されるが、ステップS205Aをさらに含む点が異なる。
ステップS205Aにおいて、生成部24は、ステップS205において生成した人工用例のうち、類似する2つの人工用例を統合する。具体的には、生成部24は、今回のステップS205において生成した人工用例と、前回までのステップS205において生成した人工用例の何れかとが類似条件を満たすか否かを判定する。類似条件を満たすと判定した場合、生成部24は、類似条件を満たす2つの人工用例を統合する。
統合処理の一例として、2つの人工用例を合成する処理が挙げられる。この場合、生成部24は、2つの人工用例を合成して1つの人工用例を生成し、類似条件を満たした元の2つの人工用例を削除する。また、統合処理の他の例として、2つの人工用例のうち一方を削除する処理が挙げられる。なお、統合処理は、類似条件を満たす2つの人工用例の代わりに、当該2つの人工用例を参照して生成した1つの人工用例を採用する処理であればよく、上述した処理に限られない。なお、人工用例を削除するとは、ステップS206でラベルを付与する対象、及びステップS208で訓練用例に追加する対象から削除することである。これにより、統合された人工用例に対して、ラベルが付与されるとともに訓練用例に追加される。
本例示的実施形態においては、生成部が、複数の人工用例を生成し、生成した複数の人工用例のうち、類似条件を満たす2つの人工用例を1つの人工用例に統合する、との構成が採用されている。
本発明の例示的実施形態4について、図面を参照して詳細に説明する。なお、例示的実施形態2にて説明した構成要素と同じ機能を有する構成要素については、同じ符号を付記してその説明を繰り返さない。本例示的実施形態は、例示的実施形態2における生成部24を次のように変形した形態である。
本例示的実施形態において、生成部24は、生成した人工用例のうち、訓練後の機械学習モデル群COM0を用いて得られる複数の予測結果にばらつきがある人工用例を出力する。ここで、ばらつきがある人工用例は、ばらつきの評価結果が「ばらつきが大きい」ことを示す人工用例である。ばらつきの評価の詳細については、上述した通りであるため、詳細を繰り返さない。換言すると、生成部24は、生成した人工用例のばらつきを、訓練後の機械学習モデル群COM0を用いて事後評価し、事後評価により予測結果にばらつきがある人工用例を採用する。
本例示的実施形態における情報処理方法S20Bについて、図12を参照して説明する。図12は、例示的実施形態4に係る情報処理方法S20Bの流れを説明するフロー図である。図12に示す情報処理方法S20Bは、例示的実施形態2に係る情報処理方法S20とほぼ同様に構成されるが、ステップS205Bをさらに含む点が異なる。
ステップS205Bにおいて、生成部24は、ステップS205において生成した人工用例を事後評価する。
本例示的実施形態においては、生成部が、生成した人工用例のうち、訓練後の機械学習モデル群を用いて得られる複数の予測結果にばらつきがある人工用例を出力する、との構成が採用されている。
本発明の例示的実施形態5について、図面を参照して詳細に説明する。なお、例示的実施形態2にて説明した構成要素と同じ機能を有する構成要素については、同じ符号を付記してその説明を繰り返さない。
図13に示すように、本例示的実施形態では、機械学習モデル群COM0は、機械学習モデルmj(j=1,2,…,M)を含む。各機械学習モデルmjは、同一の機械学習アルゴリズムによって生成されるモデルである。例えば、各機械学習モデルmjは、決定木であってもよい。
本例示的実施形態のステップS203において、訓練部22は、ステップS201で取得部21が取得した訓練用例群Tから、訓練用例群Djを抽出する。訓練用例群Djは、訓練用例群Tの一部である。例えば、訓練部22は、ランダムサンプリングにより訓練用例群Djを抽出してもよい。訓練部22は、訓練用例群Diを用いて、機械学習モデルmjを訓練することを、j=1,2,…,Mについて繰り返す。
本例示的実施形態のステップS204において、選択部23は、機械学習モデル群COM0を用いて、訓練用例群Tに含まれる各訓練用例について、予測結果のばらつきを評価する。また、選択部23は、予測結果にばらつきがある訓練用例を選択する。図13の例では、選択部23は、機械学習モデル群COM0を用いて、予測結果にばらつきがある訓練用例t1,t3,…を選択している。
本例示的実施形態は、機械学習モデル群を構成する機械学習モデルとして、全て同一の機械学習アルゴリズムによって生成されたモデルを用い、取得した訓練用例群の中から予測結果にばらつきがある訓練用例を選択する、との構成を採用している。
情報処理装置10,20の一部又は全部の機能は、集積回路(ICチップ)等のハードウェアによって実現してもよいし、ソフトウェアによって実現してもよい。
本発明は、上述した実施形態に限定されるものでなく、請求項に示した範囲で種々の変更が可能である。例えば、上述した実施形態に開示された技術的手段を適宜組み合わせて得られる実施形態についても、本発明の技術的範囲に含まれる。
上述した実施形態の一部又は全部は、以下のようにも記載され得る。ただし、本発明は、以下の記載する態様に限定されるものではない。
複数の訓練用例を取得する取得手段と、
用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、前記複数の訓練用例を用いて訓練する訓練手段と、
前記複数の訓練用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択する選択手段と、
前記複数の訓練用例のうち、前記選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成する生成手段と、
を備えた情報処理装置。
前記生成手段は、前記選択した訓練用例と、特徴量空間において前記選択した訓練用例の近傍に存在する用例とを合成して前記人工用例を生成する、付記1に記載の情報処理装置。
前記選択手段は、前記複数の訓練用例のうち、2つ以上の前記複数の予測結果にばらつきがある訓練用例を選択し、
前記生成手段は、2つ以上の前記選択した訓練用例を合成して前記人工用例を生成する、付記1に記載の情報処理装置。
前記生成手段は、
前記選択した訓練用例と、特徴量空間において前記選択した訓練用例の近傍に存在する用例とを合成する第1生成処理と、
2つ以上の前記選択した訓練用例を合成して前記人工用例を生成する第2生成処理と、
の何れかを実行することにより前記人工用例を生成する、付記1に記載の情報処理装置。
前記人工用例を前記複数の訓練用例に追加して、前記取得手段、前記訓練手段、前記選択手段、及び前記生成手段を再度機能させる、付記1から4の何れか1つに記載の情報処理装置。
前記生成手段は、
複数の前記人工用例を生成し、
複数の前記人工用例のうち類似条件を満たす2つの人工用例を1つの人工用例に統合する、付記1から4の何れか1つに記載の情報処理装置。
前記生成手段は、前記人工用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある人工用例を出力する、付記1から6の何れか1つに記載の情報処理装置。
前記機械学習モデル群は、前記人工用例を用いて訓練する訓練対象の機械学習モデルを含む、付記1から7の何れか1つに記載の情報処理装置。
前記機械学習モデル群のうち少なくとも2つは、互いに異なる機械学習アルゴリズムを用いる、付記1から8の何れか1つに記載の情報処理装置。
前記機械学習モデル群のそれぞれは、同一の機械学習アルゴリズムを用いる、付記1から8の何れか1つに記載の情報処理装置。
前記機械学習モデル群のうち少なくとも1つは決定木である、付記1から10の何れか1つに記載の情報処理装置。
前記複数の訓練用例及び前記人工用例の一部又は全部にラベルを付与するラベル付与手段をさらに備える、付記1から11の何れか1つに記載の情報処理装置。
複数の訓練用例を取得すること、
用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、前記複数の訓練用例を用いて訓練すること、
前記複数の訓練用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択すること、及び、
前記複数の訓練用例のうち、前記選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成すること、
を含む情報処理方法。
コンピュータを情報処理装置として機能させるためのプログラムであって、前記コンピュータを、
複数の訓練用例を取得する取得手段と、
用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、前記複数の訓練用例を用いて訓練する訓練手段と、
前記複数の訓練用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択する選択手段と、
前記複数の訓練用例のうち、前記選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成する生成手段と、
として機能させるプログラム。
付記14に記載のプログラムが記録された、コンピュータ読み取り可能な記録媒体。
上述した実施形態の一部又は全部は、更に、以下のように表現することもできる。
複数の訓練用例を取得する取得処理と、
用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、前記複数の訓練用例を用いて訓練する訓練処理と、
前記複数の訓練用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択する選択処理と、
前記複数の訓練用例のうち、前記選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成する生成処理と、を実行する情報処理装置。
11、21 取得部
12、22 訓練部
13、23 選択部
14、24 生成部
25 ラベル付与部
26 出力部
27 制御部
Claims (14)
- 複数の訓練用例を取得する取得手段と、
用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、前記複数の訓練用例を用いて訓練する訓練手段と、
前記複数の訓練用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択する選択手段と、
前記複数の訓練用例のうち、前記選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成する生成手段と、
を備えた情報処理装置。 - 前記生成手段は、前記選択した訓練用例と、特徴量空間において前記選択した訓練用例の近傍に存在する用例とを合成して前記人工用例を生成する、請求項1に記載の情報処理装置。
- 前記選択手段は、前記複数の訓練用例のうち、2つ以上の前記複数の予測結果にばらつきがある訓練用例を選択し、
前記生成手段は、2つ以上の前記選択した訓練用例を合成して前記人工用例を生成する、請求項1に記載の情報処理装置。 - 前記生成手段は、
前記選択した訓練用例と、特徴量空間において前記選択した訓練用例の近傍に存在する用例とを合成する第1生成処理と、
2つ以上の前記選択した訓練用例を合成して前記人工用例を生成する第2生成処理と、
の何れかを実行することにより前記人工用例を生成する、請求項1に記載の情報処理装置。 - 前記人工用例を前記複数の訓練用例に追加して、前記取得手段、前記訓練手段、前記選択手段、及び前記生成手段を再度機能させる、請求項1から4の何れか1項に記載の情報処理装置。
- 前記生成手段は、
複数の前記人工用例を生成し、
複数の前記人工用例のうち類似条件を満たす2つの人工用例を1つの人工用例に統合する、請求項1から4の何れか1項に記載の情報処理装置。 - 前記生成手段は、前記人工用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある人工用例を出力する、請求項1から6の何れか1項に記載の情報処理装置。
- 前記機械学習モデル群は、前記人工用例を用いて訓練する訓練対象の機械学習モデルを含む、請求項1から7の何れか1項に記載の情報処理装置。
- 前記機械学習モデル群のうち少なくとも2つは、互いに異なる機械学習アルゴリズムを用いる、請求項1から8の何れか1項に記載の情報処理装置。
- 前記機械学習モデル群のそれぞれは、同一の機械学習アルゴリズムを用いる、請求項1から8の何れか1項に記載の情報処理装置。
- 前記機械学習モデル群のうち少なくとも1つは決定木である、請求項1から10の何れか1項に記載の情報処理装置。
- 前記複数の訓練用例及び前記人工用例の一部又は全部にラベルを付与するラベル付与手段をさらに備える、請求項1から11の何れか1項に記載の情報処理装置。
- 複数の訓練用例を取得すること、
用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、前記複数の訓練用例を用いて訓練すること、
前記複数の訓練用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択すること、及び、
前記複数の訓練用例のうち、前記選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成すること、
を含む情報処理方法。 - コンピュータを情報処理装置として機能させるためのプログラムであって、前記コンピュータを、
複数の訓練用例を取得する取得手段と、
用例を入力として予測結果を出力する機械学習モデルを複数含む機械学習モデル群を、前記複数の訓練用例を用いて訓練する訓練手段と、
前記複数の訓練用例のうち、訓練後の前記機械学習モデル群を用いて得られる複数の予測結果にばらつきがある訓練用例を選択する選択手段と、
前記複数の訓練用例のうち、前記選択した訓練用例を含む2つ以上の訓練用例を合成して人工用例を生成する生成手段と、
として機能させるプログラム。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2023523860A JPWO2022249392A1 (ja) | 2021-05-27 | 2021-05-27 | |
PCT/JP2021/020174 WO2022249392A1 (ja) | 2021-05-27 | 2021-05-27 | 情報処理装置、情報処理方法、及びプログラム |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/JP2021/020174 WO2022249392A1 (ja) | 2021-05-27 | 2021-05-27 | 情報処理装置、情報処理方法、及びプログラム |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022249392A1 true WO2022249392A1 (ja) | 2022-12-01 |
Family
ID=84229594
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2021/020174 WO2022249392A1 (ja) | 2021-05-27 | 2021-05-27 | 情報処理装置、情報処理方法、及びプログラム |
Country Status (2)
Country | Link |
---|---|
JP (1) | JPWO2022249392A1 (ja) |
WO (1) | WO2022249392A1 (ja) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019022136A1 (ja) * | 2017-07-25 | 2019-01-31 | 国立大学法人東京大学 | 学習方法、学習プログラム、学習装置及び学習システム |
US20190080205A1 (en) * | 2016-09-16 | 2019-03-14 | Deep Learning Analytics, LLC | Systems and Methods for Deep Model Translation Generation |
-
2021
- 2021-05-27 WO PCT/JP2021/020174 patent/WO2022249392A1/ja active Application Filing
- 2021-05-27 JP JP2023523860A patent/JPWO2022249392A1/ja active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190080205A1 (en) * | 2016-09-16 | 2019-03-14 | Deep Learning Analytics, LLC | Systems and Methods for Deep Model Translation Generation |
WO2019022136A1 (ja) * | 2017-07-25 | 2019-01-31 | 国立大学法人東京大学 | 学習方法、学習プログラム、学習装置及び学習システム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2022249392A1 (ja) | 2022-12-01 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3483767B1 (en) | Device for detecting variant malicious code on basis of neural network learning, method therefor, and computer-readable recording medium in which program for executing same method is recorded | |
CN108351984B (zh) | 硬件高效的深度卷积神经网络 | |
EP3605407A1 (en) | Information processing device, information processing method, and computer-readable storage medium | |
KR20170030923A (ko) | 음향 모델 생성 장치 및 방법, 음성 인식 장치 및 방법 | |
KR20190091806A (ko) | 생성적 적대 네트워크를 이용한 비디오 시퀀스 생성 시스템 및 그 방법 | |
CN114118384B (zh) | 神经网络模型的量化方法、可读介质和电子设备 | |
JP2007298564A (ja) | 信号分離装置、信号分離方法、ならびに、プログラム | |
WO2016157275A1 (ja) | 計算機及びグラフデータ生成方法 | |
Mahmud et al. | Quantum dimension reduction for pattern recognition in high-resolution spatio-spectral data | |
WO2022249392A1 (ja) | 情報処理装置、情報処理方法、及びプログラム | |
CN110069284A (zh) | 一种基于opu指令集的编译方法及编译器 | |
WO2022003991A1 (ja) | 2次元マップ生成装置、2次元マップ生成方法および2次元マップ生成用プログラム | |
CN113449587A (zh) | 一种人体行为识别和身份认证方法、装置和电子设备 | |
CN112528650A (zh) | 一种Bert模型预训练方法、系统及计算机设备 | |
WO2022249373A1 (ja) | 情報処理装置、情報処理方法及びプログラム | |
JPWO2022249392A5 (ja) | ||
KR20200134813A (ko) | 기계 학습을 위한 이미지 처리 장치 및 방법 | |
JPWO2022249373A5 (ja) | ||
US20240242123A1 (en) | Information processing apparatus, information processing method, and storage medium | |
JP2022151502A (ja) | プログラム、情報処理装置、及び方法 | |
Xue et al. | Removing watermarks for image processing networks via referenced subspace attention | |
De Luca et al. | Towards a GPU parallel software for environmental data fitting | |
CN115798453A (zh) | 语音重建方法、装置、计算机设备和存储介质 | |
JP7003435B2 (ja) | 情報処理装置、プログラム、情報処理方法及びデータ構造 | |
JP6842436B2 (ja) | 情報処理装置、情報処理方法、及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21943041 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 18561354 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2023523860 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21943041 Country of ref document: EP Kind code of ref document: A1 |