WO2020170803A1 - Dispositif, procédé et programme d'augmentation - Google Patents

Dispositif, procédé et programme d'augmentation Download PDF

Info

Publication number
WO2020170803A1
WO2020170803A1 PCT/JP2020/004113 JP2020004113W WO2020170803A1 WO 2020170803 A1 WO2020170803 A1 WO 2020170803A1 JP 2020004113 W JP2020004113 W JP 2020004113W WO 2020170803 A1 WO2020170803 A1 WO 2020170803A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
class
data set
generator
classifier
Prior art date
Application number
PCT/JP2020/004113
Other languages
English (en)
Japanese (ja)
Inventor
真弥 山口
毅晴 江田
沙那恵 村松
Original Assignee
日本電信電話株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 日本電信電話株式会社 filed Critical 日本電信電話株式会社
Publication of WO2020170803A1 publication Critical patent/WO2020170803A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Definitions

  • the present invention relates to an expansion device, an expansion method, and an expansion program.
  • the preparation of learning data includes not only the collection of learning data but also the addition of annotations such as labels to the learning data.
  • rule-based data extension has been known as a technique for reducing the cost of maintaining learning data.
  • a method is known in which another learning data is generated by adding changes to an image used as learning data according to a specific rule such as inversion, scaling, noise addition, rotation, etc. See Patent Document 1 or 2.
  • the learning data is voice or text, similar rule-based data expansion may be performed.
  • the conventional technology has a problem that it may be difficult to efficiently obtain high-quality learning data that greatly improves the accuracy of the model by data expansion.
  • it is difficult to increase variations in the attributes of learning data, which limits the improvement of model accuracy.
  • an image in which the attributes of “front of the window”, “cat”, and “front” of the image of a cat facing the front of the window are changed is generated. Is difficult to do.
  • the expansion device includes a first data set, which is a set of data belonging to a first class, and a second data, which is a set of data belonging to a second class. Belongs to a third class from the second data set based on a metric calculated from the similarity between the second data set and the degree of uniformity of data included in the second data set.
  • the generator causes the first class to be determined by the classifier.
  • the generator the identifying so as to generate data that is likely to be identified as data of a data set and is calculated by the classifier such that the probabilities of belonging to the plurality of third classes are non-zero and equal.
  • Generating a data by specifying the first class to a learning unit that learns a generative model including a generator and the classifier, and to the generator of the generative model that has been learned by the learning unit
  • an extraction unit that extracts, from the data generated by the generation unit, data identified as data of the first data set by the identifier as extension data, and an extraction unit that extracts the data.
  • a assigning unit that assigns a label indicating that the extension data belongs to the first class.
  • high-quality learning data that greatly improves the accuracy of the model can be efficiently obtained by data expansion.
  • FIG. 1 is a diagram for explaining the learning system according to the first embodiment.
  • FIG. 2 is a diagram illustrating a configuration example of the expansion device according to the first embodiment.
  • FIG. 3 is a diagram for explaining a classifier that also has a function as a classifier.
  • FIG. 4 is a diagram for explaining a learning process for external data.
  • FIG. 5 is a diagram for explaining the learning process for the target data.
  • FIG. 6 is a diagram for explaining extraction of extension data.
  • FIG. 7 is a diagram for explaining the expanded data set.
  • FIG. 8 is a flowchart showing a processing flow of the expansion device according to the first embodiment.
  • FIG. 9 is a flowchart showing a flow of parameter update processing of the expansion device according to the first embodiment.
  • FIG. 10 is a diagram showing data used in the experiment.
  • FIG. 11 is a diagram showing the results of the experiment.
  • FIG. 12 is a diagram showing the results of the experiment.
  • FIG. 13 is a diagram showing the results of the experiment.
  • FIG. 14 is a diagram showing conditions of the additional experiment.
  • FIG. 15 is a diagram showing the result of the additional experiment.
  • FIG. 16 is a diagram illustrating an example of a computer that executes the extension program.
  • FIG. 1 is a diagram for explaining the learning system according to the first embodiment.
  • the learning system 1 includes an expansion device 10 and a learning device 20.
  • the expansion device 10 performs data expansion of the target data set S T using the external data set S O and outputs the expanded data set S′ gen .
  • the learning device 20 also learns the target model 21 using the expanded data set S′ gen .
  • the target model 21 may be a known model that performs machine learning.
  • the objective model 21 is a class classifier such as Resnet-152.
  • each data set in FIG. 1 is labeled data used in the target model 21. That is, each data set is a combination of data and a label indicating the class to which the data belongs.
  • each data set is a combination of image data and a label.
  • the target model 21 may be a voice recognition model or a natural language recognition model. In that case, each data set is voice data with a label or text data with a label.
  • the target data set S T is assumed to be a combination of the target data X T and the target label y T attached to the target data X T.
  • the external data set S O is assumed to be a combination of the external data X O and the external label y O attached to the external data X O.
  • the target label y T is a label to be learned by the target model 21.
  • the target label y T is an ID for identifying the person shown in the image of the target data.
  • the target label y T is a text in which the voice of the target data is transcribed.
  • the external data set S O is a data set for extending the target data set S T.
  • the external dataset S O may be a dataset in a domain different from that of the target dataset S T.
  • the domain is a characteristic peculiar to the data set, and is represented by the data, the label, and the generation distribution.
  • a domain of a data set whose data is X 0 and whose label is y 0 is represented as (X 0 , y 0 , P(X 0 , y 0 )).
  • the target model 21 is an image recognition model
  • the learning device 20 learns the target model 21 so that the image of the person whose ID is “0002” can be recognized from the image.
  • the target data set S T is a combination of the label “ID:0002” and an image in which the person is known to be reflected.
  • the external data set S O is a combination of a label indicating an ID other than “0002” and an image in which it is known that the person corresponding to the ID is shown.
  • the expansion device 10 outputs an expanded data set S′ gen in which an attribute that the data of the target data set S T does not have is taken in from the external data set S O. This makes it possible to obtain variation data that could not be obtained from the target data set S T alone. For example, according to the expansion device 10, even when the target data set S T includes only the image of the back of a certain person, it is possible to obtain the image of the front of the person. Become.
  • the generation model 121 is a model based on GAN (Generative Adversarial Networks).
  • GAN Geneative Adversarial Networks
  • both the generator G and the discriminator D are neural networks.
  • the generator G When the target class label y T or the outer class label y O is specified together with the noise z, the generator G generates an image based on the specified label.
  • the purpose class is a class to which the purpose data X T belongs.
  • the external class is a class to which the external data set X O belongs.
  • the discriminator D discriminates whether the image generated by the generator G is a real product (Real) or a fake product (Fake). For example, the discriminator D receives the image X gen generated by the generator G and the image XT +O of the target data set and the external data, and which of the two images is X gen (fake), And which is XT+O (genuine: Real).
  • the discriminator D also has a function as a classifier. That is, the classifier D can calculate the probability that the image belongs to each class. Details of the function of the discriminator D as a classifier will be described later.
  • the expansion device 10 learns the generative model 121 using a predetermined data set selected from the external data set S O. Further, the expansion device 10 reduces the error (Adversarial Loss) when the discriminator D discriminates whether the image is a genuine image or a fake image, and the error (OLSR Loss) regarding the function as the classifier, Learning is performed by updating the parameters of the discriminator D. Note that updating of each parameter in learning is performed by, for example, an error back propagation method (Backpropagation).
  • Backpropagation an error back propagation method
  • the expansion data is not the generation data X gen generated by the generator G based on the label y T of the target data set S T , but the generated data X gen. include X'gen extracted from the X gen to the extended set of data S'gen.
  • FIG. 2 is a diagram illustrating a configuration example of the expansion device according to the first embodiment.
  • the expansion device 10 includes an input/output unit 11, a storage unit 12, and a control unit 13.
  • the input/output unit 11 is an interface for receiving data input and outputting data.
  • the input/output unit 11 may be a communication module that communicates data with an external device.
  • the storage unit 12 is a storage device such as an HDD (Hard Disk Drive), SSD (Solid State Drive), and optical disk.
  • the storage unit 12 may be a rewritable semiconductor memory such as RAM (Random Access Memory), flash memory, NVSRAM (Non Volatile Static Random Access Memory).
  • the storage unit 12 stores an OS (Operating System) and various programs executed by the expansion device 10. Further, the storage unit 12 stores various information used in executing the program.
  • the storage unit 12 also stores the generation model 121. Specifically, the storage unit 12 stores parameters used in each process by the generative model 121.
  • the control unit 13 controls the entire expansion device 10.
  • the control unit 13 includes, for example, electronic circuits such as CPU (Central Processing Unit), MPU (Micro Processing Unit), and GPU (Graphics Processing Unit), ASIC (Application Specific Integrated Circuit), FPGA (Field Programmable Gate Array), and the like. It is an integrated circuit.
  • the control unit 13 has an internal memory for storing programs and control data defining various processing procedures, and executes each process using the internal memory. Further, the control unit 13 functions as various processing units by operating various programs.
  • the control unit 13 has a selection unit 131, a learning unit 132, a generation unit 133, an extraction unit 134, and an addition unit 135.
  • the selection unit 131 determines the similarity between the target data set, which is a set of data belonging to the target class, and the external data set, which is a set of data belonging to the external class, and the uniformity of the data included in the external data set.
  • a predetermined data set is selected from the external data sets based on a metric calculated from the degree.
  • the data set selected by the selection unit 131 is an example of the third data set.
  • the label of the data set selected by the selection unit 131 is an example of the third label.
  • the selection unit 131 has selected from the external data sets based on the metric calculated from the degree of similarity and the degree of uniformity calculated using the statistical value of the brightness value of the image data included in each data set. External datasets can be selected. For example, the selection unit 131 calculates the metric M by the equation (1). At this time, the selection unit 131 selects a data set whose calculated metric M is equal to or greater than the threshold value.
  • the FID of the expression (1) is the degree of similarity and is represented by the expression (2) (reference: Non-Patent Document 9).
  • ⁇ XT and ⁇ XO are average values of the luminance values of the image included in the target data set and the image included in the external data set.
  • ⁇ XT and ⁇ XT are matrices representing the brightness values of the image of the target data set and the image of the external data set. The FID decreases as the degree of similarity between the target data set and the external data set increases.
  • the MS-SSIM of the formula (1) is the degree of uniformity and is represented by the formula (3) (reference: Non-Patent Document 10).
  • I M (x 1 , x 2 ) is a function that increases as the difference in brightness between the images x 1 and x 2 increases.
  • c M (x 1 , x 2 ) is a function that increases as the difference in contrast between the images x 1 and x 2 increases.
  • s M (x 1 , x 2 ) is a function that increases as the difference in structure between the images x 1 and x 2 increases.
  • the learning unit 132 will be described. Note that the external data set and the external label in the description of the learning unit 132 mean the external data set and the label of the external data selected by the selection unit 131.
  • the learning unit 132 easily identifies the generator G as the data of the target data set by the classifier D when the target class is specified in the generator G that generates data based on the specified class. To learn.
  • the discriminator D also has a function as a classifier (reference: Non-Patent Document 17).
  • FIG. 3 is a diagram for explaining a classifier that also has a function as a classifier.
  • the classifier D when the classifier D functions as a classifier, the classifier D may be referred to as the classifier C.
  • L Adv. Is the error of the discrimination result of the discriminator D.
  • L OLSR in FIG. 3 is an error of the classification result of the classifier C.
  • OLSR is an abbreviation for Outer Label Smoothing Regularization.
  • the learning unit 132 includes a generator G, a classifier D, and a classifier C so as to generate data calculated by the classifier C such that the probabilities of belonging to a plurality of outer classes are non-zero and even. Train the model.
  • FIG. 4 is a diagram for explaining a learning process for external data.
  • FIG. 5 is a diagram for explaining the learning process for the target data.
  • the generator G generates the image data X gen when the external label y O is specified together with the noise z.
  • the external data set S O includes image data X O.
  • the discriminator D discriminates whether the image data X gen and the image data X O are genuine or counterfeit.
  • the learning unit 132 updates the parameter of the discriminator D so that the binary discrimination error of whether it is a genuine article or a fake one becomes small, and conversely, the parameter of the generator G so that the binary discrimination error becomes large. To update.
  • the learning unit 132 calculates the error according to the equation (4) where Z is 0.
  • the generator G when the target label y T is specified together with the noise z, the generator G generates the image data X gen . Further, the target data set S T includes the image data X T. The discriminator D discriminates whether the image data X gen and the image data X T are genuine or counterfeit. Then, the learning unit 132 updates the parameter of the discriminator D so that the binary discrimination error between the genuine and the fake is small, and conversely updates the generator G so that the binary discrimination error is large. To do.
  • the learning unit 132 calculates the error according to the equation (4) where Z is 1.
  • the parameters of the entire generative model 121 including the generator G are updated. That is, according to the learning process performed by the learning unit 132, the classifier C calculates the data of the target data set so that the probability of belonging to each outer class is the reciprocal of the number of classes included in the outer class.
  • the label specified for the generator G is a unique expression of each class.
  • noise is a common expression common to all classes.
  • the label represents the type of animal
  • the type of animal shown in the image is determined by the label
  • the background and posture of the image are determined by noise. Therefore, for example, if noise is common between the case of specifying a label indicating "dog" and the case of specifying a label indicating "cat", in each case, the generator G determines the image of the dog and the dog. It is conceivable to generate an image of a cat having the same background and posture as the image of.
  • the learning unit 132 performs the learning of the classifier C as described above in order to obtain useful information for improving the quality of generated data even from a label different from the target label.
  • the generator G can generate “an image of a dog sitting facing left indoors” that could not be generated conventionally. Is possible. As a result, the generator G can generate various kinds of data, so that the quality of the extension data is improved.
  • the generation unit 133 generates data by designating the target class in the generator G of the generation model learned by the learning unit 132. At this time, the generation unit 133 specifies noise together with the target class.
  • the noise may be a value generated according to a normal distribution of N(0,1).
  • the extraction unit 134 sets, as the extension data X′ gen , the data identified by the identifier D as the data (Good) of the target data set from the data generated by the generation unit 133. Extract.
  • FIG. 6 is a diagram for explaining extraction of extension data. It is conceivable that the generator G may generate data that is discriminated as an imitation (Bad) by the discriminator D even if it has been learned. Therefore, the extraction unit 134 extracts the extension data by using, for example, DRS (Discriminator Rejection Sampling) (reference: Non-Patent Document 13).
  • DRS Discriminator Rejection Sampling
  • the extraction unit 134 further excludes, from the extension data, data whose Euclidean distance from the target data set is a predetermined value or more. For example, when the Euclidean distance between the feature vector of the extracted image of the extension data and the feature vector of the image of the target data set is equal to or greater than the threshold value, the extraction unit 134 excludes the extension data.
  • the assigning unit 135 assigns a label indicating that it belongs to the target class to the extension data extracted by the extracting unit 134.
  • FIG. 7 is a diagram for explaining the expanded data set. As shown in FIG. 7, the expanded data set S′ T is obtained by adding a target label y T to the target data X T and the expansion data X′ gen .
  • FIG. 8 is a flowchart showing a processing flow of the expansion device according to the first embodiment.
  • the expansion device 10 receives an input of a target data set and an external data set (step S11).
  • the expansion device 10 evaluates the external data set by a metric based on the similarity between the data sets and the degree of uniformity within the external data set, and selects a predetermined data set (step S12).
  • the expansion device 10 uses the generation model to generate an image from the target data set and the external data set (step S13). At this time, the expansion device 10 specifies the target label and the external label together with noise in the generation model. Here, the expansion device 10 updates the parameters of the generation model based on the generated image (step S14).
  • the parameter updating process (learning process) in step S14 will be described later with reference to FIG.
  • the expansion device 10 specifies the label of the target data set in the generation model (step S15), and generates an image based on the specified label (step S16).
  • the expansion device 10 extracts an image for expansion from the generated image (step S17). At this time, the expansion device 10 can extract the image for expansion according to the identification criterion of the identifier of the generation model.
  • the expansion device 10 integrates the image of the target data set and the image for expansion and gives the label of the target data set (step S18). Further, the expansion device 10 outputs the expanded data set to which the target label has been added (step S19) and transfers it to the learning device 20.
  • FIG. 9 is a flowchart showing a flow of parameter update processing of the expansion device according to the first embodiment.
  • the expansion device 10 first updates the parameters of the generative model 121 based on the binary determination error (step S141).
  • the expansion device 10 selects an unselected label (step S142).
  • the selected label is the target label (step S143, target label)
  • the expansion device 10 updates the parameter so that the label is smoothed (step S144). That is, the expansion device 10 is a generation model such that the probability calculated by the classifier C and the probability that the target data is classified into each outer class is 1/K (K is the number of outer classes).
  • the parameters of 121 are updated (step S144).
  • step S143 external label
  • the parameters of the generated model 121 are updated so that the external class is correctly classified (step S145).
  • step S146, Yes If there is an unselected label (step S146, Yes), the expansion device 10 returns to step S142 and repeats the processing. On the other hand, when there is no unselected label (step S146, No), the expansion device 10 ends the parameter updating process.
  • the extension device 10 determines the similarity between the target data set, which is a set of data belonging to the target class, and the external data set, which is a set of data belonging to the external class, and the external data set.
  • a predetermined data set is selected from the external data sets based on a metric calculated from the degree of uniformity of the included data.
  • the extension device 10 easily identifies the generator G as the data of the target data set by the discriminator D when the target class is designated in the generator G that generates data based on the designated class.
  • the expansion device 10 generates data by designating a target class to the generator G of the learned generation model.
  • the expansion device 10 extracts, from the generated data, the data identified as the data of the target data set by the identifier D as the extension data.
  • the extension device 10 adds a label indicating that it belongs to the target class to the extracted extension data.
  • the expansion device 10 preselects an input external data set according to a predetermined criterion, performs learning so that useful information obtained from the external data set is not lost, and generates the generated data of the discriminator. Extract by standard. Therefore, according to the first embodiment, it is possible to efficiently obtain high-quality learning data that greatly improves the accuracy of the model by data expansion.
  • the expansion device 10 has already selected from the external data sets based on the metric calculated from the degree of similarity and the uniformity calculated using the statistic of the brightness value of the image data included in each data set. Select an external dataset. As a result, data that does not contribute to improving the quality of learning data can be excluded in advance, and the quality of learning data can be improved.
  • the expansion device 10 performs a calculation so that the classifier C, with respect to the data of the target data set, has the probability of belonging to each of the outer classes of the selected outer data set to be the reciprocal of the number of classes included in the selected outer class. Then, the data of the external data set is calculated so that the probability of belonging to the class to which the data belongs is 1. This makes it possible to generate learning data having useful information included in the label.
  • the extension device 10 further excludes, from the extension data, data whose Euclidean distance from the target data set is a predetermined value or more. As a result, outliers can be excluded from the extension data, and the quality of the learning data can be improved.
  • FIG. 10 is a diagram showing data used in the experiment. Then, the selection unit 131 changed the threshold value of the metric for selecting the external data set, and compared the accuracy (Top-1 Accuracy) of the target model 21 with the FID of the expanded data set.
  • FIG. 11 is a diagram showing the results of the experiment. In other words, the similarity between the expanded data set and the target data set increased as the metric threshold increased. Therefore, it can be said that the selection of data by the metrics improves the quality of the learning data.
  • FIG. 12 is a diagram showing the results of the experiment. Therefore, it can be said that the selection of the data by the metrics improves the quality of the learning data and, as a result, the accuracy of the target model 21.
  • FIG. 13 is a diagram showing the results of the experiment.
  • Baseline is a conventional rule-based data extension method.
  • FIG. 14 is a diagram showing conditions of the additional experiment.
  • FIG. 15 is a diagram showing the result of the additional experiment. As shown in FIG. 15, in the method of the first embodiment, even if the target data is reduced to 1/10, the accuracy close to that of the conventional method when the target data is not reduced is obtained.
  • the learning function of the objective model 21 is provided in the learning device 20 different from the expansion device 10.
  • the extension device 10 may be provided with an objective model learning unit that causes the objective model 21 to learn the extended data set S′ gen .
  • the expansion device 10 can suppress resource consumption due to data transfer between the devices, and efficiently execute data expansion and learning of the target model as a series of processes.
  • each constituent element of each illustrated device is functionally conceptual, and does not necessarily have to be physically configured as illustrated. That is, the specific form of distribution and integration of each device is not limited to that shown in the figure, and all or part of the device may be functionally or physically distributed in arbitrary units according to various loads or usage conditions, or It can be integrated and configured. Furthermore, all or arbitrary parts of the processing functions performed by each device may be realized by a CPU and a program that is analyzed and executed by the CPU, or may be realized as hardware by a wired logic.
  • the expansion device 10 can be implemented by installing an expansion program for executing the above data expansion as package software or online software in a desired computer.
  • the information processing apparatus can be caused to function as the expansion apparatus 10 by causing the information processing apparatus to execute the above-described expansion program.
  • the information processing device includes a desktop or notebook personal computer.
  • the information processing apparatus includes in its category a mobile communication terminal such as a smartphone, a mobile phone or a PHS (Personal Handyphone System), and a slate terminal such as a PDA (Personal Digital Assistant).
  • the expansion device 10 can be implemented as a terminal device used by a user as a client and as an expansion server device that provides the client with the service related to the above data expansion.
  • the extended server device is implemented as a server device that provides an extended service in which target data is input and extended data is output.
  • the expansion server device may be implemented as a Web server, or may be implemented as a cloud that provides the above-mentioned service related to data expansion by outsourcing.
  • FIG. 16 is a diagram illustrating an example of a computer that executes an extension program.
  • the computer 1000 has, for example, a memory 1010 and a CPU 1020.
  • the computer 1000 also has a hard disk drive interface 1030, a disk drive interface 1040, a serial port interface 1050, a video adapter 1060, and a network interface 1070. These units are connected by a bus 1080.
  • the memory 1010 includes a ROM (Read Only Memory) 1011 and a RAM 1012.
  • the ROM 1011 stores, for example, a boot program such as BIOS (Basic Input Output System).
  • BIOS Basic Input Output System
  • the hard disk drive interface 1030 is connected to the hard disk drive 1090.
  • the disk drive interface 1040 is connected to the disk drive 1100.
  • a removable storage medium such as a magnetic disk or an optical disk is inserted into the disk drive 1100.
  • the serial port interface 1050 is connected to, for example, a mouse 1110 and a keyboard 1120.
  • the video adapter 1060 is connected to the display 1130, for example.
  • the hard disk drive 1090 stores, for example, an OS 1091, an application program 1092, a program module 1093, and program data 1094. That is, the program that defines each process of the expansion device 10 is implemented as the program module 1093 in which the code executable by the computer is described.
  • the program module 1093 is stored in the hard disk drive 1090, for example.
  • a program module 1093 for executing the same processing as the functional configuration of the expansion device 10 is stored in the hard disk drive 1090.
  • the hard disk drive 1090 may be replaced by SSD.
  • the setting data used in the processing of the above-described embodiment is stored as the program data 1094 in, for example, the memory 1010 or the hard disk drive 1090. Then, the CPU 1020 reads the program module 1093 and the program data 1094 stored in the memory 1010 or the hard disk drive 1090 into the RAM 1012 as necessary, and executes the processing of the above-described embodiment.
  • the program module 1093 and the program data 1094 are not limited to being stored in the hard disk drive 1090, but may be stored in, for example, a removable storage medium and read by the CPU 1020 via the disk drive 1100 or the like. Alternatively, the program module 1093 and the program data 1094 may be stored in another computer connected via a network (LAN (Local Area Network), WAN (Wide Area Network), etc.). The program module 1093 and the program data 1094 may be read by the CPU 1020 from another computer via the network interface 1070.
  • LAN Local Area Network
  • WAN Wide Area Network

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

La présente invention concerne un dispositif d'augmentation qui sélectionne un ensemble de données prescrit à partir d'un ensemble de données externe sur la base d'une matrice calculée à partir du degré de similarité entre un ensemble de données cible et l'ensemble de données externe et à partir du degré d'uniformité de l'ensemble de données externe. Le dispositif d'augmentation effectue un apprentissage d'un modèle de génération, de telle sorte que, lorsqu'une classe cible est désignée pour un générateur, le générateur génère des données qui sont facilement identifiées par un identifiant comme étant authentiques et qui sont calculées par un classificateur de telle sorte que la probabilité d'appartenir à une pluralité de classes externes ne soit pas nulle et soit uniforme. Le dispositif d'augmentation désigne une classe cible pour le générateur du modèle de génération appris et génère des données. Le dispositif d'augmentation extrait, à partir des données générées et en tant que données d'augmentation, des données identifiées comme étant falsifiées par l'identifiant. Le dispositif d'augmentation fixe une étiquette cible aux données d'augmentation extraites.
PCT/JP2020/004113 2019-02-20 2020-02-04 Dispositif, procédé et programme d'augmentation WO2020170803A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2019-028851 2019-02-20
JP2019028851A JP2020135485A (ja) 2019-02-20 2019-02-20 拡張装置、拡張方法及び拡張プログラム

Publications (1)

Publication Number Publication Date
WO2020170803A1 true WO2020170803A1 (fr) 2020-08-27

Family

ID=72144650

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2020/004113 WO2020170803A1 (fr) 2019-02-20 2020-02-04 Dispositif, procédé et programme d'augmentation

Country Status (2)

Country Link
JP (1) JP2020135485A (fr)
WO (1) WO2020170803A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107464301A (zh) * 2017-08-07 2017-12-12 青岛松立软件信息技术股份有限公司 一种停车场出口收费系统及其收费管理方法
WO2022249415A1 (fr) * 2021-05-27 2022-12-01 日本電信電話株式会社 Dispositif de fourniture d'informations, procédé de fourniture d'informations et programme de fourniture d'informations

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
KONO, YOHEI ET AL.: "Data Expansion Using GAN", IPSJ SIG TECHNICAL REPORT, 3 May 2017 (2017-05-03), pages 2 - 5, ISSN: 2188-8701, Retrieved from the Internet <URL:https://ipsj.ixsq.nii.ac.jp/ej/?action=repository_uri&item_id=178747&file_id=1&file_no=1> [retrieved on 20200313] *
YAMAGUCHI, SHINYA ET AL.: "Study of Data Expansion Methods by GAN Using Multiple Data Sets", IBIS2018 DISCUSSION TRACK, 6 November 2018 (2018-11-06), Retrieved from the Internet <URL:http://ibisml.org/ibis2018/discussion> [retrieved on 20200313] *
YAMAGUCHI, SHINYA: "Data Expansion by Antagonistic Generation Network Multiple Data Set Simultaneous Learning", DEIM 2019 FINAL PROCEEDINGS, 4 March 2019 (2019-03-04), Retrieved from the Internet <URL:https://db-event.jpn.org/deim2019/post/papers/20.pdf> [retrieved on 20200313] *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107464301A (zh) * 2017-08-07 2017-12-12 青岛松立软件信息技术股份有限公司 一种停车场出口收费系统及其收费管理方法
WO2022249415A1 (fr) * 2021-05-27 2022-12-01 日本電信電話株式会社 Dispositif de fourniture d'informations, procédé de fourniture d'informations et programme de fourniture d'informations

Also Published As

Publication number Publication date
JP2020135485A (ja) 2020-08-31

Similar Documents

Publication Publication Date Title
WO2020045236A1 (fr) Dispositif, procédé et programme d&#39;augmentation
CN110362677B (zh) 文本数据类别的识别方法及装置、存储介质、计算机设备
JP7266674B2 (ja) 画像分類モデルの訓練方法、画像処理方法及び装置
JP2019028839A (ja) 分類器、分類器の学習方法、分類器における分類方法
JP5214760B2 (ja) 学習装置、方法及びプログラム
JP6158882B2 (ja) 生成装置、生成方法、及び生成プログラム
US11551080B2 (en) Learning dataset generation method, new learning dataset generation device and learning method using generated learning dataset
WO2020170803A1 (fr) Dispositif, procédé et programme d&#39;augmentation
WO2020045313A1 (fr) Dispositif d&#39;estimation de masque, procédé d&#39;estimation de masque, et programme d&#39;estimation de masque
CN111783873A (zh) 基于增量朴素贝叶斯模型的用户画像方法及装置
US10832036B2 (en) Meta-learning for facial recognition
WO2023088174A1 (fr) Procédé et appareil de détection de cible
JP6535134B2 (ja) 作成装置、作成プログラム、および作成方法
JP2010009517A (ja) パターン検出器の学習装置、学習方法及びプログラム
CN112801186A (zh) 一种验证图像生成方法、装置及设备
JP6214073B2 (ja) 生成装置、生成方法、及び生成プログラム
JP2013058095A (ja) 情報処理装置、推定機生成方法、及びプログラム
JP5491430B2 (ja) クラス分類装置、クラス分類方法、及びクラス分類プログラム
Esmaeili et al. Transduction with matrix completion using smoothed rank function
JPWO2018151124A1 (ja) パラメタ算出装置、パラメタ算出方法、及び、パラメタ算出プログラム
CN114912540A (zh) 迁移学习方法、装置、设备及存储介质
JP7047664B2 (ja) 学習装置、学習方法および予測システム
CN113239203A (zh) 一种基于知识图谱的筛选方法及装置
JP6930408B2 (ja) 推定装置、推定方法および推定プログラム
CN113822112A (zh) 用于确定标签权重的方法和装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20758783

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20758783

Country of ref document: EP

Kind code of ref document: A1