WO2024034196A1 - Trained model selection method, trained model selection device, and trained model selection program - Google Patents

Trained model selection method, trained model selection device, and trained model selection program Download PDF

Info

Publication number
WO2024034196A1
WO2024034196A1 PCT/JP2023/016265 JP2023016265W WO2024034196A1 WO 2024034196 A1 WO2024034196 A1 WO 2024034196A1 JP 2023016265 W JP2023016265 W JP 2023016265W WO 2024034196 A1 WO2024034196 A1 WO 2024034196A1
Authority
WO
WIPO (PCT)
Prior art keywords
model
trained
data
language
analysis target
Prior art date
Application number
PCT/JP2023/016265
Other languages
French (fr)
Japanese (ja)
Inventor
将隆 窪内
拓磨 西本
Original Assignee
堺化学工業株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 堺化学工業株式会社 filed Critical 堺化学工業株式会社
Publication of WO2024034196A1 publication Critical patent/WO2024034196A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/38Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/383Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Definitions

  • the present invention relates to a technique for selecting a trained model for analyzing an analysis target from a plurality of trained models.
  • trained models generated by deep learning etc. have been used to analyze data.
  • it is preferable to generate a trained model with generalization performance but in general, compared to a specialized trained model, a trained model with high generalization performance is suitable for each task. accuracy is poor. Therefore, multiple trained models are often generated that are specialized for the task.
  • Patent Document 1 when data (character data in image data, etc.) is input, similar test data with the highest degree of similarity to the input data is calculated from a similar database prepared in advance, and the similar test data is It is disclosed that the trained model is processed using a plurality of trained models and the trained model with the highest accuracy rate is selected.
  • test data user's face image, etc.
  • attribute information user's age, gender, etc.
  • the present invention was made to solve the above problem, and an object of the present invention is to quickly select a trained model suitable for an analysis target from a plurality of trained models.
  • a trained model selection method is a trained model selection method for selecting a trained model for analyzing an analysis target from a plurality of trained models, the method comprising: a trained model corresponding to each of the plurality of trained models; A model language vector that is a language vector, a model feature that is a feature of each of the plurality of trained models, or a link between the training data used for machine learning of each of the plurality of trained models and the analysis target.
  • a trained model corresponding to the analysis target is selected using the degree of similarity.
  • the model language vector is generated from language labels assigned to each of the trained models.
  • the model language vector is generated from a language label given to teacher data used for machine learning of each of the trained models.
  • the learned model selection method includes a receiving step of accepting the analysis target or language data corresponding to the analysis target as search data, and converting the search data into a language vector into a search language vector.
  • a language vector conversion step a language vector comparison step of comparing the search language vector with each of the model language vectors, and learning at least one from the plurality of trained models based on the comparison results of the language vector comparison step.
  • a selection step of selecting a completed model.
  • a trained model corresponding to a model language vector having the highest degree of similarity to the search language vector is selected.
  • the learned model selection method includes a receiving step of accepting the analysis target or language data corresponding to the analysis target as search data, and converting the search data into a language vector into a search language vector.
  • a trained model corresponding to a model feature having the highest degree of similarity to the search data feature is selected.
  • the learned model selection method includes a reception step of accepting the analysis target or language data corresponding to the analysis target as search data; a conversion step of converting the search data into data in the same format as the teacher data, a comparison step of comparing the search data with the teacher data, and at least one of the plurality of trained models based on the comparison result of the comparison step. a selection step of selecting a trained model.
  • the teacher data constitutes a plurality of teacher data sets respectively corresponding to the plurality of trained models
  • the comparison step the plurality of teacher data sets are sequentially selected and selected.
  • the teacher data of the selected teacher data set and the search data are compared, and the degree of similarity between the teacher data and the search data is calculated for each of the teacher data sets, and in the selection step, the degree of similarity is used to select the plurality of At least one teacher data set is selected from the teacher data sets, and a trained model corresponding to the selected teacher data set is selected.
  • the early search data is an image.
  • a trained model selection device is a trained model selection device that selects a trained model for analyzing an analysis target from a plurality of trained models, and wherein a trained model is assigned to each of the plurality of trained models.
  • a model language vector that is a language vector generated from a language label, a model feature that is a feature of each of the plurality of trained models, or a teacher used for machine learning of each of the plurality of trained models.
  • a trained model corresponding to the analysis target is selected using the degree of similarity between the data and the analysis target.
  • the trained model selection program is a trained model selection program that selects a trained model for analyzing an analysis target from a plurality of trained models, and is a trained model selection program that selects a trained model for analyzing an analysis target from a plurality of trained models, and is a trained model selection program that corresponds to each of the plurality of trained models.
  • a model language vector that is a language vector, a model feature that is a feature of each of the plurality of trained models, or a link between the training data used for machine learning of each of the plurality of trained models and the analysis target.
  • a computer is caused to execute a process of selecting a trained model according to the analysis target.
  • a trained model suitable for an analysis target can be quickly selected from a plurality of trained models.
  • FIG. 1 is a block diagram of a learned model selection device according to Embodiment 1 of the present invention.
  • FIG. This is an example of a vector space.
  • 3 is a flowchart showing a processing procedure of a learned model selection method according to Embodiment 1 of the present invention. This is an example of language data for search.
  • 7 is a flowchart showing a processing procedure of a learned model selection method according to a modification of Embodiment 1 of the present invention.
  • FIG. 2 is an explanatory diagram showing an example of generation of a model language vector.
  • FIG. 2 is a block diagram of a learned model selection device according to a second embodiment of the present invention. This is an example of a vector space.
  • FIG. 7 is a flowchart showing a processing procedure of a learned model selection method according to Embodiment 2 of the present invention.
  • FIG. 3 is a block diagram of a learned model selection device according to Embodiment 3 of the present invention.
  • 12 is a flowchart showing a processing procedure of a learned model selection method according to Embodiment 3 of the present invention.
  • Embodiment 1 of the present invention will be described below.
  • FIG. 1 is a block diagram of a learned model selection device 1 according to the first embodiment.
  • the trained model selection device 1 has a function of selecting a trained model for analyzing an analysis target from a plurality of trained models.
  • the analysis target is an image and the analysis method is segmentation, but the analysis target and analysis method are not particularly limited.
  • the trained model selection device 1 can be configured with a general-purpose computer, and its hardware configuration includes a processor such as a CPU or GPU, a main storage device such as a DRAM or SRAM (not shown), and an HDD or SSD. It is equipped with an auxiliary storage device 10.
  • the auxiliary storage device 10 includes a plurality of learned models M1 to Mn machine-learned using images, model language vectors V1 to Vn, as well as a learned model selection device 1 such as a learned model selection program. Contains various programs.
  • the learned models M1 to Mn are artificial intelligence models that have been machine learned using different learning data sets.
  • each learning data set may include the same teacher image in common, but it is preferable that each learning data set is composed of teacher images having different characteristics.
  • trained model M1 is machine learned using a training dataset that includes many teacher images of round particles
  • trained model M2 is machine learned using a training dataset that includes many teacher images of cherries. ing.
  • the number of trained models M1 to Mn is n (n ⁇ 2), but the number is not particularly limited.
  • the model language vectors V1 to Vn are generated from the language labels assigned to each of the trained models M1 to Mn.
  • the trained model M2 has been machine learned using a group of images that includes many images of cherries, and can analyze images of cherries and objects similar to cherries with high accuracy.
  • Color, coordinates" "Round, small, red, whatever” is given a language label.
  • the language label may be assigned by a human or by an algorithm such as image caption, mirror GAN, image to text, etc.
  • a model language vector V2 in which (shape, size, color, coordinates) is encoded is generated.
  • the number of dimensions of the model language vector is four, but the number of dimensions is not particularly limited, and may be several hundred dimensions, for example.
  • the k-dimensional model language vector can be represented by a vector space V shown in FIG. 2, for example. Further, in the above example, the model language vector included only elements that humans can recognize, but it may also include elements that humans cannot recognize.
  • the trained model selection device 1 includes a reception section 11, a language vector conversion section 12, a language vector comparison section 13, a selection section 14, and an analysis section 15 as functional blocks.
  • each of these units is realized in software by the processor of the learned model selection device 1 reading out the learned model selection program into the main storage device and executing it.
  • FIG. 3 is a flowchart showing the processing procedure of the learned model selection method according to the present embodiment. These steps S1 to S8 are executed by the trained model selection device 1. Note that from the viewpoint of the final purpose, steps S1 to S8 are processing steps of the image analysis method, and are divided into steps S1 to S3, which are preprocessing steps, and steps S4 to S8, which are analysis steps. Note that the calculation load can be reduced by creating a database for the preprocessing process.
  • step S1 the trained models M1 to Mn are stored in the auxiliary storage device 10.
  • the learned models M1 to Mn may be stored in an external storage device or cloud.
  • step S2 the language labels assigned to each of the trained models M1 to Mn are converted into model language vectors V1 to Vn.
  • Step S2 may be executed by the language vector converter 12.
  • step S3 the model language vectors V1 to Vn are stored in the auxiliary storage device 10.
  • the model language vectors V1 to Vn may be stored in an external storage device or cloud.
  • step S4 the reception unit 11 receives the analysis target or the language data corresponding to the analysis target as search data.
  • the reception unit 11 receives language data corresponding to the analysis target as search data.
  • the linguistic data for example, when an image of a cherry is to be analyzed, the user performs a search by inputting the linguistic data "small round red particles" into the search field R, as shown in FIG.
  • the language data is not particularly limited as long as it expresses the content or characteristics of the analysis target, and may be the name of the analysis target (for example, "cherry").
  • the search data may be the analysis target itself.
  • the analysis data is an image
  • the user performs a search by uploading the image data to the learned model selection device 1.
  • Uploaded images are converted into linguistic data using algorithms such as image captions, mirror GAN, and image to text.
  • step S6 the language vector comparison unit 13 compares the search language vector Va converted in step S5 with each of the model language vectors V1 to Vn.
  • the language vector comparison unit 13 calculates the similarity of the search language vector Va to each of the model language vectors V1 to Vn by numerical calculation.
  • the degree of similarity can be determined by a method using an algorithm such as cosine similarity or pattern matching, or by a known technique of inference using a trained model that has learned a data set in which the degree of similarity has been evaluated subjectively by a person.
  • step S7 the selection unit 14 selects at least one learned model from the learned models M1 to Mn based on the comparison result in step S6.
  • the selection unit 14 selects the trained model corresponding to the model language vector having the highest degree of similarity to the search language vector Va from among the model language vectors V1 to Vn as the trained model suitable for the analysis target. do.
  • the selection unit 14 may select a plurality of trained models as long as they are trained models that correspond to model language vectors that have a high degree of similarity to the search language vector Va.
  • trained models may be ranked in descending order of similarity to the search language vector Va, and the user may select one of the trained models.
  • step S8 the analysis unit 15 analyzes the analysis target using the trained model selected in step S7.
  • FIG. 5 is a flowchart showing the processing procedure of a learned model selection method according to a modification of the first embodiment, and these processes are executed by the learned model selection device 1 shown in FIG. 1.
  • This modification is characterized in that the model language vectors V1 to Vn stored in the auxiliary storage device 10 are generated from the language labels given to the teacher data used for machine learning of each of the trained models M1 to Mn. , is different from the above. That is, the flowchart shown in FIG. 5 differs from the flowchart shown in FIG. 3 in that step S2 is replaced with step S2'. Therefore, the explanation of steps S1, S3 to S8 will be omitted below.
  • step S2' the language labels given to the teacher data used for machine learning of each of the trained models M1 to Mn are converted into model language vectors V1 to Vn.
  • each of the training data T1 to Tm for generating the learned model Mx (2 ⁇ x ⁇ n) is given a linguistic label such as “round” or “square”.
  • the language label may be assigned by a human or by an algorithm such as image caption, mirror GAN, image to text, etc.
  • a model language vector Vx is generated by converting these language labels into language vectors.
  • ⁇ One-hot vector A method of storing words as individual vectors
  • ⁇ Distributed representation A method of representing words as low-dimensional vectors
  • ⁇ Bag-of-words A method of using the number of times a word appears in a sentence as an element of a vector
  • a trained model corresponding to the analysis target is selected from the trained models M1 to Mn by using the model language vectors V1 to Vn corresponding to each of the trained models M1 to Mn. ing. Since language vectors require less calculation to compare vectors, a trained model suitable for an analysis target can be selected more quickly than in the prior art.
  • Embodiment 2 of the present invention will be described below.
  • members having the same functions as those in Embodiment 1 described above are designated by the same reference numerals, and detailed description thereof will be omitted.
  • FIG. 7 is a block diagram of a trained model selection device 1' according to the second embodiment.
  • the trained model selection device 1' has a function of selecting a trained model for analyzing an analysis target from a plurality of trained models.
  • the analysis target is an image and the analysis method is segmentation, but the analysis target and the analysis method are not particularly limited.
  • the hardware configuration of the trained model selection device 1' is the same as that of the trained model selection device 1.
  • the auxiliary storage device 10 of the trained model selection device 1' includes n trained models M1 to Mn, model features F1 to Fn, feature conversion models MC, and trained model selection programs. Various programs for operating the model selection device 1' are stored.
  • the model feature quantities F1 to Fn are the feature quantities of each of the learned models M1 to Mn, respectively.
  • the features of the trained model are latent variables that accompany the trained model.
  • the trained model is a neural network
  • the hyperparameters and feature filters used to generate the trained model It is.
  • the model feature amount can be represented by a vector space W shown in FIG. 8, for example.
  • the feature quantity conversion model MC is an artificial intelligence model in which the relationship (probability distribution) between the model language vectors V1 to Vn (FIG. 1) and the model feature quantities F1 to Fn is machine-learned in advance.
  • the machine learning method is not particularly limited, but for example, linear regression or random forest can be applied.
  • the trained model selection device 1' includes a reception section 11, a language vector conversion section 12, a selection section 14, an analysis section 15, a feature amount conversion section 16, and a feature amount comparison section 17 as functional blocks. ing. That is, the trained model selection device 1' is the trained model selection device 1 shown in FIG. In this embodiment, each of these parts is realized in software by the processor of the learned model selection device 1' reading out the learned model selection program into the main storage and executing it.
  • FIG. 9 is a flowchart showing the processing procedure of the trained model selection method according to the present embodiment, and these steps S11 to S21 are executed by the trained model selection device 1'.
  • steps S11 to S21 are processing steps of the image analysis method, and are divided into steps S11 to S15, which are preprocessing steps, and steps S16 to S21, which are analysis steps. Note that the calculation load can be reduced by creating a database for the preprocessing process.
  • step S11 the trained models M1 to Mn are stored in the auxiliary storage device 10. Step S11 is similar to step S1 shown in FIG.
  • step S12 the language labels assigned to each of the learned models M1 to Mn are converted into model language vectors V1 to Vn.
  • Step S12 is similar to step S2 shown in FIG. 3, but may be similar to step S2' shown in FIG.
  • step S13 model feature quantities F1 to Fn of learned models M1 to Mn are calculated.
  • step S14 machine learning is performed on the relationship between the model language vectors V1 to Vn generated in step S12 and the model feature quantities F1 to Fn generated in step S13. Thereby, when a language vector is input, a feature amount conversion model MC is generated that outputs a feature amount corresponding to the language vector.
  • step S15 the model feature quantities F1 to Fn and the machine-learned feature quantity conversion model MC are stored in the auxiliary storage device 10. Note that these may be saved in an external storage device or cloud.
  • step S16 (reception step), the reception unit 11 receives the analysis target or the language data corresponding to the analysis target as search data.
  • Step S16 is similar to step S4 shown in FIG.
  • step S17 (language vector conversion step), the language vector conversion unit 12 converts the search data into a language vector and converts it into a search language vector Va.
  • Step S17 is similar to step S5 shown in FIG.
  • step S18 feature amount conversion step
  • the feature amount conversion unit 16 converts the search language vector Va converted in step S17 into the search data feature amount Fa using the feature amount conversion model MC.
  • the feature conversion model MC performs machine learning on the relationship between the model language vectors V1 to Vn of the trained models M1 to Mn and the model features F1 to Fn, so when the search language vector Va is input, the search A search data feature amount Fa, which is a feature amount corresponding to the language vector Va, is output.
  • step S19 feature amount comparison step
  • the feature amount comparison unit 17 compares the search data feature amount Fa converted in step S18 with each of the model feature amounts F1 to Fn.
  • the feature comparison unit 17 calculates the degree of similarity of the search data feature Fa to each of the model features F1 to Fn by numerical calculation.
  • the degree of similarity can be determined by a method using an algorithm such as cosine similarity or pattern matching, or by a known technique of inference using a trained model that has learned a data set in which the degree of similarity has been evaluated subjectively by a person.
  • step S20 the selection unit 14 selects at least one learned model from the learned models M1 to Mn based on the comparison result in step S19.
  • the selection unit 14 selects a trained model corresponding to a model feature having the highest degree of similarity to the search data feature Fa among the model features F1 to Fn as a trained model suitable for the analysis target. select. Note that the selection unit 14 may select a plurality of trained models as long as they correspond to model features having a high degree of similarity to the search data feature Fa.
  • step S21 the analysis unit 15 analyzes the analysis target using the learned model selected in step S20.
  • model feature quantities F1 to Fn which are the feature quantities of each of the trained models M1 to Mn
  • a trained model according to the analysis target is generated from the trained models M1 to Mn. Selected. Similar to language vectors, feature quantities require a small amount of calculation to compare feature quantities with each other, so compared to conventional techniques, a trained model suitable for an analysis target can be selected more quickly.
  • Embodiment 3 of the present invention will be described below.
  • members having the same functions as those in Embodiments 1 and 2 described above are denoted by the same reference numerals, and detailed description thereof will be omitted.
  • FIG. 10 is a block diagram of a trained model selection device 1'' according to the third embodiment.
  • the trained model selection device 1'' is similar to the trained model selection devices 1 and 1' shown in FIGS. 1 and 7. , has a function of selecting a trained model for analyzing an analysis target from a plurality of trained models.
  • the analysis target and teacher data are images, and the analysis method is segmentation, but the data formats and analysis method of the analysis target and teacher data are not particularly limited.
  • the hardware configuration of the learned model selection device 1'' is the same as that of the learned model selection devices 1 and 1'.
  • the auxiliary storage device 10 of the learned model selection device 1'' stores n learned models M1 to M1.
  • various programs for operating the learned model selection device 1'' such as a learned model selection program, are stored.
  • the teacher data sets S1 to Sn correspond to learned models M1 to Mn, respectively, and the teacher data used for machine learning of the learned models M1 to Mn constitute the teacher data sets S1 to Sn. That is, the teacher data set Sk (1 ⁇ k ⁇ n) is composed of m pieces of teacher data k-1 to km (m is an undefined integer) used for machine learning of the trained model Mk. . Note that in recent machine learning, transfer learning is often performed, so m is approximately several hundred.
  • the learned model selection device 1'' includes a reception section 11, a selection section 14, an analysis section 15, a search data conversion section 18, and a search data comparison section 19 as functional blocks.
  • the model selection device 1'' is the learned model selection device 1 shown in FIG. 1 in which the language vector comparison section 13 is replaced with a search data conversion section 18 and a search data comparison section 19.
  • each of these parts is realized in software by the processor of the learned model selection device 1'' reading out the learned model selection program into the main storage and executing it.
  • FIG. 11 is a flowchart showing the processing procedure of the learned model selection method according to the present embodiment, and these steps S31 to S37 is executed by the trained model selection device 1''.
  • steps S31 to S37 are processing steps of the image analysis method, and are divided into step S31, which is a preprocessing step, and steps S32 to S37, which are analysis steps. Note that the calculation load can be reduced by creating a database for the preprocessing process.
  • step S31 the trained models M1 to Mn and the teacher data sets S1 to Sn are stored in the auxiliary storage device 10.
  • step S32 reception step
  • the reception unit 11 receives the analysis target or the language data corresponding to the analysis target as search data.
  • step S33 If the search data is in the same format as the teacher data (image in this embodiment) (YES in step S33), the process moves to step S35. If the search data is in a format different from the teacher data (for example, language data) (NO in step S33), the process moves to step S34.
  • step S34 conversion step
  • the search data conversion unit 18 converts the search data into a teacher data format (image).
  • An algorithm such as mirror GAN can be used for conversion to an image.
  • step S35 the search data comparison unit 19 compares the search data (image) with the teacher data. Specifically, the search data comparison unit 19 sequentially selects the teacher data sets S1 to Sn, compares each teacher data of the selected teacher data sets with the search data, and compares the teacher data and the search data for each teacher data set. Calculate the degree of similarity with the data. In this embodiment, the search data comparison unit 19 calculates the average value or maximum value of similarity.
  • the degree of similarity between the training data and the search data can be determined by known techniques such as algorithmic methods such as cosine similarity and pattern matching, and inference using a trained model trained on a data set that evaluates the degree of similarity based on human subjectivity. can. Note that the search data comparison unit 19 does not need to compare all the teacher data of each teacher data set.
  • step S36 selection step
  • the selection unit 14 selects at least one learned model from the learned models M1 to Mn based on the comparison result in step S35.
  • the selection unit 14 selects the trained model corresponding to the teaching data set with the largest average or maximum similarity between the teaching data and the search data as the trained model suitable for the analysis target.
  • the selection unit 14 may select a plurality of trained models as long as they are trained models that correspond to a teacher data set with a large average value or maximum value of similarity between each teacher data and the search data.
  • step S37 the analysis unit 15 analyzes the analysis target using the trained model selected in step S36.
  • the analysis target is an image and the search data is language data
  • the language data is converted to an image and compared with the teacher data
  • the present invention is not limited to this. If the analysis target is in a format other than an image and the search data is language data, the language data is converted to data in the same format as the teacher data and compared with the teacher data.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Medical Informatics (AREA)
  • Databases & Information Systems (AREA)
  • Artificial Intelligence (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Mathematical Physics (AREA)
  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

A trained model selection method for selecting a trained model for analyzing an analysis target from among a plurality of trained models M1 to Mn, said trained model selection method selecting the trained model for the analysis target using model language vectors V1 to Vn, which are language vectors corresponding to the trained models M1 to Mn respectively, model feature quantities F1 to Fn, which are feature quantities of the trained models M1 to Mn respectively, or the degree of similarity between the analysis target and the training data that has been used for machine learning of each of the trained models M1 to Mn.

Description

学習済モデル選択方法、学習済モデル選択装置および学習済モデル選択プログラムTrained model selection method, trained model selection device, and trained model selection program
 本発明は、複数の学習済モデルから解析対象を解析するための学習済モデルを選択する技術に関する。 The present invention relates to a technique for selecting a trained model for analyzing an analysis target from a plurality of trained models.
 画像などのデータ解析において、深層学習などによって生成された学習済モデルを用いて解析が行われてきた。あらゆる課題に対応できることを重視する場合は、汎化性能をもつ学習済モデルを生成することが好ましいが、一般に、特化した学習済モデルと比較し、汎化性能が高い学習済モデルは各課題に対する精度が劣る。そのため、課題に合わせて特化した複数の学習済モデルを生成することが多い。 In analyzing data such as images, trained models generated by deep learning etc. have been used to analyze data. When it is important to be able to deal with all types of tasks, it is preferable to generate a trained model with generalization performance, but in general, compared to a specialized trained model, a trained model with high generalization performance is suitable for each task. accuracy is poor. Therefore, multiple trained models are often generated that are specialized for the task.
 従来では、複数の学習済モデルがある場合、課題に対してどの学習済モデルを使えばいいかを判断するのは人であり、トライアル&エラーが必要だった。そのため、最適な学習済モデルを選択するのに時間がかかっていた。 Previously, when there were multiple trained models, it was up to humans to decide which trained model to use for the task, requiring trial and error. Therefore, it took time to select the optimal trained model.
 そこで、課題に対して最適な学習済モデルを自動的に選択する技術が提案されている。例えば、特許文献1では、データ(画像データ中の文字データ等)が入力されると、あらかじめ用意された類似データベースから、入力データと最も類似度が高い類似テストデータを算出して、類似テストデータを複数の学習済モデルで処理し、最も正解率が高い学習済モデルを選択することが開示されている。また、特許文献2では、属性情報(ユーザの年齢、性別等)を付与したテストデータ(ユーザの顔画像等)をあらかじめ用意した複数の学習済モデルで処理し、最も正解率が高い学習済モデルを選択することが開示されている。 Therefore, a technology has been proposed that automatically selects the optimal trained model for the task. For example, in Patent Document 1, when data (character data in image data, etc.) is input, similar test data with the highest degree of similarity to the input data is calculated from a similar database prepared in advance, and the similar test data is It is disclosed that the trained model is processed using a plurality of trained models and the trained model with the highest accuracy rate is selected. In addition, in Patent Document 2, test data (user's face image, etc.) with attribute information (user's age, gender, etc.) is processed by multiple trained models prepared in advance, and the trained model with the highest accuracy rate is used. It is disclosed that it is possible to select.
特開2019-40417号公報JP 2019-40417 Publication 国際公開第2018/142766号International Publication No. 2018/142766
 しかしながら、特許文献1、2に記載の従来技術では、用意された複数の学習済モデルにテストデータを入力して、各学習済モデルの正解率(性能)を評価する必要があるため、演算処理量が多くなり、学習済モデルの選択に時間がかかるという問題がある。 However, in the conventional techniques described in Patent Documents 1 and 2, it is necessary to input test data to a plurality of prepared trained models and evaluate the accuracy rate (performance) of each trained model, so the calculation process There is a problem that the amount of training models increases and it takes time to select a trained model.
 本発明は、上記問題を解決するためになされたものであって、複数の学習済モデルから解析対象に適した学習済モデルを迅速に選択することを課題とする。 The present invention was made to solve the above problem, and an object of the present invention is to quickly select a trained model suitable for an analysis target from a plurality of trained models.
 本発明に係る学習済モデル選択方法は、複数の学習済モデルから解析対象を解析するための学習済モデルを選択する学習済モデル選択方法であって、前記複数の学習済モデルの各々に対応する言語ベクトルであるモデル言語ベクトル、前記複数の学習済モデルの各々の特徴量であるモデル特徴量、または、前記複数の学習済モデルの各々の機械学習に用いられた教師データと前記解析対象との類似度を用いて、前記解析対象に応じた学習済モデルを選択する。 A trained model selection method according to the present invention is a trained model selection method for selecting a trained model for analyzing an analysis target from a plurality of trained models, the method comprising: a trained model corresponding to each of the plurality of trained models; A model language vector that is a language vector, a model feature that is a feature of each of the plurality of trained models, or a link between the training data used for machine learning of each of the plurality of trained models and the analysis target. A trained model corresponding to the analysis target is selected using the degree of similarity.
 好ましい実施形態によれば、前記モデル言語ベクトルは、前記学習済モデルの各々に付与された言語ラベルから生成される。 According to a preferred embodiment, the model language vector is generated from language labels assigned to each of the trained models.
 好ましい実施形態によれば、前記モデル言語ベクトルは、前記学習済モデルの各々の機械学習に用いられた教師データに付与された言語ラベルから生成される。 According to a preferred embodiment, the model language vector is generated from a language label given to teacher data used for machine learning of each of the trained models.
 好ましい実施形態によれば、前記学習済モデル選択方法は、前記解析対象または前記解析対象に対応する言語データを検索データとして受け付ける受付ステップと、前記検索データを言語ベクトル化して検索言語ベクトルに変換する言語ベクトル変換ステップと、前記検索言語ベクトルを、前記モデル言語ベクトルの各々と比較する言語ベクトル比較ステップと、前記言語ベクトル比較ステップの比較結果に基づいて、前記複数の学習済モデルから少なくとも1つの学習済モデルを選択する選択ステップと、を備える。 According to a preferred embodiment, the learned model selection method includes a receiving step of accepting the analysis target or language data corresponding to the analysis target as search data, and converting the search data into a language vector into a search language vector. a language vector conversion step, a language vector comparison step of comparing the search language vector with each of the model language vectors, and learning at least one from the plurality of trained models based on the comparison results of the language vector comparison step. a selection step of selecting a completed model.
 好ましい実施形態によれば、前記選択ステップでは、前記検索言語ベクトルに対する類似度が最も大きいモデル言語ベクトルに対応する学習済モデルを選択する。 According to a preferred embodiment, in the selection step, a trained model corresponding to a model language vector having the highest degree of similarity to the search language vector is selected.
 好ましい実施形態によれば、前記学習済モデル選択方法は、前記解析対象または前記解析対象に対応する言語データを検索データとして受け付ける受付ステップと、前記検索データを言語ベクトル化して検索言語ベクトルに変換する言語ベクトル変換ステップと、前記モデル言語ベクトルと前記モデル特徴量との関係を機械学習した特徴量変換用モデルを用いて、前記検索言語ベクトルを検索データ特徴量に変換する特徴量変換ステップと、前記検索データ特徴量を、前記モデル特徴量の各々と比較する特徴量比較ステップと、前記特徴量比較ステップの比較結果に基づいて、前記複数の学習済モデルから少なくとも1つの学習済モデルを選択する選択ステップと、を備える。 According to a preferred embodiment, the learned model selection method includes a receiving step of accepting the analysis target or language data corresponding to the analysis target as search data, and converting the search data into a language vector into a search language vector. a feature amount conversion step of converting the search language vector into a search data feature using a feature conversion model obtained by machine learning the relationship between the model language vector and the model feature; a feature comparison step of comparing the search data feature with each of the model features; and a selection of selecting at least one trained model from the plurality of trained models based on the comparison result of the feature comparison step. and a step.
 好ましい実施形態によれば、前記選択ステップでは、前記検索データ特徴量に対する類似度が最も大きいモデル特徴量に対応する学習済モデルを選択する。 According to a preferred embodiment, in the selection step, a trained model corresponding to a model feature having the highest degree of similarity to the search data feature is selected.
 好ましい実施形態によれば、前記学習済モデル選択方法は、前記解析対象または前記解析対象に対応する言語データを検索データとして受け付ける受付ステップと、前記検索データが言語データである場合に、前記言語データを前記教師データと同一形式のデータに変換する変換ステップと、前記検索データを前記教師データと比較する比較ステップと、前記比較ステップの比較結果に基づいて、前記複数の学習済モデルから少なくとも1つの学習済モデルを選択する選択ステップと、を備える。 According to a preferred embodiment, the learned model selection method includes a reception step of accepting the analysis target or language data corresponding to the analysis target as search data; a conversion step of converting the search data into data in the same format as the teacher data, a comparison step of comparing the search data with the teacher data, and at least one of the plurality of trained models based on the comparison result of the comparison step. a selection step of selecting a trained model.
 好ましい実施形態によれば、前記教師データは、前記複数の学習済モデルにそれぞれ対応する複数の教師データセットを構成しており、前記比較ステップでは、前記複数の教師データセットを順次選択し、選択した教師データセットの教師データと前記検索データとを比較し、前記教師データセットごとに、教師データと検索データとの類似度を算出し、前記選択ステップでは、前記類似度を用いて前記複数の教師データセットから少なくとも1つの教師データセットを選択し、選択された教師データセットに対応する学習済モデルを選択する。 According to a preferred embodiment, the teacher data constitutes a plurality of teacher data sets respectively corresponding to the plurality of trained models, and in the comparison step, the plurality of teacher data sets are sequentially selected and selected. The teacher data of the selected teacher data set and the search data are compared, and the degree of similarity between the teacher data and the search data is calculated for each of the teacher data sets, and in the selection step, the degree of similarity is used to select the plurality of At least one teacher data set is selected from the teacher data sets, and a trained model corresponding to the selected teacher data set is selected.
 好ましい実施形態によれば、前期検索データは、画像である。 According to a preferred embodiment, the early search data is an image.
 本発明に係る学習済モデル選択装置は、複数の学習済モデルから解析対象を解析するための学習済モデルを選択する学習済モデル選択装置であって、前記複数の学習済モデルの各々に付与された言語ラベルから生成された言語ベクトルであるモデル言語ベクトル、前記複数の学習済モデルの各々の特徴量であるモデル特徴量、または、前記複数の学習済モデルの各々の機械学習に用いられた教師データと前記解析対象との類似度を用いて、前記解析対象に応じた学習済モデルを選択する。 A trained model selection device according to the present invention is a trained model selection device that selects a trained model for analyzing an analysis target from a plurality of trained models, and wherein a trained model is assigned to each of the plurality of trained models. a model language vector that is a language vector generated from a language label, a model feature that is a feature of each of the plurality of trained models, or a teacher used for machine learning of each of the plurality of trained models. A trained model corresponding to the analysis target is selected using the degree of similarity between the data and the analysis target.
 本発明に係る学習済モデル選択プログラムは、複数の学習済モデルから解析対象を解析するための学習済モデルを選択する学習済モデル選択プログラムであって、前記複数の学習済モデルの各々に対応する言語ベクトルであるモデル言語ベクトル、前記複数の学習済モデルの各々の特徴量であるモデル特徴量、または、前記複数の学習済モデルの各々の機械学習に用いられた教師データと前記解析対象との類似度を用いて、前記解析対象に応じた学習済モデルを選択する処理をコンピュータに実行させる。 The trained model selection program according to the present invention is a trained model selection program that selects a trained model for analyzing an analysis target from a plurality of trained models, and is a trained model selection program that selects a trained model for analyzing an analysis target from a plurality of trained models, and is a trained model selection program that corresponds to each of the plurality of trained models. A model language vector that is a language vector, a model feature that is a feature of each of the plurality of trained models, or a link between the training data used for machine learning of each of the plurality of trained models and the analysis target. Using the degree of similarity, a computer is caused to execute a process of selecting a trained model according to the analysis target.
 本発明によれば、複数の学習済モデルから解析対象に適した学習済モデルを迅速に選択することができる。 According to the present invention, a trained model suitable for an analysis target can be quickly selected from a plurality of trained models.
本発明の実施形態1に係る学習済モデル選択装置のブロック図である。1 is a block diagram of a learned model selection device according to Embodiment 1 of the present invention. FIG. ベクトル空間の一例である。This is an example of a vector space. 本発明の実施形態1に係る学習済モデル選択方法の処理手順を示すフローチャートである。3 is a flowchart showing a processing procedure of a learned model selection method according to Embodiment 1 of the present invention. 検索のための言語データの一例である。This is an example of language data for search. 本発明の実施形態1の変形例に係る学習済モデル選択方法の処理手順を示すフローチャートである。7 is a flowchart showing a processing procedure of a learned model selection method according to a modification of Embodiment 1 of the present invention. モデル言語ベクトルの生成する一例を示す説明図である。FIG. 2 is an explanatory diagram showing an example of generation of a model language vector. 本発明の実施形態2に係る学習済モデル選択装置のブロック図である。FIG. 2 is a block diagram of a learned model selection device according to a second embodiment of the present invention. ベクトル空間の一例である。This is an example of a vector space. 本発明の実施形態2に係る学習済モデル選択方法の処理手順を示すフローチャートである。7 is a flowchart showing a processing procedure of a learned model selection method according to Embodiment 2 of the present invention. 本発明の実施形態3に係る学習済モデル選択装置のブロック図である。FIG. 3 is a block diagram of a learned model selection device according to Embodiment 3 of the present invention. 本発明の実施形態3に係る学習済モデル選択方法の処理手順を示すフローチャートである。12 is a flowchart showing a processing procedure of a learned model selection method according to Embodiment 3 of the present invention.
 以下、本発明の実施形態について添付図面を参照して説明する。なお、本発明は、下記の実施形態に限定されるものではない。 Hereinafter, embodiments of the present invention will be described with reference to the accompanying drawings. Note that the present invention is not limited to the embodiments described below.
 [実施形態1]
 以下、本発明の実施形態1について説明する。
[Embodiment 1]
Embodiment 1 of the present invention will be described below.
 (学習済モデル選択装置)
 図1は、実施形態1に係る学習済モデル選択装置1のブロック図である。学習済モデル選択装置1は、複数の学習済モデルから解析対象を解析するための学習済モデルを選択する機能を有している。本実施形態において、解析対象は画像であり、解析方法はセグメンテーションであるが、解析対象および解析方法は特に限定されない。
(Learned model selection device)
FIG. 1 is a block diagram of a learned model selection device 1 according to the first embodiment. The trained model selection device 1 has a function of selecting a trained model for analyzing an analysis target from a plurality of trained models. In this embodiment, the analysis target is an image and the analysis method is segmentation, but the analysis target and analysis method are not particularly limited.
 学習済モデル選択装置1は、汎用のコンピュータで構成することができ、ハードウェア構成として、CPUやGPUなどのプロセッサ、DRAMやSRAMなどの主記憶装置(図示省略)、および、HDDやSSDなどの補助記憶装置10を備えている。補助記憶装置10には、画像を用いて機械学習された複数の学習済モデルM1~Mn、モデル言語ベクトルV1~Vnの他、学習済モデル選択プログラム等の学習済モデル選択装置1を動作させるための各種プログラムが格納されている。 The trained model selection device 1 can be configured with a general-purpose computer, and its hardware configuration includes a processor such as a CPU or GPU, a main storage device such as a DRAM or SRAM (not shown), and an HDD or SSD. It is equipped with an auxiliary storage device 10. The auxiliary storage device 10 includes a plurality of learned models M1 to Mn machine-learned using images, model language vectors V1 to Vn, as well as a learned model selection device 1 such as a learned model selection program. Contains various programs.
 学習済モデルM1~Mnは、互いに異なる学習データセットを用いて機械学習された人工知能モデルである。なお、各学習データセットは、同一の教師画像を共通に含んでもよいが、互いに特徴の異なる教師画像で構成されていることが好ましい。例えば、学習済モデルM1は、丸い粒子の教師画像を多く含む学習データセットを用いて機械学習されており、学習済モデルM2は、サクランボの教師画像を多く含む学習データセットを用いて機械学習されている。学習済モデルM1~Mnの数はn(n≧2)であるが、その数は特に限定されない。 The learned models M1 to Mn are artificial intelligence models that have been machine learned using different learning data sets. Note that each learning data set may include the same teacher image in common, but it is preferable that each learning data set is composed of teacher images having different characteristics. For example, trained model M1 is machine learned using a training dataset that includes many teacher images of round particles, and trained model M2 is machine learned using a training dataset that includes many teacher images of cherries. ing. The number of trained models M1 to Mn is n (n≧2), but the number is not particularly limited.
 本実施形態において、モデル言語ベクトルV1~Vnはそれぞれ、学習済モデルM1~Mnの各々に付与された言語ラベルから生成されたものである。例えば、学習済モデルM2は、サクランボの画像を多く含む画像群を用いて機械学習されており、サクランボおよびサクランボに類似する物体の画像を高精度に解析可能であるため、「形、大きさ、色、座標」=「丸い、小さい、赤色、なんでもいい」という言語ラベルが付与されている。言語ラベルは、人間によって付与されてもよいし、画像キャプション、mirror GAN、image to textなどのアルゴリズムによって付与されてもよい。この言語ラベルを言語ベクトル化することで、(形、大きさ、色、座標)が符号化されたモデル言語ベクトルV2が生成される。 In this embodiment, the model language vectors V1 to Vn are generated from the language labels assigned to each of the trained models M1 to Mn. For example, the trained model M2 has been machine learned using a group of images that includes many images of cherries, and can analyze images of cherries and objects similar to cherries with high accuracy. Color, coordinates" = "Round, small, red, whatever" is given a language label. The language label may be assigned by a human or by an algorithm such as image caption, mirror GAN, image to text, etc. By converting this language label into a language vector, a model language vector V2 in which (shape, size, color, coordinates) is encoded is generated.
 なお、上記の例では、モデル言語ベクトルの次元数は4次元であるが、次元数は特に限定されず、例えば数百次元であってもよい。k次元のモデル言語ベクトルは、例えば図2に示すベクトル空間Vで表すことができる。また、上記の例では、モデル言語ベクトルは人間が認識できる要素のみを含んでいたが、人間が認識できない要素を含んでもよい。 Note that in the above example, the number of dimensions of the model language vector is four, but the number of dimensions is not particularly limited, and may be several hundred dimensions, for example. The k-dimensional model language vector can be represented by a vector space V shown in FIG. 2, for example. Further, in the above example, the model language vector included only elements that humans can recognize, but it may also include elements that humans cannot recognize.
 学習済モデル選択装置1は、機能ブロックとして、受付部11と、言語ベクトル変換部12と、言語ベクトル比較部13と、選択部14と、解析部15とを備えている。本実施形態において、これらの各部は、学習済モデル選択装置1のプロセッサが学習済モデル選択プログラムを主記憶装置に読み出して実行することによってソフトウェア的に実現される。 The trained model selection device 1 includes a reception section 11, a language vector conversion section 12, a language vector comparison section 13, a selection section 14, and an analysis section 15 as functional blocks. In this embodiment, each of these units is realized in software by the processor of the learned model selection device 1 reading out the learned model selection program into the main storage device and executing it.
 (学習済モデル選択方法)
 学習済モデル選択装置1の上記各部の機能について、図3に基づいて説明する。図3は、本実施形態に係る学習済モデル選択方法の処理手順を示すフローチャートである。これらのステップS1~S8は、学習済モデル選択装置1によって実行される。なお、最終目的の観点では、ステップS1~S8は、画像解析方法の処理工程であり、前処理工程であるステップS1~S3と、解析工程であるステップS4~S8に区分される。なお、前処理工程はデータベース化することで計算負荷を軽減することも可能である。
(Trained model selection method)
The functions of the above-mentioned parts of the trained model selection device 1 will be explained based on FIG. 3. FIG. 3 is a flowchart showing the processing procedure of the learned model selection method according to the present embodiment. These steps S1 to S8 are executed by the trained model selection device 1. Note that from the viewpoint of the final purpose, steps S1 to S8 are processing steps of the image analysis method, and are divided into steps S1 to S3, which are preprocessing steps, and steps S4 to S8, which are analysis steps. Note that the calculation load can be reduced by creating a database for the preprocessing process.
 ステップS1では、学習済モデルM1~Mnを補助記憶装置10に保存する。なお、学習済モデルM1~Mnは、外部の記憶装置やクラウドに保存されてもよい。 In step S1, the trained models M1 to Mn are stored in the auxiliary storage device 10. Note that the learned models M1 to Mn may be stored in an external storage device or cloud.
 ステップS2では、学習済モデルM1~Mnの各々に付与された言語ラベルをモデル言語ベクトルV1~Vnに変換する。ステップS2は、言語ベクトル変換部12が実行してもよい。 In step S2, the language labels assigned to each of the trained models M1 to Mn are converted into model language vectors V1 to Vn. Step S2 may be executed by the language vector converter 12.
 ステップS3では、モデル言語ベクトルV1~Vnを補助記憶装置10に保存する。なお、モデル言語ベクトルV1~Vnは、外部の記憶装置やクラウドに保存されてもよい。 In step S3, the model language vectors V1 to Vn are stored in the auxiliary storage device 10. Note that the model language vectors V1 to Vn may be stored in an external storage device or cloud.
 ステップS4(受付ステップ)では、受付部11が、解析対象または解析対象に対応する言語データを検索データとして受け付ける。本実施形態では、受付部11は、解析対象に対応する言語データを検索データとして受け付ける。言語データは、例えば、サクランボの画像を解析対象とする場合、ユーザは、図4に示すように、「丸くて小さな赤い粒子」という言語データを検索欄Rに入力することにより検索を行う。言語データは、解析対象の内容や特徴を表現するものであれば特に限定されず、解析対象の名称(例えば「サクランボ」)であってもよい。 In step S4 (reception step), the reception unit 11 receives the analysis target or the language data corresponding to the analysis target as search data. In this embodiment, the reception unit 11 receives language data corresponding to the analysis target as search data. As for the linguistic data, for example, when an image of a cherry is to be analyzed, the user performs a search by inputting the linguistic data "small round red particles" into the search field R, as shown in FIG. The language data is not particularly limited as long as it expresses the content or characteristics of the analysis target, and may be the name of the analysis target (for example, "cherry").
 また、検索データは、解析対象そのものであってもよい。解析データが画像である場合、ユーザは画像データを学習済モデル選択装置1にアップロードすることにより検索を行う。アップロードされた画像は、画像キャプション、mirror GAN、image to textなどのアルゴリズムによって言語データに変換される。 Additionally, the search data may be the analysis target itself. When the analysis data is an image, the user performs a search by uploading the image data to the learned model selection device 1. Uploaded images are converted into linguistic data using algorithms such as image captions, mirror GAN, and image to text.
 ステップS5(言語ベクトル変換ステップ)では、言語ベクトル変換部12が、検索データを言語ベクトル化して検索言語ベクトルVaに変換する。例えば、検索データが、「丸くて小さな赤い粒子」である場合、言語ベクトル変換部12は、「形、大きさ、色、座標」=「丸い、小さい、赤色、なんでもいい」という検索言語ベクトルVaに変換する。検索データを検索言語ベクトルに変換する手法は、特に限定されず、単語をベクトル化する辞書や、言語データをベクトル化するように機械学習された人工知能モデルを用いてもよい。なお、検索ワードに対応する検索言語ベクトルは可変長でもよい。 In step S5 (language vector conversion step), the language vector conversion unit 12 converts the search data into a language vector and converts it into a search language vector Va. For example, if the search data is "round and small red particles", the language vector conversion unit 12 converts the search language vector Va such that "shape, size, color, coordinates" = "round, small, red, whatever". Convert to The method of converting search data into a search language vector is not particularly limited, and a dictionary that vectorizes words or an artificial intelligence model that has been machine learned to vectorize language data may be used. Note that the search language vector corresponding to the search word may have a variable length.
 ステップS6(言語ベクトル比較ステップ)では、言語ベクトル比較部13が、ステップS5で変換された検索言語ベクトルVaを、モデル言語ベクトルV1~Vnの各々と比較する。本実施形態では、言語ベクトル比較部13は、数値計算により検索言語ベクトルVaのモデル言語ベクトルV1~Vnの各々に対する類似度を算出する。類似度は、コサイン類似度やパターンマッチングなどアルゴリズムによる方法、人の主観により類似度を評価したデータセットを学習した学習済みモデルによる推論の公知の技術によって求めることができる。 In step S6 (language vector comparison step), the language vector comparison unit 13 compares the search language vector Va converted in step S5 with each of the model language vectors V1 to Vn. In this embodiment, the language vector comparison unit 13 calculates the similarity of the search language vector Va to each of the model language vectors V1 to Vn by numerical calculation. The degree of similarity can be determined by a method using an algorithm such as cosine similarity or pattern matching, or by a known technique of inference using a trained model that has learned a data set in which the degree of similarity has been evaluated subjectively by a person.
 ステップS7(選択ステップ)では、選択部14が、ステップS6の比較結果に基づいて、学習済モデルM1~Mnから少なくとも1つの学習済モデルを選択する。本実施形態では、選択部14は、モデル言語ベクトルV1~Vnのうち、検索言語ベクトルVaに対する類似度が最も大きいモデル言語ベクトルに対応する学習済モデルを、解析対象に適した学習済モデルとして選択する。なお、選択部14は、検索言語ベクトルVaに対する類似度が大きいモデル言語ベクトルに対応する学習済モデルであれば、学習済モデルを複数選択してもよい。また、検索言語ベクトルVaに対する類似度が大きい順に学習済モデルをランキング化し、それらの学習済モデルからユーザが選んでもよい。 In step S7 (selection step), the selection unit 14 selects at least one learned model from the learned models M1 to Mn based on the comparison result in step S6. In the present embodiment, the selection unit 14 selects the trained model corresponding to the model language vector having the highest degree of similarity to the search language vector Va from among the model language vectors V1 to Vn as the trained model suitable for the analysis target. do. Note that the selection unit 14 may select a plurality of trained models as long as they are trained models that correspond to model language vectors that have a high degree of similarity to the search language vector Va. Alternatively, trained models may be ranked in descending order of similarity to the search language vector Va, and the user may select one of the trained models.
 ステップS8では、解析部15が、ステップS7で選択された学習済モデルを用いて、解析対象の解析を行う。 In step S8, the analysis unit 15 analyzes the analysis target using the trained model selected in step S7.
 (変形例)
 図5は、実施形態1の変形例に係る学習済モデル選択方法の処理手順を示すフローチャートであり、これらの処理は、図1に示す学習済モデル選択装置1によって実行される。本変形例は、補助記憶装置10に保存されたモデル言語ベクトルV1~Vnが、学習済モデルM1~Mnの各々の機械学習に用いられた教師データに付与された言語ラベルから生成される点で、上記と異なっている。すなわち、図5に示すフローチャートは、図3に示すフローチャートにおいて、ステップS2をステップS2’に置き換えた点で異なっている。そのため、以下ではステップS1、S3~S8の説明は省略する。
(Modified example)
FIG. 5 is a flowchart showing the processing procedure of a learned model selection method according to a modification of the first embodiment, and these processes are executed by the learned model selection device 1 shown in FIG. 1. This modification is characterized in that the model language vectors V1 to Vn stored in the auxiliary storage device 10 are generated from the language labels given to the teacher data used for machine learning of each of the trained models M1 to Mn. , is different from the above. That is, the flowchart shown in FIG. 5 differs from the flowchart shown in FIG. 3 in that step S2 is replaced with step S2'. Therefore, the explanation of steps S1, S3 to S8 will be omitted below.
 ステップS2’では、学習済モデルM1~Mnの各々の機械学習に用いられた教師データに付与された言語ラベルをモデル言語ベクトルV1~Vnに変換する。例えば、図6に示すように、学習済モデルMx(2≦x≦n)を生成するための教師データT1~Tmの各々には、「丸い」、「四角い」等の言語ラベルが付与されている。言語ラベルは、人間によって付与されてもよいし、画像キャプション、mirror GAN、image to textなどのアルゴリズムによって付与されてもよい。これらの言語ラベルを、言語ベクトル化することにより、モデル言語ベクトルVxが生成される。 In step S2', the language labels given to the teacher data used for machine learning of each of the trained models M1 to Mn are converted into model language vectors V1 to Vn. For example, as shown in FIG. 6, each of the training data T1 to Tm for generating the learned model Mx (2≦x≦n) is given a linguistic label such as “round” or “square”. There is. The language label may be assigned by a human or by an algorithm such as image caption, mirror GAN, image to text, etc. A model language vector Vx is generated by converting these language labels into language vectors.
 言語ベクトル化の具体的手法は特に限定されないが、例えば、以下の手法が挙げられる。
・one-hotベクトル:単語をそれぞれのベクトルで格納する方法
・分散表現:単語を低次元のベクトルで表現する手法
・Bag-of-words:文章内の単語出現回数をベクトルの要素とする方法
これらの手法については、例えば、https://deepage.net/bigdata/machine_learning/2016/09/02/word2vec_power_of_word_vector.html、https://deepage.net/machine_learning/2017/01/08/doc2vec.htmlを参照されたい。
Although the specific method of language vectorization is not particularly limited, examples include the following methods.
・One-hot vector: A method of storing words as individual vectors ・Distributed representation: A method of representing words as low-dimensional vectors ・Bag-of-words: A method of using the number of times a word appears in a sentence as an element of a vector For methods, see, for example, https://deepage.net/bigdata/machine_learning/2016/09/02/word2vec_power_of_word_vector.html, https://deepage.net/machine_learning/2017/01/08/doc2vec.html I want to be
 また、上記手法は単語をベクトル化する手法であるが、image to textやGAN(Generative Adversarial Networks)などで文章が生成される場合は、文章のベクトル化も可能である(例えば、Doc2Vec、BERT)。 Additionally, although the above method vectorizes words, if the text is generated using image to text or GAN (Generative Adversarial Networks), it is also possible to vectorize the text (e.g., Doc2Vec, BERT). .
 (小括)
 以上のように、実施形態1では、学習済モデルM1~Mnの各々に対応するモデル言語ベクトルV1~Vnを用いることにより、学習済モデルM1~Mnから解析対象に応じた学習済モデルを選択している。言語ベクトルは、ベクトル同士の比較のための演算量が少ないため、従来技術に比べ、解析対象に適した学習済モデルを迅速に選択することができる。
(Brief Summary)
As described above, in the first embodiment, a trained model corresponding to the analysis target is selected from the trained models M1 to Mn by using the model language vectors V1 to Vn corresponding to each of the trained models M1 to Mn. ing. Since language vectors require less calculation to compare vectors, a trained model suitable for an analysis target can be selected more quickly than in the prior art.
 [実施形態2]
 以下、本発明の実施形態2について説明する。なお、実施形態2において、上述の実施形態1におけるものと同様の機能を有する部材については、同一の符号を付し、詳細な説明を省略する。
[Embodiment 2]
Embodiment 2 of the present invention will be described below. In Embodiment 2, members having the same functions as those in Embodiment 1 described above are designated by the same reference numerals, and detailed description thereof will be omitted.
 (学習済モデル選択装置)
 図7は、実施形態2に係る学習済モデル選択装置1’のブロック図である。学習済モデル選択装置1’は、図1に示す学習済モデル選択装置1と同様に、複数の学習済モデルから解析対象を解析するための学習済モデルを選択する機能を有している。本実施形態においても、解析対象は画像であり、解析方法はセグメンテーションであるが、解析対象および解析方法は特に限定されない。
(Learned model selection device)
FIG. 7 is a block diagram of a trained model selection device 1' according to the second embodiment. Like the trained model selection device 1 shown in FIG. 1, the trained model selection device 1' has a function of selecting a trained model for analyzing an analysis target from a plurality of trained models. In this embodiment as well, the analysis target is an image and the analysis method is segmentation, but the analysis target and the analysis method are not particularly limited.
 学習済モデル選択装置1’のハードウェア構成は、学習済モデル選択装置1と同様である。学習済モデル選択装置1’の補助記憶装置10には、n個の学習済モデルM1~Mn、モデル特徴量F1~Fn、特徴量変換用モデルMCの他、学習済モデル選択プログラム等の学習済モデル選択装置1’を動作させるための各種プログラムが格納されている。 The hardware configuration of the trained model selection device 1' is the same as that of the trained model selection device 1. The auxiliary storage device 10 of the trained model selection device 1' includes n trained models M1 to Mn, model features F1 to Fn, feature conversion models MC, and trained model selection programs. Various programs for operating the model selection device 1' are stored.
 モデル特徴量F1~Fnはそれぞれ、学習済モデルM1~Mnの各々の特徴量である。本実施形態において、学習済モデルの特徴量は、学習済モデルに付随する潜在変数であり、例えば、学習済モデルがニューラルネットワークである場合、学習済モデルを生成する際のハイパーパラメータおよび特徴量フィルタである。モデル特徴量は、例えば図8に示すベクトル空間Wで表すことができる。 The model feature quantities F1 to Fn are the feature quantities of each of the learned models M1 to Mn, respectively. In this embodiment, the features of the trained model are latent variables that accompany the trained model. For example, if the trained model is a neural network, the hyperparameters and feature filters used to generate the trained model It is. The model feature amount can be represented by a vector space W shown in FIG. 8, for example.
 特徴量変換用モデルMCは、モデル言語ベクトルV1~Vn(図1)とモデル特徴量F1~Fnとの関係(確率分布)をあらかじめ機械学習した人工知能モデルである。機械学習方法は特に限定されないが、例えば、線形回帰やランダムフォレストを適用できる。 The feature quantity conversion model MC is an artificial intelligence model in which the relationship (probability distribution) between the model language vectors V1 to Vn (FIG. 1) and the model feature quantities F1 to Fn is machine-learned in advance. The machine learning method is not particularly limited, but for example, linear regression or random forest can be applied.
 学習済モデル選択装置1’は、機能ブロックとして、受付部11と、言語ベクトル変換部12と、選択部14と、解析部15と、特徴量変換部16と、特徴量比較部17とを備えている。すなわち、学習済モデル選択装置1’は、図1に示す学習済モデル選択装置1において、言語ベクトル比較部13を特徴量変換部16および特徴量比較部17に置き換えたものである。本実施形態において、これらの各部は、学習済モデル選択装置1’のプロセッサが学習済モデル選択プログラムを主記憶装置に読み出して実行することによってソフトウェア的に実現される。 The trained model selection device 1' includes a reception section 11, a language vector conversion section 12, a selection section 14, an analysis section 15, a feature amount conversion section 16, and a feature amount comparison section 17 as functional blocks. ing. That is, the trained model selection device 1' is the trained model selection device 1 shown in FIG. In this embodiment, each of these parts is realized in software by the processor of the learned model selection device 1' reading out the learned model selection program into the main storage and executing it.
 (学習済モデル選択方法)
 学習済モデル選択装置1’の上記各部の機能について、図9に基づいて説明する。図9は、本実施形態に係る学習済モデル選択方法の処理手順を示すフローチャートであり、これらのステップS11~S21は、学習済モデル選択装置1’によって実行される。なお、最終目的の観点では、ステップS11~S21は、画像解析方法の処理工程であり、前処理工程であるステップS11~S15と、解析工程であるステップS16~S21に区分される。なお、前処理工程はデータベース化することで計算負荷を軽減することも可能である。
(Trained model selection method)
The functions of the above-mentioned parts of the trained model selection device 1' will be explained based on FIG. 9. FIG. 9 is a flowchart showing the processing procedure of the trained model selection method according to the present embodiment, and these steps S11 to S21 are executed by the trained model selection device 1'. Note that from the viewpoint of the final purpose, steps S11 to S21 are processing steps of the image analysis method, and are divided into steps S11 to S15, which are preprocessing steps, and steps S16 to S21, which are analysis steps. Note that the calculation load can be reduced by creating a database for the preprocessing process.
 ステップS11では、学習済モデルM1~Mnを補助記憶装置10に保存する。ステップS11は、図3に示すステップS1と同様である。 In step S11, the trained models M1 to Mn are stored in the auxiliary storage device 10. Step S11 is similar to step S1 shown in FIG.
 ステップS12では、学習済モデルM1~Mnの各々に付与された言語ラベルをモデル言語ベクトルV1~Vnに変換する。ステップS12は、図3に示すステップS2と同様であるが、図5に示すステップS2’と同様であってもよい。 In step S12, the language labels assigned to each of the learned models M1 to Mn are converted into model language vectors V1 to Vn. Step S12 is similar to step S2 shown in FIG. 3, but may be similar to step S2' shown in FIG.
 ステップS13では、学習済モデルM1~Mnのモデル特徴量F1~Fnを演算する。 In step S13, model feature quantities F1 to Fn of learned models M1 to Mn are calculated.
 ステップS14では、ステップS12で生成されたモデル言語ベクトルV1~Vnと、ステップS13で生成されたモデル特徴量F1~Fnとの関係を機械学習する。これにより、言語ベクトルを入力すると、当該言語ベクトルに対応する特徴量を出力する特徴量変換用モデルMCが生成される。 In step S14, machine learning is performed on the relationship between the model language vectors V1 to Vn generated in step S12 and the model feature quantities F1 to Fn generated in step S13. Thereby, when a language vector is input, a feature amount conversion model MC is generated that outputs a feature amount corresponding to the language vector.
 ステップS15では、モデル特徴量F1~Fnと、機械学習された特徴量変換用モデルMCを補助記憶装置10に保存する。なお、これらは外部の記憶装置やクラウドに保存されてもよい。 In step S15, the model feature quantities F1 to Fn and the machine-learned feature quantity conversion model MC are stored in the auxiliary storage device 10. Note that these may be saved in an external storage device or cloud.
 ステップS16(受付ステップ)では、受付部11が、解析対象または解析対象に対応する言語データを検索データとして受け付ける。ステップS16は、図3に示すステップS4と同様である。 In step S16 (reception step), the reception unit 11 receives the analysis target or the language data corresponding to the analysis target as search data. Step S16 is similar to step S4 shown in FIG.
 ステップS17(言語ベクトル変換ステップ)では、言語ベクトル変換部12が、検索データを言語ベクトル化して検索言語ベクトルVaに変換する。ステップS17は、図3に示すステップS5と同様である。 In step S17 (language vector conversion step), the language vector conversion unit 12 converts the search data into a language vector and converts it into a search language vector Va. Step S17 is similar to step S5 shown in FIG.
 ステップS18(特徴量変換ステップ)では、特徴量変換部16が、特徴量変換用モデルMCを用いて、ステップS17で変換された検索言語ベクトルVaを検索データ特徴量Faに変換する。特徴量変換用モデルMCは、学習済モデルM1~Mnのモデル言語ベクトルV1~Vnとモデル特徴量F1~Fnとの関係を機械学習しているため、検索言語ベクトルVaが入力されると、検索言語ベクトルVaに対応する特徴量である検索データ特徴量Faを出力する。 In step S18 (feature amount conversion step), the feature amount conversion unit 16 converts the search language vector Va converted in step S17 into the search data feature amount Fa using the feature amount conversion model MC. The feature conversion model MC performs machine learning on the relationship between the model language vectors V1 to Vn of the trained models M1 to Mn and the model features F1 to Fn, so when the search language vector Va is input, the search A search data feature amount Fa, which is a feature amount corresponding to the language vector Va, is output.
 ステップS19(特徴量比較ステップ)では、特徴量比較部17が、ステップS18で変換された検索データ特徴量Faを、モデル特徴量F1~Fnの各々と比較する。本実施形態では、特徴量比較部17は、数値計算により検索データ特徴量Faのモデル特徴量F1~Fnの各々に対する類似度を算出する。類似度は、コサイン類似度やパターンマッチングなどアルゴリズムによる方法、人の主観により類似度を評価したデータセットを学習した学習済みモデルによる推論の公知の技術によって求めることができる。 In step S19 (feature amount comparison step), the feature amount comparison unit 17 compares the search data feature amount Fa converted in step S18 with each of the model feature amounts F1 to Fn. In this embodiment, the feature comparison unit 17 calculates the degree of similarity of the search data feature Fa to each of the model features F1 to Fn by numerical calculation. The degree of similarity can be determined by a method using an algorithm such as cosine similarity or pattern matching, or by a known technique of inference using a trained model that has learned a data set in which the degree of similarity has been evaluated subjectively by a person.
 ステップS20(選択ステップ)では、選択部14が、ステップS19の比較結果に基づいて、学習済モデルM1~Mnから少なくとも1つの学習済モデルを選択する。本実施形態では、選択部14は、モデル特徴量F1~Fnのうち、検索データ特徴量Faに対する類似度が最も大きいモデル特徴量に対応する学習済モデルを、解析対象に適した学習済モデルとして選択する。なお、選択部14は、検索データ特徴量Faに対する類似度が大きいモデル特徴量に対応する学習済モデルであれば、学習済モデルを複数選択してもよい。 In step S20 (selection step), the selection unit 14 selects at least one learned model from the learned models M1 to Mn based on the comparison result in step S19. In the present embodiment, the selection unit 14 selects a trained model corresponding to a model feature having the highest degree of similarity to the search data feature Fa among the model features F1 to Fn as a trained model suitable for the analysis target. select. Note that the selection unit 14 may select a plurality of trained models as long as they correspond to model features having a high degree of similarity to the search data feature Fa.
 ステップS21では、解析部15が、ステップS20で選択された学習済モデルを用いて、解析対象の解析を行う。 In step S21, the analysis unit 15 analyzes the analysis target using the learned model selected in step S20.
 (小括)
 以上のように、実施形態2では、学習済モデルM1~Mnの各々の特徴量であるモデル特徴量F1~Fnを用いることにより、学習済モデルM1~Mnから解析対象に応じた学習済モデルを選択している。特徴量も言語ベクトルと同様、特徴量同士の比較のための演算量が少ないため、従来技術に比べ、解析対象に適した学習済モデルを迅速に選択することができる。
(Brief Summary)
As described above, in the second embodiment, by using the model feature quantities F1 to Fn, which are the feature quantities of each of the trained models M1 to Mn, a trained model according to the analysis target is generated from the trained models M1 to Mn. Selected. Similar to language vectors, feature quantities require a small amount of calculation to compare feature quantities with each other, so compared to conventional techniques, a trained model suitable for an analysis target can be selected more quickly.
 [実施形態3]
 以下、本発明の実施形態3について説明する。なお、実施形態3において、上述の実施形態1および2におけるものと同様の機能を有する部材については、同一の符号を付し、詳細な説明を省略する。
[Embodiment 3]
Embodiment 3 of the present invention will be described below. In Embodiment 3, members having the same functions as those in Embodiments 1 and 2 described above are denoted by the same reference numerals, and detailed description thereof will be omitted.
 (学習済モデル選択装置)
 図10は、実施形態3に係る学習済モデル選択装置1”のブロック図である。学習済モデル選択装置1”は、図1および図7に示す学習済モデル選択装置1、1’と同様に、複数の学習済モデルから解析対象を解析するための学習済モデルを選択する機能を有している。本実施形態において、解析対象および教師データは画像であり、解析方法はセグメンテーションであるが、解析対象および教師データのデータ形式、並びに解析方法は特に限定されない。
(Learned model selection device)
FIG. 10 is a block diagram of a trained model selection device 1'' according to the third embodiment.The trained model selection device 1'' is similar to the trained model selection devices 1 and 1' shown in FIGS. 1 and 7. , has a function of selecting a trained model for analyzing an analysis target from a plurality of trained models. In this embodiment, the analysis target and teacher data are images, and the analysis method is segmentation, but the data formats and analysis method of the analysis target and teacher data are not particularly limited.
 学習済モデル選択装置1”のハードウェア構成は、学習済モデル選択装置1、1’と同様である。学習済モデル選択装置1”の補助記憶装置10には、n個の学習済モデルM1~Mn、n個の教師データセットS1~Snの他、学習済モデル選択プログラム等の学習済モデル選択装置1”を動作させるための各種プログラムが格納されている。 The hardware configuration of the learned model selection device 1'' is the same as that of the learned model selection devices 1 and 1'.The auxiliary storage device 10 of the learned model selection device 1'' stores n learned models M1 to M1. In addition to Mn and n teacher data sets S1 to Sn, various programs for operating the learned model selection device 1'', such as a learned model selection program, are stored.
 教師データセットS1~Snはそれぞれ、学習済モデルM1~Mnに対応しており、学習済モデルM1~Mnの機械学習に用いられた教師データは、教師データセットS1~Snを構成している。すなわち、教師データセットSk(1≦k≦n)は、学習済モデルMkの機械学習に用いられたm個の教師データk-1~k-m(mは不定の整数)で構成されている。なお、近年の機械学習では転移学習が行われることが多いため、mは数百程度である。 The teacher data sets S1 to Sn correspond to learned models M1 to Mn, respectively, and the teacher data used for machine learning of the learned models M1 to Mn constitute the teacher data sets S1 to Sn. That is, the teacher data set Sk (1≦k≦n) is composed of m pieces of teacher data k-1 to km (m is an undefined integer) used for machine learning of the trained model Mk. . Note that in recent machine learning, transfer learning is often performed, so m is approximately several hundred.
 学習済モデル選択装置1”は、機能ブロックとして、受付部11と、選択部14と、解析部15と、検索データ変換部18と、検索データ比較部19とを備えている。すなわち、学習済モデル選択装置1”は、図1に示す学習済モデル選択装置1において、言語ベクトル比較部13を検索データ変換部18および検索データ比較部19に置き換えたものである。本実施形態において、これらの各部は、学習済モデル選択装置1”のプロセッサが学習済モデル選択プログラムを主記憶装置に読み出して実行することによってソフトウェア的に実現される。 The learned model selection device 1'' includes a reception section 11, a selection section 14, an analysis section 15, a search data conversion section 18, and a search data comparison section 19 as functional blocks. The model selection device 1'' is the learned model selection device 1 shown in FIG. 1 in which the language vector comparison section 13 is replaced with a search data conversion section 18 and a search data comparison section 19. In this embodiment, each of these parts is realized in software by the processor of the learned model selection device 1'' reading out the learned model selection program into the main storage and executing it.
 (学習済モデル選択方法)
 学習済モデル選択装置1”の上記各部の機能について、図11に基づいて説明する。図11は、本実施形態に係る学習済モデル選択方法の処理手順を示すフローチャートであり、これらのステップS31~S37は、学習済モデル選択装置1”によって実行される。なお、最終目的の観点では、ステップS31~S37は、画像解析方法の処理工程であり、前処理工程であるステップS31と、解析工程であるステップS32~S37に区分される。なお、前処理工程はデータベース化することで計算負荷を軽減することも可能である。
(Trained model selection method)
The functions of the above-mentioned parts of the learned model selection device 1'' will be explained based on FIG. 11. FIG. 11 is a flowchart showing the processing procedure of the learned model selection method according to the present embodiment, and these steps S31 to S37 is executed by the trained model selection device 1''. Note that from the viewpoint of the final purpose, steps S31 to S37 are processing steps of the image analysis method, and are divided into step S31, which is a preprocessing step, and steps S32 to S37, which are analysis steps. Note that the calculation load can be reduced by creating a database for the preprocessing process.
 ステップS31では、学習済モデルM1~Mnおよび教師データセットS1~Snを補助記憶装置10に保存する。 In step S31, the trained models M1 to Mn and the teacher data sets S1 to Sn are stored in the auxiliary storage device 10.
 ステップS32(受付ステップ)では、受付部11が、解析対象または解析対象に対応する言語データを検索データとして受け付ける。 In step S32 (reception step), the reception unit 11 receives the analysis target or the language data corresponding to the analysis target as search data.
 検索データが教師データと同じ形式(本実施形態では画像)である場合(ステップS33でYES)、ステップS35に移行する。検索データが教師データと異なる形式(例えば、言語データ)である場合(ステップS33でNO)、ステップS34に移行する。 If the search data is in the same format as the teacher data (image in this embodiment) (YES in step S33), the process moves to step S35. If the search data is in a format different from the teacher data (for example, language data) (NO in step S33), the process moves to step S34.
 ステップS34(変換ステップ)では、検索データ変換部18が、検索データを教師データの形式(画像)に変換する。画像への変換には、mirror GAN等のアルゴリズムを用いることができる。 In step S34 (conversion step), the search data conversion unit 18 converts the search data into a teacher data format (image). An algorithm such as mirror GAN can be used for conversion to an image.
 ステップS35(比較ステップ)では、検索データ比較部19が、検索データ(画像)を教師データと比較する。具体的には、検索データ比較部19は、教師データセットS1~Snを順次選択し、選択した教師データセットの各教師データと検索データとを比較し、教師データセットごとに、教師データと検索データとの類似度を算出する。本実施形態では、検索データ比較部19は、類似度の平均値または最大値を算出する。教師データと検索データとの類似度は、コサイン類似度やパターンマッチングなどアルゴリズムによる方法、人の主観により類似度を評価したデータセットを学習した学習済みモデルによる推論等の公知の技術によって求めることができる。なお、検索データ比較部19は、各教師データセットの全ての教師データを比較対象とする必要はない。 In step S35 (comparison step), the search data comparison unit 19 compares the search data (image) with the teacher data. Specifically, the search data comparison unit 19 sequentially selects the teacher data sets S1 to Sn, compares each teacher data of the selected teacher data sets with the search data, and compares the teacher data and the search data for each teacher data set. Calculate the degree of similarity with the data. In this embodiment, the search data comparison unit 19 calculates the average value or maximum value of similarity. The degree of similarity between the training data and the search data can be determined by known techniques such as algorithmic methods such as cosine similarity and pattern matching, and inference using a trained model trained on a data set that evaluates the degree of similarity based on human subjectivity. can. Note that the search data comparison unit 19 does not need to compare all the teacher data of each teacher data set.
 ステップS36(選択ステップ)では、選択部14が、ステップS35の比較結果に基づいて、学習済モデルM1~Mnから少なくとも1つの学習済モデルを選択する。本実施形態では、選択部14は、教師データと検索データとの類似度の平均値または最大値が最も大きい教師データセットに対応する学習済モデルを、解析対象に適した学習済モデルとして選択する。なお、選択部14は、各教師データと検索データとの類似度の平均値または最大値が大きい教師データセットに対応する学習済モデルであれば、学習済モデルを複数選択してもよい。 In step S36 (selection step), the selection unit 14 selects at least one learned model from the learned models M1 to Mn based on the comparison result in step S35. In the present embodiment, the selection unit 14 selects the trained model corresponding to the teaching data set with the largest average or maximum similarity between the teaching data and the search data as the trained model suitable for the analysis target. . Note that the selection unit 14 may select a plurality of trained models as long as they are trained models that correspond to a teacher data set with a large average value or maximum value of similarity between each teacher data and the search data.
 ステップS37では、解析部15が、ステップS36で選択された学習済モデルを用いて、解析対象の解析を行う。 In step S37, the analysis unit 15 analyzes the analysis target using the trained model selected in step S36.
 (小括)
 以上のように、実施形態3では、学習済モデルM1~Mnの各々の機械学習に用いられた教師データと解析対象との類似度を用いることにより、学習済モデルM1~Mnから解析対象に応じた学習済モデルを選択している。本実施形態では、学習済モデルにテストデータを入力する必要がないため、従来技術に比べ、解析対象に適した学習済モデルを迅速に選択することができる。
(Brief Summary)
As described above, in the third embodiment, by using the degree of similarity between the training data used for machine learning of each of the trained models M1 to Mn and the analysis target, A trained model is selected. In this embodiment, since there is no need to input test data to a trained model, a trained model suitable for an analysis target can be selected more quickly than in the prior art.
 なお、本実施形態では、解析対象は画像であり、検索データが言語データである場合は、言語データを画像に変換して教師データと比較していたが、本発明はこれに限定されない。解析対象が画像以外の形式であり、検索データが言語データである場合は、言語データを教師データと同一形式のデータに変換して教師データと比較する。 Note that in this embodiment, when the analysis target is an image and the search data is language data, the language data is converted to an image and compared with the teacher data, but the present invention is not limited to this. If the analysis target is in a format other than an image and the search data is language data, the language data is converted to data in the same format as the teacher data and compared with the teacher data.
 (付記事項)
 以上、本発明の実施形態について説明したが、本発明は上記実施形態に限定されるものではなく、その趣旨を逸脱しない限りにおいて、種々の変更が可能である。
(Additional notes)
Although the embodiments of the present invention have been described above, the present invention is not limited to the above embodiments, and various changes can be made without departing from the spirit thereof.
1  学習済モデル選択装置
1’ 学習済モデル選択装置
1” 学習済モデル選択装置
10 補助記憶装置
11 受付部
12 言語ベクトル変換部
13 言語ベクトル比較部
14 選択部
15 解析部
16 特徴量変換部
17 特徴量比較部
18 検索データ変換部
19 検索データ比較部
F1~Fn モデル特徴量
Fa 検索データ特徴量
M1~Mn 学習済モデル
MC 特徴量変換用モデル
S1~Sn 教師データセット
T1~Tm 教師データ
T1-1~T1-m 教師データ
Tn-1~Tn-m 教師データ
V1~Vn モデル言語ベクトル
Va 検索言語ベクトル
1 Trained model selection device 1' Trained model selection device 1'' Trained model selection device 10 Auxiliary storage device 11 Reception unit 12 Language vector conversion unit 13 Language vector comparison unit 14 Selection unit 15 Analysis unit 16 Feature amount conversion unit 17 Features Quantity comparison section 18 Search data conversion section 19 Search data comparison section F1 to Fn Model feature quantity Fa Search data feature quantity M1 to Mn Learned model MC Feature quantity conversion model S1 to Sn Teacher data set T1 to Tm Teacher data T1-1 ~T1-m Teacher data Tn-1~Tn-m Teacher data V1~Vn Model language vector Va Search language vector

Claims (12)

  1.  複数の学習済モデルから解析対象を解析するための学習済モデルを選択する学習済モデル選択方法であって、
     前記複数の学習済モデルの各々に対応する言語ベクトルであるモデル言語ベクトル、前記複数の学習済モデルの各々の特徴量であるモデル特徴量、または、前記複数の学習済モデルの各々の機械学習に用いられた教師データと前記解析対象との類似度を用いて、前記解析対象に応じた学習済モデルを選択する、学習済モデル選択方法。
    A trained model selection method for selecting a trained model for analyzing an analysis target from a plurality of trained models, the method comprising:
    A model language vector that is a language vector corresponding to each of the plurality of trained models, a model feature quantity that is a feature quantity of each of the plurality of trained models, or a machine learning method for each of the plurality of trained models. A trained model selection method that selects a trained model according to the analysis target using a degree of similarity between used teacher data and the analysis target.
  2.  前記モデル言語ベクトルは、前記学習済モデルの各々に付与された言語ラベルから生成される、請求項1に記載の学習済モデル選択方法。 The trained model selection method according to claim 1, wherein the model language vector is generated from a language label given to each of the trained models.
  3.  前記モデル言語ベクトルは、前記学習済モデルの各々の機械学習に用いられた教師データに付与された言語ラベルから生成される、請求項1に記載の学習済モデル選択方法。 The learned model selection method according to claim 1, wherein the model language vector is generated from a language label given to teacher data used for machine learning of each of the learned models.
  4.  前記解析対象または前記解析対象に対応する言語データを検索データとして受け付ける受付ステップと、
     前記検索データを言語ベクトル化して検索言語ベクトルに変換する言語ベクトル変換ステップと、
     前記検索言語ベクトルを、前記モデル言語ベクトルの各々と比較する言語ベクトル比較ステップと、
     前記言語ベクトル比較ステップの比較結果に基づいて、前記複数の学習済モデルから少なくとも1つの学習済モデルを選択する選択ステップと、
    を備える、請求項2または3に記載の学習済モデル選択方法。
    a reception step of accepting the analysis target or language data corresponding to the analysis target as search data;
    a language vector conversion step of converting the search data into a language vector and converting it into a search language vector;
    a language vector comparison step of comparing the search language vector with each of the model language vectors;
    a selection step of selecting at least one trained model from the plurality of trained models based on the comparison result of the language vector comparison step;
    The learned model selection method according to claim 2 or 3, comprising:
  5.  前記選択ステップでは、前記検索言語ベクトルに対する類似度が最も大きいモデル言語ベクトルに対応する学習済モデルを選択する、請求項4に記載の学習済モデル選択方法。 5. The trained model selection method according to claim 4, wherein in the selection step, a trained model corresponding to a model language vector having the highest degree of similarity to the search language vector is selected.
  6.  前記解析対象または前記解析対象に対応する言語データを検索データとして受け付ける受付ステップと、
     前記検索データを言語ベクトル化して検索言語ベクトルに変換する言語ベクトル変換ステップと、
     前記モデル言語ベクトルと前記モデル特徴量との関係を機械学習した特徴量変換用モデルを用いて、前記検索言語ベクトルを検索データ特徴量に変換する特徴量変換ステップと、
     前記検索データ特徴量を、前記モデル特徴量の各々と比較する特徴量比較ステップと、
     前記特徴量比較ステップの比較結果に基づいて、前記複数の学習済モデルから少なくとも1つの学習済モデルを選択する選択ステップと、
    を備える、請求項1に記載の学習済モデル選択方法。
    a reception step of accepting the analysis target or language data corresponding to the analysis target as search data;
    a language vector conversion step of converting the search data into a language vector and converting it into a search language vector;
    a feature quantity conversion step of converting the search language vector into a search data feature quantity using a feature quantity conversion model obtained by machine learning the relationship between the model language vector and the model feature quantity;
    a feature comparison step of comparing the search data feature with each of the model feature;
    a selection step of selecting at least one trained model from the plurality of trained models based on the comparison result of the feature value comparison step;
    The learned model selection method according to claim 1, comprising:
  7.  前記選択ステップでは、前記検索データ特徴量に対する類似度が最も大きいモデル特徴量に対応する学習済モデルを選択する、請求項6に記載の学習済モデル選択方法。 The trained model selection method according to claim 6, wherein in the selection step, a trained model corresponding to a model feature having the highest degree of similarity to the search data feature is selected.
  8.  前記解析対象または前記解析対象に対応する言語データを検索データとして受け付ける受付ステップと、
     前記検索データが言語データである場合に、前記言語データを前記教師データと同一形式のデータに変換する変換ステップと、
     前記検索データを前記教師データと比較する比較ステップと、
     前記比較ステップの比較結果に基づいて、前記複数の学習済モデルから少なくとも1つの学習済モデルを選択する選択ステップと、
    を備える、請求項1に記載の学習済モデル選択方法。
    a reception step of accepting the analysis target or language data corresponding to the analysis target as search data;
    When the search data is linguistic data, a conversion step of converting the linguistic data into data in the same format as the teacher data;
    a comparison step of comparing the search data with the teacher data;
    a selection step of selecting at least one trained model from the plurality of trained models based on the comparison result of the comparison step;
    The learned model selection method according to claim 1, comprising:
  9.  前記教師データは、前記複数の学習済モデルにそれぞれ対応する複数の教師データセットを構成しており、
     前記比較ステップでは、前記複数の教師データセットを順次選択し、選択した教師データセットの教師データと前記検索データとを比較し、前記教師データセットごとに、教師データと検索データとの類似度を算出し、
     前記選択ステップでは、前記類似度を用いて前記複数の教師データセットから少なくとも1つの教師データセットを選択し、選択された教師データセットに対応する学習済モデルを選択する、請求項8に記載の学習済モデル選択方法。
    The training data constitutes a plurality of training data sets respectively corresponding to the plurality of trained models,
    In the comparison step, the plurality of teacher data sets are sequentially selected, the teacher data of the selected teacher data set and the search data are compared, and the degree of similarity between the teacher data and the search data is determined for each of the teacher data sets. Calculate,
    9. The method according to claim 8, wherein in the selection step, at least one teacher data set is selected from the plurality of teacher data sets using the similarity, and a trained model corresponding to the selected teacher data set is selected. Trained model selection method.
  10.  前期検索データは、画像である、請求項9に記載の学習済みモデル選択方法。 The trained model selection method according to claim 9, wherein the first-term search data is an image.
  11.  複数の学習済モデルから解析対象を解析するための学習済モデルを選択する学習済モデル選択装置であって、
     前記複数の学習済モデルの各々に付与された言語ラベルから生成された言語ベクトルであるモデル言語ベクトル、前記複数の学習済モデルの各々の特徴量であるモデル特徴量、または、前記複数の学習済モデルの各々の機械学習に用いられた教師データと前記解析対象との類似度を用いて、前記解析対象に応じた学習済モデルを選択する、学習済モデル選択装置。
    A trained model selection device that selects a trained model for analyzing an analysis target from a plurality of trained models,
    A model language vector that is a language vector generated from a language label assigned to each of the plurality of trained models, a model feature that is a feature of each of the plurality of trained models, or a model feature that is a feature of each of the plurality of trained models, or a plurality of trained models. A trained model selection device that selects a trained model according to the analysis target using a degree of similarity between training data used for machine learning of each model and the analysis target.
  12.  複数の学習済モデルから解析対象を解析するための学習済モデルを選択する学習済モデル選択プログラムであって、
     前記複数の学習済モデルの各々に対応する言語ベクトルであるモデル言語ベクトル、前記複数の学習済モデルの各々の特徴量であるモデル特徴量、または、前記複数の学習済モデルの各々の機械学習に用いられた教師データと前記解析対象との類似度を用いて、前記解析対象に応じた学習済モデルを選択する処理をコンピュータに実行させる学習済モデル選択プログラム。
    A trained model selection program that selects a trained model for analyzing an analysis target from multiple trained models,
    A model language vector that is a language vector corresponding to each of the plurality of trained models, a model feature quantity that is a feature quantity of each of the plurality of trained models, or a machine learning method for each of the plurality of trained models. A trained model selection program that causes a computer to execute a process of selecting a trained model corresponding to the analysis target using a degree of similarity between used teaching data and the analysis target.
PCT/JP2023/016265 2022-08-09 2023-04-25 Trained model selection method, trained model selection device, and trained model selection program WO2024034196A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2022-127374 2022-08-09
JP2022127374A JP2024024504A (en) 2022-08-09 2022-08-09 Learned model selection method, learned model selection device, and learned model selection program

Publications (1)

Publication Number Publication Date
WO2024034196A1 true WO2024034196A1 (en) 2024-02-15

Family

ID=89851509

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2023/016265 WO2024034196A1 (en) 2022-08-09 2023-04-25 Trained model selection method, trained model selection device, and trained model selection program

Country Status (2)

Country Link
JP (1) JP2024024504A (en)
WO (1) WO2024034196A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021095693A1 (en) * 2019-11-14 2021-05-20 キヤノン株式会社 Information processing device, information processing method, and program
JP2021144348A (en) * 2020-03-10 2021-09-24 富士通株式会社 Information processing device and information processing method
JP2022064214A (en) * 2020-10-13 2022-04-25 株式会社ブルーブックス Data management system, method for managing data, and mechanical learning data management program

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2021095693A1 (en) * 2019-11-14 2021-05-20 キヤノン株式会社 Information processing device, information processing method, and program
JP2021144348A (en) * 2020-03-10 2021-09-24 富士通株式会社 Information processing device and information processing method
JP2022064214A (en) * 2020-10-13 2022-04-25 株式会社ブルーブックス Data management system, method for managing data, and mechanical learning data management program

Also Published As

Publication number Publication date
JP2024024504A (en) 2024-02-22

Similar Documents

Publication Publication Date Title
Xia et al. Complete random forest based class noise filtering learning for improving the generalizability of classifiers
Wainer Comparison of 14 different families of classification algorithms on 115 binary datasets
US11521110B2 (en) Learning apparatus, learning method, and non-transitory computer readable storage medium
US11328125B2 (en) Method and server for text classification using multi-task learning
Zhu et al. Effective supervised discretization for classification based on correlation maximization
US20210056127A1 (en) Method for multi-modal retrieval and clustering using deep cca and active pairwise queries
US20210209514A1 (en) Machine learning method for incremental learning and computing device for performing the machine learning method
US11720788B2 (en) Calculation scheme decision system, calculation scheme decision device, calculation scheme decision method, and storage medium
Rutkowski et al. On explainable fuzzy recommenders and their performance evaluation
Chien et al. Learning discriminant functions with fuzzy attributes for classification using genetic programming
Farazi et al. Accuracy vs. complexity: a trade-off in visual question answering models
JP3568181B2 (en) Neural network analyzer, storage medium
CN117494051A (en) Classification processing method, model training method and related device
Schulz et al. Uncertainty quantification of surrogate explanations: an ordinal consensus approach
WO2024034196A1 (en) Trained model selection method, trained model selection device, and trained model selection program
Harari et al. Automatic features generation and selection from external sources: a DBpedia use case
US11983633B2 (en) Machine learning predictions by generating condition data and determining correct answers
JPWO2020144736A1 (en) Semantic relationship learning device, semantic relationship learning method, and semantic relationship learning program
CN114997155A (en) Fact verification method and device based on table retrieval and entity graph reasoning
CN110633363B (en) Text entity recommendation method based on NLP and fuzzy multi-criterion decision
JPH08190578A (en) Design support device
CN113157892A (en) User intention processing method and device, computer equipment and storage medium
Özpolat et al. Exploring the Potential of Quantum-Based Machine Learning: A Comparative Study of QSVM and Classical Machine Learning Algorithms
US20190065586A1 (en) Learning method, method of using result of learning, generating method, computer-readable recording medium and learning device
JP7168963B2 (en) Natural language processing device and natural language processing method

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23852182

Country of ref document: EP

Kind code of ref document: A1