WO2022237215A1 - Model training method and system, and device and computer-readable storage medium - Google Patents

Model training method and system, and device and computer-readable storage medium Download PDF

Info

Publication number
WO2022237215A1
WO2022237215A1 PCT/CN2022/071325 CN2022071325W WO2022237215A1 WO 2022237215 A1 WO2022237215 A1 WO 2022237215A1 CN 2022071325 W CN2022071325 W CN 2022071325W WO 2022237215 A1 WO2022237215 A1 WO 2022237215A1
Authority
WO
WIPO (PCT)
Prior art keywords
sample data
information
model
category
training
Prior art date
Application number
PCT/CN2022/071325
Other languages
French (fr)
Chinese (zh)
Inventor
李明磊
怀宝兴
袁晶
Original Assignee
华为云计算技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 华为云计算技术有限公司 filed Critical 华为云计算技术有限公司
Publication of WO2022237215A1 publication Critical patent/WO2022237215A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology

Abstract

Provided in the present application are a model training method and system, and a device and a computer-readable storage medium. The method comprises the following steps: acquiring an original sample set, wherein the original sample set comprises a plurality of pieces of sample data; receiving, by means of a labelling interface, a labelling result of a user for each piece of sample data, so as to acquire a training sample set, wherein the training sample set comprises the plurality of pieces of sample data and the labelling result for each piece of sample data, the labelling result for each piece of sample data comprises category information of each piece of sample data and association information of the category information of each piece of sample data; and training a classification model according to the training sample set. By using the method, the training efficiency of a classification model can be improved, and the training costs of the classification model can be reduced.

Description

模型训练方法、系统、设备及计算机可读存储介质Model training method, system, device and computer-readable storage medium
本申请要求于2021年5月11日提交中国国家知识产权局、申请号为202110513038.9、发明名称为“模型训练方法、系统、设备及计算机可读存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed with the State Intellectual Property Office of China on May 11, 2021, with the application number 202110513038.9, and the title of the invention is "model training method, system, device, and computer-readable storage medium", all of which The contents are incorporated by reference in this application.
技术领域technical field
本申请涉及人工智能(artificial intelligence,AI)技术领域,具体涉及一种模型训练方法、系统、设备及计算机可读存储介质。This application relates to the technical field of artificial intelligence (AI), in particular to a model training method, system, device, and computer-readable storage medium.
背景技术Background technique
近年来,随着AI技术的快速发展,深度学习已经在多个领域取得了斐然的成绩,尤其是分类领域,比如说,文本分类、图像分类或者语音分类等。在分类学习的过程,需要使用到大量的标注数据作为样本对分类模型进行训练。目前,为了保证分类模型的正确率,仍然需要人工对样本数据进行分类标注。但由于人工标注的效率较低且成本较高,因此分类模型的训练成本高。In recent years, with the rapid development of AI technology, deep learning has achieved remarkable results in many fields, especially in the field of classification, such as text classification, image classification or speech classification. In the process of classification learning, it is necessary to use a large amount of labeled data as samples to train the classification model. At present, in order to ensure the accuracy of the classification model, it is still necessary to manually classify and label the sample data. However, due to the low efficiency and high cost of manual labeling, the training cost of classification models is high.
发明内容Contents of the invention
本申请提供了一种模型训练方法、系统、设备及计算机可读存储介质,利用该方法能够提高分类模型的训练效率,减少分类模型的训练成本。The present application provides a model training method, system, device and computer-readable storage medium, which can improve the training efficiency of the classification model and reduce the training cost of the classification model.
第一方面,本申请提供了一种模型训练方法,该方法包括如下步骤:In a first aspect, the present application provides a method for model training, which includes the following steps:
获取原始样本集,原始样本集包括多个样本数据;Obtain an original sample set, the original sample set includes a plurality of sample data;
通过标注接口接收用户对每个样本数据的标注结果以获取训练样本集,其中,训练样本集包括上述多个样本数据以及每个样本数据的标注结果,每个样本数据的标注结果包括每个样本数据的类别信息以及每个样本数据的类别信息的关联信息;Receive the user's labeling results for each sample data through the labeling interface to obtain the training sample set, wherein the training sample set includes the above-mentioned multiple sample data and the labeling results of each sample data, and the labeling results of each sample data include each sample The category information of the data and the associated information of the category information of each sample data;
根据训练样本集对分类模型进行训练。The classification model is trained according to the training sample set.
实施第一方面所描述的方法,用户不仅可以标注出样本数据的类别信息,还可以标注出样本数据的类别信息的关联信息,这样在训练分类模型时就可以不仅根据样本数据的类别信息,还可以根据样本数据的类别信息的关联信息,从而提高分类模型的训练效率,减少分类模型的训练成本。Implementing the method described in the first aspect, the user can not only mark the category information of the sample data, but also mark the associated information of the category information of the sample data, so that when training the classification model, not only the category information of the sample data, but also the According to the association information of the category information of the sample data, the training efficiency of the classification model can be improved and the training cost of the classification model can be reduced.
在第一方面的一种可能的实现方式中,上述多个样本数据中第一样本数据的类别信息的关联信息包含于第一样本数据中。可选的,第一样本数据可以是上述多个样本数据中的一部分样本数据,也可以是上述多个样本数据中的全部样本数据。In a possible implementation manner of the first aspect, the association information of the category information of the first sample data among the plurality of sample data is included in the first sample data. Optionally, the first sample data may be a part of the above-mentioned multiple sample data, or may be all of the above-mentioned multiple sample data.
在第一方面的一种可能的实现方式中,第一样本数据的类别信息的关联信息包括类别解释信息和类别解释信息的元数据,类别解释信息是第一样本数据中体现第一样本数据的类别信息的部分。类别解释信息的元数据包括类别解释信息的个数、类别解释信息在第一样本数据中的位置等。In a possible implementation manner of the first aspect, the associated information of the category information of the first sample data includes category explanation information and metadata of the category explanation information, and the category explanation information is the first item in the first sample data The category information section of this data. The metadata of category explanation information includes the number of category explanation information, the position of category explanation information in the first sample data, and the like.
在第一方面的一种可能的实现方式中,上述多个样本数据中第二样本数据的类别信息的 关联信息不包含于第二样本数据中。可选的,第二样本数据可以是上述多个样本数据中的一部分样本数据,也可以是上述多个样本数据中的全部样本数据。In a possible implementation manner of the first aspect, the association information of the category information of the second sample data among the plurality of sample data is not included in the second sample data. Optionally, the second sample data may be a part of the above-mentioned multiple sample data, or may be all of the above-mentioned multiple sample data.
可以看出,用户在标注样本数据的类别信息的关联信息时,可以在样本数据中标注出其类别信息的关联信息,也可以不在样本数据中标注其类别信息的关联信息。而且,当样本数据的类别信息的关联信息包括多个部分(类别解释信息和类别解释信息的元数据)时,一部分的类别信息的关联信息可以标注在样本数据中,一部分的类别信息的关联信息可以不标注在样本数据中。如此,用户可以根据自己的需求来标注的样本数据的类别信息的关联信息。It can be seen that when the user marks the associated information of the category information of the sample data, the user may mark the associated information of the category information in the sample data, or may not mark the associated information of the category information in the sample data. Moreover, when the associated information of the category information of the sample data includes multiple parts (category explanation information and metadata of the category explanation information), the associated information of a part of the category information can be marked in the sample data, and the associated information of a part of the category information May not be marked in the sample data. In this way, the user can mark the associated information of the category information of the sample data according to his own needs.
在第一方面的一种可能的实现方式中,上述每个样本数据的类别信息的关联信息是用户生成该样本数据的类别信息的过程中确认的。如此,用户标注一个样本数据的所需的时间不会太多。In a possible implementation manner of the first aspect, the association information of the above category information of each sample data is confirmed by the user during the process of generating the category information of the sample data. In this way, the time required for the user to label a sample data will not be too much.
在第一方面的一种可能的实现方式中,上述根据训练样本集对分类模型进行训练,包括:将上述多个样本数据输入分类模型,得到每个样本数据的预测的类别信息以及预测的类别信息的关联信息;根据损失函数调整分类模型的参数,直至损失函数的输出满足阈值;其中,损失函数包括第一损失函数和第二损失函数,第一损失函数用于指示每个样本数据的类别信息与每个样本的预测的类别信息之间的差异,第二损失函数用于指示每个样本数据的类别信息的关联信息与每个样本数据的预测的类别信息的关联信息之间的差异。In a possible implementation of the first aspect, the training of the classification model based on the training sample set includes: inputting the above-mentioned multiple sample data into the classification model, and obtaining the predicted category information and the predicted category of each sample data Information related to information; adjust the parameters of the classification model according to the loss function until the output of the loss function meets the threshold; where the loss function includes a first loss function and a second loss function, and the first loss function is used to indicate the category of each sample data information and the predicted class information of each sample, the second loss function is used to indicate the difference between the associated information of the class information of each sample data and the associated information of the predicted class information of each sample data.
通过上述方法,使得分类模型不仅可以根据样本数据的类别信息进行训练,还可以根据样本数据的类别信息的关联信息进行训练,由于样本数据的类别信息的关联信息与样本数据的类别信息是相关的,因此,根据样本数据的类别信息的关联信息进行分类模型的训练相当于从另一个维度训练分类模型,有助于提高分类模型的正确率。那么,通过上述方法便可以提高分类模型的训练效率,减少分类模型的训练成本。Through the above method, the classification model can not only be trained according to the category information of the sample data, but also can be trained according to the associated information of the category information of the sample data, because the associated information of the category information of the sample data is related to the category information of the sample data , therefore, training the classification model based on the association information of the category information of the sample data is equivalent to training the classification model from another dimension, which helps to improve the accuracy of the classification model. Then, the above method can improve the training efficiency of the classification model and reduce the training cost of the classification model.
在第一方面的一种可能的实现方式中,上述分类模型包括编码模型、第一任务模型和第二任务模型,编码模型用于提取上述多个样本数据的特征,第一任务模型用于根据上述多个样本数据的特征确定上述每个样本数据的类别信息,第二任务模型用于根据上述多个样本数据的特征确定上述每个样本数据的类别信息的关联信息。如此设计,可以令数据的分类过程与数据的类别信息的关联信息的预测过程共享编码模型提取到的特征,同时又可以单独地输出数据的分类结果和类别信息的关联信息。In a possible implementation of the first aspect, the above classification model includes an encoding model, a first task model and a second task model, the encoding model is used to extract the features of the above multiple sample data, and the first task model is used to The characteristics of the plurality of sample data determine the category information of each sample data, and the second task model is used to determine the associated information of the category information of each sample data according to the characteristics of the plurality of sample data. With such a design, the data classification process and the prediction process of the associated information of the category information of the data can share the features extracted by the encoding model, and at the same time, the classification result of the data and the associated information of the category information can be output separately.
在第一方面的一种可能的实现方式中,上述方法还包括:获取另一个原始样本集,另一个原始样本集也包括多个样本数据;然后,通过标注接口接收用户对另一个原始样本集中的每个样本数据标注的类别信息,从而得到上述训练样本集。该训练样本集包括两个部分,其一,上述原始样本集、原始样本集中每个样本数据的类别信息以及类别信息的关联信息;其二,上述另一个原始样本集,以及另一个原始样本集中每个样本数据的类别信息。然后,利用该训练样本集对上述分类模型进行训练。也就是说,用户在标注样本数据时,可以选择标注样本数据的类别信息的关联信息,也可以选择不标注样本数据的类别信息的关联信息。In a possible implementation of the first aspect, the above method further includes: acquiring another original sample set, which also includes a plurality of sample data; The category information labeled by each sample data of , so as to obtain the above training sample set. The training sample set includes two parts, one, the above-mentioned original sample set, the category information of each sample data in the original sample set, and the associated information of the category information; second, the above-mentioned another original sample set, and another original sample set Class information for each sample data. Then, use the training sample set to train the above classification model. That is to say, when labeling the sample data, the user may choose to label the associated information of the category information of the sample data, or may choose not to label the associated information of the category information of the sample data.
第二方面,本申请提供了一种模型训练系统,该系统包括:In a second aspect, the present application provides a model training system, which includes:
训练数据标注模块,用于获取原始样本集,原始样本集包括多个样本数据;The training data labeling module is used to obtain an original sample set, and the original sample set includes a plurality of sample data;
训练数据标注模块还用于通过标注接口接收用户对每个样本数据的标注结果以获取训练样本集,其中,训练样本集包括上述多个样本数据以及每个样本数据的标注结果,每个样本数据的标注结果包括每个样本数据的类别信息以及每个样本数据的类别信息的关联信息;The training data labeling module is also used to receive the user's labeling results for each sample data through the labeling interface to obtain a training sample set, wherein the training sample set includes the above-mentioned multiple sample data and the labeling results of each sample data, and each sample data The labeling results of include the category information of each sample data and the associated information of the category information of each sample data;
模型训练模块,用于根据训练样本集对分类模型进行训练。The model training module is used to train the classification model according to the training sample set.
在第二方面的一种可能的实现方式中,上述多个样本数据中第一样本数据的类别信息的 关联信息包含于第一样本数据中。可选的,第一样本数据可以是上述多个样本数据中的一部分样本数据,也可以是上述多个样本数据中的全部样本数据。In a possible implementation manner of the second aspect, the association information of the category information of the first sample data among the plurality of sample data is included in the first sample data. Optionally, the first sample data may be a part of the above-mentioned multiple sample data, or may be all of the above-mentioned multiple sample data.
在第二方面的一种可能的实现方式中,上述第一样本数据的类别信息的关联信息包括类别解释信息和所述类别解释信息的元数据,类别解释信息是第一样本数据中体现第一样本数据的类别信息的部分。类别解释信息的元数据包括类别解释信息的个数、类别解释信息在第一样本数据中的位置等。In a possible implementation manner of the second aspect, the associated information of the category information of the first sample data includes category explanation information and metadata of the category explanation information, and the category explanation information is embodied in the first sample data. Part of the category information of the first sample data. The metadata of category explanation information includes the number of category explanation information, the position of category explanation information in the first sample data, and the like.
在第二方面的一种可能的实现方式中,上述多个样本数据中第二样本数据的类别信息的关联信息不包含于第二样本数据中。可选的,第二样本数据可以是上述多个样本数据中的一部分样本数据,也可以是上述多个样本数据中的全部样本数据。In a possible implementation manner of the second aspect, the association information of the category information of the second sample data among the plurality of sample data is not included in the second sample data. Optionally, the second sample data may be a part of the above-mentioned multiple sample data, or may be all of the above-mentioned multiple sample data.
在第二方面的一种可能的实现方式中,上述每个样本数据的类别信息的关联信息是用户生成该样本数据的类别信息的过程中确认的。In a possible implementation manner of the second aspect, the association information of the above category information of each sample data is confirmed by the user during the process of generating the category information of the sample data.
在第二方面的一种可能的实现方式中,上述模型训练模块具体用于:将上述多个样本数据输入分类模型,得到每个样本数据的预测的类别信息以及预测的类别信息的关联信息;根据损失函数调整分类模型的参数,直至损失函数的输出满足阈值;其中,损失函数包括第一损失函数和第二损失函数,第一损失函数用于指示每个样本数据的类别信息与每个样本的预测的类别信息之间的差异,第二损失函数用于指示每个样本数据的类别信息的关联信息与每个样本数据的预测的类别信息的关联信息之间的差异。In a possible implementation of the second aspect, the above-mentioned model training module is specifically configured to: input the above-mentioned multiple sample data into the classification model, and obtain the predicted category information of each sample data and the associated information of the predicted category information; Adjust the parameters of the classification model according to the loss function until the output of the loss function meets the threshold; where the loss function includes a first loss function and a second loss function, and the first loss function is used to indicate that the category information of each sample data is consistent with each sample The difference between the predicted class information of each sample data, the second loss function is used to indicate the difference between the associated information of the class information of each sample data and the associated information of the predicted class information of each sample data.
在第二方面的一种可能的实现方式中,上述分类模型包括编码模型、第一任务模型和第二任务模型,编码模型用于提取上述多个样本数据的特征,第一任务模型用于根据上述多个样本数据的特征确定上述每个样本数据的类别信息,第二任务模型用于根据上述多个样本数据的特征确定上述每个样本数据的类别信息的关联信息。In a possible implementation of the second aspect, the classification model includes an encoding model, a first task model, and a second task model, the encoding model is used to extract the features of the plurality of sample data, and the first task model is used to The characteristics of the plurality of sample data determine the category information of each sample data, and the second task model is used to determine the associated information of the category information of each sample data according to the characteristics of the plurality of sample data.
在第二方面的一种可能的实现方式中,上述训练数据标注模块还用于:获取另一个原始样本集,另一个原始样本集也包括多个样本数据;然后,通过标注接口接收用户对另一个原始样本集中的每个样本数据标注的类别信息,从而得到上述训练样本集。该训练样本集包括两个部分,其一,上述原始样本集、原始样本集中每个样本数据的类别信息以及类别信息的关联信息;其二,上述另一个原始样本集,以及另一个原始样本集中每个样本数据的类别信息。该训练样本集也可以被用于对上述分类模型进行训练。In a possible implementation of the second aspect, the above-mentioned training data labeling module is also used to: obtain another original sample set, which also includes multiple sample data; The category information labeled by each sample data in an original sample set, so as to obtain the above training sample set. The training sample set includes two parts, one, the above-mentioned original sample set, the category information of each sample data in the original sample set, and the associated information of the category information; second, the above-mentioned another original sample set, and another original sample set Class information for each sample data. The training sample set can also be used to train the above classification model.
第三方面,本申请提供了一种计算设备,该计算设备包括处理器和存储器,存储器存储计算机指令,处理器执行计算机指令,以使计算设备执行前述第一方面或第一方面的任意一种可能的实现方式中的方法。In a third aspect, the present application provides a computing device, the computing device includes a processor and a memory, the memory stores computer instructions, and the processor executes the computer instructions, so that the computing device performs any one of the aforementioned first aspect or the first aspect method in a possible implementation.
第四方面,本申请提供了一种计算机可读存储介质,该计算机可读存储介质存储有计算机程序代码,当计算机程序代码被计算设备执行时,计算设备执行前述第一方面或第一方面的任意一种可能的实现方式中的方法。In a fourth aspect, the present application provides a computer-readable storage medium, the computer-readable storage medium stores computer program code, and when the computer program code is executed by a computing device, the computing device executes the aforementioned first aspect or the first aspect. A method in any one of the possible implementations.
附图说明Description of drawings
为了更清楚地说明本申请提供的技术方案,下面将对本申请中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions provided by this application, the accompanying drawings that need to be used in this application will be briefly introduced below. Obviously, the accompanying drawings in the following description are some embodiments of this application. For those skilled in the art As far as people are concerned, other drawings can also be obtained based on these drawings on the premise of not paying creative work.
图1是本申请提供的一种模型训练系统的结构示意图;Fig. 1 is a schematic structural diagram of a model training system provided by the present application;
图2是本申请提供的一种模型训练方法的流程示意图;Fig. 2 is a schematic flow chart of a model training method provided by the present application;
图3是本申请提供的一种GUI的示意图;Fig. 3 is a schematic diagram of a GUI provided by the present application;
图4是本申请提供的一种分类模型的结构示意图;Fig. 4 is a schematic structural diagram of a classification model provided by the present application;
图5是本申请提供的一种模型训练系统的部署方式;Fig. 5 is a deployment mode of a model training system provided by the present application;
图6是本申请提供的另一种模型训练系统的部署方式;Fig. 6 is the deployment mode of another model training system provided by the present application;
图7是本申请提供的一种计算设备的结构示意图;FIG. 7 is a schematic structural diagram of a computing device provided by the present application;
图8是本申请提供的一种计算设备系统的结构示意图。FIG. 8 is a schematic structural diagram of a computing device system provided by the present application.
具体实施方式Detailed ways
下面将结合附图对本申请提供的技术方案进行详细介绍。The technical solutions provided by the present application will be described in detail below in conjunction with the accompanying drawings.
深度学习(deep learning)是一类基于深层次神经网络算法的机器学习技术,其主要特征是使用多重非线性变换来对数据进行处理和分析。近年来,深度学习在分类领域(例如,文本分类、图像分类、语音分类等)取得了极大的成功,这一成功主要得益于分类模型的准确性。Deep learning is a kind of machine learning technology based on deep neural network algorithm, its main feature is to use multiple nonlinear transformations to process and analyze data. In recent years, deep learning has achieved great success in the field of classification (eg, text classification, image classification, speech classification, etc.), and this success is mainly due to the accuracy of the classification model.
一般地,分类模型在被用于执行分类任务之前都需要被训练。在训练分类模型时,需要人工对大量的样本数据进行分类标注,然后将样本数据作为分类模型的输入,将样本数据对应的类别标签(label)作为分类模型输出值的参考,然后利用损失函数(loss function)计算分类模型输出值与样本数据对应的类别标签的损失值,并根据损失值反复调整分类模型中的参数,直至分类模型可以根据输入的样本数据输出与样本数据对应的类别标签非常接近的值。之后,便可以利用训练好的分类模型来预测数据的类别。Generally, classification models need to be trained before being used to perform classification tasks. When training a classification model, it is necessary to manually classify and label a large number of sample data, then use the sample data as the input of the classification model, and use the category label (label) corresponding to the sample data as a reference for the output value of the classification model, and then use the loss function ( loss function) calculates the loss value of the classification model output value and the category label corresponding to the sample data, and repeatedly adjusts the parameters in the classification model according to the loss value until the classification model can output the category label corresponding to the sample data according to the input sample data very close value. After that, the trained classification model can be used to predict the category of the data.
但需要注意的一点是,目前在对样本数据进行分类标注时,通常都是标注人员读完一个样本数据后,根据提供的类别标签来标注样本数据的类别,然后继续读取下一个样本数据,直至将所有的样本数据都标注完成。例如,一个文本分类的例子:为了训练一个用于识别文本数据带有的情绪的分类模型,类别标签可以设置为以下两个:正面情绪标签和负面情绪标签。然后,由标注人员对搜集到的样本数据一一进行情绪标注,如果标注人员认为样本数据表达了正面情绪,则为该样本数据添加正面情绪标签,反之,为该样本数据添加负面情绪标签。最后,利用标注好的样本数据对情绪分类模型进行训练。不难看出,上述方式会使得训练出的分类模型的准确性主要依靠类别标签。而且,人工标注的效率低下,但成本却非常高,这就会导致需要较高的成本才能训练出一个性能较好的分类模型。However, one thing to note is that currently, when classifying and labeling sample data, the labeler usually marks the category of the sample data according to the provided category label after reading a sample data, and then continues to read the next sample data. Until all the sample data are marked. For example, an example of text classification: In order to train a classification model for identifying sentiments associated with text data, the class labels can be set to the following two: positive sentiment labels and negative sentiment labels. Then, the annotators will label the collected sample data one by one. If the annotators think that the sample data expresses positive emotions, they will add positive emotional labels to the sample data, otherwise, add negative emotional labels to the sample data. Finally, use the labeled sample data to train the emotion classification model. It is not difficult to see that the above method will make the accuracy of the trained classification model mainly rely on category labels. Moreover, the efficiency of manual labeling is low, but the cost is very high, which will lead to a higher cost to train a classification model with better performance.
针对上述问题,本申请提供了一种模型训练系统,能够提高分类模型的训练效率,减少了分类模型的训练成本。In view of the above problems, the present application provides a model training system, which can improve the training efficiency of the classification model and reduce the training cost of the classification model.
请参见图1,图1示出了一种可能的模型训练系统的结构示意图。如图1所示,模型训练系统100包括训练数据标注模块110和模型训练模块120。可选的,模型训练系统100还可以包括训练数据存储模块130、模型存储模块140和模型选择模块150。下面将简要地介绍上述各个模块的功能:Please refer to FIG. 1 , which shows a schematic structural diagram of a possible model training system. As shown in FIG. 1 , the model training system 100 includes a training data labeling module 110 and a model training module 120 . Optionally, the model training system 100 may further include a training data storage module 130 , a model storage module 140 and a model selection module 150 . The following will briefly introduce the functions of each of the above modules:
训练数据标注模块110:用于获取原始样本集,其中,原始样本集包括多个样本数据。训练数据标注模块110也用于接收第一用户对原始样本集中的每个样本数据的标注结果以获取训练样本集。Training data labeling module 110: for obtaining an original sample set, wherein the original sample set includes a plurality of sample data. The training data labeling module 110 is also configured to receive a labeling result of each sample data in the original sample set by the first user to obtain a training sample set.
在一具体的实施例中,原始样本集包括第一原始样本集,第一原始样本集包括多个样本数据。训练数据标注模块110具体用于:接收第一用户对第一原始样本集中的每个样本数据的标注结果,其中,第一用户对第一原始样本集中的每个样本数据的标注结果包括第一原始样本集中的每个样本数据的类别信息以及每个样本数据的类别信息的关联信息。那么,训练样本集包括第一原始样本集中的多个样本数据,以及用户对第一原始样本集中的每个样本数 据的标注结果。In a specific embodiment, the original sample set includes a first original sample set, and the first original sample set includes a plurality of sample data. The training data labeling module 110 is specifically configured to: receive the first user's labeling results for each sample data in the first original sample set, wherein the first user's labeling results for each sample data in the first original sample set include the first Category information of each sample data in the original sample set and associated information of the category information of each sample data. Then, the training sample set includes a plurality of sample data in the first original sample set, and a labeling result of each sample data in the first original sample set by the user.
可选的,原始样本集还包括第二原始样本集,第二原始样本集包括多个样本数据。训练数据标注模块110还用于接收第一用户对第二原始样本集中的每个样本数据的标注结果,其中,第一用户对第二原始样本集中的每个样本数据的标注结果包括第二原始样本集中的每个样本数据的类别信息。在这种情况下,训练样本集包括两个部分,一部分是第一原始样本集中的多个样本数据、第一用户对第一原始样本集中的每个样本数据的标注结果;另一部分是第二原始样本集中的多个样本数据、第一用户对第二原始样本集中的每个样本数据的标注结果。也就是说,第一用户可以选择对样本数据的类别信息的关联信息进行标注,也可以选择不对样本数据的类别信息的关联信息进行标注。Optionally, the original sample set further includes a second original sample set, and the second original sample set includes a plurality of sample data. The training data labeling module 110 is also used to receive the first user's labeling results for each sample data in the second original sample set, wherein the first user's labeling results for each sample data in the second original sample set include the second original The category information of each sample data in the sample set. In this case, the training sample set includes two parts, one part is a plurality of sample data in the first original sample set, and the labeling result of each sample data in the first original sample set by the first user; the other part is the second A plurality of sample data in the original sample set, and an annotation result of each sample data in the second original sample set by the first user. That is to say, the first user may choose to mark the associated information of the category information of the sample data, or may choose not to mark the associated information of the category information of the sample data.
可选的,训练数据标注模块110还用于将训练样本集发送给训练数据存储模块130。训练数据标注模块110也可以将上述原始样本集发送给训练数据存储模块130。Optionally, the training data labeling module 110 is also configured to send the training sample set to the training data storage module 130 . The training data labeling module 110 may also send the above-mentioned original sample set to the training data storage module 130 .
在一种可能的实现方式中,训练数据标注模块110可采用图形用户界面(graphical user interface,GUI)实现。例如,GUI可以向第一用户展示原始样本集中的每个样本数据,GUI上还提供标注接口,第一用户可以通过标注接口对原始样本集中的每个样本数据进行类别信息以及类别信息的关联信息的标注。GUI在第一用户对原始样本集中所有的样本数据标注完成后,将标注后得到的训练样本集存储至训练数据存储模块130。In a possible implementation manner, the training data labeling module 110 may be implemented using a graphical user interface (graphical user interface, GUI). For example, the GUI can display each sample data in the original sample set to the first user, and a labeling interface is also provided on the GUI, and the first user can perform category information and associated information of the category information on each sample data in the original sample set through the labeling interface. label. After the first user marks all the sample data in the original sample set, the GUI stores the marked training sample set in the training data storage module 130 .
模型训练模块120:用于从训练数据存储模块130读取训练样本集,并根据训练样本集对分类模型进行训练,从而得到训练好的分类模型。Model training module 120: used to read the training sample set from the training data storage module 130, and train the classification model according to the training sample set, so as to obtain a trained classification model.
可选的,模型训练模块120还用于接收第二用户输入的对训练好的分类模型的预期效果(包括准确率、运行时间等),以使得模型训练模块120可以根据第二用户的预期效果来训练分类模型。其中,第二用户可以是需要训练分类模型的人员,第一用户和第二用户可以是同一用户,也可以是不同的用户。Optionally, the model training module 120 is also used to receive the expected effect (including accuracy, running time, etc.) of the trained classification model input by the second user, so that the model training module 120 can be based on the expected effect of the second user to train the classification model. Wherein, the second user may be a person who needs to train the classification model, and the first user and the second user may be the same user or different users.
在一些实施例中,模型训练系统100还可以包括模型存储模块140。模型存储模块140用于存储经过模型训练模块120训练得到的训练好的分类模型。In some embodiments, the model training system 100 may further include a model storage module 140 . The model storage module 140 is used for storing the trained classification model trained by the model training module 120 .
在模型训练模块120训练分类模型之前,还需获取需要进行训练的分类模型。模型训练模块120获取分类模型的方式多种多样,例如:Before the model training module 120 trains the classification model, it is also necessary to acquire the classification model to be trained. The model training module 120 acquires classification models in various ways, for example:
方式一:模型训练模块120获取第二用户上传的分类模型。Way 1: The model training module 120 acquires the classification model uploaded by the second user.
方式二:模型存储模块140还存储了多个未训练的模型。模型选择模块150向第二用户展示模型存储模块140中存储的多个未训练的模型,并接收第二用户选择的模型作为分类模型,然后发送至模型训练模块140。Way 2: The model storage module 140 also stores multiple untrained models. The model selection module 150 presents a plurality of untrained models stored in the model storage module 140 to the second user, and receives the model selected by the second user as a classification model, and then sends it to the model training module 140 .
方式三:模型选择模块150接收第二用户输入的需求,并根据第二用户的需求从模型存储模块140中选择合适的模型以作为分类模型,并发送给模型训练模块120。可选的,用户输入的需求可以包括用户期望完成的分类任务,例如,如果用户期望完成的分类任务为文本分类任务,则模型选择模块150会从模型存储模块140中获取用于实现文本分类的模型(例如卷积神经网络(convolutional neural networks,CNN))。可选的,用户输入的需求还可以包括用户对分类模型的参数的要求,例如,CNN的卷积核大小、卷积层数、激活函数、池化层个数等。Method 3: The model selection module 150 receives the requirement input by the second user, and selects an appropriate model from the model storage module 140 as the classification model according to the requirement of the second user, and sends it to the model training module 120 . Optionally, the requirement input by the user may include the classification task that the user expects to complete. For example, if the classification task that the user expects to complete is a text classification task, the model selection module 150 will obtain the text classification task from the model storage module 140. Models (such as convolutional neural networks (CNN)). Optionally, the requirements input by the user may also include the user's requirements on the parameters of the classification model, for example, the convolution kernel size, the number of convolution layers, the activation function, the number of pooling layers, etc. of CNN.
需要说明的是,上述模型训练系统可以是一个与用户交互的系统,这个系统可以是软件系统也可以是硬件系统,也可以是软硬件结合的系统,本申请对此不作具体限定。还需说明的是,图1仅是示例性地展示了模型训练系统的一种结构化示意图,在实际应用中可以根据具体情况对图1示出的模型训练系统进行相应的变换。It should be noted that the above-mentioned model training system may be a system that interacts with users. This system may be a software system, a hardware system, or a system combining software and hardware, which is not specifically limited in this application. It should also be noted that FIG. 1 is only an exemplary structural diagram of the model training system, and the model training system shown in FIG. 1 can be transformed accordingly in practical applications according to specific situations.
下面将结合图2对上述模型训练系统100进行模型训练的过程进行详细描述。The process of model training performed by the above-mentioned model training system 100 will be described in detail below with reference to FIG. 2 .
请参见图2,图2示出了本申请提供的一种模型训练的方法的流程示意图,该方法由图1示出的模型训练系统100执行。如图2所示,该方法包括但不限于以下步骤:Please refer to FIG. 2 . FIG. 2 shows a schematic flowchart of a model training method provided by the present application, and the method is executed by the model training system 100 shown in FIG. 1 . As shown in Figure 2, the method includes but is not limited to the following steps:
S101:模型训练系统100获取原始样本集。S101: The model training system 100 acquires an original sample set.
在一具体的实施例中,原始样本集包括多个样本数据。模型训练系统100可以通过以下方式获取原始样本集:In a specific embodiment, the original sample set includes a plurality of sample data. The model training system 100 can obtain the original sample set in the following ways:
方式一:获取第二用户上传的原始样本集。具体地,模型训练系统100提供数据上传界面,数据上传界面包括数据上传接口,那么,第二用户可以通过点击数据上传接口将预先准备好的原始样本集上传至模型训练系统100。Method 1: Obtain the original sample set uploaded by the second user. Specifically, the model training system 100 provides a data upload interface, and the data upload interface includes a data upload interface. Then, the second user can upload the pre-prepared original sample set to the model training system 100 by clicking on the data upload interface.
方式二:获取第二用户的业务需求,并根据第二用户的业务需求从数据库(包括本地数据库或其他设备的数据库)中搜索符合要求的样本数据,从而得到原始样本集。例如,如果第二用户的业务需求为文本的情感分类,那么模型训练系统100可以从本地数据库或者联网获取大量的文本数据作为原始样本集。又例如,如果第二用户的业务需求为人脸检测,那么可以从本地数据库或联网获取大量的图像作为原始样本集。Method 2: Obtain the business requirements of the second user, and search for sample data meeting the requirements from the database (including local databases or databases of other devices) according to the business requirements of the second user, so as to obtain the original sample set. For example, if the business requirement of the second user is text sentiment classification, then the model training system 100 may acquire a large amount of text data from a local database or the Internet as an original sample set. For another example, if the business requirement of the second user is face detection, a large number of images can be obtained from a local database or the Internet as an original sample set.
S102:模型训练系统100通过标注接口接收第一用户对原始样本集中的每个样本数据的标注结果,以获取训练样本集。S102: The model training system 100 receives a labeling result of each sample data in the original sample set by the first user through the labeling interface, so as to obtain a training sample set.
其中,原始样本集包括第一原始样本集,第一原始样本集包括多个样本数据。第一用户对第一原始样本集中的每个样本数据的标注结果包括第一原始样本集中的每个样本数据的类别信息以及每个样本数据的类别信息的关联信息。样本数据的类别信息用于指示该样本数据的类别,样本数据的类别信息的关联信息是第一用户生成该样本数据的类别信息的过程中确认的信息。如此,第一用户可以无需消耗较多的时间便可以标注出样本数据的类别信息的关联信息。Wherein, the original sample set includes a first original sample set, and the first original sample set includes a plurality of sample data. The tagging result of each sample data in the first original sample set by the first user includes category information of each sample data in the first original sample set and associated information of the category information of each sample data. The category information of the sample data is used to indicate the category of the sample data, and the associated information of the category information of the sample data is information confirmed by the first user during the process of generating the category information of the sample data. In this way, the first user can mark out the associated information of the category information of the sample data without consuming much time.
在一具体的实施例中,第一用户为对样本数据进行标注的人员,第二用户为需要训练分类模型的人员。可选的,第一用户与第二用户可以是同一用户,也可以是不同用户。In a specific embodiment, the first user is a person who labels sample data, and the second user is a person who needs to train a classification model. Optionally, the first user and the second user may be the same user or different users.
可选的,多个样本数据包括第一样本数据,第一样本数据的类别信息的关联信息包含于第一样本数据中。其中,第一样本数据可以是多个样本数据中的一部分样本数据,也可以是多个样本数据中的全部样本数据。Optionally, the plurality of sample data includes first sample data, and the associated information of the category information of the first sample data is included in the first sample data. Wherein, the first sample data may be a part of the sample data in the multiple sample data, or may be all the sample data in the multiple sample data.
可选的,多个样本数据还包括第二样本数据,第二样本数据的类别信息的关联信息不包含于第二样本数据中。第二样本数据可以是多个样本数据中的一部分样本数据,也可以是多个样本数据中的全部样本数据。Optionally, the plurality of sample data further includes second sample data, and the associated information of the category information of the second sample data is not included in the second sample data. The second sample data may be a part of the sample data, or all the sample data in the multiple sample data.
在一具体的实施例中,样本数据的类别信息的关联信息可以包括类别解释信息和类别解释信息的元数据。可选的,类别解释信息可以是样本数据中体现样本数据的类别信息的部分。也可以是标注在样本数据外,体现样本数据的类别信息的部分,此次不作具体限定。In a specific embodiment, the associated information of the category information of the sample data may include category explanation information and metadata of the category explanation information. Optionally, the category explanation information may be a part of the sample data that reflects the category information of the sample data. It may also be a part marked outside the sample data to reflect the category information of the sample data, which is not specifically limited this time.
进一步地,类别解释信息可以是用户为样本数据标注对应的类别信息的原因,该原因可以是从正面直接反应样本数据的类别信息,也可以是从反面间接反应样本数据的类别信息。例如,一个文本数据为“这个手机太好看了”,当第一用户需要为该文本数据标注类别信息(包括正面情绪和负面情绪)时,用户可以根据该文本数据中的“太好看”确定该文本数据表达了正面情绪,此时用户可以将该文本数据标注为正面情绪,那么,“太好看”可以作为该文本数据中的类别解释信息。或者,第一用户还可以在样本数据外,为该样本数据添加“不好看”标注以作为该样本数据的类别解释信息。Further, the category explanation information may be the reason why the user marks the corresponding category information for the sample data, and the reason may directly reflect the category information of the sample data from the positive side, or indirectly reflect the category information of the sample data from the negative side. For example, a piece of text data is "this mobile phone is so beautiful", when the first user needs to label category information (including positive emotions and negative emotions) for the text data, the user can determine the mobile phone according to the "too good-looking" in the text data. The text data expresses positive emotions. At this time, the user can mark the text data as positive emotions. Then, "too good-looking" can be used as the category explanation information in the text data. Alternatively, the first user may also add an "unattractive" label to the sample data as category explanation information of the sample data.
应理解,样本数据不同,样本数据的类别解释信息的表现形式可能不同。例如,文本数据的类别解释信息可以是该文本数据中的关键字、关键词或关键句;图像数据的类别解释信息可以是该图像数据中的某个或某些区域;音频数据的类别解释信息可以是该段音频数据中的某个或某些片段。It should be understood that the sample data is different, and the representation form of the category interpretation information of the sample data may be different. For example, the category explanation information of text data can be keywords, key words or key sentences in the text data; the category explanation information of image data can be one or some regions in the image data; the category explanation information of audio data It can be one or some fragments in this piece of audio data.
进一步地,类别解释信息的元数据能够描述类别解释信息,类别解释信息的元数据可以包括类别解释信息的数量、类别解释信息在样本数据中的位置等,本申请不作具体限定。应理解,类别解释信息的元数据也可以反应样本数据的类别信息。例如,一个文本数据为“虽然这个手机的颜色不好看,但是我认为这个手机的性能好,而且外观独特,所以我很喜欢”,该文本数据中既有表示负面情绪的关键词(即“不好看”),也有表示正面情绪的关键词(即“性能好”、“外观独特”、“很喜欢”),但考虑到表示正面情绪的关键词的数量(即类别解释信息的数量)大于表示负面情绪的关键词的数量,且表示正面情绪的关键词位于关键词“但是”后的位置(即类别解释信息在样本数据中的位置),因此,该文本数据的类别信息应为正面情绪。Further, the metadata of the category explanation information can describe the category explanation information, and the metadata of the category explanation information may include the quantity of the category explanation information, the position of the category explanation information in the sample data, etc., which are not specifically limited in this application. It should be understood that the metadata of the category explanation information may also reflect the category information of the sample data. For example, a piece of text data is "Although the color of this mobile phone is not good-looking, but I think this mobile phone has good performance and unique appearance, so I like it very much", there are keywords expressing negative emotions (ie "no Good-looking"), there are also keywords expressing positive emotions (i.e., "good performance", "unique appearance", "like it very much"), but considering that the number of keywords expressing positive emotions (i.e., the amount of category explanatory information) is greater than that expressing The number of keywords of negative emotions, and the keywords representing positive emotions are located after the keyword "but" (that is, the position of category explanation information in the sample data), therefore, the category information of the text data should be positive emotions.
在一具体的实施例中,模型训练系统100提供数据标注界面,数据标注界面上可以显示样本数据,数据标注界面也可以接收用户对该样本数据标注的类别信息和类别信息的关联信息。以图3为例,数据标注界面200包括工具栏210和标注区220,工具栏210上承载了类别标签211和类别信息的关联信息的标注工具212,标注区220上显示了样本数据。这样,用户可以在标注区220查看样本数据,并利用工具栏210实现对样本数据的标注。具体地,用户读取样本数据后,从标注工具212中选取合适的标注工具在样本数据中标注出类别信息的关联信息,然后从类别标签211中点击该样本数据对应的类别标签。最后,点击保存,完成该样本数据的标注。如图3所示,当用户看到样本数据为“这个手机太好看了”,这时,用户可以从标注工具212中选取划线工具,然后,在“太好看”下面划线,然后点击正面情绪标签,点击保存键,从而完成对该样本数据的标注。In a specific embodiment, the model training system 100 provides a data labeling interface, on which sample data can be displayed, and the data labeling interface can also receive the category information and the associated information of the category information marked by the user on the sample data. Taking FIG. 3 as an example, the data labeling interface 200 includes a tool bar 210 and a labeling area 220 . The tool bar 210 carries the labeling tool 212 related to category labels 211 and category information, and the labeling area 220 displays sample data. In this way, the user can view the sample data in the annotation area 220 and use the tool bar 210 to mark the sample data. Specifically, after the user reads the sample data, he selects a suitable labeling tool from the labeling tool 212 to mark the related information of the category information in the sample data, and then clicks the category label corresponding to the sample data from the category label 211 . Finally, click Save to complete the labeling of the sample data. As shown in Figure 3, when the user sees the sample data as "this mobile phone is too good-looking", at this time, the user can select the line tool from the annotation tool 212, then draw a line under "too good-looking", and then click on the front Emotion label, click the save button to complete the labeling of the sample data.
应理解,图3仅仅作为一种举例,在实际应用中,数据标注界面200上还可以显示已标注的样本数据的个数、待标注的样本数据的个数、标注进度等,还可以提供返回键、放大缩小键等,本申请不作具体限定。It should be understood that FIG. 3 is only an example. In practical applications, the data labeling interface 200 can also display the number of sample data that has been marked, the number of sample data to be marked, the progress of marking, etc., and can also provide a return keys, zoom in and out keys, etc., which are not specifically limited in this application.
可选的,原始样本集还可以包括第二原始样本集,第一用户对第二原始样本集中的每个样本数据的标注结果包括第二原始样本集中的每个样本数据的类别信息。在这种情况下,训练样本集不仅包括第一原始样本集、用户对第一原始样本集中的每个样本数据的标注结果,还包括第二原始样本集及用户对第二原始样本集中的每个样本数据的标注结果。也就是说,第一用户可以选择对原始样本集中的样本数据标注类别信息的关联信息,也可以选择不对原始样本集中的样本数据标注类别信息的关联信息。Optionally, the original sample set may further include a second original sample set, and the labeling result of each sample data in the second original sample set by the first user includes category information of each sample data in the second original sample set. In this case, the training sample set not only includes the first original sample set and the user's labeling results for each sample data in the first original sample set, but also includes the second original sample set and the user's labeling results for each sample data in the second original sample set. Labeling results of sample data. That is to say, the first user may choose to mark the sample data in the original sample set with the associated information of category information, or may choose not to mark the sample data in the original sample set with the associated information of category information.
S103:模型训练系统100根据训练样本集对分类模型进行训练。S103: The model training system 100 trains the classification model according to the training sample set.
以训练样本集包括第一原始样本集、用户对第一原始样本集中的每个样本数据的标注结果,但不包括第二原始样本集及用户对第二原始样本集中的每个样本数据的标注结果为例,模型训练系统100根据训练样本集对分类模型进行训练,具体过程包括:模型训练系统100将多个样本数据输入分类模型,得到每个样本数据的预测的类别信息(即预测得到的每个样本数据的类别信息)和预测的类别信息的关联信息(即预测得到的每个样本数据的类别信息的关联信息),然后根据损失函数调整分类模型的网络参数,直至损失函数的输出满足预置,从而完成对分类模型的训练。其中,损失函数包括第一损失函数和第二损失函数,第一损失函数指示用户标注的每个样本数据的类别信息与预测得到的每个样本数据的类别信息之间的 差异,第二损失函数指示用户标注的每个样本数据的类别信息的关联信息与预测得到的每个样本数据的类别信息的关联信息之间的差异。The training sample set includes the first original sample set, the user's labeling results for each sample data in the first original sample set, but does not include the second original sample set and the user's labeling results for each sample data in the second original sample set Take the result as an example, the model training system 100 trains the classification model according to the training sample set, and the specific process includes: the model training system 100 inputs a plurality of sample data into the classification model, and obtains the predicted category information of each sample data (that is, the predicted category information). The category information of each sample data) and the association information of the predicted category information (that is, the association information of the category information of each sample data predicted), and then adjust the network parameters of the classification model according to the loss function until the output of the loss function satisfies Preset to complete the training of the classification model. Among them, the loss function includes a first loss function and a second loss function, the first loss function indicates the difference between the category information of each sample data marked by the user and the predicted category information of each sample data, and the second loss function Indicates the difference between the associated information of the category information of each sample data marked by the user and the predicted associated information of the category information of each sample data.
如图4所示,图4示出了一个分类模型的结构示意图。其中,分类模型300包括输入模块310、编码模块320、第一任务模块330以及第二任务模块340。As shown in FIG. 4, FIG. 4 shows a schematic structural diagram of a classification model. Wherein, the classification model 300 includes an input module 310 , an encoding module 320 , a first task module 330 and a second task module 340 .
输入模块310:用于获取训练样本集,以及将训练样本集发送给编码模块320。可选的,输入模块310还可以用于对训练样本集进行预处理。应理解,对于不同的样本数据需要进行不同的预处理过程,例如,对于图像样本,输入模块310可以对图像样本进行图像颜色空间的变换、图像裁剪、图像缩放等预处理;对于文本样本,输入模块310可以对文本样本进行去除非文本数据、去掉停用词等预处理;对于音频样本,输入模块310可以对音频样本进行音频格式的变换、分帧、加窗等预处理。可选的,输入模块310还可以用于根据样本数据的标注结果对样本数据进行预处理,例如,当某个样本数据没有对应的类别信息时,可以删除该样本数据。The input module 310 : used to obtain the training sample set, and send the training sample set to the encoding module 320 . Optionally, the input module 310 can also be used to preprocess the training sample set. It should be understood that different preprocessing processes are required for different sample data. For example, for image samples, the input module 310 can perform preprocessing such as image color space transformation, image cropping, and image scaling on the image samples; for text samples, input Module 310 can perform preprocessing such as removing non-text data and removing stop words on text samples; for audio samples, input module 310 can perform preprocessing such as audio format conversion, framing, and windowing on audio samples. Optionally, the input module 310 can also be used to preprocess the sample data according to the labeling result of the sample data, for example, when a certain sample data has no corresponding category information, the sample data can be deleted.
编码模块320:用于从输入模块310读取训练样本集,并对训练样本集中的每个样本数据进行编码处理,提取到每个样本数据的编码特征。Coding module 320: used to read the training sample set from the input module 310, and perform coding processing on each sample data in the training sample set, and extract the coding features of each sample data.
在一具体的实施例中,编码模块320可以包括编码模型321。在实际应用中,针对不同的样本数据,编码模块320可以采用不同的编码模型,比如说,对于图像样本,编码模块320可以采用业界已有的具有较好图像特征提取能力的神经网络模型(例如:CNN模型、VGG网络模型等)作为编码模型;对于文本样本,编码模块320可以采用业界已有的具有较好文本特征提取能力的神经网络模型(例如:长短期记忆网络(long short-term memory,LSTM)模型、基于变换器的双向编码器表示技术(bidirectional encoder representations from transformers,BERT)模型或Transformer模型等)作为编码模型;对于音频样本,编码模块320可以采用业界已有的具有较好音频特征提取能力的神经网络模型(例如:时延神经网络(time delay neural network,TDNN)模型、CNN模型等)作为编码模型。In a specific embodiment, the encoding module 320 may include an encoding model 321 . In practical applications, the encoding module 320 can adopt different encoding models for different sample data. For example, for image samples, the encoding module 320 can adopt an existing neural network model (such as : CNN model, VGG network model, etc.) as the encoding model; for text samples, the encoding module 320 can adopt an existing neural network model (for example: long short-term memory network (long short-term memory) with better text feature extraction capabilities in the industry , LSTM) model, converter-based bidirectional encoder representations from transformers (bidirectional encoder representations from transformers, BERT) model or Transformer model, etc.) as the encoding model; The neural network model with feature extraction capability (for example: time delay neural network (TDNN) model, CNN model, etc.) is used as the encoding model.
第一任务模块330:包括第一任务模型331,第一任务模块330用于接收编码模块320提取到的每个样本数据的编码特征,并将每个样本数据的编码特征作为第一任务模型331的输入,经过第一任务模型331的分类学习,得到每个样本数据的预测的类别信息。The first task module 330: including the first task model 331, the first task module 330 is used to receive the encoding feature of each sample data extracted by the encoding module 320, and use the encoding feature of each sample data as the first task model 331 After the classification learning of the first task model 331, the predicted category information of each sample data is obtained.
第二任务模块340:包括第二任务模型341,第二任务模块340用于接收编码模块320提取到的每个样本数据的编码特征,并将每个样本数据的编码特征作为第二任务模型341的输入,经过第二任务模型341的学习,得到每个样本数据的预测的类别信息的关联信息。The second task module 340: includes a second task model 341, the second task module 340 is used to receive the encoding feature of each sample data extracted by the encoding module 320, and use the encoding feature of each sample data as the second task model 341 After the input of the second task model 341, the association information of the predicted category information of each sample data is obtained.
应理解,第二任务模型341的选取应结合具体的分类任务、训练样本集中的样本数据及样本数据的标注结果等的需求。例如,对于文本分类任务,当用户标注的文本数据的类别信息的关联信息包含于文本数据中时,第二任务模型341可以是用于识别出文本数据中能够体现文本数据的类别信息的关键字或关键词的序列标注模型,例如,条件随机场(conditional random field,CRF)、双向LSTM-CRF(bidirectional LSTM-CRF,BiLSTM-CRF)、隐马尔可夫模型(hidden markov model,HMM)等。又例如,对于文本分类任务,当用户标注的文本数据的类别信息的关联信息不包含于文本数据中时,第二任务模型341可以是用于生成能够体现文本数据的类别信息的关键字或关键词的文本摘要模型,例如,注意力模型、LSTM模型等。又例如,对于图像分类任务,当用户标注的图像数据的类别信息的关联信息包含于图像数据中时,第二任务模型341可以是用于识别出图像数据中能够体现图像数据的类别信息的区域目标检测模型,例如,一阶段统一实时目标检测(you only look once:unified,Yolo)模型、单镜头多盒检测器(single shot multi box detector,SSD)模型、区域卷积神经网络(region convolutional  neural network,RCNN)模型。又例如,对于音频分类任务,当用户标注的音频数据的类别信息的关联信息包含于音频数据中时,第二任务模型341可以是用于识别出音频数据中能够表征音频数据的类别信息的片段,其本质是识别音频序列,因此,第二任务模型341也可以采用序列标注模型。It should be understood that the selection of the second task model 341 should be based on the requirements of specific classification tasks, sample data in the training sample set, labeling results of the sample data, and the like. For example, for the text classification task, when the associated information of the category information of the text data marked by the user is included in the text data, the second task model 341 can be used to identify the keywords in the text data that can reflect the category information of the text data Or keyword sequence tagging models, for example, conditional random field (conditional random field, CRF), bidirectional LSTM-CRF (bidirectional LSTM-CRF, BiLSTM-CRF), hidden Markov model (hidden markov model, HMM), etc. For another example, for a text classification task, when the associated information of the category information of the text data marked by the user is not included in the text data, the second task model 341 may be used to generate keywords or key words that can reflect the category information of the text data. Text summarization models for words, such as attention models, LSTM models, etc. For another example, for an image classification task, when the associated information of the category information of the image data marked by the user is included in the image data, the second task model 341 may be used to identify the region in the image data that can reflect the category information of the image data Target detection models, for example, one-stage unified real-time target detection (you only look once:unified, Yolo) model, single shot multi box detector (SSD) model, region convolutional neural network (region convolutional neural network) network, RCNN) model. For another example, for an audio classification task, when the associated information of the category information of the audio data marked by the user is included in the audio data, the second task model 341 may be used to identify a segment of the category information in the audio data that can characterize the audio data , its essence is to recognize audio sequences, therefore, the second task model 341 can also use a sequence labeling model.
在一具体的实施例中,上述编码模型321、第一任务模型331以及第二任务模型341可以是用户上传至模型训练系统100的。可选的,模型训练系统100可以存储有多个模型,因此上述编码模型321、第一任务模型331以及第二任务模型341还可以是用户从模型训练系统100中选取的。可选的,上述编码模型321、第一任务模型331以及第二任务模型341还可以是模型训练系统100根据用户的需求选取的,此次不作具体限定。其中,用户的需求可以包括用户期望完成的分类任务、用户对编码模型321、第一任务模型331以及第二任务模型341中的初始参数的要求等。In a specific embodiment, the coding model 321 , the first task model 331 and the second task model 341 may be uploaded by the user to the model training system 100 . Optionally, the model training system 100 may store a plurality of models, so the coding model 321 , the first task model 331 and the second task model 341 may also be selected by the user from the model training system 100 . Optionally, the encoding model 321 , the first task model 331 and the second task model 341 may also be selected by the model training system 100 according to the user's requirements, which are not specifically limited this time. Wherein, the user's requirement may include the classification task that the user expects to complete, the user's requirement for the initial parameters in the coding model 321 , the first task model 331 and the second task model 341 , and the like.
基于图4所示的分类模型,模型训练系统100根据训练样本集训练分类模型的具体过程为:模型训练系统100将训练样本集输入分类模型300的输入模块310,然后经由输入模块310到达编码模块320,经过编码模块320的处理可以得到每个样本数据的编码特征,然后将每个样本数据的编码特征和用户标注的每个样本数据的类别信息输入第一任务模块330,第一任务模块330接收到每个样本数据的编码特征和用户标注的每个样本数据的类别信息后,将每个样本数据的编码特征输入第一任务模型331,得到每个样本数据的预测的类别信息,然后利用第一损失函数计算用户标注的每个样本数据的类别信息与预测得到的每个样本数据的类别信息之间的第一损失值。同时,编码模块320还将每个样本数据的编码特征和每个样本数据的类别信息的关联信息输入第二任务模块340。第二任务模块340接收到每个样本数据的编码特征和每个样本数据类别信息的关联信息后,将每个样本数据的编码特征输入第二任务模型341,预测得到每个样本数据的类别信息的关联信息,然后利用第二损失函数计算用户标注的每个样本数据的类别信息的关联信息与预测得到的每个样本数据的类别信息的关联信息之间的第二损失值。然后,根据第一损失值和第二损失值(例如,第一损失值和第二损失值的和、乘积等)调整第一任务模型331、第二任务模型341和编码模型321中的参数,得到调整后的分类模型。然后,重复上述过程,直至分类模型根据输入的每个样本数据预测得到的每个样本数据的类别信息与用户标注的每个样本数据的类别信息之间的差异满足第一阈值,并且,分类模型根据输入的每个样本数据预测得到的每个样本数据的类别信息的关联信息与用户标注的每个样本数据的类别信息的关联信息之间的差异满足第二阈值。其中,第一阈值和第二阈值可以是用户设定的,也可以是模型训练系统100根据实际情况动态调整的,此处不作具体限定。Based on the classification model shown in FIG. 4 , the specific process of the model training system 100 training the classification model according to the training sample set is: the model training system 100 inputs the training sample set into the input module 310 of the classification model 300, and then reaches the encoding module via the input module 310 320, after the processing of the encoding module 320, the encoding features of each sample data can be obtained, and then the encoding features of each sample data and the category information of each sample data marked by the user are input into the first task module 330, the first task module 330 After receiving the coding features of each sample data and the category information of each sample data marked by the user, input the coding features of each sample data into the first task model 331 to obtain the predicted category information of each sample data, and then use The first loss function calculates a first loss value between the category information of each sample data marked by the user and the predicted category information of each sample data. At the same time, the encoding module 320 also inputs the associated information of the encoding feature of each sample data and the category information of each sample data into the second task module 340 . After the second task module 340 receives the coding features of each sample data and the associated information of the category information of each sample data, it inputs the coding features of each sample data into the second task model 341, and predicts the category information of each sample data Then use the second loss function to calculate the second loss value between the association information of the category information of each sample data marked by the user and the predicted association information of the category information of each sample data. Then, adjust the parameters in the first task model 331, the second task model 341 and the encoding model 321 according to the first loss value and the second loss value (for example, the sum and product of the first loss value and the second loss value, etc.), Get the adjusted classification model. Then, the above process is repeated until the difference between the category information of each sample data predicted by the classification model based on each input sample data and the category information of each sample data marked by the user satisfies the first threshold, and the classification model The difference between the association information of the category information of each sample data predicted according to the input of each sample data and the association information of the category information of each sample data marked by the user satisfies the second threshold. Wherein, the first threshold and the second threshold may be set by the user, or may be dynamically adjusted by the model training system 100 according to actual conditions, which are not specifically limited here.
在上述分类模型300的训练过程中,不仅会根据每个样本数据的类别信息训练编码模型321,还会根据每个样本数据的类别信息的关联信息训练编码模型321,如此,可以使得编码模型321提取到的特征既能够表征样本数据的类别信息的关联信息,也能够表征样本数据的类别信息。考虑到样本数据的类别信息的关联信息与样本数据的类别信息相关,因此,经过模型训练可以提高编码模型321提取到样本数据中有关样本数据的类别信息的特征,从而进一步提高第一任务模型331进行分类学习的准确率,也就是提高了分类模型300执行分类任务的准确率。In the training process of the above classification model 300, not only the coding model 321 will be trained according to the category information of each sample data, but also the coding model 321 will be trained according to the associated information of the category information of each sample data, so that the coding model 321 can be The extracted features can not only represent the association information of the category information of the sample data, but also represent the category information of the sample data. Considering that the association information of the category information of the sample data is related to the category information of the sample data, the features of the category information of the sample data extracted by the coding model 321 can be improved after model training, thereby further improving the first task model 331 The accuracy of the classification learning is to improve the accuracy of the classification model 300 performing the classification task.
本实施例中,当分类模型训练完成后,可以利用训练好的分类模型执行分类任务。具体地,在利用训练好的分类模型执行分类任务之前,可以直接利用训练好的分类模型执行分类任务,这样不仅可以预测得到输入数据的类别信息,还可以预测得到输入数据的类别解释信 息,这样可以增加预测得到的类别信息的可信度。另外,也可以先将训练好的分类模型中的第一任务模块删除,然后再执行分类任务,这样可以提高分类任务的效率,节省计算资源。In this embodiment, after the classification model is trained, the trained classification model can be used to perform the classification task. Specifically, before using the trained classification model to perform the classification task, the trained classification model can be directly used to perform the classification task, so that not only the category information of the input data can be predicted, but also the category interpretation information of the input data can be predicted, so that The reliability of the predicted category information can be increased. In addition, the first task module in the trained classification model can also be deleted first, and then the classification task is performed, which can improve the efficiency of the classification task and save computing resources.
应理解,当训练样本集不仅包括第一原始样本集、用户对第一原始样本集中的每个样本数据的标注结果,还包括第二原始样本集及用户对第二原始样本集中的每个样本数据的标注结果时,模型训练系统100在训练分类模型时,可以将仅标注有类别信息的样本数据的类别信息的关联信息设置为0,然后利用上述损失函数(包括第一损失函数和第二损失函数)调整分类模型的参数,从而完成分类模型的训练。当训练样本集包括第二原始样本集及用户对第二原始样本集中的每个样本数据的标注结果,但不包括第一原始样本集、用户对第一原始样本集中的每个样本数据的标注结果时,模型训练系统100在训练分类模型时,可以仅利用第一损失函数调整分类模型的参数,从而完成分类模型的训练。It should be understood that when the training sample set includes not only the first original sample set and the user's labeling results for each sample data in the first original sample set, but also includes the second original sample set and the user's labeling results for each sample data in the second original sample set When labeling the data, the model training system 100 can set the association information of the category information of the sample data labeled with category information to 0 when training the classification model, and then use the above loss function (including the first loss function and the second Loss function) to adjust the parameters of the classification model to complete the training of the classification model. When the training sample set includes the second original sample set and the user's labeling results for each sample data in the second original sample set, but does not include the first original sample set, the user's labeling results for each sample data in the first original sample set As a result, when training the classification model, the model training system 100 can only use the first loss function to adjust the parameters of the classification model, thereby completing the training of the classification model.
需要说明的是,上述模型训练方法具体可以由模型训练系统100中的一个或多个模块共同实现。具体地,训练数据标注模块110用于实现步骤S101-S102。训练数据存储模块130用于将训练数据标注模块110经过步骤S101-S102后获取到的原始样本集以及训练样本集进行存储。模型训练模块120用于实现步骤S103。模型存储模块140用于存储模型训练模块120经过步骤S103后得到的训练好的分类模型。可选的,模型存储模块140还用于存储编码模型、第一任务模型和第二任务模型。可选的,模型选择模块150用于根据用户的需求选择出上述编码模型321、第一任务模型331和第二任务模型341。It should be noted that, the above-mentioned model training method may be implemented jointly by one or more modules in the model training system 100 . Specifically, the training data labeling module 110 is used to implement steps S101-S102. The training data storage module 130 is used to store the original sample set and the training sample set acquired by the training data labeling module 110 after steps S101-S102. The model training module 120 is used to implement step S103. The model storage module 140 is used to store the trained classification model obtained by the model training module 120 after step S103. Optionally, the model storage module 140 is also used to store the coding model, the first task model and the second task model. Optionally, the model selection module 150 is configured to select the above-mentioned coding model 321 , first task model 331 and second task model 341 according to user requirements.
当模型训练系统100执行上述模型训练方法时,第一用户可以不仅对样本数据的类别信息进行标注,还可以对样本数据的类别信息的关联信息进行标注。由于样本数据的类别信息的关联信息能够体现样本数据的类别信息,因此,相较于只利用样本数据的类别信息训练分类模型,利用样本数据的类别信息和样本数据的类别信息的关联信息训练出的分类模型的准确率更高。而且,样本数据的类别信息的关联信息是第一用户在生成该样本数据的类别信息的过程中确认并标注的,因此第一用户标注一个样本数据的类别信息的关联信息所需的时间不会太多。那么,相较于用户仅标注样本数据的类别信息这种方式,通过本申请提供的方法用户只需标注更少的样本数据、使用更少的时间便可以训练出一个具有相同准确率的分类模型,从而提高了分类模型的训练效率,减少了分类模型的训练成本。When the model training system 100 executes the above model training method, the first user can not only mark the category information of the sample data, but also mark the associated information of the category information of the sample data. Since the association information of the category information of the sample data can reflect the category information of the sample data, compared with only using the category information of the sample data to train the classification model, using the category information of the sample data and the category information of the sample data to train the classification model The accuracy of the classification model is higher. Moreover, the associated information of the category information of the sample data is confirmed and marked by the first user during the process of generating the category information of the sample data, so the time required for the first user to mark the associated information of the category information of a sample data will not too much. Then, compared with the method in which the user only labels the category information of the sample data, the method provided by this application allows the user to train a classification model with the same accuracy only by labeling less sample data and using less time , thus improving the training efficiency of the classification model and reducing the training cost of the classification model.
前述内容详细介绍了本申请提供的模型训练系统100,以及利用该系统如何实现模型训练的过程,下面结合图5-图6介绍模型训练系统100的部署方式和应用场景。The preceding content introduces the model training system 100 provided by this application in detail, and how to use the system to implement the process of model training. The following describes the deployment mode and application scenarios of the model training system 100 in conjunction with FIGS. 5-6 .
上述模型训练系统100的部署灵活,具体可以部署在云环境,云环境是云计算模式下利用基础资源向用户提供云服务的实体。云环境包括云数据中心和云服务平台,所述云数据中心包括云服务提供商拥有的大量基础资源(包括计算资源、存储资源和网络资源),云数据中心包括的计算资源可以是大量的计算设备(例如服务器)。模型训练系统可以是云数据中心中用于进行模型训练的服务器;模型训练系统也可以是创建在云数据中心中的用于进行模型训练的虚拟机;模型训练系统还可以是部署在云数据中心中的服务器或者虚拟机上的软件装置,该软件装置实现模型的训练,该软件装置可以分布式地部署在多个服务器上、或者分布式地部署在多个虚拟机上、或者分布式地部署在虚拟机和服务器上。The above-mentioned model training system 100 can be deployed flexibly, specifically, it can be deployed in a cloud environment, which is an entity that uses basic resources to provide users with cloud services under the cloud computing mode. The cloud environment includes a cloud data center and a cloud service platform. The cloud data center includes a large number of basic resources (including computing resources, storage resources and network resources) owned by the cloud service provider. The computing resources included in the cloud data center can be a large number of computing resources. devices (such as servers). The model training system can be a server used for model training in the cloud data center; the model training system can also be a virtual machine created in the cloud data center for model training; the model training system can also be deployed in the cloud data center The software device on the server or virtual machine in the server, the software device realizes the training of the model, and the software device can be deployed on multiple servers in a distributed manner, or deployed on multiple virtual machines in a distributed manner, or deployed in a distributed manner on virtual machines and servers.
当模型训练系统100部署在云环境时,模型训练系统100可以由云服务提供商在云服务平台(例如,AI开发平台)抽象为一种云服务(以下称为模型训练的云服务)提供给上述第二用户。那么,第二用户可以通过在云服务平台上购买这项模型训练的云服务,从而可以在模型训练系统100上完成分类模型的训练。其中,第二用户购买模型训练的云服务的方式可 以有多种,例如,第二用户可以按周期(例如,小时、月)付费;又例如,第二用户可以按需付费,也就是第二用户可以预先充值,并在使用完上述云服务后由云服务平台根据最终资源的使用情况进行付费。When the model training system 100 is deployed in a cloud environment, the model training system 100 can be abstracted as a cloud service (hereinafter referred to as a cloud service for model training) by a cloud service provider on a cloud service platform (for example, an AI development platform) and provided to the above-mentioned second user. Then, the second user can complete the training of the classification model on the model training system 100 by purchasing the cloud service of this model training on the cloud service platform. Among them, the second user can purchase the cloud service of model training in various ways, for example, the second user can pay by cycle (for example, hour, month); another example, the second user can pay on demand, that is, the second Users can recharge in advance, and after using the above cloud services, the cloud service platform will pay according to the usage of the final resources.
以图5为例,图5示出了一种模型训练系统100部署在云环境时的应用场景。如图5所示,第二用户在云服务平台上购买模型训练的云服务后,第二用户通过云服务平台提供的应用程序接口(Application Programming Interface,API)将原始样本集上传至模型训练系统100,然后第一用户在模型训练系统100上为原始样本集中的每个样本数据标注类别信息以及类别信息的关联信息,得到训练样本集。之后,模型训练系统100自动根据训练样本集训练分类模型,从而得到训练好的分类模型。另外,模型训练系统100还可以通过云服务平台提供的API向第二用户返回训练好的分类模型,以使得第二用户可以利用训练好的分类模型完成相应的分类任务。Taking FIG. 5 as an example, FIG. 5 shows an application scenario when the model training system 100 is deployed in a cloud environment. As shown in Figure 5, after the second user purchases the cloud service for model training on the cloud service platform, the second user uploads the original sample set to the model training system through the Application Programming Interface (API) provided by the cloud service platform 100, and then the first user marks category information and associated information of category information for each sample data in the original sample set on the model training system 100 to obtain a training sample set. Afterwards, the model training system 100 automatically trains the classification model according to the training sample set, so as to obtain a trained classification model. In addition, the model training system 100 can also return the trained classification model to the second user through the API provided by the cloud service platform, so that the second user can use the trained classification model to complete the corresponding classification task.
模型训练系统100还可以部署在边缘环境。边缘环境是指距离终端计算设备较近的边缘数据中心或者边缘计算设备(例如,边缘服务器、具有计算能力的边缘小站等)的集合。终端计算设备包括终端服务器、智能手机、笔记本电脑、平板电脑、个人台式电脑、智能摄像机等设备。当模型训练系统100部署在边缘环境时,模型训练系统100可以是单独地部署在边缘环境中的一个边缘服务器或一个虚拟机上,也可以分布式地部署在边缘环境中的多个边缘服务器上、或多个虚拟机上,或一部分部署在边缘服务器上,一部分部署在虚拟机上。The model training system 100 can also be deployed in an edge environment. The edge environment refers to a collection of edge data centers or edge computing devices (for example, edge servers, edge stations with computing capabilities, etc.) that are closer to terminal computing devices. Terminal computing devices include terminal servers, smart phones, notebook computers, tablet computers, personal desktop computers, smart cameras and other devices. When the model training system 100 is deployed in the edge environment, the model training system 100 can be deployed separately on an edge server or a virtual machine in the edge environment, and can also be deployed in a distributed manner on multiple edge servers in the edge environment , or multiple virtual machines, or a part is deployed on the edge server and a part is deployed on the virtual machines.
模型训练系统100还可以部署在一个或多个终端计算设备上。The model training system 100 can also be deployed on one or more terminal computing devices.
由于模型训练系统100可以在逻辑上划分为多个功能模块(如图1所示),因此,模型训练系统100还可以分布式地部署在不同的环境中,不同的环境可以包括上述云环境、上述边缘环境和上述终端计算设备。以图6为例,图6示出了一种模型训练系统100分布式地部署在不同环境时的应用场景。如图6所示,模型训练系统100中的训练数据标注模块110可以部署在终端计算设备上,模型训练模块120部署在云环境上。可选的,当模型训练系统100包括训练数据存储模块130、模型存储模块140以及模型选择模块150时,训练数据存储模块130、模型存储模块140以及模型选择模块150也可以部署在云环境中。Since the model training system 100 can be logically divided into multiple functional modules (as shown in FIG. 1 ), the model training system 100 can also be deployed in different environments in a distributed manner. Different environments can include the above-mentioned cloud environment, The aforementioned edge environment and the aforementioned terminal computing device. Taking FIG. 6 as an example, FIG. 6 shows an application scenario when the model training system 100 is distributed and deployed in different environments. As shown in FIG. 6 , the training data labeling module 110 in the model training system 100 can be deployed on a terminal computing device, and the model training module 120 can be deployed on a cloud environment. Optionally, when the model training system 100 includes the training data storage module 130, the model storage module 140, and the model selection module 150, the training data storage module 130, the model storage module 140, and the model selection module 150 may also be deployed in a cloud environment.
当模型训练系统100单独地部署在任意环境中的一个计算设备上(例如,单独部署在一个终端计算设备上)时,部署有模型训练系统100的计算设备可以是如图7所示的计算设备。如图7所示,图7示出了部署有模型训练系统100的计算设备400的硬件结构示意图。其中,计算设备400包括存储器410、处理器420、通信接口430以及总线440。其中,存储器410、处理器420、通信接口430通过总线440实现彼此之间的通信连接。When the model training system 100 is deployed separately on a computing device in any environment (for example, separately deployed on a terminal computing device), the computing device deployed with the model training system 100 may be a computing device as shown in FIG. 7 . As shown in FIG. 7 , FIG. 7 shows a schematic diagram of a hardware structure of a computing device 400 deployed with the model training system 100 . Wherein, the computing device 400 includes a memory 410 , a processor 420 , a communication interface 430 and a bus 440 . Wherein, the memory 410 , the processor 420 , and the communication interface 430 are connected to each other through the bus 440 .
存储器410可以是只读存储器(read only memory,ROM),静态存储设备、动态存储设备或者随机存取存储器(random access memory,RAM)。存储器410可以存储程序,例如,训练数据标注模块110中的程序、模型训练模块120中的程序等。当存储器410中存储的程序被处理器420执行时,处理器420和通信接口430用于执行上述步骤S101-S103所述的部分或全部方法。存储器410还可以存储数据,例如:存储器410中的一部分存储资源可用于存储训练数据存储模块130中存储的原始样本集和训练样本集,一部分存储资源可用于存储模型存储模块140中存储的各个模型,一部分存储资源用于存储处理器420在执行过程中产生的中间数据或结果数据,例如,分类模型的参数等。The memory 410 may be a read only memory (read only memory, ROM), a static storage device, a dynamic storage device or a random access memory (random access memory, RAM). The memory 410 may store programs, for example, programs in the training data labeling module 110, programs in the model training module 120, and the like. When the program stored in the memory 410 is executed by the processor 420, the processor 420 and the communication interface 430 are used to execute part or all of the methods described in the above steps S101-S103. The memory 410 can also store data, for example: a part of the storage resources in the memory 410 can be used to store the original sample set and the training sample set stored in the training data storage module 130, and a part of the storage resources can be used to store each model stored in the model storage module 140 , a part of the storage resource is used to store intermediate data or result data generated by the processor 420 during execution, for example, parameters of the classification model, and the like.
处理器420可以采用通用的中央处理器(central processing unit,CPU),微处理器,专用集成电路(application specific integrated circuit,ASIC),图形处理器(graphics processing unit,GPU) 或者一个或多个集成电路。The processor 420 may adopt a general-purpose central processing unit (central processing unit, CPU), a microprocessor, an application specific integrated circuit (application specific integrated circuit, ASIC), a graphics processing unit (graphics processing unit, GPU) or one or more integrated circuit.
处理器420还可以是一种集成电路芯片,具有信号的处理能力。在实现过程中,上述模型训练系统100的部分或全部功能可用通过处理器420中的硬件的集成逻辑电路或者软件形式的指令完成。处理器420还可以是通用处理器、数据信号处理器(digital signal process,DSP)、现场可编程逻辑门阵列(field programmable gate array,FPGA)或者其他可编程逻辑器件,分立门或者晶体管逻辑器件,分立硬件组件,可以实现或者执行本申请实施例中公开的方法、步骤及逻辑框图。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等,结合本申请实施例所公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器、闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介质位于存储器410,处理器420读取存储器410中的信息,结合其硬件完成上述模型训练系统100的部分或全部功能。The processor 420 may also be an integrated circuit chip, which has a signal processing capability. During implementation, part or all of the functions of the above-mentioned model training system 100 can be implemented by hardware integrated logic circuits in the processor 420 or instructions in the form of software. The processor 420 can also be a general-purpose processor, a data signal processor (digital signal process, DSP), a field programmable logic gate array (field programmable gate array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, The discrete hardware components can realize or execute the methods, steps and logic block diagrams disclosed in the embodiments of the present application. The general-purpose processor can be a microprocessor or the processor can also be any conventional processor, etc., and the steps of the method disclosed in conjunction with the embodiments of the present application can be directly embodied as a hardware decoding processor to execute and complete, or use decoding processing The combination of hardware and software modules in the device is completed. The software module can be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, register. The storage medium is located in the memory 410, and the processor 420 reads the information in the memory 410, and combines with its hardware to complete part or all of the functions of the model training system 100 described above.
通信接口430使用例如但不限于收发器一类的收发模块,来实现计算设备400与其他设备或通信网络之间的通信。例如,可以通过通信接口430获取用户上传的原始样本集,还可以通过通信接口430将训练好的分类模型发送给其他设备。The communication interface 430 uses a transceiver module such as but not limited to a transceiver to implement communication between the computing device 400 and other devices or communication networks. For example, the original sample set uploaded by the user can be obtained through the communication interface 430 , and the trained classification model can also be sent to other devices through the communication interface 430 .
总线440可以包括在计算设备400中的各个部件(例如,存储器410、处理器420、通信接口430)之间传送信息的通路。Bus 440 may comprise a pathway for communicating information between various components in computing device 400 (eg, memory 410 , processor 420 , communication interface 430 ).
当上述模型训练系统100中的各个模块分布式地部署在同一环境或不同环境中的多个计算设备上时,部署有模型训练系统100的多个计算设备可以构成如图8所示的计算设备系统。如图8所示,图8示出了部署有模型训练系统100的计算设备系统500的硬件结构示意图。其中,计算设备系统500包括多个计算设备600,计算设备系统500中的多个计算设备600可以通过内部处理器执行计算机指令协同地实现模型训练系统100的功能。When the various modules in the above-mentioned model training system 100 are distributed and deployed on multiple computing devices in the same environment or in different environments, the multiple computing devices deployed with the model training system 100 can constitute a computing device as shown in FIG. 8 system. As shown in FIG. 8 , FIG. 8 shows a schematic diagram of a hardware structure of a computing device system 500 deployed with the model training system 100 . Wherein, the computing device system 500 includes multiple computing devices 600 , and the multiple computing devices 600 in the computing device system 500 can cooperatively implement the functions of the model training system 100 through the execution of computer instructions by an internal processor.
如图8所示,每个计算设备600包括存储器610、处理器620、通信接口630以及总线640。其中,存储器610、处理器620、通信接口630通过总线640实现彼此之间的通信连接。As shown in FIG. 8 , each computing device 600 includes a memory 610 , a processor 620 , a communication interface 630 and a bus 640 . Wherein, the memory 610 , the processor 620 , and the communication interface 630 are connected to each other through the bus 640 .
存储器610可以是ROM、RAM、静态存储设备或者动态存储设备。存储器610可以存储计算机指令,当存储器610中存储的计算机指令被处理器620执行时,处理器620和通信接口430用于执行上述步骤S101-S103所述的部分或全部方法。存储器610还可以存储数据,例如:存储器610中的一部分存储资源可用于存储训练数据存储模块130中存储的原始样本集和训练样本集,一部分存储资源可用于存储模型存储模块140中存储的各个模型,一部分存储资源用于存储处理器620在执行过程中产生的中间数据或结果数据,例如,分类模型的参数等。Memory 610 may be ROM, RAM, static storage, or dynamic storage. The memory 610 may store computer instructions. When the computer instructions stored in the memory 610 are executed by the processor 620, the processor 620 and the communication interface 430 are used to execute part or all of the methods described in the above steps S101-S103. The memory 610 can also store data, for example: a part of the storage resources in the memory 610 can be used to store the original sample set and the training sample set stored in the training data storage module 130, and a part of the storage resources can be used to store each model stored in the model storage module 140 , a part of the storage resource is used to store intermediate data or result data generated by the processor 620 during execution, for example, parameters of the classification model and the like.
处理器620可以采用通用的CPU、GPU、ASIC、微处理器或者一个或多个集成电路。处理器620还可以是一种集成电路芯片,具有信号的处理能力。在实现过程中,本申请的模型训练系统的部分或全部功能可用通过处理器620中的硬件的集成逻辑电路或者软件形式的指令完成。处理器620还可以是DSP、FPGA、其他可编程逻辑器件、通用处理器、分立门、分立硬件组件或者晶体管逻辑器件,可以实现或者执行本申请实施例中公开的方法、步骤及逻辑框图。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等,结合本申请实施例所公开的方法的步骤可以直接体现为硬件译码处理器执行完成,或者用译码处理器中的硬件及软件模块组合执行完成。软件模块可以位于随机存储器、闪存、只读存储器,可编程只读存储器或者电可擦写可编程存储器、寄存器等本领域成熟的存储介质中。该存储介 质位于存储器610,处理器620读取存储器610中的信息,结合其硬件完成上述模型训练系统100的部分功能。The processor 620 may adopt a general-purpose CPU, GPU, ASIC, microprocessor, or one or more integrated circuits. The processor 620 may also be an integrated circuit chip, which has a signal processing capability. During implementation, part or all of the functions of the model training system of the present application can be implemented by hardware integrated logic circuits in the processor 620 or instructions in the form of software. The processor 620 can also be DSP, FPGA, other programmable logic devices, general-purpose processors, discrete gates, discrete hardware components or transistor logic devices, and can realize or execute the methods, steps and logic diagrams disclosed in the embodiments of the present application. The general-purpose processor can be a microprocessor or the processor can also be any conventional processor, etc., and the steps of the method disclosed in conjunction with the embodiments of the present application can be directly embodied as a hardware decoding processor to execute and complete, or use decoding processing The combination of hardware and software modules in the device is completed. The software module can be located in a mature storage medium in the field such as random access memory, flash memory, read-only memory, programmable read-only memory or electrically erasable programmable memory, register. The storage medium is located in the memory 610, and the processor 620 reads the information in the memory 610, and completes part of the functions of the above-mentioned model training system 100 in combination with its hardware.
通信接口6300使用例如但不限于收发器一类的收发模块,来实现计算设备600与其他设备或通信网络之间的通信。例如,可以通过通信接口630获取用户上传的待标注的样本数据集等。The communication interface 6300 uses a transceiver module such as but not limited to a transceiver to implement communication between the computing device 600 and other devices or communication networks. For example, a sample data set to be marked uploaded by a user may be obtained through the communication interface 630 .
总线640可包括在计算设备600各个部件(例如,存储器610、处理器620、通信接口630)之间传送信息的通路。Bus 640 may include pathways for communicating information between various components of computing device 600 (eg, memory 610 , processor 620 , communication interface 630 ).
上述每个计算设备600间通过通信网络建立通信通路。每个计算设备600上运行模型训练系统100中的一部分(例如:运行模型训练系统100中的训练数据标注模块110、模型训练模块120、训练数据存储模块130、模型存储模块140以及模型选择模块150中的一个或多个模块)。任一计算设备600可以为云数据中心中的服务器,或边缘数据中心中的计算设备,或终端计算设备。A communication path is established between each of the aforementioned computing devices 600 through a communication network. Run a part of the model training system 100 on each computing device 600 (for example: run the training data labeling module 110, the model training module 120, the training data storage module 130, the model storage module 140 and the model selection module 150 in the model training system 100 one or more modules in ). Any computing device 600 may be a server in a cloud data center, or a computing device in an edge data center, or a terminal computing device.
上述各个附图对应的流程的描述各有侧重,某个流程中没有详细描述的部分,可以参见其他流程的相关描述。The descriptions of the processes corresponding to each of the above-mentioned drawings have their own emphasis. For the parts that are not described in detail in a certain process, you can refer to the relevant descriptions of other processes.
在上述实施例中,可以全部或部分地通过软件、硬件或者其组合来实现。当使用软件实现时,可以全部或部分地以计算机程序产品的形式实现。提供模型训练系统的计算机程序产品包括一个或多个由模型训练系统执行的计算指令,在计算机上加载和执行这些计算机程序指令时,全部或部分地产生按照本申请实施例图所述的流程或功能。In the above-mentioned embodiments, all or part may be implemented by software, hardware or a combination thereof. When implemented using software, it may be implemented in whole or in part in the form of a computer program product. The computer program product that provides the model training system includes one or more calculation instructions executed by the model training system. When these computer program instructions are loaded and executed on the computer, all or part of the flow or process described in the embodiment diagram of the present application will be generated. Function.
上述计算机可以是通用计算机、专用计算机、计算机网络、或者其他可编程装置。上述计算机指令可以存储在计算机可读存储介质中,或者从一个计算机可读存储介质向另一个计算机可读存储介质传输,例如,上述计算机指令可以从一个网站站点、计算机、服务器或数据中心通过有线(例如,同轴电缆、光纤、双绞线或无线(例如,红外、无线、微波)等)方式向另一个网站站点、计算机、服务器或数据中心进行传输。上述计算机可读存储介质存储有提供模型训练系统的计算机程序指令。所述计算机可读存储介质可以是计算机能够存取的任何可用介质或者是包含一个或多个介质集成的服务器、数据中心等数据存储设备。上述可用介质可以是磁性介质(例如,软盘、硬盘、磁带)、光介质(例如,光盘)、或者半导体介质(例如,固态硬盘(solid state disk,SSD))。The above-mentioned computers may be general-purpose computers, special-purpose computers, computer networks, or other programmable devices. The above-mentioned computer instructions may be stored in a computer-readable storage medium, or transmitted from one computer-readable storage medium to another computer-readable storage medium. (eg, coaxial cable, optical fiber, twisted pair, or wireless (eg, infrared, wireless, microwave), etc.) to another website site, computer, server, or data center. The above-mentioned computer-readable storage medium stores computer program instructions providing a model training system. The computer-readable storage medium may be any available medium that can be accessed by a computer, or a data storage device such as a server or a data center integrated with one or more media. The above-mentioned usable medium may be a magnetic medium (for example, a floppy disk, a hard disk, or a magnetic tape), an optical medium (for example, an optical disk), or a semiconductor medium (for example, a solid state disk (solid state disk, SSD)).

Claims (16)

  1. 一种模型训练方法,其特征在于,包括:A model training method, characterized in that, comprising:
    获取原始样本集,所述原始样本集包括多个样本数据;obtaining an original sample set, the original sample set including a plurality of sample data;
    通过标注接口接收用户对每个样本数据的标注结果以获取训练样本集,其中,所述训练样本集包括所述多个样本数据以及所述每个样本数据的标注结果,所述每个样本数据的标注结果包括所述每个样本数据的类别信息以及所述每个样本数据的类别信息的关联信息;Receive the user's labeling results for each sample data through the labeling interface to obtain a training sample set, wherein the training sample set includes the plurality of sample data and the labeling results of each sample data, and each sample data The labeling result includes the category information of each sample data and the associated information of the category information of each sample data;
    根据所述训练样本集对分类模型进行训练。The classification model is trained according to the training sample set.
  2. 根据权利要求1所述的方法,其特征在于,所述多个样本数据中第一样本数据的类别信息的关联信息包含于所述第一样本数据中。The method according to claim 1, wherein the association information of the category information of the first sample data among the plurality of sample data is included in the first sample data.
  3. 根据权利要求1所述的方法,其特征在于,所述第一样本数据的类别信息的关联信息包括类别解释信息和所述类别解释信息的元数据,所述类别解释信息是所述第一样本数据中体现所述第一样本数据的类别信息的部分。The method according to claim 1, wherein the associated information of the category information of the first sample data includes category explanation information and metadata of the category explanation information, and the category explanation information is the first The part of the sample data embodies the category information of the first sample data.
  4. 根据权利要求1至3任一所述的方法,其特征在于,所述多个样本数据中第二样本数据的类别信息的关联信息不包含于所述第二样本数据中。The method according to any one of claims 1 to 3, characterized in that the associated information of the category information of the second sample data in the plurality of sample data is not included in the second sample data.
  5. 根据权利要求1至4任一所述的方法,其特征在于,所述每个样本数据的类别信息的关联信息是所述用户生成该样本数据的类别信息的过程中确认的。The method according to any one of claims 1 to 4, wherein the association information of the category information of each sample data is confirmed during the process of generating the category information of the sample data by the user.
  6. 根据权利要求1至5任一项所述的方法,其特征在于,所述根据所述训练样本集对分类模型进行训练,包括:The method according to any one of claims 1 to 5, wherein the training of the classification model according to the training sample set includes:
    将所述多个样本数据输入所述分类模型,得到所述每个样本数据的预测的类别信息以及预测的类别信息的关联信息;inputting the plurality of sample data into the classification model to obtain the predicted category information of each sample data and the associated information of the predicted category information;
    根据损失函数调整所述分类模型的参数,直至所述损失函数的输出满足阈值;Adjusting the parameters of the classification model according to the loss function until the output of the loss function meets the threshold;
    其中,所述损失函数包括第一损失函数和第二损失函数,所述第一损失函数用于指示所述每个样本数据的类别信息与所述每个样本的预测的类别信息之间的差异,所述第二损失函数用于指示所述每个样本数据的类别信息的关联信息与所述每个样本数据的预测的类别信息的关联信息之间的差异。Wherein, the loss function includes a first loss function and a second loss function, and the first loss function is used to indicate the difference between the class information of each sample data and the predicted class information of each sample , the second loss function is used to indicate the difference between the associated information of the category information of each sample data and the predicted associated information of the category information of each sample data.
  7. 根据权利要求1至6任一项所述的方法,其特征在于,所述分类模型包括编码模型、第一任务模型和第二任务模型,所述编码模型用于提取所述多个样本数据的特征,所述第一任务模型用于根据所述多个样本数据的特征确定所述每个样本数据的类别信息,所述第二任务模型用于根据所述多个样本数据的特征确定所述每个样本数据的类别信息的关联信息。The method according to any one of claims 1 to 6, wherein the classification model includes an encoding model, a first task model and a second task model, and the encoding model is used to extract the plurality of sample data features, the first task model is used to determine the category information of each sample data according to the characteristics of the multiple sample data, and the second task model is used to determine the Association information of category information for each sample data.
  8. 一种模型训练系统,其特征在于,包括:A model training system, characterized in that it comprises:
    训练数据标注模块,用于获取原始样本集,所述原始样本集包括多个样本数据;The training data labeling module is used to obtain an original sample set, and the original sample set includes a plurality of sample data;
    所述训练数据标注模块还用于通过标注接口接收用户对每个样本数据的标注结果以获取训练样本集,其中,所述训练样本集包括所述多个样本数据以及所述每个样本数据的标注结果,所述每个样本数据的标注结果包括所述每个样本数据的类别信息以及所述每个样本数据的类别信息的关联信息;The training data labeling module is also used to receive the user's labeling results for each sample data through the labeling interface to obtain a training sample set, wherein the training sample set includes the plurality of sample data and each sample data labeling results, the labeling results of each sample data include the category information of each sample data and the associated information of the category information of each sample data;
    模型训练模块,用于根据所述训练样本集对分类模型进行训练。The model training module is used to train the classification model according to the training sample set.
  9. 根据权利要求8所述的系统,其特征在于,所述多个样本数据中第一样本数据的类别信息的关联信息包含于所述第一样本数据中。The system according to claim 8, wherein the association information of the category information of the first sample data among the plurality of sample data is included in the first sample data.
  10. 根据权利要求8所述的系统,其特征在于,所述第一样本数据的类别信息的关联信息包括类别解释信息和所述类别解释信息的元数据,所述类别解释信息是所述第一样本数据中体现所述第一样本数据的类别信息的部分。The system according to claim 8, wherein the associated information of the category information of the first sample data includes category explanation information and metadata of the category explanation information, and the category explanation information is the first The part of the sample data embodies the category information of the first sample data.
  11. 根据权利要求8至10任一所述的系统,其特征在于,所述多个样本数据中第二样本数据的类别信息的关联信息不包含于所述第二样本数据中。The system according to any one of claims 8 to 10, wherein the association information of the category information of the second sample data among the plurality of sample data is not included in the second sample data.
  12. 根据权利要求8至11任一所述的系统,其特征在于,所述每个样本数据的类别信息的关联信息是所述用户生成该样本数据的类别信息的过程中确认的。The system according to any one of claims 8 to 11, wherein the association information of the category information of each sample data is confirmed by the user during the process of generating the category information of the sample data.
  13. 根据权利要求8至12任一项所述的系统,其特征在于,所述模型训练模块具体用于:The system according to any one of claims 8 to 12, wherein the model training module is specifically used for:
    将所述多个样本数据输入所述分类模型,得到所述每个样本数据的预测的类别信息以及预测的类别信息的关联信息;inputting the plurality of sample data into the classification model to obtain the predicted category information of each sample data and the associated information of the predicted category information;
    根据损失函数调整所述分类模型的参数,直至所述损失函数的输出满足阈值;Adjusting the parameters of the classification model according to the loss function until the output of the loss function meets the threshold;
    其中,所述损失函数包括第一损失函数和第二损失函数,所述第一损失函数用于指示所述每个样本数据的类别信息与所述每个样本的预测的类别信息之间的差异,所述第二损失函数用于指示所述每个样本数据的类别信息的关联信息与所述每个样本数据的预测的类别信息的关联信息之间的差异。Wherein, the loss function includes a first loss function and a second loss function, and the first loss function is used to indicate the difference between the class information of each sample data and the predicted class information of each sample , the second loss function is used to indicate the difference between the associated information of the category information of each sample data and the predicted associated information of the category information of each sample data.
  14. 根据权利要求8至12任一项所述的系统,其特征在于,所述分类模型包括编码模型、第一任务模型和第二任务模型,所述编码模型用于提取所述多个样本数据的特征,所述第一任务模型用于根据所述多个样本数据的特征确定所述每个样本数据的类别信息,所述第二任务模型用于根据所述多个样本数据的特征确定所述每个样本数据的类别信息的关联信息。The system according to any one of claims 8 to 12, wherein the classification model includes an encoding model, a first task model and a second task model, and the encoding model is used to extract the plurality of sample data features, the first task model is used to determine the category information of each sample data according to the characteristics of the multiple sample data, and the second task model is used to determine the Association information of category information for each sample data.
  15. 一种计算设备,其特征在于,所述计算设备包括处理器和存储器,所述存储器存储计算机指令,所述处理器执行所述计算机指令,以使所述计算设备执行前述权利要求1至7任一项所述的方法。A computing device, characterized in that the computing device includes a processor and a memory, the memory stores computer instructions, the processor executes the computer instructions, so that the computing device performs any of the preceding claims 1 to 7 one of the methods described.
  16. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质存储有计算机程序代码,当所述计算机程序代码被计算设备执行时,所述计算设备执行前述权利要求1至7任一项所述的方法。A computer-readable storage medium, characterized in that the computer-readable storage medium stores computer program codes, and when the computer program codes are executed by a computing device, the computing device executes any one of the preceding claims 1 to 7. method described in the item.
PCT/CN2022/071325 2021-05-11 2022-01-11 Model training method and system, and device and computer-readable storage medium WO2022237215A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110513038.9A CN115329825A (en) 2021-05-11 2021-05-11 Model training method, system, device and computer readable storage medium
CN202110513038.9 2021-05-11

Publications (1)

Publication Number Publication Date
WO2022237215A1 true WO2022237215A1 (en) 2022-11-17

Family

ID=83912904

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/071325 WO2022237215A1 (en) 2021-05-11 2022-01-11 Model training method and system, and device and computer-readable storage medium

Country Status (2)

Country Link
CN (1) CN115329825A (en)
WO (1) WO2022237215A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103198052A (en) * 2013-04-09 2013-07-10 山东师范大学 Active learning method based on support vector machine
CN109492549A (en) * 2018-10-24 2019-03-19 杭州睿琪软件有限公司 A kind of processing of training sample set, model training method and system
US20200279136A1 (en) * 2019-03-01 2020-09-03 Royal Bank Of Canada System and method for multi-type mean field reinforcement machine learning
CN112328823A (en) * 2020-11-25 2021-02-05 Oppo广东移动通信有限公司 Training method and device for multi-label classification model, electronic equipment and storage medium
CN112434811A (en) * 2019-08-26 2021-03-02 华为技术有限公司 Knowledge graph construction method and device, computing equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103198052A (en) * 2013-04-09 2013-07-10 山东师范大学 Active learning method based on support vector machine
CN109492549A (en) * 2018-10-24 2019-03-19 杭州睿琪软件有限公司 A kind of processing of training sample set, model training method and system
US20200279136A1 (en) * 2019-03-01 2020-09-03 Royal Bank Of Canada System and method for multi-type mean field reinforcement machine learning
CN112434811A (en) * 2019-08-26 2021-03-02 华为技术有限公司 Knowledge graph construction method and device, computing equipment and storage medium
CN112328823A (en) * 2020-11-25 2021-02-05 Oppo广东移动通信有限公司 Training method and device for multi-label classification model, electronic equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
LI MINGLEI, WANG DA, LU QIN, LONG YUNFEI: "Event Based Emotion Classification for News Articles", PROCEEDINGS OF THE 30TH PACIFIC ASIA CONFERENCE ON LANGUAGE, INFORMATION AND COMPUTATION, 30 October 2016 (2016-10-30), pages 153 - 162, XP093003892 *

Also Published As

Publication number Publication date
CN115329825A (en) 2022-11-11

Similar Documents

Publication Publication Date Title
WO2022142014A1 (en) Multi-modal information fusion-based text classification method, and related device thereof
WO2020125445A1 (en) Classification model training method, classification method, device and medium
WO2020114429A1 (en) Keyword extraction model training method, keyword extraction method, and computer device
US20180365258A1 (en) Artificial intelligence-based searching method and apparatus, device and computer-readable storage medium
WO2021190174A1 (en) Information determining method and apparatus, computer device, and storage medium
CN106919652B (en) Short-sighted frequency automatic marking method and system based on multi-source various visual angles transductive learning
JP7334395B2 (en) Video classification methods, devices, equipment and storage media
CN112084334B (en) Label classification method and device for corpus, computer equipment and storage medium
CN111680159A (en) Data processing method and device and electronic equipment
WO2020147409A1 (en) Text classification method and apparatus, computer device, and storage medium
CN107832338B (en) Method and system for recognizing core product words
CN112328761B (en) Method and device for setting intention label, computer equipment and storage medium
CN111950279B (en) Entity relationship processing method, device, equipment and computer readable storage medium
CN111144120A (en) Training sentence acquisition method and device, storage medium and electronic equipment
CN111191445B (en) Advertisement text classification method and device
CN115982376B (en) Method and device for training model based on text, multimode data and knowledge
CN111866610A (en) Method and apparatus for generating information
CN111177319A (en) Risk event determination method and device, electronic equipment and storage medium
WO2023045605A1 (en) Data processing method and apparatus, computer device, and storage medium
CN112231569A (en) News recommendation method and device, computer equipment and storage medium
CN113051914A (en) Enterprise hidden label extraction method and device based on multi-feature dynamic portrait
US20230214423A1 (en) Video generation
WO2021114634A1 (en) Text annotation method, device, and storage medium
WO2024021685A1 (en) Reply content processing method and media content interactive content interaction method
Andriyanov Combining Text and Image Analysis Methods for Solving Multimodal Classification Problems

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22806193

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE