CN111680517A - Method, apparatus, device and storage medium for training a model - Google Patents

Method, apparatus, device and storage medium for training a model Download PDF

Info

Publication number
CN111680517A
CN111680517A CN202010524463.3A CN202010524463A CN111680517A CN 111680517 A CN111680517 A CN 111680517A CN 202010524463 A CN202010524463 A CN 202010524463A CN 111680517 A CN111680517 A CN 111680517A
Authority
CN
China
Prior art keywords
intention
training
sample data
sample
labeling
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010524463.3A
Other languages
Chinese (zh)
Other versions
CN111680517B (en
Inventor
韩磊
张红阳
孙叔琦
孙珂
李婷婷
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN202010524463.3A priority Critical patent/CN111680517B/en
Publication of CN111680517A publication Critical patent/CN111680517A/en
Application granted granted Critical
Publication of CN111680517B publication Critical patent/CN111680517B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Mathematical Physics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Databases & Information Systems (AREA)
  • Manipulator (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The application discloses a method, a device, equipment and a storage medium for training a model, and relates to the fields of natural language processing and deep learning. The specific implementation scheme is as follows: acquiring a first training sample set and a second training sample set; predicting the first sample data by using a first intention recognition model, and determining confidence degrees corresponding to all prediction intents; determining a third training sample set according to the first labeling intention, the confidence degrees corresponding to the prediction intents, the second sample data and the corresponding second labeling intention; the second intent recognition model is trained based on a third set of training samples. The implementation mode can make full use of historical data used for establishing the man-machine conversation robot, reduce the cost of the newly-built man-machine conversation robot, and improve the intention recognition effect of the newly-built man-machine conversation robot under the condition of a small amount of labeled samples.

Description

Method, apparatus, device and storage medium for training a model
Technical Field
The present application relates to the field of computer technologies, and in particular, to the field of natural language processing and deep learning technologies, and in particular, to a method, an apparatus, a device, and a storage medium for training a model.
Background
Intent recognition is one of the core functions of a human-machine dialog robot, typically implemented using an intent recognition model. While the effectiveness of the intent recognition model depends heavily on the quantity and quality of the training data. The more samples in the training data are, the higher the labeling quality is, and the better the effect of the obtained intention recognition model is. Manually labeled intention dialogue samples have high data quality, but are difficult to scale due to the high labeling cost.
With the rise of human-computer interaction robots, various human-computer interaction robots are established for different applications, and the intentions of the human-computer interaction robots are deeply customized on application scenes basically. Therefore, the human-machine interaction robots in different application scenarios have different intention classification granularities, different intention names, and the like. Thus, when a new application scene is faced, the data of the man-machine interaction robot established before is difficult to be reused.
Disclosure of Invention
A method, apparatus, device, and storage medium for training a model are provided.
According to a first aspect, there is provided a method for training a model, comprising: acquiring a first training sample set and a second training sample set, wherein the first training sample set comprises first sample data and a corresponding first labeling intention, and the second training sample set comprises second sample data and a corresponding second labeling intention; acquiring a trained first intention recognition model and a second intention recognition model to be trained; predicting the first sample data by using a first intention recognition model, and determining confidence degrees corresponding to all prediction intents; determining a third training sample set according to the first labeling intention, the confidence degrees corresponding to the prediction intents, the second sample data and the corresponding second labeling intention; the second intent recognition model is trained based on a third set of training samples.
According to a second aspect, there is provided an apparatus for training a model, comprising: the device comprises a first acquisition unit, a second acquisition unit and a third acquisition unit, wherein the first acquisition unit is configured to acquire a first training sample set and a second training sample set, the first training sample set comprises first sample data and corresponding first labeling intents, and the second training sample set comprises second sample data and corresponding second labeling intents; a second acquisition unit configured to acquire the trained first intention recognition model and a second intention recognition model to be trained; a confidence degree determining unit configured to predict the first sample data by using the first intention recognition model, and determine a confidence degree corresponding to each prediction intention; the sample generation unit is configured to determine a third training sample set according to the first labeling intention, the confidence degree corresponding to each prediction intention, the second sample data and the corresponding second labeling intention; a first model training unit configured to train a second intent recognition model according to a third set of training samples.
According to a third aspect, there is provided an electronic device comprising: at least one processor; and a memory communicatively coupled to the at least one processor; wherein the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method as described in the first aspect.
According to a fourth aspect, there is provided a non-transitory computer readable storage medium having stored thereon computer instructions for causing a computer to perform the method as described in the first aspect.
According to the technology of the application, historical data used for establishing the man-machine conversation robot can be fully utilized, the cost of the newly-built man-machine conversation robot is reduced, and the intention recognition effect of the newly-built man-machine conversation robot is improved under the condition of a small number of labeled samples.
It should be understood that the statements in this section do not necessarily identify key or critical features of the embodiments of the present disclosure, nor do they limit the scope of the present disclosure. Other features of the present disclosure will become apparent from the following description.
Drawings
The drawings are included to provide a better understanding of the present solution and are not intended to limit the present application. Wherein:
FIG. 1 is an exemplary system architecture diagram in which one embodiment of the present application may be applied;
FIG. 2 is a flow diagram of one embodiment of a method for training a model according to the present application;
FIG. 3 is a schematic diagram of an application scenario of a method for training a model according to the present application;
FIG. 4 is a flow diagram of another embodiment of a method for training a model according to the present application;
FIG. 5 is a flow diagram of yet another embodiment of a method for training a model according to the present application;
FIG. 6 is a schematic block diagram of one embodiment of an apparatus for training models according to the present application;
FIG. 7 is a block diagram of an electronic device for implementing a method for training a model according to an embodiment of the present application.
Detailed Description
The following description of the exemplary embodiments of the present application, taken in conjunction with the accompanying drawings, includes various details of the embodiments of the application for the understanding of the same, which are to be considered exemplary only. Accordingly, those of ordinary skill in the art will recognize that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the present application. Also, descriptions of well-known functions and constructions are omitted in the following description for clarity and conciseness.
It should be noted that the embodiments and features of the embodiments in the present application may be combined with each other without conflict. The present application will be described in detail below with reference to the embodiments with reference to the attached drawings.
Fig. 1 illustrates an exemplary system architecture 100 to which embodiments of the present method for training a model or apparatus for training a model may be applied.
As shown in fig. 1, the system architecture 100 may include intelligent end devices 101, 102, 103, a network 104, and a server 105. The network 104 serves as a medium for providing communication links between the intelligent terminal devices 101, 102, 103 and the server 105. Network 104 may include various connection types, such as wired, wireless communication links, or fiber optic cables, to name a few.
The user may use the intelligent terminal device 101, 102, 103 to interact with the server 105 via the network 104 to receive or send messages or the like. Various communication client applications, such as a voice recognition application, a shopping application, a search application, an instant messaging tool, a mailbox client, social platform software, and the like, may be installed on the intelligent terminal devices 101, 102, 103.
The intelligent terminal devices 101, 102, 103 may be hardware or software. When the smart terminal 101, 102, 103 is hardware, it can be various electronic devices with voice recognition function, including but not limited to smart phones, smart speakers, smart robots, etc. When the smart terminal 101, 102, 103 is software, it can be installed in the electronic devices listed above. It may be implemented as multiple pieces of software or software modules (e.g., to provide distributed services) or as a single piece of software or software module. And is not particularly limited herein.
The server 105 may be a server that provides various services, such as a background server that provides a human-machine conversation robot for the intelligent terminal devices 101, 102, 103. The background server may analyze and perform other processing on the data such as the voice received by the terminal devices 101, 102, and 103, and feed back the processing result (for example, response data) to the intelligent terminal devices 101, 102, and 103.
The server 105 may be hardware or software. When the server 105 is hardware, it may be implemented as a distributed server cluster composed of a plurality of servers, or may be implemented as a single server. When the server 105 is software, it may be implemented as multiple pieces of software or software modules (e.g., to provide distributed services), or as a single piece of software or software module. And is not particularly limited herein.
It should be noted that the method for training the model provided in the embodiment of the present application is generally performed by the server 105. Accordingly, the means for training the model is typically located in the server 105.
It should be understood that the number of intelligent end devices, networks, and servers in fig. 1 is merely illustrative. There may be any number of intelligent end devices, networks, and servers, as desired for implementation.
With continued reference to FIG. 2, a flow 200 of one embodiment of a method for training a model according to the present application is shown. The method for training the model of the embodiment comprises the following steps:
step 201, a first training sample set and a second training sample set are obtained.
In this embodiment, an executing agent (e.g., the server 105 shown in fig. 1) of the method for training the model may obtain a first training sample set and a second training sample set. The first training sample set comprises first sample data and corresponding first labeling intents, and the second training sample set comprises second sample data and corresponding second labeling intents. Here, the sample data in the first training sample set may be sample data for a new domain, and the number thereof may be small. Because the manual annotation data consumption period is long and the cost is high, only a small amount of annotation data is acquired, so that the cost is reduced, and the efficiency is improved. The second set of training samples may include general sample data or historical sample data used in other fields, and the amount of sample data included in the second set of training samples may be very large. The sample data may be a plurality of dialog turns or a text. Each sample data corresponds to a labeling intention. The annotation intent can include audio and video requirements, search requirements, and the like.
It will be appreciated that the first set of training samples and the second set of training samples may include data that partially overlap, and as such, the first annotation intent may partially overlap the second annotation intent.
Step 202, obtaining a trained first intention recognition model and a second intention recognition model to be trained.
In this embodiment, the executing subject may further obtain a trained first intention recognition model and a second intention recognition model to be trained. Here, the first intention recognition model may be a general intention recognition model, and may also be an intention recognition model of other application fields. The second intent recognition model may be an intent recognition model applied to the new domain. It can be appreciated that the first intent recognition model is less accurate when applied to a new domain.
Step 203, the first sample data is predicted by using the first intention recognition model, and each prediction intention and the corresponding confidence coefficient are determined.
In this embodiment, the first intention recognition model may be utilized to predict the sample data, that is, the sample data in the first sample training sample set is input into the first intention recognition model. The confidence corresponding to each prediction intention of the sample data can be obtained. For example, for sample data a, the results output by the first intent recognition model may include: the intent 1 confidence is 0.2, the intent 2 confidence is 0.4, the intent 3 confidence is 0.3, and the intent 4 confidence is 0.3. For sample data B, the results output by the first intent recognition model may include: intent 5 confidence is 0.7. Each of the predicted intents may be considered herein as an intent tag for the first sample data.
Conventional supervised learning typically assumes that each sample data is associated with a unique one of the labels. However, in many practical tasks, one sample data typically possesses multiple tags. For example, in text classification, if a document is classified, the olympics belong to both business and sports; in image annotation, an image in a paris scene is associated with both the tower and the sky. In this embodiment, each intention label corresponding to the first sample data can be obtained by the first intention recognition model.
It is understood that the number of resulting prediction intents is at least one. The first annotation intention may or may not be included in each predicted intention.
And 204, determining a third training sample set according to the first labeling intention, the confidence degrees corresponding to the prediction intents, the second sample data and the corresponding second labeling intention.
In this embodiment, after obtaining the confidence degrees corresponding to the prediction intents, a third training sample set may be determined by combining the first annotation intention, the second sample data, and the corresponding second annotation intention. Specifically, the prediction intent with the highest confidence may be determined first. And then, taking out second sample data corresponding to a second labeling intention which is the same as the prediction intention in the second training sample set. And modifying the second labeling intention of the taken sample data into the first labeling intention to obtain a third training sample set. For example, the first labeled intention is intention 1, and the confidence degrees corresponding to the prediction intents are: the intent 1 confidence is 0.2, the intent 2 confidence is 0.4, the intent 3 confidence is 0.3, and the intent 4 confidence is 0.3. It may be determined that the confidence of intent 2 is highest in the predicted intent. Then, sample data corresponding to fig. 2 in the second training sample set may be fetched. And modifying the labeling intention of the taken sample data into intention 1. Thus, a new sample data corresponding to fig. 1, i.e. a third training sample set, is obtained. Here, the third set of training samples is an extension of the first set of training samples.
Step 205, training a second intention recognition model according to a third training sample set.
After the third set of training samples is obtained, the second intent recognition model may be trained using the third set of training samples. It is to be understood that each sample data in the third set of training samples corresponds to at least one intention tag. Thus, the result output by the second intent recognition model also includes at least one intent tag.
Referring to FIG. 3, a schematic diagram of an application scenario of the method for training a model of the present application is shown. In the application scenario of fig. 3, the server 301 obtains historical annotation data applied to the medical field, the architectural field, and the financial field from the database 302. Meanwhile, the server 301 acquires a small amount of annotation data applied to a new field (legal field) from the terminal device 303. Through the processing of the steps 202-205, an intention recognition model applied to the legal field is obtained through training. The server 301 may return the above-described intention recognition model to the terminal device 303 to cause the technician to apply the above-described intention recognition model to the human-machine conversation robot.
The method for training the model provided by the embodiment of the application can make full use of historical data used for establishing the man-machine interactive robot, reduce the cost of the newly-built man-machine interactive robot, and improve the intention recognition effect of the newly-built man-machine interactive robot under the condition of a small number of labeled samples.
With continued reference to FIG. 4, a flow 400 of another embodiment of a method for training a model according to the present application is shown. As shown in fig. 4, the method for training a model of the present embodiment may include the following steps:
step 401, a first training sample set and a second training sample set are obtained.
In this embodiment, the first training sample set may include a plurality of subsets, and sample data in each subset corresponds to the same labeling intention. Similarly, the second training sample set may also include a plurality of subsets, and the sample data in each subset corresponds to the same labeling intention.
Step 402, processing a second training sample set according to second sample data corresponding to each second labeling intention.
In this embodiment, the execution subject may first process the sample data in the second training sample set. The above processing may include merging, deleting, modifying, and the like. Specifically, the execution subject may consider the number of second sample data corresponding to each second annotation intention when performing processing. For example, the second annotation intention and the corresponding second sample data, the number of which is less than the preset threshold, are deleted.
In some optional implementations of this embodiment, the executing entity may process the second set of training samples by:
step 4021, selecting a first training sample subset and a second training sample subset from the second training sample set.
In this implementation, the first training sample subset and the second training sample subset may be selected from the second training sample set at first. And the second labeling intentions corresponding to the first training sample subset and the second training sample subset are different, and the quantity of the second sample data in the first training sample subset is greater than that of the second sample data in the second training sample subset. That is, the first subset of training samples and the second subset of training samples are a set of second sample data corresponding to different second annotation intents. And the number of second sample data in the first subset of training samples is larger.
Step 4022, determining the number of common sample data of the first subset of training samples and the second subset of training samples.
After selecting the first subset of training samples and the second subset of training samples, the number of sample data common to the first subset of training samples and the second subset of training samples may be determined. Here, the common sample data is second sample data that exists in both the first training sample subset and the second training sample subset. The execution subject may count the common sample data to determine the number thereof.
And step 4023, determining whether the preset condition is met according to the number of the common sample data and the number of the second sample data in the second training sample subset.
Next, the number of sample data that the execution subject can share, the number of samples in the second subset of training samples, and whether the preset condition is satisfied is determined. The preset conditions may include, but are not limited to: the number of samples in the second subset of training samples is greater than or equal to a certain number threshold, and the number of common sample data is greater than or equal to a certain number threshold.
In some alternative implementations, whether the preset condition is satisfied may be determined by the following steps not shown in fig. 4: determining the ratio of the number of the common sample data to the number of the second sample data in the second training sample subset; in response to determining that the number of second sample data in the second subset of training samples is less than a first preset number threshold and the ratio is greater than a first preset proportion threshold, or that the ratio is greater than a second preset proportion threshold, determining that a preset condition is satisfied.
In this implementation, the executing agent may first determine a ratio of the number of common sample data to the number of second sample data in the second subset of training samples. The preset condition is considered to be satisfied if the number of second sample data in the second subset of training samples is greater than a first preset number threshold (e.g., 50) while the ratio is greater than a first preset proportion threshold. Or, the ratio is larger than a second preset ratio threshold value, and the preset condition is determined to be met.
Step 4024, in response to the preset condition being met, modifying the second labeling intention corresponding to the second training sample subset into the second labeling intention corresponding to the first training sample subset.
If it is determined that the preset condition is satisfied, the execution subject may modify the second labeling intention corresponding to the second subset of training samples into the second labeling intention corresponding to the first subset of training samples. In this way, the reduction and merging of the second sample data in the second training sample set can be realized.
In some optional implementations of this embodiment, the processing may further include the following steps not shown in fig. 4: and deleting the training sample subset of which the number of the second sample data is less than a second preset number threshold.
In this implementation, if the number of second sample data included in a certain subset in the second training sample set is too small, the subset is considered to be unavailable, and the second sample data in the subset and the corresponding second annotation intention can be deleted directly.
Through the processing, more sample data corresponding to each labeling intention in the second training sample set can be obtained, and the method is convenient for subsequent use.
And 403, training the initial intention recognition model by using the processed second training sample set to obtain a first intention recognition model.
The executing subject may train the initial intention recognition model using the processed second training sample set, resulting in a first intention recognition model. In particular, the executing agent may use the second sample data as an input of the initial intent recognition model and use the corresponding second annotation intent as an output of the initial intent recognition model. Here, the initial intention recognition model may be an initialized intention recognition model, that is, parameters in the intention recognition model are initialized values, and may also be a pre-trained intention recognition model.
It should be noted that the first steps 402 to 403 may be executed by an executing entity of the method for training a model of the present application, or may be implemented by other electronic devices. If implemented by other electronic devices, the other electronic devices may send the first intent recognition model to the executing agent.
Step 404, a second intention recognition model to be trained is obtained.
In this embodiment, the second intention recognition model may be an intention recognition model having the same structure as the first intention recognition model but different parameters, or the second intention recognition model may be an intention recognition model having a different structure from the first intention recognition model. The domain to which the second intent recognition model applies may be different from the domain to which the first intent recognition model applies.
Step 405, predicting the first sample data by using the first intention recognition model, and determining confidence degrees corresponding to the prediction intents.
And 406, determining a third training sample set according to the first labeling intention, the confidence degrees corresponding to the prediction intents, the second sample data and the corresponding second labeling intention.
Step 407, train the second intention recognition model according to the third training sample set.
The principle of steps 405 to 407 is similar to that of steps 203 to 205, and is not described herein again.
According to the method for training the model provided by the embodiment of the application, the intention recognition model can be trained by using the historical annotation data, and then a new intention recognition model is trained based on the intention recognition model, so that the utilization rate of the historical annotation data is improved.
With continued reference to FIG. 5, a flow 500 of another embodiment of a method for training a model according to the present application is shown. As shown in fig. 5, the method for training a model of the present embodiment may include the following steps:
step 501, a first training sample set and a second training sample set are obtained.
Step 502, obtaining a trained first intention recognition model and a second intention recognition model to be trained.
Step 503, the first sample data is predicted by using the first intention recognition model, and the confidence degree corresponding to each prediction intention is determined.
Step 504, according to the first labeling intention and the confidence degrees corresponding to the prediction intents, a second labeling intention similar to the first labeling intention is determined from the second training sample set.
In this embodiment, the execution subject may determine, from the second training sample set, a second annotation intention that is similar to the first annotation intention according to the first annotation intention and the confidence degrees corresponding to the prediction intents. Specifically, the executing subject may search for a second prediction intention with a confidence degree greater than a preset threshold value in the training sample set, and use the searched second annotation intention as a second annotation intention similar to the first annotation intention.
In some alternative implementations of the embodiment, the performing agent may determine the second annotation intent that is similar to the first annotation intent by: for each first annotation intention, sorting the prediction intents in a high-to-low order according to the confidence degrees of the prediction intents of the first sample data corresponding to the first annotation intention; and regarding a second labeling intention related to the first preset number of prediction intents in the sequence in the second training sample set as a second labeling intention similar to the first labeling intention.
In this implementation, for each first annotation intention, the executing subject may first determine first sample data corresponding to the first annotation intention, and then determine a confidence of each prediction intention of each first sample data. And sequencing the confidence degrees according to the sequence from high to low. Then take the pre-set number (N) of predicted intents. And determining second labeling intents related to the first N predicted intents, and finally, regarding the second labeling intents as second labeling intents similar to the first labeling intents. Here, correlation means the same as or similar to the first N prediction intents.
And 505, selecting partial second sample data from the second sample data corresponding to the determined second labeling intention.
After determining the second labeling intents that are close to each other, the execution subject may select a part of second sample data from the second sample data corresponding to each second labeling intention. In the selection, the selection may be performed randomly, or multiple rounds of dialog pattern data may be selected preferentially. The number of the second sample data selected from the second sample data corresponding to each second labeling intention may be the same or different.
In some optional implementations of this embodiment, the execution subject may select part of the second sample data by: and for each determined second labeling intention, selecting second sample data from second sample data corresponding to the second labeling intention according to the ratio of the confidence degree corresponding to the prediction intention related to the second labeling intention to the sum of the confidence degrees corresponding to the prediction intents.
In this implementation, the executing agent may first calculate the sum of the confidence levels of the prediction intents, denoted as P. The ratio of the confidence of each prediction intent to the above sum is then calculated. Then, the execution subject may calculate a sum of the second sample data corresponding to each second annotation intention, which is denoted as M. For each determined second annotation intention, the executive body can select sample data according to the ratio corresponding to the prediction intention related to the second annotation intention. Specifically, the number of the second sample data is selected from the second sample data corresponding to the second labeling intention to be equal to P × M/P. Wherein p is the confidence of the prediction intent associated with the second annotation intent.
Step 506, determining a third training sample set according to the selected second sample data and the first training sample set.
After the second sample data is selected, the execution subject may determine a third training sample set in combination with the second sample data and the first training sample set. Specifically, the execution subject may use the second annotation intention and the first training sample set corresponding to the selected second sample data as a third training sample set.
In some optional implementations of this embodiment, the performing agent may determine the third set of training samples by: modifying the second labeling intention of the selected second sample data into a similar first labeling intention; and determining a third training sample set according to the marking intention corresponding to the modified second sample data and the first training sample set.
In this implementation manner, the execution subject may modify the second annotation intention of the selected second sample data into the first annotation intention similar to the second annotation intention. In this way, the number of sample data corresponding to the first labeling intention is increased, and a third training sample set is obtained by combining the first sample data corresponding to each first labeling intention in the first training sample set. It is to be understood that the labeling intents included in the third training sample set are the same as the labeling intents in the first training sample set, and thus can be used for training of the second intention recognition model.
Step 507, training a second intention recognition model according to the third training sample set.
The method for training the model provided by the embodiment of the application can expand a small amount of labeled data in a new field by using historical labeled data, so that the utilization rate of the historical labeled data is improved.
With further reference to fig. 6, as an implementation of the method shown in the above figures, the present application provides an embodiment of an apparatus for training a model, which corresponds to the embodiment of the method shown in fig. 2, and which can be applied in various electronic devices.
As shown in fig. 6, the training model apparatus 600 of the present embodiment includes: a first acquisition unit 601, a second acquisition unit 602, a confidence determination unit 603, a sample generation unit 604, and a first model training unit 605.
A first obtaining unit 601 configured to obtain a first training sample set and a second training sample set. The first training sample set comprises first sample data and corresponding first labeling intents, and the second training sample set comprises second sample data and corresponding second labeling intents.
A second obtaining unit 602 configured to obtain the trained first intention recognition model and a second intention recognition model to be trained.
A confidence determining unit 603 configured to predict the first sample data using the first intention recognition model, and determine a confidence corresponding to each prediction intention;
the sample generating unit 604 is configured to determine a third training sample set according to the first annotation intention, the confidence degree corresponding to each prediction intention, and the second sample data and the corresponding second annotation intention.
A first model training unit 605 configured to train a second intent recognition model based on the third set of training samples.
In some optional implementations of this embodiment, the apparatus 600 may further include, not shown in fig. 6: a sample processing unit and a second model training unit.
And the sample processing unit is configured to process the second training sample set according to the second sample data corresponding to each second labeling intention.
And the second model training unit is configured to train the initial intention recognition model by using the processed second training sample set to obtain the first intention recognition model.
In some optional implementations of this embodiment, the sample processing unit may further include one or more of the following components not shown in fig. 6: the system comprises a subset determining module, a common sample determining module, a judging module and an intention modifying module.
The subset determining module is configured to select a first training sample subset and a second training sample subset from a second training sample set, wherein second labeling intents corresponding to the first training sample subset and the second training sample subset are different, and the quantity of second sample data in the first training sample subset is greater than that of the second sample data in the second training sample subset;
a common sample determination module configured to determine a number of common sample data of the first subset of training samples and the second subset of training samples;
the judging module is configured to determine whether a preset condition is met according to the number of the common sample data and the number of second sample data in the second training sample subset;
and the intention modification module is configured to modify a second labeling intention corresponding to the second training sample subset into a second labeling intention corresponding to the first training sample subset in response to the preset condition being met.
In some optional implementations of this embodiment, the determining module is further configured to: determining the ratio of the number of the common sample data to the number of the second sample data in the second training sample subset; in response to determining that the number of second sample data in the second subset of training samples is less than a first preset number threshold and the ratio is greater than a first preset proportion threshold, or that the ratio is greater than a second preset proportion threshold, determining that a preset condition is satisfied.
In some optional implementations of this embodiment, the sample processing unit may further include a deletion module, not shown in fig. 6, configured to delete the subset of training samples whose number of second sample data is smaller than the second preset number threshold.
In some optional implementations of this embodiment, the sample generating unit may further include one or more elements not shown in fig. 6: the device comprises a similar intention determining module, a sample data selecting module and a sample generating module.
And the similarity intention determining module is configured to determine a second annotation intention which is similar to the first annotation intention from the second training sample set according to the first annotation intention and the confidence degrees corresponding to the prediction intentions.
And the sample data selecting module is configured to select partial second sample data from the second sample data corresponding to the determined second labeling intention.
And the sample generation module is configured to determine a third training sample set according to the selected second sample data and the first training sample set.
In some optional implementations of this embodiment, the similar intent determination module is further configured to: for each first annotation intention, sorting the prediction intents in a high-to-low order according to the confidence degrees of the prediction intents of the first sample data corresponding to the first annotation intention; and regarding a second labeling intention related to the first preset number of prediction intents in the sequence in the second training sample set as a second labeling intention similar to the first labeling intention.
In some optional implementations of this embodiment, the sample data selecting module is further configured to: and for each determined second labeling intention, selecting second sample data from second sample data corresponding to the second labeling intention according to the ratio of the confidence degree corresponding to the prediction intention related to the second labeling intention to the sum of the confidence degrees corresponding to the prediction intents.
In some optional implementations of this embodiment, the sample generation module is further configured to: modifying the second labeling intention of the selected second sample data into a similar first labeling intention; and determining a third training sample set according to the marking intention corresponding to the modified second sample data and the first training sample set.
It should be understood that units 601 to 605 recited in the apparatus 600 for training a model correspond to respective steps in the method described with reference to fig. 2. Thus, the operations and features described above for the method for training a model are equally applicable to the apparatus 600 and the units contained therein and will not be described in detail here.
According to an embodiment of the present application, an electronic device and a readable storage medium are also provided.
As shown in fig. 7, a block diagram of an electronic device performing a method for training a model according to an embodiment of the present application is shown. Electronic devices are intended to represent various forms of digital computers, such as laptops, desktops, workstations, personal digital assistants, servers, blade servers, mainframes, and other appropriate computers. The electronic device may also represent various forms of mobile devices, such as personal digital processing, cellular phones, smart phones, wearable devices, and other similar computing devices. The components shown herein, their connections and relationships, and their functions, are meant to be examples only, and are not meant to limit implementations of the present application that are described and/or claimed herein.
As shown in fig. 7, the electronic apparatus includes: one or more processors 701, a memory 702, and interfaces for connecting the various components, including a high-speed interface and a low-speed interface. The various components are interconnected using different buses and may be mounted on a common motherboard or in other manners as desired. The processor may process instructions for execution within the electronic device, including instructions stored in or on the memory to display graphical information of a GUI on an external input/output apparatus (such as a display device coupled to the interface). In other embodiments, multiple processors and/or multiple buses may be used, along with multiple memories and multiple memories, as desired. Also, multiple electronic devices may be connected, with each device providing portions of the necessary operations (e.g., as a server array, a group of blade servers, or a multi-processor system). In fig. 7, one processor 701 is taken as an example.
The memory 702 is a non-transitory computer readable storage medium as provided herein. Wherein the memory stores instructions executable by at least one processor to cause the at least one processor to perform the methods provided herein for training a model. A non-transitory computer readable storage medium of the present application stores computer instructions for causing a computer to perform the methods provided herein for training a model.
The memory 702, as a non-transitory computer readable storage medium, may be used to store non-transitory software programs, non-transitory computer executable programs, and modules, such as program instructions/modules corresponding to the method for training a model (e.g., the first acquisition unit 601, the second acquisition unit 602, the confidence determination unit 603, the sample generation unit 604, and the first model training unit 605 shown in fig. 6) in the embodiments of the present application. The processor 701 executes various functional applications of the server and data processing by executing non-transitory software programs, instructions, and modules stored in the memory 702, that is, implements the method for training a model in the above method embodiment.
The memory 702 may include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program required for at least one function; the storage data area may store data created from use of an electronic device that performs training of the model, and the like. Further, the memory 702 may include high speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, memory 702 may optionally include memory located remotely from processor 701, which may be connected via a network to an electronic device executing software for training models. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The electronic device performing the method for training a model may further include: an input device 703 and an output device 704. The processor 701, the memory 702, the input device 703 and the output device 704 may be connected by a bus or other means, and fig. 7 illustrates an example of a connection by a bus.
The input device 703 may receive input numeric or character information and generate key signal inputs related to performing user settings and function control of the electronic apparatus used to train the model, such as a touch screen, a keypad, a mouse, a track pad, a touch pad, a pointer stick, one or more mouse buttons, a track ball, a joystick, or other input devices. The output devices 704 may include a display device, auxiliary lighting devices (e.g., LEDs), and tactile feedback devices (e.g., vibrating motors), among others. The display device may include, but is not limited to, a Liquid Crystal Display (LCD), a Light Emitting Diode (LED) display, and a plasma display. In some implementations, the display device can be a touch screen.
Various implementations of the systems and techniques described here can be realized in digital electronic circuitry, integrated circuitry, application specific ASICs (application specific integrated circuits), computer hardware, firmware, software, and/or combinations thereof. These various embodiments may include: implemented in one or more computer programs that are executable and/or interpretable on a programmable system including at least one programmable processor, which may be special or general purpose, receiving data and instructions from, and transmitting data and instructions to, a storage system, at least one input device, and at least one output device.
These computer programs (also known as programs, software applications, or code) include machine instructions for a programmable processor, and may be implemented using high-level procedural and/or object-oriented programming languages, and/or assembly/machine languages. As used herein, the terms "machine-readable medium" and "computer-readable medium" refer to any computer program product, apparatus, and/or device (e.g., magnetic discs, optical disks, memory, Programmable Logic Devices (PLDs)) used to provide machine instructions and/or data to a programmable processor, including a machine-readable medium that receives machine instructions as a machine-readable signal. The term "machine-readable signal" refers to any signal used to provide machine instructions and/or data to a programmable processor.
To provide for interaction with a user, the systems and techniques described here can be implemented on a computer having: a display device (e.g., a CRT (cathode ray tube) or LCD (liquid crystal display) monitor) for displaying information to a user; and a keyboard and a pointing device (e.g., a mouse or a trackball) by which a user can provide input to the computer. Other kinds of devices may also be used to provide for interaction with a user; for example, feedback provided to the user can be any form of sensory feedback (e.g., visual feedback, auditory feedback, or tactile feedback); and input from the user may be received in any form, including acoustic, speech, or tactile input.
The systems and techniques described here can be implemented in a computing system that includes a back-end component (e.g., as a data server), or that includes a middleware component (e.g., an application server), or that includes a front-end component (e.g., a user computer having a graphical user interface or a web browser through which a user can interact with an implementation of the systems and techniques described here), or any combination of such back-end, middleware, or front-end components. The components of the system can be interconnected by any form or medium of digital data communication (e.g., a communication network). Examples of communication networks include: local Area Networks (LANs), Wide Area Networks (WANs), and the Internet.
The computer system may include clients and servers. A client and server are generally remote from each other and typically interact through a communication network. The relationship of client and server arises by virtue of computer programs running on the respective computers and having a client-server relationship to each other.
According to the technical scheme of the embodiment of the application, historical data used for establishing the man-machine conversation robot can be fully utilized, the cost of the newly-built man-machine conversation robot is reduced, and the intention recognition effect of the newly-built man-machine conversation robot is improved under the condition of a small number of labeled samples.
It should be understood that various forms of the flows shown above may be used, with steps reordered, added, or deleted. For example, the steps described in the present application may be executed in parallel, sequentially, or in different orders, and the present invention is not limited thereto as long as the desired results of the technical solutions disclosed in the present application can be achieved.
The above-described embodiments should not be construed as limiting the scope of the present application. It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and substitutions may be made in accordance with design requirements and other factors. Any modification, equivalent replacement, and improvement made within the spirit and principle of the present application shall be included in the protection scope of the present application.

Claims (20)

1. A method for training a model, comprising:
acquiring a first training sample set and a second training sample set, wherein the first training sample set comprises first sample data and a corresponding first labeling intention, and the second training sample set comprises second sample data and a corresponding second labeling intention;
acquiring a trained first intention recognition model and a second intention recognition model to be trained;
predicting the first sample data by using the first intention recognition model, and determining confidence degrees corresponding to all prediction intents;
determining a third training sample set according to the first labeling intention, the confidence degrees corresponding to the prediction intents, second sample data and the corresponding second labeling intention;
training the second intent recognition model according to the third set of training samples.
2. The method of claim 1, wherein the method further comprises:
processing the second training sample set according to second sample data corresponding to each second labeling intention;
and training an initial intention recognition model by utilizing the processed second training sample set to obtain the first intention recognition model.
3. The method according to claim 2, wherein the processing the second training sample set according to the second sample data corresponding to each second labeling intention includes:
selecting a first training sample subset and a second training sample subset from the second training sample set, wherein second labeling intents corresponding to the first training sample subset and the second training sample subset are different, and the quantity of second sample data in the first training sample subset is greater than that of second sample data in the second training sample subset;
determining a number of common sample data of the first subset of training samples and the second subset of training samples;
determining whether a preset condition is met according to the number of the common sample data and the number of second sample data in the second training sample subset;
and in response to the preset condition being met, modifying the second labeling intention corresponding to the second training sample subset into the second labeling intention corresponding to the first training sample subset.
4. The method of claim 3, wherein said determining whether a preset condition is met according to the number of common sample data, the number of samples in the second subset of training samples, comprises:
determining a ratio of the number of the common sample data to the number of second sample data in the second subset of training samples;
in response to determining that the number of second sample data in the second subset of training samples is less than a first preset number threshold and the ratio is greater than a first preset proportion threshold, or that the ratio is greater than a second preset proportion threshold, determining that the preset condition is met.
5. The method according to claim 2, wherein the processing the second training sample set according to the second sample data corresponding to each second labeling intention includes:
and deleting the training sample subset of which the number of the second sample data is less than a second preset number threshold.
6. The method of claim 1, wherein determining a third set of training samples according to the first annotation intention, the confidence degree corresponding to each prediction intention, and the second sample data and the corresponding second annotation intention comprises:
determining a second labeling intention similar to the first labeling intention from the second training sample set according to the first labeling intention and the confidence degrees corresponding to the prediction intents;
selecting part of second sample data from the second sample data corresponding to the determined second labeling intention;
and determining the third training sample set according to the selected second sample data and the first training sample set.
7. The method of claim 6, wherein the determining, from the second training sample set, a second annotation intent that is similar to the annotation intent according to the first annotation intent and the confidence level corresponding to each predicted intent comprises:
for each first labeling intention, sorting the prediction intents in a high-to-low order according to the confidence degrees of the prediction intents of the first sample data corresponding to the first labeling intention;
and taking the second labeling intention related to the first preset number of prediction intents in the sequence in the second training sample set as a second labeling intention similar to the first labeling intention.
8. The method of claim 6, wherein said selecting a portion of second sample data from the second sample data corresponding to the determined second annotation intent comprises:
and for each determined second labeling intention, selecting second sample data from second sample data corresponding to the second labeling intention according to the ratio of the confidence degree corresponding to the prediction intention related to the second labeling intention to the sum of the confidence degrees corresponding to the prediction intents.
9. The method according to claim 6, wherein said determining the third set of training samples from the selected second sample data and the first set of training samples comprises:
modifying the second labeling intention of the selected second sample data into a similar first labeling intention;
and determining the third training sample set according to the marking intention corresponding to the modified second sample data and the first training sample set.
10. An apparatus for training a model, comprising:
a first obtaining unit configured to obtain a first training sample set and a second training sample set, wherein the first training sample set comprises first sample data and a corresponding first labeling intention, and the second training sample set comprises second sample data and a corresponding second labeling intention;
a second acquisition unit configured to acquire the trained first intention recognition model and a second intention recognition model to be trained;
a confidence degree determining unit configured to predict the first sample data by using the first intention recognition model, and determine a confidence degree corresponding to each prediction intention;
the sample generation unit is configured to determine a third training sample set according to the first annotation intention, the confidence degrees corresponding to the prediction intents, second sample data and corresponding second annotation intention;
a first model training unit configured to train the second intent recognition model according to the third set of training samples.
11. The apparatus of claim 10, wherein the apparatus further comprises:
a sample processing unit configured to process the second training sample set according to second sample data corresponding to each of the second annotation intents;
and the second model training unit is configured to train the initial intention recognition model by using the processed second training sample set to obtain the first intention recognition model.
12. The apparatus of claim 11, wherein the sample processing unit comprises:
a subset determining module configured to select a first training sample subset and a second training sample subset from the second training sample set, where the first training sample subset and the second training sample subset correspond to different second labeling intents, and a quantity of second sample data in the first training sample subset is greater than a quantity of second sample data in the second training sample subset;
a common sample determination module configured to determine a number of common sample data of the first subset of training samples and the second subset of training samples;
a judging module configured to determine whether a preset condition is met according to the number of the common sample data and the number of second sample data in the second training sample subset;
and the intention modification module is configured to modify a second labeling intention corresponding to the second training sample subset into a second labeling intention corresponding to the first training sample subset in response to the preset condition being met.
13. The apparatus of claim 12, wherein the determining module is further configured to:
determining a ratio of the number of the common sample data to the number of second sample data in the second subset of training samples;
in response to determining that the number of second sample data in the second subset of training samples is less than a first preset number threshold and the ratio is greater than a first preset proportion threshold, or that the ratio is greater than a second preset proportion threshold, determining that the preset condition is met.
14. The apparatus of claim 12, wherein the sample processing unit comprises:
a deletion module configured to delete the subset of training samples for which the number of the second sample data is less than a second preset number threshold.
15. The apparatus of claim 10, wherein the sample generation unit comprises:
a similarity intention determining module configured to determine a second labeling intention similar to the first labeling intention from the second training sample set according to the first labeling intention and the confidence degrees corresponding to the prediction intentions;
the sample data selecting module is configured to select partial second sample data from second sample data corresponding to the determined second labeling intention;
a sample generation module configured to determine the third training sample set according to the selected second sample data and the first training sample set.
16. The apparatus of claim 15, wherein the similar intent determination module is further configured to:
for each first labeling intention, sorting the prediction intents in a high-to-low order according to the confidence degrees of the prediction intents of the first sample data corresponding to the first labeling intention;
and taking the second labeling intention related to the first preset number of prediction intents in the sequence in the second training sample set as a second labeling intention similar to the first labeling intention.
17. The apparatus of claim 15, wherein the sample data selection module is further configured to:
and for each determined second labeling intention, selecting second sample data from second sample data corresponding to the second labeling intention according to the ratio of the confidence degree corresponding to the prediction intention related to the second labeling intention to the sum of the confidence degrees corresponding to the prediction intents.
18. The apparatus of claim 15, wherein the sample generation module is further configured to:
modifying the second labeling intention of the selected second sample data into a similar first labeling intention;
and determining the third training sample set according to the marking intention corresponding to the modified second sample data and the first training sample set.
19. An electronic device, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1-9.
20. A non-transitory computer readable storage medium having stored thereon computer instructions for causing the computer to perform the method of any one of claims 1-9.
CN202010524463.3A 2020-06-10 2020-06-10 Method, apparatus, device and storage medium for training model Active CN111680517B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010524463.3A CN111680517B (en) 2020-06-10 2020-06-10 Method, apparatus, device and storage medium for training model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010524463.3A CN111680517B (en) 2020-06-10 2020-06-10 Method, apparatus, device and storage medium for training model

Publications (2)

Publication Number Publication Date
CN111680517A true CN111680517A (en) 2020-09-18
CN111680517B CN111680517B (en) 2023-05-16

Family

ID=72454836

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010524463.3A Active CN111680517B (en) 2020-06-10 2020-06-10 Method, apparatus, device and storage medium for training model

Country Status (1)

Country Link
CN (1) CN111680517B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112257458A (en) * 2020-10-21 2021-01-22 阳光保险集团股份有限公司 Intention recognition model training method, intention recognition method, device and equipment
CN112529624A (en) * 2020-12-15 2021-03-19 北京百度网讯科技有限公司 Method, device, equipment and storage medium for generating business prediction model
CN112541739A (en) * 2020-12-15 2021-03-23 平安科技(深圳)有限公司 Method, device, equipment and medium for testing question-answer intention classification model
CN112560992A (en) * 2020-12-25 2021-03-26 北京百度网讯科技有限公司 Method and device for optimizing image classification model, electronic equipment and storage medium
CN112669816A (en) * 2020-12-24 2021-04-16 北京有竹居网络技术有限公司 Model training method, speech recognition method, device, medium and equipment
CN112966108A (en) * 2021-03-08 2021-06-15 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for detecting data and training classification model
CN113569546A (en) * 2021-06-16 2021-10-29 上海淇玥信息技术有限公司 Intention labeling method and device and electronic equipment

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108920622A (en) * 2018-06-29 2018-11-30 北京奇艺世纪科技有限公司 A kind of training method of intention assessment, training device and identification device
CN109063221A (en) * 2018-11-02 2018-12-21 北京百度网讯科技有限公司 Query intention recognition methods and device based on mixed strategy
CN109086814A (en) * 2018-07-23 2018-12-25 腾讯科技(深圳)有限公司 A kind of data processing method, device and the network equipment
CN109522556A (en) * 2018-11-16 2019-03-26 北京九狐时代智能科技有限公司 A kind of intension recognizing method and device
CN110457449A (en) * 2019-07-05 2019-11-15 阿里巴巴集团控股有限公司 Method, apparatus, equipment and the storage medium of on-line training model
US20200065384A1 (en) * 2018-08-26 2020-02-27 CloudMinds Technology, Inc. Method and System for Intent Classification
CN110990576A (en) * 2019-12-24 2020-04-10 用友网络科技股份有限公司 Intention classification method based on active learning, computer device and storage medium
CN111143561A (en) * 2019-12-26 2020-05-12 北京百度网讯科技有限公司 Intention recognition model training method and device and electronic equipment

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108920622A (en) * 2018-06-29 2018-11-30 北京奇艺世纪科技有限公司 A kind of training method of intention assessment, training device and identification device
CN109086814A (en) * 2018-07-23 2018-12-25 腾讯科技(深圳)有限公司 A kind of data processing method, device and the network equipment
US20200065384A1 (en) * 2018-08-26 2020-02-27 CloudMinds Technology, Inc. Method and System for Intent Classification
CN109063221A (en) * 2018-11-02 2018-12-21 北京百度网讯科技有限公司 Query intention recognition methods and device based on mixed strategy
CN109522556A (en) * 2018-11-16 2019-03-26 北京九狐时代智能科技有限公司 A kind of intension recognizing method and device
CN110457449A (en) * 2019-07-05 2019-11-15 阿里巴巴集团控股有限公司 Method, apparatus, equipment and the storage medium of on-line training model
CN110990576A (en) * 2019-12-24 2020-04-10 用友网络科技股份有限公司 Intention classification method based on active learning, computer device and storage medium
CN111143561A (en) * 2019-12-26 2020-05-12 北京百度网讯科技有限公司 Intention recognition model training method and device and electronic equipment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MATEUSZ BUDNIK等: ""Rethinking deep active learning:Using unlabeled data at model training"" *
付博;陈毅恒;邵艳秋;刘挺;: ""基于用户自然标注的微博文本的消费意图识别"" *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112257458A (en) * 2020-10-21 2021-01-22 阳光保险集团股份有限公司 Intention recognition model training method, intention recognition method, device and equipment
CN112529624A (en) * 2020-12-15 2021-03-19 北京百度网讯科技有限公司 Method, device, equipment and storage medium for generating business prediction model
CN112541739A (en) * 2020-12-15 2021-03-23 平安科技(深圳)有限公司 Method, device, equipment and medium for testing question-answer intention classification model
WO2022126987A1 (en) * 2020-12-15 2022-06-23 平安科技(深圳)有限公司 Test method and apparatus for question-and-answer intention classification model, device and medium
CN112529624B (en) * 2020-12-15 2024-01-09 北京百度网讯科技有限公司 Method, device, equipment and storage medium for generating business prediction model
CN112669816A (en) * 2020-12-24 2021-04-16 北京有竹居网络技术有限公司 Model training method, speech recognition method, device, medium and equipment
CN112669816B (en) * 2020-12-24 2023-06-02 北京有竹居网络技术有限公司 Model training method, voice recognition method, device, medium and equipment
CN112560992A (en) * 2020-12-25 2021-03-26 北京百度网讯科技有限公司 Method and device for optimizing image classification model, electronic equipment and storage medium
CN112560992B (en) * 2020-12-25 2023-09-01 北京百度网讯科技有限公司 Method, device, electronic equipment and storage medium for optimizing picture classification model
CN112966108A (en) * 2021-03-08 2021-06-15 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for detecting data and training classification model
CN112966108B (en) * 2021-03-08 2024-03-12 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for detecting data and training classification model
CN113569546A (en) * 2021-06-16 2021-10-29 上海淇玥信息技术有限公司 Intention labeling method and device and electronic equipment

Also Published As

Publication number Publication date
CN111680517B (en) 2023-05-16

Similar Documents

Publication Publication Date Title
CN111428008B (en) Method, apparatus, device and storage medium for training a model
CN111680517B (en) Method, apparatus, device and storage medium for training model
CN111639710A (en) Image recognition model training method, device, equipment and storage medium
CN112036509A (en) Method and apparatus for training image recognition models
CN112507735B (en) Training method and device of machine translation model and electronic equipment
CN111708922A (en) Model generation method and device for representing heterogeneous graph nodes
CN111104514A (en) Method and device for training document label model
CN111078878B (en) Text processing method, device, equipment and computer readable storage medium
CN111522967A (en) Knowledge graph construction method, device, equipment and storage medium
CN112559870B (en) Multi-model fusion method, device, electronic equipment and storage medium
CN111667056A (en) Method and apparatus for searching model structure
CN111460135B (en) Method and device for generating text abstract
CN111339759A (en) Method and device for training field element recognition model and electronic equipment
CN111582477A (en) Training method and device of neural network model
CN111241838B (en) Semantic relation processing method, device and equipment for text entity
CN110717340A (en) Recommendation method and device, electronic equipment and storage medium
CN111522944A (en) Method, apparatus, device and storage medium for outputting information
CN111783013A (en) Comment information publishing method, device, equipment and computer-readable storage medium
CN111311309A (en) User satisfaction determining method, device, equipment and medium
CN111666771A (en) Semantic label extraction device, electronic equipment and readable storage medium of document
CN112232089B (en) Pre-training method, device and storage medium of semantic representation model
CN114547244A (en) Method and apparatus for determining information
CN112580723B (en) Multi-model fusion method, device, electronic equipment and storage medium
CN112270169B (en) Method and device for predicting dialogue roles, electronic equipment and storage medium
CN111325006B (en) Information interaction method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant