CN108920622B - Training method, training device and recognition device for intention recognition - Google Patents

Training method, training device and recognition device for intention recognition Download PDF

Info

Publication number
CN108920622B
CN108920622B CN201810694995.4A CN201810694995A CN108920622B CN 108920622 B CN108920622 B CN 108920622B CN 201810694995 A CN201810694995 A CN 201810694995A CN 108920622 B CN108920622 B CN 108920622B
Authority
CN
China
Prior art keywords
model
training
recognition
vector
loss function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810694995.4A
Other languages
Chinese (zh)
Other versions
CN108920622A (en
Inventor
符文君
吴友政
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing QIYI Century Science and Technology Co Ltd
Original Assignee
Beijing QIYI Century Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing QIYI Century Science and Technology Co Ltd filed Critical Beijing QIYI Century Science and Technology Co Ltd
Priority to CN201810694995.4A priority Critical patent/CN108920622B/en
Publication of CN108920622A publication Critical patent/CN108920622A/en
Application granted granted Critical
Publication of CN108920622B publication Critical patent/CN108920622B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Machine Translation (AREA)

Abstract

A training method, a device and a recognition device for intention recognition are provided, the method comprises the following steps: obtaining corpus text vectors corresponding to corpora in the corpus text; constructing a joint loss function formula of a training model; acquiring training data; segmenting the training data, mapping the segmented training data to corresponding corpus text vectors, and recording the corpus text vectors as training vectors; predicting the training sample vector by adopting a training model, and calculating the loss function value of each training model according to the prediction result; and calculating a joint loss function value based on the loss function values of the models, judging whether the joint loss function value is smaller than a set threshold value, if so, finishing the training, otherwise, updating the parameters of the models, and continuing the iterative training. Therefore, the generalization capability of intention identification can be improved as much as possible, and the problems of semantic ambiguity and fault tolerance are solved.

Description

Training method, training device and recognition device for intention recognition
Technical Field
The invention relates to the technical field of intelligence, in particular to an intention recognition training method, an intention recognition training device and an intention recognition device based on multi-task learning.
Background
"intention recognition" refers to determining the category of intention to which a user inputs a piece of information for expressing the requirements of a query. The current intention recognition technology is mainly applied to search engines, man-machine conversation systems and the like, and can be particularly divided into methods based on templates/word banks and methods based on supervised classification, wherein the method based on the templates/word banks is used for mining specific intention templates/word banks from historical input of users, and if the input of the users is matched with words in templates/word banks of corresponding categories, the input is considered to belong to the intention category; the method based on supervised classification builds an intention classification model based on historical input data and predicts the input intention category of the user. The applicant has found through research that the defects of the prior art are mainly in the following aspects:
firstly, generalization capability, a template-based method is limited by the problems of coverage rate of a template and a word bank, and a supervised classification-based method is limited by the problems of data scale and data quality of training corpora.
Secondly, ambiguity problem, semantic missing and fault tolerance problem, the short text often has the problems of incompleteness, semantic missing, input error and the like, for example, the user inputs ' playing and bear ' at night ', and actually wants to search ' the midnight palace of the teddy bear '.
In addition, some prior art attempts to classify based on a method of multitask learning. However, this method also has two problems: firstly, the method obtains a text vector based on training of other tasks, then the text vector is spliced with the text vector of a target task, and then a classifier of the task is trained, wherein errors of other tasks may bring negative effects to the task; secondly, if other tasks are irrelevant to the current task, a large amount of irrelevant external information is introduced, and the classification result may be disturbed.
Therefore, how to solve the problems of semantic missing, fault tolerance, ambiguity and generalization in the intent recognition becomes one of the technical problems to be solved urgently by those skilled in the art.
Disclosure of Invention
In view of this, embodiments of the present invention provide a training method, a training device, and a recognition device for intent recognition based on multi-task learning, so as to solve the problems of model generalization ability, ambiguity, and semantic fault tolerance during intent recognition.
In order to achieve the above purpose, the embodiments of the present invention provide the following technical solutions:
a training method of intent recognition, comprising:
mapping the corpora in the corpus text to a semantic space to obtain a low-dimensional dense vector corresponding to the corpora, and recording the low-dimensional dense vector as a corpus text vector;
constructing a joint loss function formula of the training model, wherein the training model comprises an intention recognition model, a similar short text generation model and an entity recognition model;
acquiring training data for the training model;
segmenting the training data, mapping the processed training data to corresponding corpus text vectors, and recording the mapped training data as training vectors;
inputting the training vector into the training model, and calculating loss function values of the intention recognition model, the similar short text generation model and the entity recognition model based on a prediction result output by the training model and a real result corresponding to the training vector;
substituting the loss function values of the models into the joint loss function formula to obtain joint loss function values;
and judging whether the joint loss function value is smaller than a set value, if not, adjusting training parameters of the intention recognition model, the similar short text generation model and the entity recognition model to reduce the loss function value of the models, continuing iterative training, and if so, finishing the training.
Preferably, in the training method for intention recognition, the constructing a joint loss function formula of the training model includes:
constructing a joint loss function formula of an intention recognition model, a similar short text generation model and an entity recognition model: loss _ total ═ α lossintent_recognition+β*losssim_query_generation+γ*lossentity_recognitionWherein α, β and γ are predetermined loss weighting factors, and the lossintent_recognitionLoss function for the intent recognition model, the lossentity_recognitionLoss function for entity recognition model, losssim_query_generationA loss function of the model is generated for similar text.
Preferably, in the above training method for intention recognition, the mapping corpora in the corpus text to a semantic space to obtain a low-dimensional dense vector corresponding to the corpora, includes:
segmenting the corresponding granularity of a word level, a word level or a pinyin level of the corpus in the corpus database;
training the segmented text based on a neural network model, and expressing the segmented text as a low-dimensional dense vector, wherein the low-dimensional dense vector comprises a word vector, a word vector or a pinyin vector.
Preferably, in the above training method for intention recognition, the training method for intention recognition, in which an intention recognition model is obtained by modeling an intention recognition task, includes:
and modeling aiming at the intention identification task by adopting an LSTM model to obtain an intention identification model, wherein the input of the intention identification model is query, and the output of the intention identification model is an intention category label.
Preferably, in the training method for intention recognition, modeling is performed on the similar short text to obtain a similar short text generation model, and the method includes:
and establishing a similar short text generation model aiming at the similar short text by adopting a Seq2Seq model, wherein the similar short text generation model is input by a user to input a query and output as the similar short text.
Preferably, in the training method for intention recognition, modeling an entity recognition task to obtain an entity recognition task model includes:
constructing a multi-classification model based on a convolutional neural network, and training an entity recognition task based on training data by using the multi-classification model as an entity recognition task model, wherein the training data is as follows: context text containing entities contained in the text to be recognized, the model input is short text containing entities, and the output is entity type tags.
An intent recognition training apparatus comprising:
the corpus vector training unit is used for mapping the corpus in the corpus text to a semantic space to obtain a low-dimensional dense vector corresponding to the corpus and marking the low-dimensional dense vector as a corpus text vector;
the model storage unit is used for storing an intention recognition model, a similar short text generation model and an entity recognition model;
the combined loss function formula storage unit is used for storing a combined loss function formula of a training model, and the training model comprises an intention recognition model, a similar short text generation model and an entity recognition model;
a training data acquisition unit for acquiring training data for the training model;
the training vector acquisition unit is used for segmenting the training data, mapping the processed training data to corresponding corpus text vectors and recording the mapped corpus text vectors as training vectors;
a loss function value calculation unit, configured to input the training vector to the training model, calculate a loss function value of the intention recognition model, the similar short text generation model, and the entity recognition model based on a prediction result output by the training model and a real result corresponding to the training vector, and bring the loss function value of each model into the joint loss function formula to obtain a joint loss function value;
and the parameter adjusting unit is used for judging whether the joint loss function value is smaller than a set value, if not, adjusting the training parameters of the intention recognition model, the similar short text generation model and the entity recognition model so as to reduce the loss function value of the models, continuing iterative training, and if so, finishing the training.
Preferably, in the training apparatus for intention recognition, the joint loss function formula is:
loss_total=α*lossintent_recognition+β*losssim_query_generation+γ*lossentity_recognitionwherein α, β and γ are predetermined loss weighting factors, and the lossintent_recognitionLoss function for the intent recognition model, the lossentity_recognitionLoss function for entity recognition model, losssim_query_generationA loss function of the model is generated for similar text.
Preferably, in the training apparatus for intention recognition, the corpus vector training unit is specifically configured to:
segmenting the corresponding granularity of a word level, a word level or a pinyin level of the corpus in the corpus database;
training the segmented text based on a neural network model, and expressing the segmented text as a low-dimensional dense vector, wherein the low-dimensional dense vector comprises a word vector, a word vector or a pinyin vector.
Preferably, in the intention recognition training device, the intention recognition model stored in the model storage means is:
modeling aiming at an intention recognition task by adopting an LSTM model to obtain an intention recognition model, wherein the input of the intention recognition model is a query, and the output of the intention recognition model is an intention category label;
the similar short text generation model stored in the model storage unit is as follows:
establishing a similar short text generation model aiming at the similar short text by adopting a Seq2Seq model, wherein the similar short text generation model is input by a user to input a query and output as the similar short text;
the entity recognition task model stored in the model storage unit is as follows:
the multi-classification model is constructed based on a convolutional neural network, and training data of the multi-classification model are as follows: context text containing entities contained in the text to be recognized, the model input is short text containing entities, and the output is entity type tags.
An intent recognition device comprising a memory and a processor;
the memory is stored with an intention recognition model, a similar short text generation model and an entity recognition model obtained by training with any intention recognition training method, and the processor is used for calling and executing the intention recognition model, the similar short text generation model and the entity recognition model when a query of a user is obtained.
Based on the technical scheme, the scheme provided by the embodiment of the invention can process the input text by acquiring the intention recognition model, the similar short text generation model and the entity recognition model based on a multi-task learning mode, can more effectively learn related language knowledge, improves the generalization capability of the intention recognition model, and can solve the problem of semantic ambiguity existing in intention recognition as far as possible.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the provided drawings without creative efforts.
Fig. 1 is a schematic flowchart of a training method for intention recognition disclosed in an embodiment of the present application;
FIG. 2 is an example of building an intent recognition model based on an LSTM model;
FIG. 3 is a schematic diagram of an intention recognition training apparatus provided in the present application;
fig. 4 is a schematic structural diagram of an intention identifying apparatus disclosed in an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Aiming at the problem that ambiguity often occurs when the input text is subjected to intent recognition in the prior art, the application discloses a training method for intent recognition, which can include the following steps:
step S101: training the linguistic data in the expectation database to obtain a text vector corresponding to the linguistic data;
specifically, in this step, the corpora in the corpus text are mapped to a semantic space, so as to obtain low-dimensional dense vectors corresponding to the corpora, which are recorded as corpus text vectors, the low-dimensional dense vectors may be word vectors, and pinyin vectors, and during training, the adopted training model may adopt models such as word2 vec. The corpus database stores corpus such as historical text input information, click logs and voice recognition result information of a user, the corpus is in a text form, the text input information is input data input by the user through input equipment, the click logs refer to link contents clicked by the user, for example, the user searches, when a link is clicked and opened through a mouse, the clicked contents and corresponding query are stored in the click logs, the voice information is voice contents when the user searches through voice or performs other voice interaction, and the voice information can be converted into matched characters through a voice recognition technology and then is stored in the corpus database as voice corpus. Specifically, the corpus database may capture corresponding corpora from logs of each system in the electronic device through a preset data capture application, for example, the corpora in the corpus database may be obtained from mass short message data such as a search query log, a click log, a question and answer knowledge base title, a microblog, a voice conversation log and the like through the data capture application, and the captured historical input text information, click log information, voice information and the like are stored in the corpus database.
The method specifically comprises the following steps: firstly, a certain text (corpus) in a corpus database is subjected to granularity segmentation of a corresponding character level/word level/pinyin level, then training is carried out based on a neural network model, and the segmented text is expressed as a low-dimensional dense real number vector. For example, the text is "people in china", and in this case, "chinese/country/person/people", "china/people", "zhongguo/renmin" can be obtained by segmenting the text based on the words/pinyin. The word vector corresponding to the text is obtained in the scheme, so that the problem of unknown words is solved, and the pinyin vector corresponding to the text is obtained, so that the problem of error conversion compatible with voice recognition is solved. Certainly, in order to facilitate data storage and extraction, the corpus text vectors corresponding to each corpus can be stored in a matrix form, specifically, after the low-dimensional dense vectors corresponding to the corpuses are obtained, the low-dimensional dense vectors corresponding to the corpuses are combined based on a word2vec model, and the feature vector matrix corresponding to the text is obtained. For example, in the implementation, a word vector of 100 dimensions is selected, and after training is completed, the "chinese/national/people/civilian" obtains a corresponding word vector matrix of 4 × 100, each row of the matrix corresponds to a vector of a single character, and each column represents each dimension of the word vector, as follows:
"middle" [ -0.1111,0.111, … …,0.1233]
"nation" [0.5324,0.3678, … …, -0.1111]
"human" [ -0.1111,0.1915, … …,0.4996]
"Min" [ -0.1111,0.8113, … …,0.3267 ].
Step S102: constructing a joint loss function formula of the training model, wherein the training model comprises an intention recognition model, a similar short text generation model and an entity recognition model;
before constructing the joint loss function formula of the training model, modeling needs to be performed for each task to obtain the training model, specifically:
modeling is respectively carried out on the intention recognition task, the similar short text task and the entity recognition task to obtain an intention recognition model, a similar short text generation model and an entity recognition model;
for the intention recognition task, when the intention recognition task is trained, training data are query input by a user and an intention category label corresponding to the query, after the query is subjected to word segmentation or word segmentation, each word or word is mapped to a corresponding vector based on the training result of step S101, then each word or word is mapped to all corresponding vectors to be averaged, a vector corresponding to the query is obtained and is marked as V1, then entity recognition is performed on the query, and whether the query matches a specific pattern in a preset regular pattern table is checked, for example: "play laojiu gate", laojiu gate is identified as album by the regular pattern table of presetting, match the mode in the regular pattern table: and playing album, and obtaining a corresponding K-dimensional vector V2 by the query, wherein K is the number of the modes in the regular mode table, if the ith dimension value is 1, the query matches the ith mode in the regular mode table, and if the ith dimension value is 0, the query does not match the modes in the regular mode table. V1 and V2 are then used as inputs to the preset neural network model. In this embodiment, the structure of the neural network includes an input layer, an lstm layer, a dense layer, and a skip layer. See the example of fig. 2 in particular.
In this embodiment, for example, for a VIDEO search engine, if the input query is "PLAY old nine doors", the output intention category label is "PLAY _ VIDEO", the input query is "DOWNLOAD royal glory", the output intention category label is "DOWNLOAD _ GAME", the input query is "member charge", and the output intention label is "PURCHASE".
For the similar short text generation task, the training data is a similar short text pair, and the training data specifically comprises the following three types: the method comprises the steps of firstly, querying query which clicks the same document with the query, secondly, querying query of the same session, thirdly, querying query and title clicked by the query. When modeling is carried out on a similar short text generation task, modeling is carried out on the similar short text based on a Seq2Seq model to obtain a similar short text generation model, the input of the similar short text generation model is a query of a user, and the output of the similar short text generation model is a short text similar to the query.
The goal of the Seq2Seq model is to output a sequence Y based on a given sequence X, and the loss function value is calculated using cross entropy. The Seq2Seq model consists of an encoder and a decoder, wherein the encoder neural network converts an input sequence into a vector with a fixed length, and the decoder neural network generates a new sequence according to the vector generated by the encoder neural network. During model training, vectors of the encoder neural network are shared with other tasks. When predicting text, for example: the inputs to the model are: "how to buy the member", the encoder neural network inputs the text segment into the network in sequence word by word or word by word, converts it into a vector of fixed length, and the decoder neural network converts the vector of fixed length into a new text output, for example, outputs "how to recharge the member".
For an entity recognition task, a multi-classification model is constructed based on a convolutional neural network to serve as an entity recognition model, the multi-classification model is adopted to train the entity recognition task based on training data, and the training data of the entity recognition model is as follows: the context text containing the entity, the model input as short text containing the entity, and the output as entity type tag. The specific sources of training data may be: the Internet such as encyclopedic and the like discloses corpus resources or corpus data constructed by manual labeling of application related personnel.
In this embodiment, the convolutional neural network architecture includes the following layers: the system comprises a vector search layer, a linear conversion layer 1, a Sigmoid layer, a linear conversion layer 2 and a whole sentence analysis layer. When the convolutional neural network is trained, the neural network model is trained by adopting a cross entropy loss function value based on a random gradient descent algorithm. When the entity type label of the input text is predicted, for example, for the input of 'Beijing City', at the time t, the Chinese character to be processed of the system is 'north', the 'north' is mapped into a real number vector through a vector searching layer and transmitted to a linear conversion layer, after the processing of the linear layer and a sigmoid layer, the system scores all possible marks corresponding to the 'north', the mark with the maximum labeling probability is marked, and the higher the score is, the higher the marking probability is. At the next time t +1, the system then processes the next Chinese character "Jing" in the sentence. And at a sentence analysis layer, generating a score network aiming at the processed text, wherein the nodes in the t-th column are scores of all entity type labels corresponding to Chinese characters to be processed at the time t, the connection line identification transition probability between the nodes in the t +1 column and the nodes in the t column is used for describing the conversion possibility among the labels, and finally, based on a viterbi algorithm, finding a path with the highest overall score in the network to be used as a final mark sequence. For example, the entity type label path corresponding to "Beijing City" is "B-LOC I-LOC I-LOC".
After the training models are built, the loss function value of each training model is built, wherein the loss function value of the intention recognition model is lossintent_recognitionThe loss function value of the entity recognition model is lossentity_recognitionSimilar textThe value of the loss function of the generative model is
Figure BDA0001713383960000081
The lossintent_recognition、lossentity_recognitionAnd losssim_query_generationThe value of (a) can be obtained by a mean square error method or a cross entropy method when training the training model. A specific loss weight factor is given to each loss function value, and the sum of the three loss function values after the loss weight factor is given is used as a joint loss function value of the intention recognition model, the similar short text generation model and the entity recognition model, namely, through a joint loss function formula:
loss_total=α*lossintent_recognition+β*losssim_query_generation+γ*lossentity_recognitioncalculating to obtain a combined loss function value loss _ total, wherein alpha, beta and gamma are loss respectivelyintent_recognition、lossentity_recognitionAnd losssim_query_generationThe values of the alpha, the beta and the gamma are between (0, 1), and the specific values can be adjusted according to the requirements of users. That is, when x is a user input and the loss function of each of the three models is f1(x), which corresponds to lossintent_recognitionThe loss function of a similar short text model is f2(x), which is equivalent to lossentity_recognitionThe loss function of the entity recognition model is f3(x), which corresponds to losssim_query_generationThe combined loss function is then a linear weighting of 3 loss functions, a x f1(x) + c x f2(x) + b x f3(x), where a corresponds to α, b corresponds to β, and c corresponds to γ.
Step S103: acquiring training data for the training model;
wherein the training data comprises training data for an intent recognition model, a similar short text generation model, and an entity recognition model;
in this step, the training data is some data in a preset training data set, specific data content of the training data can be selected according to user requirements, the training data is used as an input text of each model, and an actual similar short text and a category value corresponding to each training data are known.
Step S104: carrying out segmentation processing on the training data;
the method specifically comprises the following steps: performing word segmentation, word segmentation or word and word mixed segmentation processing on the training data of the intention recognition model, the similar short text generation model and the entity recognition model;
step S105: mapping the processed training data to corresponding corpus text vectors, and recording the training vectors as training vectors;
performing word segmentation or word mixing segmentation on the input text of each model, when the input text is a Chinese text, selecting a word segmentation model based on hidden Markov, characters or a conditional random field to segment the input text, when the input text is an English text, selecting punctuation and blank space as separators to segment, and if the implementation mode selects word segmentation, the input text is: "i wants to see old nine gates", the input text is segmented into "i/want/see/old nine gates". If the input text is 'I way to play game', the input text is segmented into 'I/way/to/play/game'; if the embodiment selects the segmentation, the segmentation may be performed according to characters, for example, the input text "i want to see all ages and nine doors" may be segmented into "i/want/see/old/nine/doors", and in addition, according to application requirements, the word-mixing segmentation may also be selected, for example, the input text is segmented according to a character segmentation mode for chinese in the input text, and the input text is segmented according to a word segmentation mode for english in the input text, such as: "i/want/see/billons".
After the training data input by each model is segmented, the segmented training data is mapped to the corpus text vector obtained in step S101, and the specific process is as follows: obtaining a corpus database text vector corresponding to a corpus which is the same as the training data in the corpus database, wherein the corpus database text vector is a training vector corresponding to the training data and can be a word vector, a word vector or a pinyin vector. The three models may share a set of word/term/pinyin vectors.
Step S106: inputting the training vector into the training model, and calculating loss function values of the intention recognition model, the similar short text generation model and the entity recognition model based on a prediction result output by the training model and a real result corresponding to the training vector;
in this step, when each model is trained, each task is cycled, a training vector corresponding to a part of training data is randomly extracted as a model input, and a loss function value of each model is calculated according to an output result of the model and a real result corresponding to the training data.
Step S107: calculating to obtain a joint loss function value based on the joint loss function formula and the loss function values of the models;
in this step, each loss function value calculated in step S107 is substituted into the joint loss function formula, so that the joint loss function value can be obtained.
Step S108: judging whether the joint loss function value is smaller than a set value, if so, finishing training, and if not, executing the step S109;
step S109: and updating parameters of the intention recognition model, the similar short text generation model and the entity recognition model based on a back propagation algorithm, synchronously updating the input vector shared by the three models as a model parameter, and continuously executing the step S103 for iterative training.
In the scheme, the input text is processed in a multi-task learning mode, so that the related language knowledge can be more effectively learned, the generalization capability of the intention recognition model is improved, the problems of semantic missing, fault tolerance, ambiguity, generalization and the like in intention recognition are solved as far as possible, and particularly, the parameters of the intention recognition model, the similar short text generation model and the entity recognition model can be updated based on a back propagation algorithm.
Corresponding to the above method, the present application also discloses an intention recognition training device, see fig. 3, which includes:
the corpus vector training unit 01 is used for mapping the corpus in the corpus text to a semantic space to obtain a low-dimensional dense vector corresponding to the corpus, and marking the low-dimensional dense vector as a corpus text vector;
the model storage unit 02 is used for storing an intention recognition model, a similar short text generation model and an entity recognition model;
the combined loss function formula storage unit 03 is used for storing a combined loss function formula of a training model, wherein the training model comprises an intention recognition model, a similar short text generation model and an entity recognition model;
a training data acquisition unit 04 for acquiring training data for the training model;
the training vector acquisition unit 05 is used for segmenting the training data, mapping the processed training data to corresponding corpus text vectors and recording the mapped corpus text vectors as training vectors;
a loss function value calculation unit 06, configured to input the training vector to the training model, calculate a loss function value of the intention recognition model, the similar short text generation model, and the entity recognition model based on a prediction result output by the training model and a real result corresponding to the training vector, and bring the loss function value of each model into the joint loss function formula to obtain a joint loss function value;
and the parameter adjusting unit 07 is used for judging whether the joint loss function value is smaller than a set value, if not, adjusting the training parameters of the intention recognition model, the similar short text generation model and the entity recognition model so as to reduce the loss function value of the models, continuing iterative training, and if so, finishing the training.
Corresponding to the method, the corpus database stores historical text input information, click logs and voice information of users, and can train corpus text vectors and construct training data sets of three models based on the information.
Corresponding to the above method, the corpus vector training unit is specifically configured to, when training a vector:
segmenting the corresponding granularity of a word level, a word level or a pinyin level of the corpus in the corpus database;
training the segmented text based on a neural network model, and expressing the segmented text as a low-dimensional dense vector, wherein the low-dimensional dense vector comprises a word vector, a word vector or a pinyin vector.
Corresponding to the above method, the intention recognition model stored in the model storage unit is:
modeling aiming at an intention recognition task by adopting an LSTM model to obtain an intention recognition model, wherein the input of the intention recognition model is a query, and the output of the intention recognition model is an intention category label;
the similar short text generation model stored in the model storage unit is as follows:
establishing a similar short text generation model aiming at the similar short text by adopting a Seq2Seq model, wherein the similar short text generation model is input by a user to input a query and output as the similar short text;
the entity recognition task model stored in the model storage unit is as follows:
the multi-classification model is constructed based on a convolutional neural network, and training data of the multi-classification model are as follows: context text containing entities contained in the text to be recognized, the model input is short text containing entities, and the output is entity type tags.
Corresponding to the above training method for intention recognition, the present application also discloses an intention recognition device, and referring to fig. 4, the intention recognition device disclosed in the embodiments of the present application is schematically configured, and the device may include:
a memory 100 and a processor 200;
the intention recognition device further comprises a communication interface 300 and a communication bus 400, wherein the memory 100, the processor 200 and the communication interface 300 are all in communication with each other via the communication bus 400.
The memory 100 is used for storing program codes; the program code includes computer operational instructions. Specifically, the memory stores program codes of an intention recognition model, a similar short text generation model, and an entity recognition model obtained by training the intention recognition training method disclosed in any one of the above embodiments of the present application.
Memory 100 may comprise high-speed RAM memory and may also include non-volatile memory (non-volatile memory), such as at least one disk memory.
The processor 200 may be a central processing unit CPU or an application Specific Integrated circuit asic or one or more Integrated circuits configured to implement embodiments of the present invention. The processor 200 is used to call and execute the program code. Specifically, the processor is configured to invoke and execute the intention recognition model, the similar short text generation model, and the entity recognition model when the query of the user is obtained.
For convenience of description, the above system is described with the functions divided into various modules, which are described separately. Of course, the functionality of the various modules may be implemented in the same one or more software and/or hardware implementations as the present application.
The embodiments in the present specification are described in a progressive manner, and the same and similar parts among the embodiments are referred to each other, and each embodiment focuses on the differences from the other embodiments. In particular, the system or system embodiments are substantially similar to the method embodiments and therefore are described in a relatively simple manner, and reference may be made to some of the descriptions of the method embodiments for related points. The above-described system and system embodiments are only illustrative, wherein the units described as separate parts may or may not be physically separate, and the parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment. One of ordinary skill in the art can understand and implement it without inventive effort.
Those of skill would further appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both, and that the various illustrative components and steps have been described above generally in terms of their functionality in order to clearly illustrate this interchangeability of hardware and software. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in Random Access Memory (RAM), memory, Read Only Memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art.
It is further noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
The previous description of the disclosed embodiments is provided to enable any person skilled in the art to make or use the present invention. Various modifications to these embodiments will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other embodiments without departing from the spirit or scope of the invention. Thus, the present invention is not intended to be limited to the embodiments shown herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (11)

1. A training method for intention recognition, comprising:
mapping the linguistic data in the linguistic data library text to a semantic space to obtain a low-dimensional dense vector corresponding to the linguistic data, and marking the low-dimensional dense vector as a linguistic data library text vector, wherein the linguistic data comprises historical text input information, a click log and voice recognition result information of a user;
constructing a joint loss function formula of a training model, wherein the training model comprises an intention recognition model, a similar short text generation model and an entity recognition model;
acquiring training data for the training model;
segmenting the training data, mapping the processed training data to corresponding corpus text vectors, and recording the mapped training data as training vectors;
inputting the training vector into the training model, and calculating loss function values of the intention recognition model, the similar short text generation model and the entity recognition model based on a prediction result output by the training model and a real result corresponding to the training vector; wherein the inputting the training vector to the training model comprises: circularly traversing each model included in the training model, randomly extracting a training vector corresponding to a part of training data as model input, and inputting the training vector into the model;
substituting the loss function values of the models into the joint loss function formula to obtain joint loss function values;
and judging whether the joint loss function value is smaller than a set value, if not, adjusting training parameters of the intention recognition model, the similar short text generation model and the entity recognition model to reduce the loss function value of the models, continuing iterative training, and if so, finishing the training.
2. The training method for intention recognition according to claim 1, wherein constructing the joint loss function formula of the training model comprises:
constructing a joint loss function formula of an intention recognition model, a similar short text generation model and an entity recognition model: loss _ total ═ α lossintent_recognition+β*losssim_query_generation+γ*lossentity_recognitionWherein α, β and γ are predetermined loss weighting factors, and the lossintent_recognitionLoss function for the intent recognition model, the lossentity_recognitionLoss function for entity recognition model, losssim_query_generationA loss function of the model is generated for similar text.
3. The method for training intent recognition according to claim 1, wherein the mapping corpora in the corpus text to semantic space to obtain the low-dimensional dense vector corresponding to the corpora comprises:
segmenting the corresponding granularity of a word level, a word level or a pinyin level of the corpus in the corpus database;
training the segmented text based on a neural network model, and expressing the segmented text as a low-dimensional dense vector, wherein the low-dimensional dense vector comprises a word vector, a word vector or a pinyin vector.
4. The training method for intention recognition according to claim 1, wherein modeling for the intention recognition task results in an intention recognition model, comprising:
and modeling aiming at the intention identification task by adopting an LSTM model to obtain an intention identification model, wherein the input of the intention identification model is query, and the output of the intention identification model is an intention category label.
5. The training method for intention recognition according to claim 1, wherein modeling the similar short text to obtain a similar short text generation model comprises:
and establishing a similar short text generation model aiming at the similar short text by adopting a Seq2Seq model, wherein the similar short text generation model is input by a user to input a query and output as the similar short text.
6. The training method for intention recognition according to claim 1, wherein modeling for the entity recognition task to obtain an entity recognition task model comprises:
constructing a multi-classification model based on a convolutional neural network, and training an entity recognition task based on training data by using the multi-classification model as an entity recognition task model, wherein the training data is as follows: context text containing entities contained in the text to be recognized, the model input is short text containing entities, and the output is entity type tags.
7. An intent recognition training device, comprising:
the corpus vector training unit is used for mapping the corpus in the corpus database text to a semantic space to obtain a low-dimensional dense vector corresponding to the corpus and marking the low-dimensional dense vector as a corpus database text vector, wherein the corpus comprises historical text input information, a click log and voice recognition result information of a user;
the model storage unit is used for storing an intention recognition model, a similar short text generation model and an entity recognition model;
the combined loss function formula storage unit is used for storing a combined loss function formula of a training model, and the training model comprises an intention recognition model, a similar short text generation model and an entity recognition model;
a training data acquisition unit for acquiring training data for the training model;
the training vector acquisition unit is used for segmenting the training data, mapping the processed training data to corresponding corpus text vectors and recording the mapped corpus text vectors as training vectors;
a loss function value calculation unit, configured to input the training vector to the training model, calculate a loss function value of the intention recognition model, the similar short text generation model, and the entity recognition model based on a prediction result output by the training model and a real result corresponding to the training vector, and bring the loss function value of each model into the joint loss function formula to obtain a joint loss function value; wherein the inputting the training vector to the training model comprises: circularly traversing each model included in the training model, randomly extracting a training vector corresponding to a part of training data as model input, and inputting the training vector into the model;
and the parameter adjusting unit is used for judging whether the joint loss function value is smaller than a set value, if not, adjusting the training parameters of the intention recognition model, the similar short text generation model and the entity recognition model so as to reduce the loss function value of the models, continuing iterative training, and if so, finishing the training.
8. The training device for intention recognition of claim 7, wherein the joint loss function is formulated as:
loss_total=α*lossintent_recognition+β*losssim_query_generation+γ*lossentity_recognitionwherein α, β and γ are predetermined loss weighting factors, and the lossintent_recognitionLoss function for the intent recognition model, the lossentity_recognitionLoss function for entity recognition model, losssim_query_generationA loss function of the model is generated for similar text.
9. The training device for intention recognition according to claim 7, wherein the corpus vector training unit is specifically configured to:
segmenting the corresponding granularity of a word level, a word level or a pinyin level of the corpus in the corpus database;
training the segmented text based on a neural network model, and expressing the segmented text as a low-dimensional dense vector, wherein the low-dimensional dense vector comprises a word vector, a word vector or a pinyin vector.
10. The training apparatus for intention recognition according to claim 7, wherein the intention recognition models stored in the model storage unit are:
modeling aiming at an intention recognition task by adopting an LSTM model to obtain an intention recognition model, wherein the input of the intention recognition model is a query, and the output of the intention recognition model is an intention category label;
the similar short text generation model stored in the model storage unit is as follows:
establishing a similar short text generation model aiming at the similar short text by adopting a Seq2Seq model, wherein the similar short text generation model is input by a user to input a query and output as the similar short text;
the entity recognition task model stored in the model storage unit is as follows:
the multi-classification model is constructed based on a convolutional neural network, and training data of the multi-classification model are as follows: context text containing entities contained in the text to be recognized, the model input is short text containing entities, and the output is entity type tags.
11. An intent recognition device comprising a memory and a processor;
the memory stores an intention recognition model, a similar short text generation model and an entity recognition model which are obtained by training through the intention recognition training method according to any one of claims 1 to 6, and the processor is used for calling and executing the intention recognition model, the similar short text generation model and the entity recognition model when a query of a user is obtained.
CN201810694995.4A 2018-06-29 2018-06-29 Training method, training device and recognition device for intention recognition Active CN108920622B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810694995.4A CN108920622B (en) 2018-06-29 2018-06-29 Training method, training device and recognition device for intention recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810694995.4A CN108920622B (en) 2018-06-29 2018-06-29 Training method, training device and recognition device for intention recognition

Publications (2)

Publication Number Publication Date
CN108920622A CN108920622A (en) 2018-11-30
CN108920622B true CN108920622B (en) 2021-07-20

Family

ID=64423645

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810694995.4A Active CN108920622B (en) 2018-06-29 2018-06-29 Training method, training device and recognition device for intention recognition

Country Status (1)

Country Link
CN (1) CN108920622B (en)

Families Citing this family (58)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111292752B (en) * 2018-12-06 2023-05-12 北京嘀嘀无限科技发展有限公司 User intention recognition method and device, electronic equipment and storage medium
CN109741751A (en) * 2018-12-11 2019-05-10 上海交通大学 Intension recognizing method and device towards intelligent sound control
CN109545186B (en) * 2018-12-16 2022-05-27 魔门塔(苏州)科技有限公司 Speech recognition training system and method
CN111354354B (en) * 2018-12-20 2024-02-09 深圳市优必选科技有限公司 Training method, training device and terminal equipment based on semantic recognition
CN109857844B (en) * 2018-12-29 2022-01-14 北京三快在线科技有限公司 Intent recognition method and device based on ordering dialogue text and electronic equipment
US20200226476A1 (en) * 2019-01-10 2020-07-16 Visa International Service Association System, Method, and Computer Program Product for Incorporating Knowledge from More Complex Models in Simpler Models
CN109857868A (en) * 2019-01-25 2019-06-07 北京奇艺世纪科技有限公司 Model generating method, file classification method, device and computer readable storage medium
CN111563209B (en) * 2019-01-29 2023-06-30 株式会社理光 Method and device for identifying intention and computer readable storage medium
CN111563208B (en) * 2019-01-29 2023-06-30 株式会社理光 Method and device for identifying intention and computer readable storage medium
CN114254750A (en) * 2019-01-29 2022-03-29 北京金山数字娱乐科技有限公司 Accuracy loss determination method and apparatus
CN109871485B (en) * 2019-02-13 2022-04-05 北京航空航天大学 Personalized recommendation method and device
CN111680514B (en) * 2019-02-25 2024-03-01 北京猎户星空科技有限公司 Information processing and model training method, device, equipment and storage medium
CN111612025B (en) * 2019-02-25 2023-12-12 北京嘀嘀无限科技发展有限公司 Description model training method, text description device and electronic equipment
CN109933663A (en) * 2019-02-26 2019-06-25 上海凯岸信息科技有限公司 Intention assessment algorithm based on embedding method
CN113348502A (en) * 2019-04-17 2021-09-03 深圳市欢太科技有限公司 Voice recognition method, voice recognition device, storage medium and electronic equipment
CN110147535A (en) * 2019-04-18 2019-08-20 平安科技(深圳)有限公司 Similar Text generation method, device, equipment and storage medium
CN110085308B (en) * 2019-04-23 2022-02-25 挂号网(杭州)科技有限公司 Diagnosis and treatment department classification method based on fusion deep learning
CN110287283B (en) * 2019-05-22 2023-08-01 中国平安财产保险股份有限公司 Intention model training method, intention recognition method, device, equipment and medium
CN110134969B (en) * 2019-05-27 2023-07-14 北京奇艺世纪科技有限公司 Entity identification method and device
CN110287285B (en) * 2019-05-31 2023-06-16 平安科技(深圳)有限公司 Method and device for identifying problem intention, computer equipment and storage medium
CN110222339B (en) * 2019-06-05 2023-04-28 深圳市思迪信息技术股份有限公司 Intention recognition method and device based on improved XGBoost algorithm
CN110334344A (en) * 2019-06-13 2019-10-15 腾讯科技(深圳)有限公司 A kind of semanteme intension recognizing method, device, equipment and storage medium
CN110516035A (en) * 2019-07-05 2019-11-29 同济大学 A kind of man-machine interaction method and system of mixing module
CN112185351A (en) * 2019-07-05 2021-01-05 北京猎户星空科技有限公司 Voice signal processing method and device, electronic equipment and storage medium
CN110610001B (en) * 2019-08-12 2024-01-23 大箴(杭州)科技有限公司 Short text integrity recognition method, device, storage medium and computer equipment
CN112417145A (en) * 2019-08-23 2021-02-26 武汉Tcl集团工业研究院有限公司 Text multi-classification model generation method, text processing device and medium
CN110532368B (en) * 2019-09-04 2023-03-14 达闼机器人股份有限公司 Question answering method, electronic equipment and computer readable storage medium
CN110619047B (en) * 2019-09-16 2022-09-02 出门问问信息科技有限公司 Method and device for constructing natural language model and readable storage medium
CN110674253A (en) * 2019-09-23 2020-01-10 出门问问信息科技有限公司 Semantic processing method and equipment
CN110807332B (en) 2019-10-30 2024-02-27 腾讯科技(深圳)有限公司 Training method, semantic processing method, device and storage medium for semantic understanding model
CN112749565A (en) * 2019-10-31 2021-05-04 华为终端有限公司 Semantic recognition method and device based on artificial intelligence and semantic recognition equipment
CN110852108B (en) * 2019-11-11 2022-03-29 中山大学 Joint training method, apparatus and medium for entity recognition and entity disambiguation
CN110956018B (en) * 2019-11-22 2023-04-18 腾讯科技(深圳)有限公司 Training method of text processing model, text processing method, text processing device and storage medium
CN112906370B (en) * 2019-12-04 2022-12-20 马上消费金融股份有限公司 Intention recognition model training method, intention recognition method and related device
CN113010667A (en) * 2019-12-20 2021-06-22 王道维 Training method for machine learning decision model by using natural language corpus
CN111198938B (en) * 2019-12-26 2023-12-01 深圳市优必选科技股份有限公司 Sample data processing method, sample data processing device and electronic equipment
CN111161740A (en) * 2019-12-31 2020-05-15 中国建设银行股份有限公司 Intention recognition model training method, intention recognition method and related device
CN111209383B (en) * 2020-01-06 2023-04-07 广州小鹏汽车科技有限公司 Method and device for processing multi-turn dialogue, vehicle, and storage medium
CN111341309A (en) * 2020-02-18 2020-06-26 百度在线网络技术(北京)有限公司 Voice interaction method, device, equipment and computer storage medium
CN111309915B (en) * 2020-03-03 2022-09-13 爱驰汽车有限公司 Method, system, device and storage medium for training natural language of joint learning
CN111667066A (en) * 2020-04-23 2020-09-15 北京旷视科技有限公司 Network model training and character recognition method and device and electronic equipment
CN111597804B (en) * 2020-05-15 2023-03-10 腾讯科技(深圳)有限公司 Method and related device for training entity recognition model
CN111680517B (en) * 2020-06-10 2023-05-16 北京百度网讯科技有限公司 Method, apparatus, device and storage medium for training model
CN111538894B (en) * 2020-06-19 2020-10-23 腾讯科技(深圳)有限公司 Query feedback method and device, computer equipment and storage medium
CN111737416B (en) * 2020-06-29 2022-08-19 重庆紫光华山智安科技有限公司 Case processing model training method, case text processing method and related device
CN111785350B (en) * 2020-06-30 2023-04-18 易联众信息技术股份有限公司 Information extraction method, application, device and medium
CN111949784A (en) * 2020-08-14 2020-11-17 中国工商银行股份有限公司 Outbound method and device based on intention recognition
CN112102832B (en) * 2020-09-18 2021-12-28 广州小鹏汽车科技有限公司 Speech recognition method, speech recognition device, server and computer-readable storage medium
CN112183631B (en) * 2020-09-28 2024-01-12 云知声智能科技股份有限公司 Method and terminal for establishing intention classification model
CN112256826A (en) * 2020-10-19 2021-01-22 网易(杭州)网络有限公司 Emotion analysis method, evaluation method and emotion analysis model training method and device
CN112417116B (en) * 2020-11-18 2022-03-15 四川长虹电器股份有限公司 Question understanding model training method and system based on few-sample corpus
CN112417886A (en) * 2020-11-20 2021-02-26 平安普惠企业管理有限公司 Intention entity information extraction method and device, computer equipment and storage medium
CN112528679B (en) * 2020-12-17 2024-02-13 科大讯飞股份有限公司 Method and device for training intention understanding model, and method and device for intention understanding
CN112765356B (en) * 2021-01-29 2022-07-12 思必驰科技股份有限公司 Training method and system of multi-intention recognition model
CN112905893B (en) * 2021-03-22 2024-01-12 北京百度网讯科技有限公司 Training method of search intention recognition model, search intention recognition method and device
CN113094475B (en) * 2021-06-08 2021-09-21 成都晓多科技有限公司 Dialog intention recognition system and method based on context attention flow
CN113254617B (en) * 2021-06-11 2021-10-22 成都晓多科技有限公司 Message intention identification method and system based on pre-training language model and encoder
CN113486178B (en) * 2021-07-12 2023-12-01 恒安嘉新(北京)科技股份公司 Text recognition model training method, text recognition method, device and medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106156003A (en) * 2016-06-30 2016-11-23 北京大学 A kind of question sentence understanding method in question answering system
CN106815566A (en) * 2016-12-29 2017-06-09 天津中科智能识别产业技术研究院有限公司 A kind of face retrieval method based on multitask convolutional neural networks
CN107464559A (en) * 2017-07-11 2017-12-12 中国科学院自动化研究所 Joint forecast model construction method and system based on Chinese rhythm structure and stress
CN107704866A (en) * 2017-06-15 2018-02-16 清华大学 Multitask Scene Semantics based on new neural network understand model and its application

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170193099A1 (en) * 2015-12-31 2017-07-06 Quixey, Inc. Machine Identification of Grammar Rules That Match a Search Query

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106156003A (en) * 2016-06-30 2016-11-23 北京大学 A kind of question sentence understanding method in question answering system
CN106815566A (en) * 2016-12-29 2017-06-09 天津中科智能识别产业技术研究院有限公司 A kind of face retrieval method based on multitask convolutional neural networks
CN107704866A (en) * 2017-06-15 2018-02-16 清华大学 Multitask Scene Semantics based on new neural network understand model and its application
CN107464559A (en) * 2017-07-11 2017-12-12 中国科学院自动化研究所 Joint forecast model construction method and system based on Chinese rhythm structure and stress

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
A Joint Model of Intent Determination and Slot Filling for Spoken Language Understanding;Xiaodong Zhang等;《Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence》;20160715;第2993-2999页 *

Also Published As

Publication number Publication date
CN108920622A (en) 2018-11-30

Similar Documents

Publication Publication Date Title
CN108920622B (en) Training method, training device and recognition device for intention recognition
CN109657054B (en) Abstract generation method, device, server and storage medium
CN110069709B (en) Intention recognition method, device, computer readable medium and electronic equipment
CN111753060A (en) Information retrieval method, device, equipment and computer readable storage medium
CN110727779A (en) Question-answering method and system based on multi-model fusion
CN108932342A (en) A kind of method of semantic matches, the learning method of model and server
CN109376222B (en) Question-answer matching degree calculation method, question-answer automatic matching method and device
CN109960728B (en) Method and system for identifying named entities of open domain conference information
CN111259127B (en) Long text answer selection method based on transfer learning sentence vector
CN111159485B (en) Tail entity linking method, device, server and storage medium
CN110633577B (en) Text desensitization method and device
CN111222305A (en) Information structuring method and device
CN111708878B (en) Method, device, storage medium and equipment for extracting sports text abstract
WO2023134083A1 (en) Text-based sentiment classification method and apparatus, and computer device and storage medium
CN113961666B (en) Keyword recognition method, apparatus, device, medium, and computer program product
CN116304745B (en) Text topic matching method and system based on deep semantic information
CN113821605A (en) Event extraction method
CN110597968A (en) Reply selection method and device
CN112632224A (en) Case recommendation method and device based on case knowledge graph and electronic equipment
CN112464655A (en) Word vector representation method, device and medium combining Chinese characters and pinyin
CN114417823A (en) Aspect level emotion analysis method and device based on syntax and graph convolution network
CN113486143A (en) User portrait generation method based on multi-level text representation and model fusion
CN115357720B (en) BERT-based multitasking news classification method and device
WO2023137903A1 (en) Reply statement determination method and apparatus based on rough semantics, and electronic device
CN110874408B (en) Model training method, text recognition device and computing equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant