CN113569581B - Intention recognition method, device, equipment and storage medium - Google Patents

Intention recognition method, device, equipment and storage medium Download PDF

Info

Publication number
CN113569581B
CN113569581B CN202110991234.7A CN202110991234A CN113569581B CN 113569581 B CN113569581 B CN 113569581B CN 202110991234 A CN202110991234 A CN 202110991234A CN 113569581 B CN113569581 B CN 113569581B
Authority
CN
China
Prior art keywords
sample
intention
abnormal
preset
model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110991234.7A
Other languages
Chinese (zh)
Other versions
CN113569581A (en
Inventor
黎华清
邓嘉镔
宋维林
车轮
黄小玲
刘银煜
陈丹瑶
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China United Network Communications Group Co Ltd
Original Assignee
China United Network Communications Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China United Network Communications Group Co Ltd filed Critical China United Network Communications Group Co Ltd
Priority to CN202110991234.7A priority Critical patent/CN113569581B/en
Publication of CN113569581A publication Critical patent/CN113569581A/en
Application granted granted Critical
Publication of CN113569581B publication Critical patent/CN113569581B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • G06F40/35Discourse or dialogue representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Machine Translation (AREA)

Abstract

The application provides an intention recognition method, a device, equipment and a storage medium, wherein the method is characterized in that after intention recognition is carried out on a dialogue sample to be processed by adopting a preset intention recognition model, the type of each intention recognized by the preset intention recognition model and the confidence coefficient of each intention type are obtained, further, according to the confidence coefficient of each intention type and a preset confidence coefficient threshold value, an abnormal sample is determined from the dialogue sample to be processed, the abnormal sample is subjected to sample expansion, the model generalization capability is increased, the overfitting phenomenon is effectively avoided, and in this way, an countermeasure network model is built based on the abnormal sample after sample expansion, so that the correct intention of the abnormal sample is determined according to the output result of the countermeasure network model, and the accuracy of intention recognition is improved. And the application builds an countermeasure network model based on the expanded sample, automatically selects the correct intention for the sample, and can reduce the manual processing amount and improve the efficiency.

Description

Intention recognition method, device, equipment and storage medium
Technical Field
The present application relates to the field of language processing technologies, and in particular, to an intent recognition method, apparatus, device, and storage medium.
Background
With the rapid development of artificial intelligence technology, multiple rounds of dialogue are used as important technology for researching interactive information processing in the field of natural semantic processing, and are widely applied to analysis of information in the process of human-to-human and robot-to-human communication so as to define the intention of both parties.
In the training and building of multi-round dialog models, a large amount of manual annotation data is generally required as a basis for training and building multi-round dialog system models. If the intention of the user is ambiguous, the data volume to be marked is multiplied, and the user is lost due to the fact that the reply content is not attached to the context and other problems.
At present, in order to solve the problems of large labeling data amount, non-conforming reply content to context and the like caused by ambiguous user intention, a concept of intention recognition is proposed. However, the existing intention recognition usually adopts a batch of sample training models, the model obtained by training is used for predicting the same batch of samples, and the phenomenon of severe overfitting exists, so that the accuracy of the intention recognition of the user is low, the intention of the user is difficult to accurately recognize, and the answer most fitting the context is made. Therefore, how to accurately recognize the intention of the user has become a problem to be solved.
Disclosure of Invention
The application provides an intention recognition method, an intention recognition device, intention recognition equipment and a storage medium, which effectively avoid the phenomenon of overfitting existing in the existing intention recognition and improve the accuracy of the intention recognition.
In a first aspect, the present application provides an intention recognition method comprising the steps of:
after intention recognition is carried out on a dialogue sample to be processed by adopting a preset intention recognition model, obtaining the types of each intention recognized by the preset intention recognition model and the confidence coefficient of each intention type in the types of each intention;
determining an abnormal sample from the dialog samples to be processed according to the confidence degree of each intention type and a preset confidence threshold, wherein the abnormal sample comprises a plurality of samples;
sample expansion is carried out on the abnormal samples, and an countermeasure network model is built based on the abnormal samples after sample expansion;
and determining the correct intention of the abnormal sample according to the output result of the countermeasure network model.
In one possible implementation manner, the determining, according to the confidence level of each intention type and a preset confidence threshold, an abnormal sample from the dialog samples to be processed includes:
comparing the confidence level of each intention type with the preset confidence level threshold value respectively;
if the confidence coefficient of the intention type i is smaller than the preset confidence coefficient threshold value, determining that a dialogue sample corresponding to the intention type i is the abnormal sample, and acquiring the abnormal sample from the dialogue sample to be processed, wherein the intention type i is any one intention type in the intention types, and i=1, 2, …, n and n are equal to the number of the intention types.
In one possible implementation manner, the performing sample expansion on the abnormal sample includes:
converting the abnormal sample into a sample in a preset language, and converting the sample in the preset language into a sample in a language corresponding to the abnormal sample, wherein the preset language is other languages except the language corresponding to the abnormal sample;
or alternatively
Randomly replacing the vocabulary in the abnormal sample according to the synonym library;
or alternatively
And carrying out batch copy pasting on the vocabulary in the abnormal sample.
In one possible implementation manner, the abnormal sample includes a first sample and a second sample;
the construction of the countermeasure network model based on the abnormal sample after sample expansion comprises the following steps:
adding the sample after the expansion of the first sample into the first sample to obtain a third sample, and adding the sample after the expansion of the second sample into the second sample to obtain a fourth sample;
constructing an intention recognition model based on the third sample and the fourth sample respectively;
obtaining the countermeasure network model based on the constructed intention recognition model
In one possible implementation, the determining the correct intention of the abnormal sample according to the output result of the countermeasure network model includes:
determining a model with a result accuracy gradient reaching an accuracy gradient threshold or a loss function gradient lower than a preset loss gradient threshold from the countermeasure network model according to the output result of the countermeasure network model;
and obtaining the correct intention of the abnormal sample according to the determined model.
In a second aspect, the present application provides an intention recognition apparatus, the apparatus comprising:
the information acquisition module is used for acquiring the types of each intention identified by the preset intention identification model and the confidence of each intention type in the types of each intention after the intention identification is carried out on the dialogue sample to be processed by adopting the preset intention identification model;
the sample determining module is used for determining an abnormal sample from the dialog samples to be processed according to the confidence degree of each intention type and a preset confidence threshold value, wherein the abnormal sample comprises a plurality of samples;
the sample processing module is used for carrying out sample expansion on the abnormal samples and building an countermeasure network model based on the abnormal samples after the sample expansion;
and the intention determining module is used for determining the correct intention of the abnormal sample according to the output result of the countermeasure network model.
In one possible implementation manner, the sample determining module is specifically configured to:
comparing the confidence level of each intention type with the preset confidence level threshold value respectively;
if the confidence coefficient of the intention type i is smaller than the preset confidence coefficient threshold value, determining that a dialogue sample corresponding to the intention type i is the abnormal sample, and acquiring the abnormal sample from the dialogue sample to be processed, wherein the intention type i is any one intention type in the intention types, and i=1, 2, …, n and n are equal to the number of the intention types.
In one possible implementation manner, the sample processing module is specifically configured to:
converting the abnormal sample into a sample in a preset language, and converting the sample in the preset language into a sample in a language corresponding to the abnormal sample, wherein the preset language is other languages except the language corresponding to the abnormal sample;
or alternatively
Randomly replacing the vocabulary in the abnormal sample according to the synonym library;
or alternatively
And carrying out batch copy pasting on the vocabulary in the abnormal sample.
In one possible implementation manner, the abnormal sample includes a first sample and a second sample, and the sample processing module is specifically configured to:
adding the sample after the expansion of the first sample into the first sample to obtain a third sample, and adding the sample after the expansion of the second sample into the second sample to obtain a fourth sample;
constructing an intention recognition model based on the third sample and the fourth sample respectively;
obtaining the countermeasure network model based on the constructed intention recognition model
In one possible implementation, the intention determining module includes:
determining a model with a result accuracy gradient reaching an accuracy gradient threshold or a loss function gradient lower than a preset loss gradient threshold from the countermeasure network model according to the output result of the countermeasure network model;
and obtaining the correct intention of the abnormal sample according to the determined model.
In a third aspect, the present application provides an intention recognition apparatus comprising:
a processor;
a memory; and
a computer program;
wherein the computer program is stored in the memory and configured to be executed by the processor, the computer program comprising instructions for performing the method according to the first aspect.
In a fourth aspect, the present application provides a computer readable storage medium storing a computer program for causing a server to execute the method of the first aspect.
In a fifth aspect, the application provides a computer program product comprising computer instructions for performing the method of the first aspect by a processor.
According to the intention recognition method, the device, the equipment and the storage medium, after intention recognition is carried out on the dialogue sample to be processed by adopting the preset intention recognition model, the types of the intentions recognized by the preset intention recognition model and the confidence coefficient of each intention type are obtained, further, according to the confidence coefficient of each intention type and the preset confidence coefficient threshold value, an abnormal sample is determined from the dialogue sample to be processed, sample expansion is carried out on the abnormal sample, and an countermeasure network model is built based on the abnormal sample after sample expansion, so that the correct intention of the abnormal sample is determined according to the output result of the countermeasure network model. And the application builds an countermeasure network model based on the expanded sample, automatically selects the correct intention for the sample, and can reduce the manual processing amount and improve the efficiency.
Drawings
In order to more clearly illustrate the embodiments of the application or the technical solutions of the prior art, the drawings which are used in the description of the embodiments or the prior art will be briefly described, it being obvious that the drawings in the description below are only some embodiments of the application, and that other drawings can be obtained according to these drawings without inventive faculty for a person skilled in the art.
FIG. 1 is a schematic diagram of an intent recognition system architecture according to an embodiment of the present application;
FIG. 2 is a schematic flow chart of an intent recognition method according to an embodiment of the present application;
FIG. 3 is a flowchart illustrating another method for identifying intent according to an embodiment of the present application;
FIG. 4 is a schematic diagram of sample expansion according to an embodiment of the present application;
FIG. 5 is a schematic diagram of intent recognition according to an embodiment of the present application;
fig. 6 is a schematic structural diagram of an intent recognition device according to an embodiment of the present application;
fig. 7 is a schematic diagram of a basic hardware architecture of an intent recognition device according to the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
The terms "first," "second," "third," and "fourth" and the like in the description and in the claims and in the above drawings, if any, are used for distinguishing between similar objects and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used may be interchanged where appropriate such that the embodiments of the application described herein may be implemented in sequences other than those illustrated or otherwise described herein. Furthermore, the terms "comprises," "comprising," and "having," and any variations thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
The multi-wheel dialogue model is widely applied to the fields of intelligent customer service, robots, automobiles, navigation and the like. In the training and building of the multi-round dialogue model, user intention recognition is needed to solve the problems of large labeling data volume, non-conforming reply content to the context and the like caused by the ambiguous user intention. The existing user intention recognition uses a batch of samples to train a model, and predicts the same batch of samples by using the model, so that the phenomenon of severe overfitting exists, and the accuracy of the user intention recognition is lower.
Therefore, the embodiment of the application provides an intention recognition method, which increases the generalization capability of a model and effectively avoids the phenomenon of overfitting by carrying out data expansion on training samples, so that the intention recognition is carried out based on the expanded samples, and the accuracy of the intention recognition is improved. In addition, the embodiment of the application builds the countermeasure network model based on the expanded sample, automatically selects the correct intention for the sample, and can reduce the manual processing amount and improve the efficiency.
Optionally, the method for identifying intent provided in the present application may be applied to a schematic diagram of an intent identification system shown in fig. 1, where the system may include at least one of a receiving device 101, a processing device 102 and a display device 103 as shown in fig. 1.
In a specific implementation, the receiving device 101 may be an input/output interface or a communication interface, and may be configured to receive a session sample to be processed.
The processing device 102 may obtain the to-be-processed dialog sample through the receiving device 101, further, preset an intention recognition model to recognize the to-be-processed dialog sample, obtain the types of each intention and the confidence of each intention type, determine an abnormal sample from the to-be-processed dialog sample according to the confidence, perform sample expansion on the abnormal sample, and build an countermeasure network model based on the abnormal sample after sample expansion, thereby determining the correct intention of the abnormal sample. The processing device 102 increases the generalization capability of the model and effectively avoids the over-fitting phenomenon by expanding the data of the abnormal samples, so that the intention recognition is performed based on the expanded samples, and the accuracy of the intention recognition is improved. The processing device 102 builds an countermeasure network model based on the expanded sample, automatically selects a correct intention for the sample, and can reduce the manual processing amount and improve the efficiency.
The display device 103 may be configured to display the types of the above-described respective intentions, the confidence level of each intention type, the abnormal sample, the correct intention of the abnormal sample, and the like.
The display device may also be a touch display screen for receiving user instructions while displaying the above content to enable interaction with a user.
It should be understood that the processing device may be implemented by a processor reading instructions in a memory and executing the instructions, or may be implemented by a chip circuit.
The above system is only one exemplary system, and may be set according to application requirements when implemented.
It should be understood that the architecture illustrated in the embodiments of the present application does not constitute a specific limitation on the architecture of the e-commerce system. In other possible embodiments of the present application, the architecture may include more or less components than those illustrated, or some components may be combined, some components may be split, or different component arrangements may be specifically determined according to the actual application scenario, and the present application is not limited herein. The components shown in fig. 1 may be implemented in hardware, software, or a combination of software and hardware.
In addition, the system architecture described in the embodiments of the present application is for more clearly describing the technical solution of the embodiments of the present application, and does not constitute a limitation on the technical solution provided in the embodiments of the present application, and as a person of ordinary skill in the art can know that, with the evolution of the system architecture and the appearance of a new service scenario, the technical solution provided in the embodiments of the present application is equally applicable to similar technical problems.
The following description of the present application is given by taking several embodiments as examples, and the same or similar concepts or processes may not be described in detail in some embodiments.
Fig. 2 is a schematic flow chart of an intent recognition method according to an embodiment of the present application, where an execution subject of the embodiment may be a processing apparatus in the embodiment shown in fig. 1, and as shown in fig. 2, the method may include:
s201: after intention recognition is carried out on the dialogue sample to be processed by adopting a preset intention recognition model, the types of the intention recognized by the preset intention recognition model and the confidence degree of each intention type in the types of the intention are obtained.
Here, the above-mentioned to-be-processed session sample may be determined according to actual situations, for example, the content of the above-mentioned to-be-processed sample is as follows:
customer service: you good please ask what you can help:
and (3) a client: i want to change package
Customer service: asking you want to turn that set of tweed
And (3) a client: i introduce me with a large Wang Ka and ice cream card
Customer service: wang Ka charge … …
The preset intention recognition model may be a model (for example, long Short-Term Memory (LSTM)) that the processing device trains by using a neural network.
The processing device performs intention recognition on the dialogue sample to be processed by adopting the preset intention recognition model, for example, recognizes multiple rounds of dialogue samples of the clients and the customer service, and obtains types of each intention recognized by the preset intention recognition model, for example, turning up Wang Ka and turning up an ice cream card.
Further, the processing device may further obtain a confidence coefficient of each intention type in the intention types by using the preset intention recognition model, for example, turn up Wang Ka, where the confidence coefficient is 0.45; the ice cream card was turned to a confidence level of 0.45.
S202: and determining an abnormal sample from the dialog samples to be processed according to the confidence degree of each intention type and a preset confidence threshold value, wherein the abnormal sample comprises a plurality of samples.
The preset confidence threshold may be determined according to practical situations, for example, the threshold is set to 0.5.
After obtaining the types of the intents identified by the preset intention identification model and the confidence coefficient of each intention type, the processing device can compare the confidence coefficient of each intention type with the preset confidence coefficient threshold value respectively. If the confidence coefficient of the intention type i is smaller than the preset confidence coefficient threshold value, the processing device can determine that the dialogue sample corresponding to the intention type i is an abnormal sample, and acquire the abnormal sample from the dialogue sample to be processed. The intention type i is any one of the intention types, i=1, 2, …, n and n are equal to the number of the intention types.
Here, the processing device performs abnormal sample screening on the dialog samples to be processed, for example, the preset confidence threshold is 0.5, and samples with confidence less than 0.5 are abnormal samples, and the processing device screens abnormal samples from the dialog samples of multiple rounds of customer and customer service. Turn up Wang Ka, confidence 0.45; the confidence coefficient of the ice cream turning card is 0.45 and is smaller than 0.5, and the processing device determines that the abnormal sample is a conversation sample corresponding to the turn size Wang Ka and a conversation sample corresponding to the ice cream turning card.
S203: and carrying out sample expansion on the abnormal samples, and building an countermeasure network model based on the abnormal samples after sample expansion.
In the embodiment of the application, the processing device increases the generalization capability of the model and effectively avoids the over-fitting phenomenon by expanding the data of the abnormal sample, so that the intention recognition is performed based on the expanded sample, and the accuracy of the intention recognition is improved.
Illustratively, the abnormal samples include a first sample and a second sample, for example, a session sample corresponding to the turn-up Wang Ka and a session sample corresponding to the turn-ice cream card.
When the processing device builds the countermeasure network model based on the abnormal sample after the sample expansion, the sample after the first sample expansion can be added into the first sample to obtain a third sample, the sample after the second sample expansion can be added into the second sample to obtain a fourth sample, further, the intention recognition model is built based on the third sample and the fourth sample respectively, and the countermeasure network model is obtained based on the built intention recognition model. For example, the processing device adds the expanded dialogue sample corresponding to the turn-up Wang Ka to the dialogue sample corresponding to the turn-up Wang Ka to obtain a third sample, and adds the expanded dialogue sample corresponding to the turn-up ice cream card to the dialogue sample corresponding to the turn-up ice cream card to obtain a fourth sample, namely, adds the expanded sample to the original training sample in batches for training, so as to construct an intention recognition model, thereby obtaining an countermeasure network model.
S204: and determining the correct intention of the abnormal sample according to the output result of the countermeasure network model.
Here, with the increase of the expanded samples, the training results of the countermeasure network model may show different variation trends, and the processing device may determine, according to the results, from the countermeasure network model, that the accuracy gradient of the results reaches an accuracy gradient threshold or that the gradient of the loss function is lower than a preset loss gradient threshold, where the intention corresponding to the model is the correct intention of the abnormal sample.
The processing device builds an countermeasure network model for the abnormal samples, automatically selects correct intention for the abnormal samples, and reduces manual processing capacity.
In addition, the training parameters of each model in the countermeasure network model need to be consistent, so that in order to save training time, only one epoche (when a complete data set passes through the network once and returns once, this process is called one epoch) can be trained.
According to the embodiment of the application, after the intention recognition is carried out on the dialogue sample to be processed by adopting the preset intention recognition model, the types of each intention recognized by the preset intention recognition model and the confidence coefficient of each intention type are obtained, and then, according to the confidence coefficient of each intention type and the preset confidence coefficient threshold value, an abnormal sample is determined from the dialogue sample to be processed, the abnormal sample is subjected to sample expansion, and an countermeasure network model is built based on the abnormal sample after the sample expansion, so that the correct intention of the abnormal sample is determined according to the output result of the countermeasure network model. In addition, the embodiment of the application builds the countermeasure network model based on the expanded sample, automatically selects the correct intention for the sample, and can reduce the manual processing amount and improve the efficiency.
In addition, in order to meet application requirements in different application scenes, when the abnormal samples are subjected to sample expansion, the embodiment of the application also considers that the abnormal samples are subjected to sample expansion based on different modes, for example, training samples are expanded by methods such as inter-translation, synonym replacement and the like, so that model generalization capability is improved, and the phenomenon of overfitting is effectively avoided. Fig. 3 is a flowchart illustrating another method for identifying intent according to an embodiment of the present application. As shown in fig. 3, the method includes:
s301: after intention recognition is carried out on the dialogue sample to be processed by adopting a preset intention recognition model, the types of the intention recognized by the preset intention recognition model and the confidence degree of each intention type in the types of the intention are obtained.
S302: and determining an abnormal sample from the dialog samples to be processed according to the confidence degree of each intention type and a preset confidence threshold value, wherein the abnormal sample comprises a plurality of samples.
Steps S301 to S302 are described in the above steps S201 to S202, and are not described herein.
S303: and converting the abnormal sample into a sample in a preset language, and converting the sample in the preset language into a sample in a language corresponding to the abnormal sample, wherein the preset language is other languages except the language corresponding to the abnormal sample.
The processing device may convert the abnormal sample into a sample of a language other than chinese, for example, an english sample, and then convert the english sample into a chinese sample.
S304: and randomly replacing the vocabulary in the abnormal sample according to the synonym library.
Here, the processing device may first determine the vocabulary having a higher occurrence frequency in the abnormal sample, for example, the vocabulary having an occurrence frequency exceeding a preset frequency threshold, and then replace the determined vocabulary according to the synonym library, so that the expansion effect on the abnormal sample is obvious, and thus, the subsequent processing can be performed more accurately.
S305: and carrying out batch copy pasting on the vocabulary in the abnormal sample.
The processing device may perform batch copy pasting on the vocabulary in the abnormal sample. The portion of the batch copy and paste may be determined by the processing device according to the content of the abnormal sample, for example, the core content of the abnormal sample is used as the portion of the batch copy and paste, so as to highlight the main portion of the abnormal sample, and similarly, the expansion effect on the abnormal sample is obvious, and the accuracy of the subsequent processing result is improved.
Here, the processing device may employ any one of the steps S303 to S305 or a combination of a plurality of the steps S303 to S305 for sample expansion of the abnormal sample, and specifically, may be determined according to actual conditions, and fig. 3 only uses any one of the steps S303 to S305 as an example.
In addition, when the processing apparatus uses a plurality of step combinations in steps S303 to S305 to expand the sample, the order of the steps in the step combinations may be determined according to the actual situation, for example, step S303 may be executed first, step S304 may be executed second, and step S305 may be executed last.
For example, as shown in fig. 4, two samples are included in the above-mentioned abnormal sample, for example, a first sample and a second sample, the first sample is intended to be a, and the second sample is intended to be B. The processing device performs sample expansion on the abnormal sample through the steps, for example, expands the first sample, adds the sample expanded by the first sample to obtain a third sample, and adds the sample expanded by the second sample to obtain a fourth sample. Then, the processing device may construct an intention recognition model, such as model 1 and model 2, based on the third sample and the fourth sample, respectively, and obtain an countermeasure network model based on the constructed intention recognition model. The processing device may further perform condition judgment according to an output result of the countermeasure network model, determine, from the countermeasure network model, that a result accuracy gradient reaches an accuracy gradient threshold or that a loss function gradient is lower than a preset loss gradient threshold, thereby obtaining a correct intention of the abnormal sample according to the determined model.
Here, the processing means may construct the intention recognition model, for example, model 1 and model 2, based on the third sample and the fourth sample, respectively, and may include: in the construction process, the processing device may input the third sample into the model 1, and then determine the accuracy of intention recognition according to the result of intention recognition output by the model 1 and the result of intention recognition corresponding to the third sample. If the accuracy is lower than a preset accuracy threshold, the processing device may adjust the model 1 according to the accuracy to improve the accuracy, and take the adjusted model 1 as a new model 1, and re-execute the step of inputting the third sample into the model 1, thereby finally obtaining the constructed model 1. The construction process of the model 2 refers to the construction process of the model 1 described above, and finally, the constructed model 2 is also obtained.
The above-described processing means obtains an countermeasure network model based on the constructed intention recognition model, such as the above-described model 1 and model 2, that is, the countermeasure network model includes the model 1 and the model 2 therein. With the increase of the expansion samples, the training results of the model 1 and the model 2 may show different variation trends, and the processing device may determine, according to the results, a model in which the accuracy gradient of the results reaches an accuracy gradient threshold or the loss function gradient is lower than a preset loss gradient threshold from the model 1 and the model 2, so as to obtain the correct intention of the abnormal sample according to the determined model.
S306: and building an countermeasure network model based on the abnormal samples after the sample expansion.
S307: and determining the correct intention of the abnormal sample according to the output result of the countermeasure network model.
Steps S306 to S307 refer to the related descriptions of steps S203 to S204, and are not described herein.
In order to meet application requirements in different application scenes, the embodiment of the application also considers that when the abnormal samples are subjected to sample expansion, the abnormal samples are subjected to sample expansion based on different modes, for example, training samples are expanded by methods such as mutual translation, synonym replacement and the like, so that model generalization capability is improved, and an overfitting phenomenon is effectively avoided, and therefore, intention recognition is performed based on the expanded samples, and the accuracy of the intention recognition is improved. In addition, the embodiment of the application builds the countermeasure network model based on the expanded sample, automatically selects the correct intention for the sample, and can reduce the manual processing amount and improve the efficiency.
In the embodiment of the present application, as shown in fig. 5, the processing apparatus may first perform intent recognition on a to-be-processed dialog sample using a preset intent recognition model, and then obtain types of each intent recognized by the preset intent recognition model and confidence levels of each intent type, and further determine an abnormal sample from the to-be-processed dialog sample according to the confidence levels of each intent type and a preset confidence threshold value. And carrying out sample expansion on the abnormal sample, and building an countermeasure network model based on the abnormal sample after sample expansion, so as to determine the correct intention of the abnormal sample according to the output result of the countermeasure network model.
Compared with the prior art, the processing device increases the model generalization capability by expanding the data of the sample, and effectively avoids the over-fitting phenomenon, so that the intention recognition is performed based on the expanded sample, and the accuracy of the intention recognition is improved. In addition, the embodiment of the application builds the countermeasure network model based on the expanded sample, automatically selects the correct intention for the sample, and can reduce the manual processing amount and improve the efficiency.
Fig. 6 is a schematic structural diagram of an intent recognition device according to an embodiment of the present application, corresponding to the intent recognition method of the above embodiment. For convenience of explanation, only portions relevant to the embodiments of the present application are shown. Fig. 6 is a schematic structural diagram of an intention recognition device according to an embodiment of the present application, where the intention recognition device 60 includes: an information obtaining module 601, a sample determining module 602, a sample processing module 603, and an intention determining module 604. The intention recognition means may be the processing means itself or a chip or an integrated circuit implementing the functions of the processing means. Here, the division of the information obtaining module, the sample determining module, the sample processing module, and the intention determining module is only a division of a logic function, and both may be integrated or independent physically.
The information obtaining module 601 is configured to obtain, after performing intent recognition on a dialog sample to be processed using a preset intent recognition model, types of respective intentions recognized by the preset intent recognition model, and confidence levels of each of the types of the respective intentions.
The sample determining module 602 is configured to determine an abnormal sample from the dialog samples to be processed according to the confidence level of each intention type and a preset confidence threshold, where the abnormal sample includes a plurality of samples.
The sample processing module 603 is configured to perform sample expansion on the abnormal sample, and build an countermeasure network model based on the abnormal sample after sample expansion.
An intention determining module 604, configured to determine a correct intention of the abnormal sample according to an output result of the countermeasure network model.
In one possible implementation, the sample determining module 602 is specifically configured to:
comparing the confidence level of each intention type with the preset confidence level threshold value respectively;
if the confidence coefficient of the intention type i is smaller than the preset confidence coefficient threshold value, determining that a dialogue sample corresponding to the intention type i is the abnormal sample, and acquiring the abnormal sample from the dialogue sample to be processed, wherein the intention type i is any one intention type in the intention types, and i=1, 2, …, n and n are equal to the number of the intention types.
In one possible implementation, the sample processing module 603 is specifically configured to:
converting the abnormal sample into a sample in a preset language, and converting the sample in the preset language into a sample in a language corresponding to the abnormal sample, wherein the preset language is other languages except the language corresponding to the abnormal sample;
or alternatively
Randomly replacing the vocabulary in the abnormal sample according to the synonym library;
or alternatively
And carrying out batch copy pasting on the vocabulary in the abnormal sample.
In one possible implementation, the abnormal samples include a first sample and a second sample
The sample processing module 603 is specifically configured to:
adding the sample after the expansion of the first sample into the first sample to obtain a third sample, and adding the sample after the expansion of the second sample into the second sample to obtain a fourth sample;
constructing an intention recognition model based on the third sample and the fourth sample respectively;
obtaining the countermeasure network model based on the constructed intention recognition model
In one possible implementation, the intent determination module 604 includes:
determining a model with a result accuracy gradient reaching an accuracy gradient threshold or a loss function gradient lower than a preset loss gradient threshold from the countermeasure network model according to the output result of the countermeasure network model;
and obtaining the correct intention of the abnormal sample according to the determined model.
The device provided by the embodiment of the application can be used for executing the technical scheme of the embodiment of the method, and the implementation principle and the technical effect are similar, and the embodiment of the application is not repeated here.
Alternatively, fig. 7 schematically provides one possible basic hardware architecture of the intent recognition device according to the application, respectively.
Referring to fig. 7, the intent recognition device comprises at least one processor 701 and a communication interface 703. Further optionally, a memory 702 and a bus 704 may also be included.
Wherein the number of processors 701 may be one or more in the intent recognition device, fig. 7 illustrates only one of the processors 701. Optionally, the processor 701 may be a central processing unit (central processing unit, CPU), a graphics processor (graphics processing unit, GPU) or a digital signal processor (digital signal processor, DSP). If the intention recognition apparatus has a plurality of processors 701, the types of the plurality of processors 701 may be different or may be the same. Optionally, the multiple processors 701 of the intent recognition device may also be integrated as a multi-core processor.
Memory 702 stores computer instructions and data; the memory 702 may store computer instructions and data necessary to implement the above-described intent recognition method provided by the present application, for example, the memory 702 stores instructions for implementing the steps of the above-described intent recognition method. Memory 702 may be any one or any combination of the following storage media: nonvolatile memory (e.g., read Only Memory (ROM), solid State Disk (SSD), hard disk (HDD), optical disk), volatile memory).
The communication interface 703 may provide information input/output for the at least one processor. Any one or any combination of the following devices may also be included: a network interface (e.g., ethernet interface), a wireless network card, etc., having network access functionality.
Optionally, the communication interface 703 may also be used for data communication with other computing devices or terminals by the intent recognition device.
Further alternatively, FIG. 7 illustrates bus 704 with a bold line. A bus 704 may connect the processor 701 with the memory 702 and the communication interface 703. Thus, through bus 704, processor 701 may access memory 702 and may also interact with other computing devices or terminals using communication interface 703.
In the present application, the intention recognition apparatus executes the computer instructions in the memory 702, so that the intention recognition apparatus implements the above-described intention recognition method provided by the present application, or so that the intention recognition apparatus deploys the above-described intention recognition device.
From a logical functional partitioning perspective, as illustrated in fig. 7, the memory 702 may include an information obtaining module 601, a sample determining module 602, a sample processing module 603, and an intention determining module 604. The inclusion herein is not limited to a physical structure, and may involve only the functions of the information obtaining module, the sample determining module, the sample processing module, and the intention determining module, respectively, when the instructions stored in the memory are executed.
The intention recognition device may be realized by hardware as a hardware module or as a circuit unit, in addition to the software as in fig. 7.
The present application provides a computer-readable storage medium, the computer program product comprising computer instructions that instruct a computing device to perform the above-described intent recognition method provided by the present application.
An embodiment of the present application provides a computer program product including computer instructions that are executed by a processor to perform the above-described intent recognition method provided by the present application.
The application provides a chip comprising at least one processor and a communication interface providing information input and/or output for the at least one processor. Further, the chip may also include at least one memory for storing computer instructions. The at least one processor is configured to invoke and execute the computer instructions to perform the above-described intent recognition method provided by the present application.
In the several embodiments provided by the present application, it should be understood that the disclosed apparatus and method may be implemented in other manners. For example, the apparatus embodiments described above are merely illustrative, e.g., the division of the units is merely a logical function division, and there may be additional divisions when actually implemented, e.g., multiple units or components may be combined or integrated into another system, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, which may be in electrical, mechanical or other form.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.

Claims (5)

1. An intent recognition method, comprising:
after intention recognition is carried out on a dialogue sample to be processed by adopting a preset intention recognition model, obtaining the types of each intention recognized by the preset intention recognition model and the confidence coefficient of each intention type in the types of each intention;
comparing the confidence coefficient of each intention type with a preset confidence coefficient threshold value respectively;
if the confidence coefficient of the intention type i is smaller than the preset confidence coefficient threshold value, determining that a dialogue sample corresponding to the intention type i is an abnormal sample, and acquiring the abnormal sample from the dialogue sample to be processed, wherein the intention type i is any one of the intention types, i=1, 2, …, n and n are equal to the number of the intention types, the abnormal sample comprises a plurality of samples, and the abnormal sample comprises a first sample and a second sample;
sample expansion is carried out on the abnormal sample, and
adding the sample after the expansion of the first sample into the first sample to obtain a third sample, and adding the sample after the expansion of the second sample into the second sample to obtain a fourth sample;
constructing an intention recognition model based on the third sample and the fourth sample respectively;
obtaining an countermeasure network model based on the constructed intention recognition model;
determining the correct intention of the abnormal sample according to the output result of the countermeasure network model;
the sample expansion of the abnormal sample comprises:
converting the abnormal sample into a sample in a preset language, and converting the sample in the preset language into a sample in a language corresponding to the abnormal sample, wherein the preset language is other languages except the language corresponding to the abnormal sample;
or alternatively, the process may be performed,
randomly replacing the vocabulary in the abnormal sample according to the synonym library;
or alternatively, the process may be performed,
and carrying out batch copy pasting on the vocabulary in the abnormal sample.
2. The method of claim 1, wherein determining the correct intent of the anomaly sample based on the output of the challenge network model comprises:
determining a model with a result accuracy gradient reaching an accuracy gradient threshold or a loss function gradient lower than a preset loss gradient threshold from the countermeasure network model according to the output result of the countermeasure network model;
and obtaining the correct intention of the abnormal sample according to the determined model.
3. An intent recognition device, comprising:
the information acquisition module is used for acquiring the types of each intention identified by the preset intention identification model and the confidence of each intention type in the types of each intention after the intention identification is carried out on the dialogue sample to be processed by adopting the preset intention identification model;
a sample determination module for
Comparing the confidence coefficient of each intention type with a preset confidence coefficient threshold value respectively;
if the confidence coefficient of the intention type i is smaller than the preset confidence coefficient threshold value, determining that a dialogue sample corresponding to the intention type i is an abnormal sample, and acquiring the abnormal sample from the dialogue sample to be processed, wherein the intention type i is any one of the intention types, i=1, 2, …, n and n are equal to the number of the intention types, the abnormal sample comprises a plurality of samples, and the abnormal sample comprises a first sample and a second sample;
a sample processing module for expanding the abnormal sample and
adding the sample after the expansion of the first sample into the first sample to obtain a third sample, and adding the sample after the expansion of the second sample into the second sample to obtain a fourth sample;
constructing an intention recognition model based on the third sample and the fourth sample respectively;
obtaining an countermeasure network model based on the constructed intention recognition model;
the intention determining module is used for determining the correct intention of the abnormal sample according to the output result of the countermeasure network model;
the sample processing module is specifically configured to:
converting the abnormal sample into a sample in a preset language, and converting the sample in the preset language into a sample in a language corresponding to the abnormal sample, wherein the preset language is other languages except the language corresponding to the abnormal sample;
or alternatively, the process may be performed,
randomly replacing the vocabulary in the abnormal sample according to the synonym library;
or alternatively, the process may be performed,
and carrying out batch copy pasting on the vocabulary in the abnormal sample.
4. An intent recognition device, comprising:
a processor;
a memory; and
a computer program;
wherein the computer program is stored in the memory and configured to be executed by the processor, the computer program comprising instructions for performing the method of claim 1 or 2.
5. A computer readable storage medium, characterized in that the computer readable storage medium stores a computer program, which causes a server to execute the method of claim 1 or 2.
CN202110991234.7A 2021-08-26 2021-08-26 Intention recognition method, device, equipment and storage medium Active CN113569581B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110991234.7A CN113569581B (en) 2021-08-26 2021-08-26 Intention recognition method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110991234.7A CN113569581B (en) 2021-08-26 2021-08-26 Intention recognition method, device, equipment and storage medium

Publications (2)

Publication Number Publication Date
CN113569581A CN113569581A (en) 2021-10-29
CN113569581B true CN113569581B (en) 2023-10-17

Family

ID=78172927

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110991234.7A Active CN113569581B (en) 2021-08-26 2021-08-26 Intention recognition method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN113569581B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117076654B (en) * 2023-10-18 2024-02-27 联通在线信息科技有限公司 Abnormality detection method and device for dialogue system

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109977401A (en) * 2019-03-15 2019-07-05 上海火商智能科技有限公司 A kind of method for recognizing semantics neural network based
CN111209384A (en) * 2020-01-08 2020-05-29 腾讯科技(深圳)有限公司 Question and answer data processing method and device based on artificial intelligence and electronic equipment
CN111626063A (en) * 2020-07-28 2020-09-04 浙江大学 Text intention identification method and system based on projection gradient descent and label smoothing
CN112154465A (en) * 2018-09-19 2020-12-29 华为技术有限公司 Method, device and equipment for learning intention recognition model
CN112446213A (en) * 2020-11-26 2021-03-05 电子科技大学 Text corpus expansion method
CN112818096A (en) * 2021-01-26 2021-05-18 广州三星通信技术研究有限公司 Dialog generating method and device
CN112905795A (en) * 2021-03-11 2021-06-04 证通股份有限公司 Text intention classification method, device and readable medium
CN112989035A (en) * 2020-12-22 2021-06-18 平安普惠企业管理有限公司 Method, device and storage medium for recognizing user intention based on text classification
CN113094481A (en) * 2021-03-03 2021-07-09 北京智齿博创科技有限公司 Intention recognition method and device, electronic equipment and computer readable storage medium
CN113139063A (en) * 2021-06-21 2021-07-20 平安科技(深圳)有限公司 Intention recognition method, device, equipment and storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3619620A4 (en) * 2017-06-26 2020-11-18 Microsoft Technology Licensing, LLC Generating responses in automated chatting

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112154465A (en) * 2018-09-19 2020-12-29 华为技术有限公司 Method, device and equipment for learning intention recognition model
CN109977401A (en) * 2019-03-15 2019-07-05 上海火商智能科技有限公司 A kind of method for recognizing semantics neural network based
CN111209384A (en) * 2020-01-08 2020-05-29 腾讯科技(深圳)有限公司 Question and answer data processing method and device based on artificial intelligence and electronic equipment
CN111626063A (en) * 2020-07-28 2020-09-04 浙江大学 Text intention identification method and system based on projection gradient descent and label smoothing
CN112446213A (en) * 2020-11-26 2021-03-05 电子科技大学 Text corpus expansion method
CN112989035A (en) * 2020-12-22 2021-06-18 平安普惠企业管理有限公司 Method, device and storage medium for recognizing user intention based on text classification
CN112818096A (en) * 2021-01-26 2021-05-18 广州三星通信技术研究有限公司 Dialog generating method and device
CN113094481A (en) * 2021-03-03 2021-07-09 北京智齿博创科技有限公司 Intention recognition method and device, electronic equipment and computer readable storage medium
CN112905795A (en) * 2021-03-11 2021-06-04 证通股份有限公司 Text intention classification method, device and readable medium
CN113139063A (en) * 2021-06-21 2021-07-20 平安科技(深圳)有限公司 Intention recognition method, device, equipment and storage medium

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
基于神经网络的汉语asr识别对话短文本意图匹配;周新宇;中国优秀硕士学位论文全文数据库信息科技辑(第4期);I138-1073 *
基于迁移学习的意图识别方法研究;赵鹏飞;中国优秀硕士学位论文全文数据库信息科技辑(第8期);I140-94 *

Also Published As

Publication number Publication date
CN113569581A (en) 2021-10-29

Similar Documents

Publication Publication Date Title
CN106887225B (en) Acoustic feature extraction method and device based on convolutional neural network and terminal equipment
CN110673748B (en) Method and device for providing candidate long sentences in input method
KR20220122566A (en) Text recognition model training method, text recognition method, and apparatus
CN108304376B (en) Text vector determination method and device, storage medium and electronic device
CN114942984B (en) Pre-training and image-text retrieval method and device for visual scene text fusion model
CN113836303A (en) Text type identification method and device, computer equipment and medium
CN113569581B (en) Intention recognition method, device, equipment and storage medium
CN116704519A (en) Character recognition method, character recognition device, electronic equipment and storage medium
CA3147341A1 (en) Category phrase recognition method, model training method, device and system
CN113516972B (en) Speech recognition method, device, computer equipment and storage medium
CN117648422A (en) Question-answer prompt system, question-answer prompt, library construction and model training method and device
CN112883736A (en) Medical entity relationship extraction method and device
CN112906368A (en) Industry text increment method, related device and computer program product
CN115186738B (en) Model training method, device and storage medium
CN109002498B (en) Man-machine conversation method, device, equipment and storage medium
CN114863450B (en) Image processing method, device, electronic equipment and storage medium
CN115080745A (en) Multi-scene text classification method, device, equipment and medium based on artificial intelligence
US11797277B2 (en) Neural network model conversion method server, and storage medium
CN115269767A (en) Model training method, device and storage medium
US20230342553A1 (en) Attribute and rating co-extraction
CN115881103B (en) Speech emotion recognition model training method, speech emotion recognition method and device
CN113836939B (en) Text-based data analysis method and device
CN116310682A (en) Event aggregation method, device and equipment based on multi-mode data
CN115035911A (en) Noise generation model training method, device, equipment and medium
CN117688909A (en) Method, apparatus, electronic device, and computer-readable medium for processing regulatory information

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant