CN110298391A - A kind of iterative increment dialogue intention classification recognition methods based on small sample - Google Patents

A kind of iterative increment dialogue intention classification recognition methods based on small sample Download PDF

Info

Publication number
CN110298391A
CN110298391A CN201910505469.3A CN201910505469A CN110298391A CN 110298391 A CN110298391 A CN 110298391A CN 201910505469 A CN201910505469 A CN 201910505469A CN 110298391 A CN110298391 A CN 110298391A
Authority
CN
China
Prior art keywords
model
classification
indicate
preliminary
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910505469.3A
Other languages
Chinese (zh)
Other versions
CN110298391B (en
Inventor
向阳
单光旭
贾圣宾
徐诗瑶
杨力
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tongji University
Original Assignee
Tongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tongji University filed Critical Tongji University
Priority to CN201910505469.3A priority Critical patent/CN110298391B/en
Publication of CN110298391A publication Critical patent/CN110298391A/en
Application granted granted Critical
Publication of CN110298391B publication Critical patent/CN110298391B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The present invention relates to a kind of, and the iterative increment dialogue based on small sample is intended to classification recognition methods, this method is based on Small Sample Database collection, it is trained since a preliminary classification model, with the use of model, the quantity of rudimentary model is increasing, model accuracy rate also steps up, the training method that previous deep learning model needs great amount of samples is abandoned, preliminary classification model of this method during repetitive exercise due to only needing a small amount of sample training one new every time, other existing history preliminary classification Model Weights are constant, then it will be trained in the result input of whole preliminary classification models again disaggregated model, the calculating speed of model will not be reduced with the increase of sample size, similarity screening model can be screened and be rejected to existing preliminary classification model simultaneously, performance is maintained in the case where guaranteeing accuracy rate , compared with prior art, the advantages that present invention has training samples number few, and calculated performance is stablized, model easily updated extension.

Description

A kind of iterative increment dialogue intention classification recognition methods based on small sample
Technical field
The present invention relates to field of computer technology, talk with more particularly, to a kind of iterative increment based on small sample and are intended to Classification recognition methods.
Background technique
For the quality for promoting products & services, many companies are proposed the customer service system of oneself, are helped by artificial customer service User answers a question, and so that user is enjoyed more perfect service, improves service quality and efficiency, but with product user quantity Increase, traditional artificial customer service is not able to satisfy the demand of numerous users, and artificial customer service needs special to carry out business Training and learning also brings certain cost, and customer service hotline is chronically at the state that the line is busy, will affect the Experience Degree of user.Cause This each major company is proposed the artificial customer service product of oneself, by the chat message between study history customer service and user, extracts Intent information contained in dialogue out can help user to solve relevant business faster.
Dialogue intention assessment is exactly to pass through to understand interpersonal chat corpus information, to the intent features in text into Row retrieval, filtering and classification etc. finally identify the purpose even emotion that user session is included, it is intended that the core of identification is pair Semantic understanding.Dialogue intension recognizing method based on machine learning includes rule-based and statistics dialogue intention assessment, base In the dialogue intention assessment and the dialogue intention assessment based on production model etc. of Machine learning classifiers.Telecommunications industry uses Intelligent customer service robot the business tine handled required for user quickly can be understood by the dialogue with user, provide use Family options mitigates the retrieval burden of user.
In recent years, with the gradually development of deep learning, more and more scientific & technical corporation are proposed the chat product of oneself, For example the Siri of Apple Inc., Microsoft little Na and the voice assistant of Iflytek etc., these interactive purposes are final It is the intention for being appreciated that user, gives user feedback and user is helped to enjoy preferably service, these, which all be unable to do without, is intended to user Recognizer.But often some chat robots due to itself learning ability it is poor, be difficult to understand for that user is profound to be asked Topic, cause to occur giving an irrelevant answer the phenomenon that even circulation is answered, therefore there are also to be improved and raisings for current intention assessment algorithm.
The algorithm of comparative maturity has rule-based matched intention assessment algorithm in dialogue recent years intention assessment, is based on The document classification algorithm of probability statistics model, and SVM (support vector machines) is used, KNN (k- neighborhood), the text of decision-tree model This sorting algorithm etc..Information of the rule-based algorithm often by keyword in statistics text, rough supposition user Intention, this mode classification is slower in the biggish inquiry under condition of data volume, and needs artificial mark, takes time and effort;Base It is higher for the quality and Spreading requirements of text in the document classification algorithm of probability statistics model, and in the case where small sample Classification is inaccurate;Text classification algorithm based on machine learning classification model is preferable for short text classifying quality, for long text For be difficult capture context of dialogue information, be easy to appear the case where giving an irrelevant answer, and new expectation is needed to instruct again Practice model, with increasing for number of samples, training complexity is larger.
Summary of the invention
It is an object of the present invention to overcome the above-mentioned drawbacks of the prior art and provide a kind of based on small sample Iterative increment dialogue is intended to classification recognition methods.
The purpose of the present invention can be achieved through the following technical solutions:
A kind of iterative increment dialogue intention classification recognition methods based on small sample, which includes following step It is rapid:
Step 1: segment for the text sentence in dialogue intention and training obtains term vector;
Step 2: obtained after successively being extracted using LSTM network and CNN network characterization for a part of term vector sentence to Another part term vector is input to what training finished by sentence vector by taxon to train preliminary classification model by amount Preliminary classification model obtains preliminary classification results;
Step 3: for preliminary classification results, passing through XGBoost after being screened according to the return of intensified learning model Model carries out secondary classification, and trains disaggregated model again using the method for gradient decline;
Step 4: using the classification error sample in entire disaggregated model training process re-enter training one it is new just Disaggregated model is walked, return step 3 is iterated incremental learning;
Step 5: sharp when the quantity of preliminary disaggregated model rises to preset threshold in updating iterative increment learning process Similarity calculation is carried out with the arbiter model that training finishes, and removes one of highest two models of similarity to maintain Overall model computational stability;
Step 6: circulation executes step 3 to step 5, and gradual perfection classification results simultaneously obtain finally corresponding to identification types knot Fruit.
Further, in the preliminary classification results in the step 2 each classification output probability are as follows:
In formula, P (Xi|Dk) indicate the output probability of each classification, XiIndicate actual classification as a result, Xi' indicate prediction point Class is as a result, X indicates all classification results.
Further, the loss function in the preliminary classification results in the step 2 are as follows:
In formula, y1Indicate the loss function in preliminary classification model output result, η1For adjustment parameter, P indicates true intention Class label probability, P' indicate prediction class label probability.
Further, the calculation formula of the return in the step 3 are as follows:
In formula, R indicates return, riIndicate that n and i are natural number, and ξ takes 0.95 when time return.
Further, the taxon in the step 2 is full Connection Neural Network, preliminary in the step 2 Disaggregated model includes mutual sequentially connected sentence vector extract layer, talks with vector extract layer and layer of classifying, in the step 3 Disaggregated model includes intensified learning interconnected-model discrimination module and decision tree-categorization module again, in the step 5 Arbiter model be full Connection Neural Network.
Further, preliminary classification results are directed in the step 3, the process screened according to return is corresponding Loss function are as follows:
y3=p (Ai|F4i)*log(p'(Ai|F4i))
In formula, y3Indicate the corresponding loss letter of process screened for preliminary classification model output result according to return Number, p (Ai|F4i) indicate under being corresponded in loss function for preliminary classification model output result according to the process that return is screened The probability of one movement, p'(Ai|F4i) indicate the process pair screened for preliminary classification model output result according to return Answer the probability of realistic operation in loss function.
Further, loss function is corresponded to by the process that XGBoost model carries out secondary classification in the step 3 Are as follows:
In formula, y4It indicates to correspond to loss function, O by the process that XGBoost model is further classifiedi' indicate to pass through The process that XGBoost model is further classified corresponds to the selection result of first stage in loss function, Oi" indicate to pass through The process that XGBoost model is further classified corresponds to further sorted result in loss function.
Further, arbiter model carries out similarity calculation in the step 5, and removes similarity highest two The process of one of model corresponds to loss function are as follows:
In formula, y5Indicate that the process of arbiter model progress similarity screening corresponds to loss function, K is indicated from existing instruction Practice the quantity that a part of data set is randomly selected in sample, pkIndicate that arbiter model carries out the corresponding damage of process of similarity screening Lose the legitimate reading in function, pk' indicate that the process of arbiter model progress similarity screening corresponds to the prediction in loss function As a result.
Step 1 and step 2 described above can integrate summary are as follows: carry out unsupervised instruction to participle using language model first Practice, obtain the term vector of fixed dimension, then the term vector in each sentence is input in LSTM network and carries out feature extraction, Extract obtained sentence vector and dimensionality reduction operation carried out using CNN, then using above-mentioned same process to the sentence in dialogue into Row feature extraction, the vector entirely talked with indicate, can preferably obtain dialogue by the way of secondary characteristics extraction Between contextual information;
Intensified learning method refers to step 3 in disaggregated model again: intensified learning mode first export on last stage to Amount then according to current state selects next movement as state, if the prediction effect of next small sample model compared with It is good, then it is added to prediction result concentration, is otherwise abandoned;The return of intensified learning model judges according to model prediction result, root The maximum small sample model set of integral benefit is selected according to the above method, this method can be improved the predictablity rate of model;It will The prediction result for the small sample model set that above-mentioned intensified learning model discrimination obtains is input in the model, and prediction label is most The tag along sort that dialogue is intended to eventually, the training by the way of gradient decline.
Iterative delta algorithm process in step 4 is as follows: first according to existing small sample model training preliminary classification mould Then type trains disaggregated model again for the result of preliminary classification model as the input of disaggregated model again, then has new mould When type is added, all preliminary classification models of re -training are not needed, but new preliminary classification model and original model is defeated Be input in disaggregated model again re -training disaggregated model again together out, due to single preliminary classification model training with again The training time of subseries model is short, can save the cycle of training of entire model, reduces time complexity, improves making for model Use performance;
Similarity is screened in step 5 specifically: with being increasing for new model, the training that preliminary classification model generates is tied Fruit constantly increases, and in order to control the service performance of entire model, and in the case where guarantee model accuracy rate, needs to existing mould Type is screened, therefore the similarity of result is exported by calculating different models, can choose reservation similarity higher two One of them in model enhances the generalization ability of model, reduces redundancy.
Compared with prior art, the invention has the following advantages that
(1) first two steps first are rapid in the method for the present invention: step 1: being segmented simultaneously for the text sentence in dialogue intention Training obtains term vector;Step 2: being obtained after successively being extracted using LSTM network and CNN network characterization for a part of term vector Another part term vector is input to training to train preliminary classification model by taxon by sentence vector by sentence vector The preliminary classification model finished obtains preliminary classification model output result;Training for small sample, can be preliminary from one Naive model starts, and is gradually promoted then as the use training accuracy rate of model, and combines existing model, abandoned with The training method of great amount of samples is needed toward deep learning model.
(2) screening of the intensified learning strategy for preliminary classification model result in step 3 in the method for the present invention, guarantees first Global optimum, avoids the excessive classification results to disaggregated model again of single Model Weight from impacting, equally subseries again Model adjusts the classification results of preliminary classification model, and personalized selection preliminary classification category of model is as a result, protect The robustness of model is demonstrate,proved.
(3) the iterative incremental learning model in the method for the present invention in step 4 only needs a small amount of sample training due to trained Last preliminary classification model, other Model Weights are constant to still carry out output, then further updates disaggregated model again Weight, in calculating speed will not by sample size increase and reduce, while similarity screening model can to existing model into Row screening and rejecting guarantee to maintain performance in the case that accuracy rate is stablized.
(4) learning strategy of iterative increment, mistake sample are used on the framework of the entire model in the method for the present invention in step 4 This continues to train the performance for promoting entire model, guarantees that scene rare in intention assessment scene can also make reliable meaning Figure classification, while function is completed in combination between multiple modules on framework, the degree of coupling is lower, can train in a distributed manner, single model Facilitate replacement and update, is easy to extend.
Detailed description of the invention
Fig. 1 is general frame schematic diagram of the invention;
Fig. 2 is specific implementation network structure of the invention.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete Site preparation description, it is clear that described embodiment is a part of the embodiments of the present invention, rather than whole embodiments.Based on this hair Embodiment in bright, those of ordinary skill in the art's every other reality obtained without making creative work Example is applied, all should belong to the scope of protection of the invention.
It is a kind of based on a small amount of sample the object of the invention is to be provided to solve defect existing for above-mentioned existing method Training strategy, accuracy rate constantly promotes, and the degree of coupling is lower between disparate modules, is easy to extend, and introducing intensified learning strategy can It is preferably promoted with the model to primary learning, finally the screening to existing model and filtering are able to maintain the stability of model, Technological frame is as shown in Figure 1.
The invention mainly comprises two big modules, first is that the instruction of the dialogue intention assessment model of the iterative increment based on small sample Practice module, second is that the use module of the dialogue intention assessment model of the iterative increment based on small sample.Wherein training module includes Three models: first is that for the model one tentatively classified to Small Sample Database collection, second is that preliminary classification result again into The model two of row classification, third is that carrying out the model three of screening and filtering to model.It the use of module is to the trained mould of above three Type carry out using, and can be carried out iterative increment study, step up model performance.
Specific module architectures are as follows:
First part is the training of model: model one uses the stratification feature extraction algorithm based on deep learning, first Corpus of text is input in language model and carries out unsupervised training, obtains term vector;Then M sentence in dialogue is carried out Participle, is converted into term vector < W for N number of word in sentence1,W2,...,WN>, after obtaining term vector, Small Sample Database Collection is divided into Da and Db two parts, and wherein Da data set is used for training pattern one, and training process is as follows, in wheel dialogues more in Da Term vector corresponding to single sentence is input in two-way LSTM (shot and long term neural network), in conjunction with attention mechanism, before obtaining To the output vector of LSTMWith the output vector of consequent LSTMThen by vector into Row splicing obtains feature vector corresponding to wordUse CNN (convolutional neural networks) later By sentence vector carry out dimensionality reduction operation, obtain each sentence for feature vector Ck(1≤k≤M), in entirely talking with M sentence, can by the above method carry out feature extraction, obtain sentence vector < C1,C2,...,CN>, then using complete Connection Neural Network extracts sentence, and by four layers of full articulamentum, every layer of output is F1~F4, the feature of the last layer Vector layer output is F4, represents the vector of current entire sentence, if final intention classification is X, then F4 access X ties up Quan Lian It connects layer and is divided into X class intention, training process is for classification X actual in sampleiWith the classification X of classificationi' using the side of supervised learning Formula, network function are indicated with f1, are declined the weight of reversed adjustment model by gradient, are trained single model, be denoted as model 1-1 ~model 1-n, the output of model are O=< O1,O2,...,ON>;
Model two screens the classification results in model one using intensified learning mode, it is assumed that has trained mould Db data set is input to and obtains the 4th layer of output vector F4 and final defeated in 1~n of model by n model in type one Outgoing vector O, taking 1-i here is i-th of model in model one, and 1≤i≤n-2, wherein the last layer in model 1-i is special The output vector F4 of layer is levied as current state Si, the output vector F4 in model 1- (i+1) is as next state, it is assumed that The output result of model 1- (i+1) is Xi+1', true classification results are Xi+1If Xi+1'=Xi+1, then classification is correct, when Preceding return ri=1, act Ai=+1, next state is converted into i+1, if Xi+1’≠Xi+1, then return ri=0, movement Ai=+2, next state is converted into i+2.It repeats the above process and is learnt, until obtaining maximum return R.Model Based on the state S and movement A saved in history in two training process, it is assumed that the weight of network is θ, and network function is indicated with f2, Income indicates that f2 backpropagation is trained, so that Q network reaches stable state with Q.Model two is to the above-mentioned result O' screened =< O'1,O'2,...,O'NThe classification of > further progress, wherein O' ∈ O, determines to the above results by XGBoost model Plan obtains result of decision O ", according to really talking with corresponding intention classificationThe result O " obtained with two decision of model is carried out Comparison, the same weight that two XGBoost classifier of model is adjusted by the way of gradient decline.
Model is third is that being eliminated and being screened to the single mini Mod in model one, from existing historical sample data D A part of sample Dc is screened, is input in model 1-1 and model 1-n, produce output result simultaneously obtains one taxon of model Then F4 is input in arbiter model f3 by the 4th layer of output vector F4, arbiter model is neural by four layers of full connection Network composition, input sample is the output vector F4 of two models one, it is assumed that is F4iAnd F4j, and (1≤i ≠ j≤n), sample is set It is set to (F4i, i) and (F4j, j), which model softmax layer of the output by two classification currently exports for judging from, By there is the training method of supervision, 1-1~1-n, model are trained, a more perfect arbiter model is obtained;
Second part is the use of model: sample D ' in practice being input in 1-n model one first, obtains model One classification results O and the 4th layer of full connection output vector F4, then as the input of two intensified learning model of model, K output is obtained as a result, and then obtaining final classification in the decision tree classifier of K output result access model two, it is right Words are intended in scene, are intended to the data of misinterpretation for user, need to turn artificial treatment, the intention classification of final artificial treatment Intention as processing mistake understands label, then by after the result queue of classification error, obtains new data set E, same point For Ea and Eb, model 1- (n+1) is denoted as using the sample E one new model one of training for having label, then new training sample This Eb, which is input in n+1 model one, to be obtained as a result, using the result of n+1 model as the training sample of model two into one in turn Walk training pattern two, the step for only have trained model two and model 1-n+1, original model 1-1~1-n, there is no again It updates, the training time is greatly saved;And the quantity of model one gradually increases with the accumulation of sample, waits until one quantity of model Certain threshold value δ is risen to, existing preceding n model is screened using trained model three, from 1- (n-1) a model In select with the maximum model 1-i ' of model 1-n similarity, then remove existing model 1-i, in this way holding model quantity dimension It holds at n, so that the calculated performance of model will not be reduced because model number is excessive.
The specific embodiment of iterative increment dialogue intension recognizing method based on small sample is as follows:
It is that telecommunications customer service intent data is divided first, customer service is intended to be divided into 50 according to industry actual needs Class (X takes 50), customer service data from the true customer service incoming call text that is converted into of dialogue it is anticipated that the extra greeting sentence of removal, It is 10 wheels that number is averagely taken turns in the important dialogue of the dialogue, and each round includes two sentences, therefore a customer service sample is about by 20 Sentence, an intention labels composition, each sentence are on a rough average in 5-10 words or so after participle.Here it chooses first 1000 data collection are divided into 10 groups, each group contains 100 datas, and each group of data are again by 1000 Small Sample Database collection It is secondary to be divided into Da data set and Db data set.Each group data set uses 70% training set as single model, 30% conduct The effect of verifying the set pair analysis model is assessed, until accuracy rate of the model on verifying collection reaches convergence, it is believed that model training It finishes.
The structure of model one is as shown in Fig. 2, training process in detail is as follows: first splitting into the dialogue of existing training sample Single different sentence obtains corresponding to each word using these sentences as input training one into language model Term vector is denoted as W, and the dimension of each term vector is 100 dimensions.It may finally obtain the term vector < W of entire sentence1,W2,...,WN >, N takes 10 here, and number of words is filled less than N with 0, extra direct to cut out the subsequent part of n-th word.It then will be whole The sentence vector < W of a sentence1,W2,...,WN> is input in LSTM model, and LSTM model is using double based on attention mechanism To LSTM, the vector that model each in this way obtains isThe corresponding output vector of each word is 200 dimensions, since the length of sentence is 10, so the vector dimension that entire sentence obtains in model is 2000 dimensions.2000 dimensions LSTM mode input into three layers of convolutional neural networks, the setting of the design parameter of convolutional network is as follows, first first layer general 2000 dimension vectors as input, convolution kernel size be 2*1, sliding step 2, this results in one 1000 tie up output to Amount, then by second layer pond layer, the size of pond layer network is 2*1, and sliding step 2 has obtained the defeated of one 500 dimension Outgoing vector, then by the output vector of 500 dimensions, then by there is one layer of convolution, convolution kernel size ties up 5*1, and step-length 5 obtains The output vector of one 100 dimension, terminates the feature extraction of sentence, obtains vector < C1,C2,...,CN>.
For entirely talking with, corresponding 100 dimensional vector of available each sentence through the above way, then by 20 The corresponding sentence vector < C of a sentence1,C2,...,CN> is input in another two-way LSTM based on attention mechanism, is led to The output vector that LSTM feedforward network obtains the output vector of one 100 dimension and backward network obtains one 100 dimension is crossed to combine It arrivesThen two vectors are subjected to direct splicing and obtain the vector of one 200 dimension, because By 20 sentences, whole vectors is spliced the vector of available one 4000 dimension, be then linked into full Connection Neural Network Classify in module, the design method of full Connection Neural Network is as follows: being that one layer of 2048 neuron form first Then network layer is added in dropout layers, dropout ratio is set as 0.2, then accesses the full articulamentum of one 1024 dimension, Then it is linked into the full articulamentum of 512 dimensions, after being standardized using Batch Normalize (batch standardization) layer, access In the full articulamentum tieed up to one 100, the output result obtained at this time is denoted as F4, and F4 is then linked into one 50 dimension In softmax, classification output, therefore available sample D are carried outkFor the output probability P (X of each classificationi|Dk) are as follows:
Here setting loss function is arranged are as follows:
In formula, y1Indicate the loss function in preliminary classification model output result, η1For adjustment parameter, value 0.6 is used In regulation loss function, P indicates that true intention class label probability, P' indicate prediction class label probability.
Using gradient decline by the way of model is trained, until the predictablity rate of model reach 75% or more or Model loss function during 10 iteration reaches stabilization can deconditioning.
Followed by the first module of model two, intensified learning result screening module, specific embodiment as shown in Fig. 2, By the sample data in above-mentioned training set Db, it is input in trained model 1-1 to model 1-n, passes through some mould Type 1-i, (1≤i≤n-2), the vector F4 of available 100 dimensioni, as the state of "current" model, F4i+1As current mould Next state of type 1-i, the movement A of model are to select next state for F4iOr F4i+1, according to model 1-'s (i+1) Prediction result OiWith legitimate readingComparison, if the correct so currently return r of predictioni=1, otherwise ri=0, at the same time may be used A is selected to obtain the movement of modeliIf ri=1, then next prediction result is advantageous, Ai=+1, indicate that selection is next A movement, if prediction error, Ai=+2, next step is directly skipped in expression, in order to prevent AiSelection, which has, is biased to Property, model over-fitting is prevented, A is set hereiThere is 0.01 probability to randomly select next movement.The calculating that model is finally returned Formula is following (ξ=0.95):
In formula, R indicates return, riIndicate that n and i are natural number, and ξ takes 0.95 when time return.
Assuming that the weight of network is θ, network function indicates that income is indicated with Q with f2, then:
If loss function y2=(f2-Q(Si,A';θ))2, declined by reversed gradient, until loss function convergence can stop Only train.The network parameter of lower intensified learning screening model described herein is arranged, and the present invention is pre- using 4 layers of deep neural network The return of next state is surveyed, training process is as follows, the result of classification is converted into the array of historical record preservation, each Record is made of following state, current state F4i, next state F4i+1, the movement A that is donei=+1 or Ai=+2, with And the return r of current actioni, it is denoted as four-tuple (F4i, F4i+1, ri, Ai), the overall dimension that inputs is 200 dimensions, and input vector is (F4i, F4i+1, ri), it being input in one four layers full articulamentum, number of network node is respectively set to [128,64,32,16], wherein Dropout layers are added between the second layer and third layer, ratio is set as 0.25, and the output of model is Ai, true output action It is available according to existing return, therefore the probability that can calculate next movement is p (Ai|F4i), realistic operation it is general Rate is p ' (Ai|F4i), here by the way of gradient decline, loss function are as follows:
y3=p (Ai|F4i)*log(p'(Ai|F4i))
Second modular character screening module of following training pattern two, obtains 1- in the first module of model two first Under i model, the output result O of NextStatei+1If acting A in 1-i modeli=+2, indicate the meter of model 1- (i+1) Result mistake is calculated, then the output result of next state 0 filling, i.e. Oi+1=0, the input dimension of model two is maintained in this way Unanimously, here using XGBoost decision tree for the selection result O'=< O' of two first stage of model1,O'2,...,O'N> Further classification reversely adjusts the training parameter of model also according to true prediction result O ", updates the classification power of decision tree Weight, sets loss function here as y4, calculation is as follows:
The training process of model three is as follows, and structure is embodied as shown in Fig. 2, training process is as follows, from existing training A part of data set Dc is selected in sample at random, the quantity of Dc is K, takes K=500 item here, takes a trained mould These data are input to the output vector F4 that the last layer is obtained in model one by type 1-iiAnd i, it is also fed to model 1-n In obtain output vector F4nAnd n, therefore use 500 data sets, available 1000 training samples and corresponding Label is then enter into full articulamentum neural network model three, and the network number of plies is set as [512,128,64,32], Wherein two or three layers of full articulamentum access dropout, and ratio is set as 0.35, in final softmax layers of access, obtain the defeated of 2 dimensions Then outgoing vector is classified, predict input F4 vector generate the result is that i or n, that is, judge that current training sample comes from In model 1-i or model 1-n, using the accuracy rate of classification as the similarity of model, because if the accuracy rate of category of model It is higher to indicate that the similarity between model is smaller, it is easy to distinguish, if the accuracy rate of model illustrates the phase of model 0.5 or so It is big like degree, it should not distinguish.If the result of prediction is pk, legitimate reading pk' loss function y5It calculates as follows:
After the training of three above module is completed, followed by the use of model, first in the actual use of input a part Data to 1-1~1-10 totally 10 models of model one, obtain exporting preliminary classification results O1-O10And model one is complete Connect the 4th layer of output vector F41-F410, it is then enter into trained model two, the reinforcing of model two first Learning layer can screen the classification results of above-mentioned 1-10 model, if screening has obtained the classification results of model one, Current classification results are so saved, if eliminating the classification results of model one, completion are filled with 0, this ensure that model Two input dimension is consistent, finally obtains the output vector of a N-dimensional, and N takes 20 here, is then input to 20 output results In the XGBoost categorised decision tree of model two, final result of secondary classification results as intention assessment is obtained, if meaning Figure identification model for user dialog information misinterpretation, then according to the practical class of service handled of user to classification error Corpus is marked, and is put into new training corpus set E, if the quantity count (E) of E reaches certain threshold value δ, that Data set E is divided into two training datasets of a and b, one model 1-11 of Ea data set re -training, then by Eb data Integrate to be input in model 1-1~1-11 and generates 11 outputs as a result, for the total amount of setting model as N (N=20), model two will here Above-mentioned output result is converted into the four-tuple of intensified learning model then in conjunction with experience replay new mechanism intensified learning model, from In filter out a part of classification results after, be input in the XGBoost decision tree of model two, obtain final point of model two Class adjusts the weight of network as a result, then compare with the actual result currently expected, and so far the training process of model two terminates, The iteration above process reaches 20 until the number of model, while training pattern 1-21, has trained before use Model 1-20 and other 1-1~-1-19 mode inputs into model three, obtain the similarity of each model, will be similar Degree takes the highest model 1-k of similarity according to descending sort from big to small1, it is removed from original model, 1-k1Later Model serial number subtracts 1, and then using newly trained model as 1-20, more new model two, obtains new classification results again.Subsequent mould The iteration update of type repeats the above process, and keeping the maximum number of model is 20.
For in the usage mode of above-mentioned model iteration, for the excavation of text information in model one, Integrated Understanding is single Also in relation with the corpus information between sentence and sentence on the basis of sentence characteristics, can more understand dialogue included in semanteme with Timing information.In the use process of model, since the data set that model itself uses is constantly increasing, model it is fault-tolerant Property constantly enhancing, and existing model is screened using the structure of intensified learning in model two, ensure that model Then robustness carries out secondary classification using classification results of the decision-tree model to model one, further promotes classifying quality.It is right In in practice to data, if classification error, one model of data set re -training of classification error is further used, is had The classification weakness for targetedly promoting existing model, forms the closed loop Training strategy an of positive feedback, in incremental learning sample Characteristic information.Model uses distributed training method, and the degree of coupling between model one, model two and model three is low, can With parallel distributed training, after the training of model one is completed, model two and model three can be updated and calculate simultaneously, mention The efficiency of model is risen, during iteration, only new model 1-n and model two need to update for new data set model Weight, existing model keep original state, as far as possible reduction operation time, the stabilization of model are also ensured, in pattern number Amount is screened after reaching a certain level using three arbiter of model, and disadvantage or similar model are eliminated, and keeps pattern number Amount is in stable state, is also prevented from the excessive duplicate result of generation in model one and interferes to the classification of model two.
The above description is merely a specific embodiment, but scope of protection of the present invention is not limited thereto, any Those familiar with the art in the technical scope disclosed by the present invention, can readily occur in various equivalent modifications or replace It changes, these modifications or substitutions should be covered by the protection scope of the present invention.Therefore, protection scope of the present invention should be with right It is required that protection scope subject to.

Claims (8)

1. a kind of iterative increment dialogue based on small sample is intended to classification recognition methods, which is characterized in that the recognition methods packet Include following steps:
Step 1: segment for the text sentence in dialogue intention and training obtains term vector;
Step 2: sentence vector is obtained after successively extracting using LSTM network and CNN network characterization for a part of term vector, it will Another part term vector, to train preliminary classification model, is input to preliminary point that training finishes by taxon by sentence vector Class model obtains preliminary classification results;
Step 3: for preliminary classification results, passing through XGBoost model after being screened according to the return of intensified learning model Secondary classification is carried out, and disaggregated model again is trained using the method for gradient decline;
Step 4: re-entering training one new preliminary point using the classification error sample in entire disaggregated model training process Class model, return step 3, is iterated incremental learning;
Step 5: utilizing instruction when the quantity of preliminary disaggregated model rises to preset threshold in updating iterative increment learning process Practice the arbiter model finished and carry out similarity calculation, and removes one of highest two models of similarity to remain whole Model computational stability;
Step 6: circulation executes step 3 to step 5, and gradual perfection classification results simultaneously obtain finally corresponding to identification types result.
2. a kind of iterative increment dialogue based on small sample according to claim 1 is intended to classification recognition methods, special Sign is, the output probability of each classification in the preliminary classification results in the step 2 are as follows:
In formula, P (Xi|Dk) indicate the output probability of each classification, XiIndicate actual classification as a result, Xi' indicate prediction classification knot Fruit, X indicate all classification results.
3. a kind of iterative increment dialogue based on small sample according to claim 1 is intended to classification recognition methods, special Sign is, the loss function in preliminary classification results in the step 2 are as follows:
In formula, y1Indicate the loss function in preliminary classification model output result, η1For adjustment parameter, P indicates true intention classification Label probability, P' indicate prediction class label probability.
4. a kind of iterative increment dialogue based on small sample according to claim 1 is intended to classification recognition methods, special Sign is, the calculation formula of the return in the step 3 are as follows:
In formula, R indicates return, riIndicate that n and i are natural number, and ξ takes 0.95 when time return.
5. a kind of iterative increment dialogue based on small sample according to claim 1 is intended to classification recognition methods, special Sign is that the taxon in the step 2 is full Connection Neural Network, the preliminary classification model packet in the step 2 Include mutual sequentially connected sentence vector extract layer, dialogue vector extract layer and classification layer, the classification mould again in the step 3 Type includes intensified learning interconnected-model discrimination module and decision tree-categorization module, the arbiter mould in the step 5 Type is full Connection Neural Network.
6. a kind of iterative increment dialogue based on small sample according to claim 1 is intended to classification recognition methods, special Sign is, is directed to preliminary classification results in the step 3, corresponds to loss function according to the process that return is screened Are as follows:
y3=p (Ai|F4i)*log(p'(Ai|F4i))
In formula, y3It indicates to correspond to loss function, p (A according to the process that return is screened for preliminary classification model output resulti |F4i) indicate to correspond to next in loss function move according to the process that return is screened for preliminary classification model output result The probability of work, p'(Ai|F4i) indicate the corresponding loss of process screened for preliminary classification model output result according to return The probability of realistic operation in function.
7. a kind of iterative increment dialogue based on small sample according to claim 1 is intended to classification recognition methods, special Sign is, corresponds to loss function by the process that XGBoost model carries out secondary classification in the step 3 are as follows:
In formula, y4It indicates to correspond to loss function, O by the process that XGBoost model is further classifiedi' indicate to pass through The process that XGBoost model is further classified corresponds to the selection result of first stage in loss function, Oi" indicate to pass through The process that XGBoost model is further classified corresponds to further sorted result in loss function.
8. a kind of iterative increment dialogue based on small sample according to claim 1 is intended to classification recognition methods, special Sign is, arbiter model carries out similarity calculation in the step 5, and remove highest two models of similarity wherein it One process corresponds to loss function are as follows:
In formula, y5Indicate that the process of arbiter model progress similarity screening corresponds to loss function, K is indicated from existing trained sample The quantity of a part of data set, p are randomly selected in thiskIndicate that arbiter model carries out the corresponding loss letter of process of similarity screening Legitimate reading in number, pk' indicate that the process of arbiter model progress similarity screening corresponds to the prediction result in loss function.
CN201910505469.3A 2019-06-12 2019-06-12 Iterative incremental dialogue intention type recognition method based on small sample Active CN110298391B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910505469.3A CN110298391B (en) 2019-06-12 2019-06-12 Iterative incremental dialogue intention type recognition method based on small sample

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910505469.3A CN110298391B (en) 2019-06-12 2019-06-12 Iterative incremental dialogue intention type recognition method based on small sample

Publications (2)

Publication Number Publication Date
CN110298391A true CN110298391A (en) 2019-10-01
CN110298391B CN110298391B (en) 2023-05-02

Family

ID=68027822

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910505469.3A Active CN110298391B (en) 2019-06-12 2019-06-12 Iterative incremental dialogue intention type recognition method based on small sample

Country Status (1)

Country Link
CN (1) CN110298391B (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110689878A (en) * 2019-10-11 2020-01-14 浙江百应科技有限公司 XLNET-based intelligent voice conversation intention recognition method
CN110704641A (en) * 2019-10-11 2020-01-17 零犀(北京)科技有限公司 Ten-thousand-level intention classification method and device, storage medium and electronic equipment
CN110728313A (en) * 2019-09-29 2020-01-24 北京声智科技有限公司 Classification model training method and device for intention classification recognition
CN110766086A (en) * 2019-10-28 2020-02-07 支付宝(杭州)信息技术有限公司 Method and device for fusing multiple classification models based on reinforcement learning model
CN110782008A (en) * 2019-10-16 2020-02-11 北京百分点信息科技有限公司 Training method, prediction method and device of deep learning model
CN110969006A (en) * 2019-12-02 2020-04-07 支付宝(杭州)信息技术有限公司 Training method and system of text sequencing model
CN111028244A (en) * 2019-12-04 2020-04-17 电子科技大学 Remote sensing image semantic segmentation method based on super-pixel under condition of known sample imbalance
CN111324727A (en) * 2020-02-19 2020-06-23 百度在线网络技术(北京)有限公司 User intention recognition method, device, equipment and readable storage medium
CN111339767A (en) * 2020-02-21 2020-06-26 百度在线网络技术(北京)有限公司 Conversation source data processing method and device, electronic equipment and computer readable medium
CN111414936A (en) * 2020-02-24 2020-07-14 北京迈格威科技有限公司 Determination method of classification network, image detection method, device, equipment and medium
CN111580411A (en) * 2020-04-27 2020-08-25 珠海格力电器股份有限公司 Control parameter optimization method, device and system
CN111611347A (en) * 2020-05-22 2020-09-01 上海乐言信息科技有限公司 Dialog state tracking and training method and system of task-based dialog system
CN111859903A (en) * 2020-07-30 2020-10-30 苏州思必驰信息科技有限公司 Event co-fingering model training method and event co-fingering resolution method
CN112069302A (en) * 2020-09-15 2020-12-11 腾讯科技(深圳)有限公司 Training method of conversation intention recognition model, conversation intention recognition method and device
CN112182213A (en) * 2020-09-27 2021-01-05 中润普达(十堰)大数据中心有限公司 Modeling method based on abnormal lacrimation feature cognition
CN112329475A (en) * 2020-11-03 2021-02-05 海信视像科技股份有限公司 Statement processing method and device
CN112487811A (en) * 2020-10-21 2021-03-12 上海旻浦科技有限公司 Cascading information extraction system and method based on reinforcement learning
CN112527969A (en) * 2020-12-22 2021-03-19 上海浦东发展银行股份有限公司 Incremental intention clustering method, device, equipment and storage medium
CN112734030A (en) * 2020-12-31 2021-04-30 中国科学技术大学 Unmanned platform decision learning method for empirical playback sampling by using state similarity
CN112989049A (en) * 2021-03-30 2021-06-18 广东工业大学 Small sample text classification method and device, computer equipment and storage medium
CN113077057A (en) * 2021-04-20 2021-07-06 中国科学技术大学 Unbiased machine learning method
CN113326689A (en) * 2020-02-28 2021-08-31 中国科学院声学研究所 Data cleaning method and device based on deep reinforcement learning model
CN113468326A (en) * 2021-06-16 2021-10-01 北京明略软件系统有限公司 Method and device for determining document classification
CN113569918A (en) * 2021-07-05 2021-10-29 北京淇瑀信息科技有限公司 Classification temperature adjusting method, classification temperature adjusting device, electronic equipment and medium
CN113569986A (en) * 2021-08-18 2021-10-29 网易(杭州)网络有限公司 Computer vision data classification method and device, electronic equipment and storage medium
CN113887643A (en) * 2021-10-12 2022-01-04 西安交通大学 New dialogue intention recognition method based on pseudo label self-training and source domain retraining
CN114722208A (en) * 2022-06-08 2022-07-08 成都健康医联信息产业有限公司 Automatic classification and safety level grading method for health medical texts
CN114785890A (en) * 2021-12-31 2022-07-22 北京泰迪熊移动科技有限公司 Crank call identification method and device
CN115329723A (en) * 2022-10-17 2022-11-11 广州数说故事信息科技有限公司 User circle layer mining method, device, medium and equipment based on small sample learning
CN115329776A (en) * 2022-10-18 2022-11-11 南京众智维信息科技有限公司 Semantic analysis method for network security co-processing based on less-sample learning
CN115457781A (en) * 2022-09-13 2022-12-09 内蒙古工业大学 Intelligent traffic signal lamp control method based on multi-agent deep reinforcement learning
CN115481221A (en) * 2021-05-31 2022-12-16 腾讯科技(深圳)有限公司 Method, device and equipment for enhancing dialogue data and computer storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107885853A (en) * 2017-11-14 2018-04-06 同济大学 A kind of combined type file classification method based on deep learning
CN108108351A (en) * 2017-12-05 2018-06-01 华南理工大学 A kind of text sentiment classification method based on deep learning built-up pattern
CN108364028A (en) * 2018-03-06 2018-08-03 中国科学院信息工程研究所 A kind of internet site automatic classification method based on deep learning
CN108734276A (en) * 2018-04-28 2018-11-02 同济大学 A kind of learning by imitation dialogue generation method generating network based on confrontation
CN109189925A (en) * 2018-08-16 2019-01-11 华南师范大学 Term vector model based on mutual information and based on the file classification method of CNN
CN109213851A (en) * 2018-07-04 2019-01-15 中国科学院自动化研究所 Across the language transfer method of speech understanding in conversational system
CN109508655A (en) * 2018-10-28 2019-03-22 北京化工大学 The SAR target identification method of incomplete training set based on twin network
CN109829541A (en) * 2019-01-18 2019-05-31 上海交通大学 Deep neural network incremental training method and system based on learning automaton

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107885853A (en) * 2017-11-14 2018-04-06 同济大学 A kind of combined type file classification method based on deep learning
CN108108351A (en) * 2017-12-05 2018-06-01 华南理工大学 A kind of text sentiment classification method based on deep learning built-up pattern
CN108364028A (en) * 2018-03-06 2018-08-03 中国科学院信息工程研究所 A kind of internet site automatic classification method based on deep learning
CN108734276A (en) * 2018-04-28 2018-11-02 同济大学 A kind of learning by imitation dialogue generation method generating network based on confrontation
CN109213851A (en) * 2018-07-04 2019-01-15 中国科学院自动化研究所 Across the language transfer method of speech understanding in conversational system
CN109189925A (en) * 2018-08-16 2019-01-11 华南师范大学 Term vector model based on mutual information and based on the file classification method of CNN
CN109508655A (en) * 2018-10-28 2019-03-22 北京化工大学 The SAR target identification method of incomplete training set based on twin network
CN109829541A (en) * 2019-01-18 2019-05-31 上海交通大学 Deep neural network incremental training method and system based on learning automaton

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
孙旭明: "基于半监督学习的文本分类关键技术研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 *
杨志明: "深度学习算法在问句意图分类中的应用研究", 《计算机工程与应用》 *
王广敏: "改进的多模型融合技术在客服问答系统上的应用", 《电信科学》 *
薛浩: "在线问答社区推荐算法研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 *
陆尧: "基于实体关系的问答系统的相关技术研究", 《中国优秀硕士学位论文全文数据库信息科技辑》 *

Cited By (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110728313A (en) * 2019-09-29 2020-01-24 北京声智科技有限公司 Classification model training method and device for intention classification recognition
CN110689878B (en) * 2019-10-11 2020-07-28 浙江百应科技有限公司 Intelligent voice conversation intention recognition method based on X L Net
CN110704641A (en) * 2019-10-11 2020-01-17 零犀(北京)科技有限公司 Ten-thousand-level intention classification method and device, storage medium and electronic equipment
CN110689878A (en) * 2019-10-11 2020-01-14 浙江百应科技有限公司 XLNET-based intelligent voice conversation intention recognition method
CN110782008A (en) * 2019-10-16 2020-02-11 北京百分点信息科技有限公司 Training method, prediction method and device of deep learning model
CN110766086A (en) * 2019-10-28 2020-02-07 支付宝(杭州)信息技术有限公司 Method and device for fusing multiple classification models based on reinforcement learning model
CN110766086B (en) * 2019-10-28 2022-07-22 支付宝(杭州)信息技术有限公司 Method and device for fusing multiple classification models based on reinforcement learning model
CN110969006A (en) * 2019-12-02 2020-04-07 支付宝(杭州)信息技术有限公司 Training method and system of text sequencing model
CN110969006B (en) * 2019-12-02 2023-03-21 支付宝(杭州)信息技术有限公司 Training method and system of text sequencing model
CN111028244A (en) * 2019-12-04 2020-04-17 电子科技大学 Remote sensing image semantic segmentation method based on super-pixel under condition of known sample imbalance
CN111324727A (en) * 2020-02-19 2020-06-23 百度在线网络技术(北京)有限公司 User intention recognition method, device, equipment and readable storage medium
CN111324727B (en) * 2020-02-19 2023-08-01 百度在线网络技术(北京)有限公司 User intention recognition method, device, equipment and readable storage medium
US11646016B2 (en) 2020-02-19 2023-05-09 Baidu Online Network Technology (Beijing) Co., Ltd. Method and apparatus for recognizing user intention, device, and readable storage medium
CN111339767B (en) * 2020-02-21 2023-07-21 百度在线网络技术(北京)有限公司 Dialogue source data processing method and device, electronic equipment and computer readable medium
CN111339767A (en) * 2020-02-21 2020-06-26 百度在线网络技术(北京)有限公司 Conversation source data processing method and device, electronic equipment and computer readable medium
CN111414936A (en) * 2020-02-24 2020-07-14 北京迈格威科技有限公司 Determination method of classification network, image detection method, device, equipment and medium
CN111414936B (en) * 2020-02-24 2023-08-18 北京迈格威科技有限公司 Determination method, image detection method, device, equipment and medium of classification network
CN113326689B (en) * 2020-02-28 2023-08-18 中国科学院声学研究所 Data cleaning method and device based on deep reinforcement learning model
CN113326689A (en) * 2020-02-28 2021-08-31 中国科学院声学研究所 Data cleaning method and device based on deep reinforcement learning model
CN111580411A (en) * 2020-04-27 2020-08-25 珠海格力电器股份有限公司 Control parameter optimization method, device and system
CN111611347A (en) * 2020-05-22 2020-09-01 上海乐言信息科技有限公司 Dialog state tracking and training method and system of task-based dialog system
CN111859903A (en) * 2020-07-30 2020-10-30 苏州思必驰信息科技有限公司 Event co-fingering model training method and event co-fingering resolution method
CN111859903B (en) * 2020-07-30 2024-01-12 思必驰科技股份有限公司 Event same-index model training method and event same-index resolution method
CN112069302A (en) * 2020-09-15 2020-12-11 腾讯科技(深圳)有限公司 Training method of conversation intention recognition model, conversation intention recognition method and device
CN112069302B (en) * 2020-09-15 2024-03-08 腾讯科技(深圳)有限公司 Training method of conversation intention recognition model, conversation intention recognition method and device
CN112182213A (en) * 2020-09-27 2021-01-05 中润普达(十堰)大数据中心有限公司 Modeling method based on abnormal lacrimation feature cognition
CN112182213B (en) * 2020-09-27 2022-07-05 中润普达(十堰)大数据中心有限公司 Modeling method based on abnormal lacrimation feature cognition
CN112487811A (en) * 2020-10-21 2021-03-12 上海旻浦科技有限公司 Cascading information extraction system and method based on reinforcement learning
CN112329475B (en) * 2020-11-03 2022-05-20 海信视像科技股份有限公司 Statement processing method and device
CN112329475A (en) * 2020-11-03 2021-02-05 海信视像科技股份有限公司 Statement processing method and device
CN112527969A (en) * 2020-12-22 2021-03-19 上海浦东发展银行股份有限公司 Incremental intention clustering method, device, equipment and storage medium
CN112527969B (en) * 2020-12-22 2022-11-15 上海浦东发展银行股份有限公司 Incremental intention clustering method, device, equipment and storage medium
CN112734030B (en) * 2020-12-31 2022-09-02 中国科学技术大学 Unmanned platform decision learning method for empirical playback sampling by using state similarity
CN112734030A (en) * 2020-12-31 2021-04-30 中国科学技术大学 Unmanned platform decision learning method for empirical playback sampling by using state similarity
CN112989049A (en) * 2021-03-30 2021-06-18 广东工业大学 Small sample text classification method and device, computer equipment and storage medium
CN113077057A (en) * 2021-04-20 2021-07-06 中国科学技术大学 Unbiased machine learning method
CN113077057B (en) * 2021-04-20 2022-09-06 中国科学技术大学 Unbiased machine learning method
CN115481221A (en) * 2021-05-31 2022-12-16 腾讯科技(深圳)有限公司 Method, device and equipment for enhancing dialogue data and computer storage medium
CN113468326A (en) * 2021-06-16 2021-10-01 北京明略软件系统有限公司 Method and device for determining document classification
CN113569918A (en) * 2021-07-05 2021-10-29 北京淇瑀信息科技有限公司 Classification temperature adjusting method, classification temperature adjusting device, electronic equipment and medium
CN113569986B (en) * 2021-08-18 2023-06-30 网易(杭州)网络有限公司 Computer vision data classification method, device, electronic equipment and storage medium
CN113569986A (en) * 2021-08-18 2021-10-29 网易(杭州)网络有限公司 Computer vision data classification method and device, electronic equipment and storage medium
CN113887643A (en) * 2021-10-12 2022-01-04 西安交通大学 New dialogue intention recognition method based on pseudo label self-training and source domain retraining
CN114785890A (en) * 2021-12-31 2022-07-22 北京泰迪熊移动科技有限公司 Crank call identification method and device
CN114722208A (en) * 2022-06-08 2022-07-08 成都健康医联信息产业有限公司 Automatic classification and safety level grading method for health medical texts
CN115457781A (en) * 2022-09-13 2022-12-09 内蒙古工业大学 Intelligent traffic signal lamp control method based on multi-agent deep reinforcement learning
CN115457781B (en) * 2022-09-13 2023-07-11 内蒙古工业大学 Intelligent traffic signal lamp control method based on multi-agent deep reinforcement learning
CN115329723A (en) * 2022-10-17 2022-11-11 广州数说故事信息科技有限公司 User circle layer mining method, device, medium and equipment based on small sample learning
CN115329776B (en) * 2022-10-18 2023-02-07 南京众智维信息科技有限公司 Semantic analysis method for network security co-processing based on less-sample learning
CN115329776A (en) * 2022-10-18 2022-11-11 南京众智维信息科技有限公司 Semantic analysis method for network security co-processing based on less-sample learning

Also Published As

Publication number Publication date
CN110298391B (en) 2023-05-02

Similar Documents

Publication Publication Date Title
CN110298391A (en) A kind of iterative increment dialogue intention classification recognition methods based on small sample
CN110263323A (en) Keyword abstraction method and system based on the long Memory Neural Networks in short-term of fence type
CN107766929B (en) Model analysis method and device
CN110110080A (en) Textual classification model training method, device, computer equipment and storage medium
CN112685504B (en) Production process-oriented distributed migration chart learning method
CN109902740B (en) Re-learning industrial control intrusion detection method based on multi-algorithm fusion parallelism
CN108549658A (en) A kind of deep learning video answering method and system based on the upper attention mechanism of syntactic analysis tree
CN110399490A (en) A kind of barrage file classification method, device, equipment and storage medium
CN109243494A (en) Childhood emotional recognition methods based on the long memory network in short-term of multiple attention mechanism
CN109800795A (en) A kind of fruit and vegetable recognition method and system
CN110059191A (en) A kind of text sentiment classification method and device
CN113378913B (en) Semi-supervised node classification method based on self-supervised learning
CN114092742B (en) Multi-angle-based small sample image classification device and method
CN111353313A (en) Emotion analysis model construction method based on evolutionary neural network architecture search
CN112395393A (en) Remote supervision relation extraction method based on multitask and multiple examples
CN116503676B (en) Picture classification method and system based on knowledge distillation small sample increment learning
CN105930792A (en) Human action classification method based on video local feature dictionary
CN110009025A (en) A kind of semi-supervised additive noise self-encoding encoder for voice lie detection
CN113505120B (en) Double-stage noise cleaning method for large-scale face data set
CN111460097A (en) Small sample text classification method based on TPN
CN111310918A (en) Data processing method and device, computer equipment and storage medium
CN117150026B (en) Text content multi-label classification method and device
CN111783688A (en) Remote sensing image scene classification method based on convolutional neural network
CN112163069A (en) Text classification method based on graph neural network node feature propagation optimization
CN112036179A (en) Electric power plan information extraction method based on text classification and semantic framework

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant