CN109241288A - Update training method, device and the equipment of textual classification model - Google Patents

Update training method, device and the equipment of textual classification model Download PDF

Info

Publication number
CN109241288A
CN109241288A CN201811192187.4A CN201811192187A CN109241288A CN 109241288 A CN109241288 A CN 109241288A CN 201811192187 A CN201811192187 A CN 201811192187A CN 109241288 A CN109241288 A CN 109241288A
Authority
CN
China
Prior art keywords
training
layer
sample text
classification model
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
CN201811192187.4A
Other languages
Chinese (zh)
Inventor
许开河
杨坤
王少军
肖京
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ping An Technology Shenzhen Co Ltd
Original Assignee
Ping An Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ping An Technology Shenzhen Co Ltd filed Critical Ping An Technology Shenzhen Co Ltd
Priority to CN201811192187.4A priority Critical patent/CN109241288A/en
Priority to PCT/CN2018/125250 priority patent/WO2020073531A1/en
Publication of CN109241288A publication Critical patent/CN109241288A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis

Abstract

This disclosure relates to field of artificial intelligence, specifically disclose update training method, device and the equipment of a kind of textual classification model, textual classification model includes extraction of semantics layer and classification layer, before newly-increased knowledge point, the training of extraction of semantics layer and layer of classifying is completed according to the sample data of original knowledge point, comprising: the mark label for obtaining the newly-increased corresponding sample text in knowledge point and being labeled to sample text;By the feature vector for completing the extraction of semantics layer building sample text of training according to sample data;The update training of classification layer is carried out, according to the corresponding mark label of the feature vector of sample text and sample text to realize the update training of textual classification model.When needing to be updated trained to textual classification model, the update training for only carrying out classification layer realizes timely updating for textual classification model so as to substantially shorten the time that textual classification model updates training.

Description

Update training method, device and the equipment of textual classification model
Technical field
This disclosure relates to field of artificial intelligence, in particular to the update training method of a kind of textual classification model, dress It sets and equipment.
Background technique
Textual classification model in existing customer service robot question answering system increases new product newly in the knowledge base of customer service robot Behind relevant knowledge point or the relevant knowledge point of newly-increased hot issue, need to textual classification model carry out re -training, one As one textual classification model of re -training need long time, so as to cause textual classification model update not in time, customer service Robot can not answer the problem of newly-increased knowledge point correlation.
So while textual classification model training time long the problem of causing textual classification model to update not in time, need It solves.
Summary of the invention
In order to solve the problems, such as present in the relevant technologies, present disclose provides a kind of update training sides of textual classification model Method and device.
A kind of update training method of textual classification model, the textual classification model include extraction of semantics layer and classification Layer completes the instruction of the extraction of semantics layer and the classification layer according to the sample data of original knowledge point before newly-increased knowledge point Practice, the update training method of the textual classification model includes:
The mark label for obtaining the newly-increased corresponding sample text in knowledge point and the sample text being labeled;
By according to the sample data complete training the extraction of semantics layer building described in sample text feature to Amount;
The classification is carried out according to the feature vector of the sample text and the corresponding mark label of the sample text The update training of layer, to realize the update training of the textual classification model.
A kind of update training device of textual classification model, the textual classification model include extraction of semantics layer and classification Layer completes the instruction of the extraction of semantics layer and the classification layer according to the sample data of original knowledge point before newly-increased knowledge point Practice, the update training device of the textual classification model includes:
Module is obtained, is configured as executing: obtaining the newly-increased corresponding sample text in knowledge point and to the sample text The mark label being labeled;
Feature vector constructs module, is configured as executing: the semanteme by completing training according to the sample data Extract the feature vector of sample text described in layer building;
Training module is updated, is configured as executing: according to the feature vector of the sample text and the sample text Corresponding mark label carries out the update training of the classification layer, to realize the update training of the textual classification model.
In one embodiment, described eigenvector building module includes:
Participle unit is configured as executing: the semanteme by completing training according to the sample data of original knowledge point Extract layer segments the sample text;
Feature vector construction unit is configured as executing: according to each word in the sample text it is corresponding coding with And the semantic weight of each word constructs the feature vector of the sample text.
In one embodiment, described device further include:
Tag along sort complementary module is configured as executing: according to the corresponding mark label supplement of the sample text The tag along sort of classification layer;
Tag along sort set update module, is configured as executing: updating the classification layer according to the tag along sort supplemented Tag along sort set.
In one embodiment, the update training module includes:
Tag along sort predicting unit is configured as executing: using the classification layer according to the feature of the sample text to Amount prediction obtains tag along sort corresponding to the sample text;
Judging unit is configured as executing: carrying out the obtained tag along sort and the corresponding mark of the sample text Infuse the consistency judgement of label;
Adjustment unit is configured as executing: if inconsistent, adjusting the parameter of the classification layer until obtained described Tag along sort is consistent with the mark label.
In one embodiment, tag along sort predicting unit includes:
Probability prediction unit is configured as executing: according to described eigenvector predicting to obtain using the classification layer described Feature vector corresponds to the probability of each tag along sort in the updated tag along sort set;
Tag along sort determination unit is configured as executing: the probability of traversal each tag along sort, with most probable value Corresponding tag along sort is as the corresponding tag along sort of the sample text.
In one embodiment, described device further include:
Class test module is configured as executing: by the updated textual classification model to several test samples Classify;
Nicety of grading computing module is configured as executing: the text after updating training is calculated according to classification results Nicety of grading of this disaggregated model to several test samples;
Training ending module is updated, is configured as executing: if the nicety of grading reaches designated precision, terminating the text The update training of this disaggregated model.
A kind of update training equipment of textual classification model, comprising:
Processor;
Memory for storage processor executable instruction;
Wherein, the processor is configured to executing the process described above.
A kind of computer readable storage medium, is stored thereon with computer program, and the computer program is held by processor The process described above is realized when row.
By the technical solution of the application, it has been trained according to the sample data of original knowledge point in textual classification model On the basis of, when needing to be updated trained to textual classification model, the update training of classification layer is only carried out, realizes text point The update training of class model realizes textual classification model so as to substantially shorten the time that textual classification model updates training Timely update, and then can be used to carry out newly-increased knowledge point in time related for customer service robot in field of artificial intelligence etc. The reply of problem.
It should be understood that the above general description and the following detailed description are merely exemplary, this can not be limited It is open.
Detailed description of the invention
The drawings herein are incorporated into the specification and forms part of this specification, and shows and meets implementation of the invention Example, and in specification together principle for explaining the present invention.
Fig. 1 is the schematic diagram of the implementation environment according to involved in the disclosure;
Fig. 2 is a kind of block diagram of server shown according to an exemplary embodiment;
Fig. 3 is a kind of flow chart of the update training method of textual classification model shown according to an exemplary embodiment;
Fig. 4 is the flow chart of the step S130 of embodiment illustrated in fig. 3;
Fig. 5 be embodiment illustrated in fig. 3 step S150 before step flow chart;
Fig. 6 is the flow chart of the step S150 of embodiment illustrated in fig. 3;
Fig. 7 is the flow chart of the step S151 of embodiment illustrated in fig. 6;
Fig. 8 be embodiment illustrated in fig. 3 step S150 after step flow chart;
Fig. 9 is a kind of block diagram of the update training device of textual classification model shown according to an exemplary embodiment;
Figure 10 is a kind of block diagram of the update training equipment of textual classification model shown according to an exemplary embodiment.
Specific embodiment
Here will the description is performed on the exemplary embodiment in detail, the example is illustrated in the accompanying drawings.Following description is related to When attached drawing, unless otherwise indicated, the same numbers in different drawings indicate the same or similar elements.Following exemplary embodiment Described in embodiment do not represent all embodiments consistented with the present invention.On the contrary, they be only with it is such as appended The example of device and method being described in detail in claims, some aspects of the invention are consistent.
Fig. 1 is the schematic diagram of the implementation environment according to involved in the disclosure.The implementation environment includes: server 200 and extremely A few terminal 100.
Wherein terminal 100 can be smart phone, tablet computer, laptop, desktop computer etc. can be with server 200 establish network connection and can run the electronic equipment of client, herein without specifically limiting.Terminal 100 and server Wireless or wired network connection has been pre-established between 200, thus, it is realized by the client run on the terminal 100 Terminal 100 is interacted with server 200.
Based on the interaction between server 200 and terminal 100, server 200 can get user on the terminal 100 Then the sample text of input constructs the feature vector of the sample text, carries out classification prediction realization text point to feature vector Update training of class model etc..Terminal 100 can receive the tag along sort for sample text that server 200 is returned.
It should be noted that the disclosure classification method, be not limited to dispose to handle accordingly in server 200 to patrol Volume, it is also possible to the processing logic being deployed in other machines.For example, being deployed in the terminal device for having computing capability The processing logic etc. of the update training of row textual classification model.
Fig. 2 is a kind of block diagram of server shown according to an exemplary embodiment.Server with this hardware configuration It can be used for carrying out the update training of textual classification model and be deployed in implementation environment shown in FIG. 1.
It should be noted that the server is the example for adapting to the disclosure, it must not believe that there is provided to this Any restrictions of open use scope.The server can not be construed to need to rely on or must have shown in Figure 2 One or more component in illustrative server 200.
The hardware configuration of the server can generate biggish difference due to the difference of configuration or performance, as shown in Fig. 2, clothes Business device 200 includes: power supply 210, interface 230, at least a memory 250 and at least central processing unit (CPU, a Central Processing Units)270。
Wherein, power supply 210 is used to provide operating voltage for each hardware device on server 200.
Interface 230 includes an at least wired or wireless network interface 231, at least a string and translation interface 233, at least one defeated Enter output interface 235 and at least USB interface 237 etc., be used for and external device communication, such as carries out data with terminal 100 Transmission.
The carrier that memory 250 is stored as resource, can be read-only memory, random access memory, disk or CD Deng the resource stored thereon includes operating system 251, application program 253 and data 255 etc., and storage mode can be of short duration It stores or permanently stores.Wherein, operating system 251 is for managing and each hardware device in control server 200 and answering It can be Windows with program 253 to realize calculating and processing of the central processing unit 270 to mass data 255 ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM etc..Application program 253 be based on operating system 251 it The upper computer program for completing at least one particular job, may include an at least module (being not shown in Fig. 2), each module The series of computation machine readable instruction to server 200 can be separately included.Data 255 can be stored in disk Sample data etc..
Central processing unit 270 may include the processor of one or more or more, and be set as through bus and memory 250 communications, for the mass data 255 in operation and processing memory 250.
As described in detail above, the server 200 for being applicable in the disclosure will read memory by central processing unit 270 The form of the series of computation machine readable instruction stored in 250 is trained to complete the update of textual classification model.
In the exemplary embodiment, server 200 can be by one or more application specific integrated circuit At (Application Specific Integrated Circuit, abbreviation ASIC), digital signal processor, digital signal Manage equipment, programmable logic device, field programmable gate array, controller, microcontroller, microprocessor or other electronic components It realizes, for executing following file classification methods.Therefore, realize the present invention is not limited to any specific hardware circuit, software with And the combination of the two.
Fig. 3 is a kind of flow chart of the update training method of textual classification model shown according to an exemplary embodiment. The update training method of text disaggregated model can be executed with the server 200 of implementation environment shown in Fig. 1.Implementation shown in Fig. 3 In example, textual classification model includes extraction of semantics layer and classification layer, before newly-increased knowledge point, according to the sample number of original knowledge point According to the training for completing extraction of semantics layer and layer of classifying, the update training method of text disaggregated model the following steps are included:
Step S110, the mark mark for obtaining the newly-increased corresponding sample text in knowledge point and sample text being labeled Label.
The textual classification model of the application is constructed by neural network, and wherein textual classification model can pass through convolution Neural network (CNN), Recognition with Recurrent Neural Network (RNN) etc. can carry out the neural network building of text classification, can also be by more The neural network ensemble of seed type forms, herein without specifically limiting.
After the training for completing extraction of semantics layer and layer of classifying according to the sample data of original knowledge point, by the training, The parameter of extraction of semantics layer and layer of classifying is determined, so that ask relevant to original knowledge point may be implemented in text disaggregated model Topic is classified, i.e., extraction of semantics layer can construct the feature vector of text, and classification layer can be with text based feature vector pair Text is classified.
After textual classification model completes training according to the sample data of original knowledge point, the sample number of original knowledge point According to constituting the database of text disaggregated model.For the textual classification model in different application scene, sample data is different, The database of corresponding textual classification model is not also identical.
Wherein increasing knowledge point newly can be the knowledge point for not including in the database of textual classification model, or for original The knowledge point for thering is the knowledge point in database to modify, herein without specifically limiting.For newly-increased knowledge point, need to utilize The sample text of newly-increased knowledge point and the update for carrying out textual classification model to the mark label that sample text is labeled are instructed Practice.
For example, such as applied to the textual classification model in the customer service robot of insurance field, newly-increased knowledge point can Corresponding to be the new insurance business developed, it is relevant to the new insurance business for increasing the sample text of knowledge point newly Problem, such as the insurance handle process, handle material, handle the relevant issues such as condition, Claims Resolution process;Newly-increased knowledge point can be with It is original settlement of insurance claim reschedualing, it is corresponding, increase settlement of insurance claim stream of the sample text of knowledge point as with the change newly The problem of Cheng Xiangguan.To, the update for carrying out textual classification model after training, customer service robot can be mentioned for user Classify out about the problem of newly-increased knowledge point, and then searches for answer according to classification results and answered to what user's presentation was searched for Case.
The mark label of sample text is the label by manually being classified to the sample text, is being embodied It, can be by being manually labeled to obtain mark label to sample text in example, and save the mark label of mark.
Step S130, the feature vector of the extraction of semantics layer building sample text by completing training according to sample data.
By being described above it is found that being completed after training in extraction of semantics layer according to sample data, the parameter of extraction of semantics layer It determines, in step s 130, according to the feature vector of the extraction of semantics layer building sample text of determined parameter.And later In step, the parameter for adjusting extraction of semantics layer again is not needed, i.e., without the update training of extraction of semantics layer.
Customer service robot especially in field of artificial intelligence, due to customer service robot carry out online service before, Using the textual classification model of a large amount of sample data training customer service robot, the extraction of semantics layer building text of textual classification model The perfect in shape and function of this feature vector.So before by online service after training, when increasing knowledge point newly, extraction of semantics Layer can also construct the feature vector of text.
In one exemplary embodiment, as shown in figure 4, step S130 includes:
Step S131, by according to the sample data of original knowledge point complete training extraction of semantics layer to sample text into Row participle.
Step S132 constructs sample according to the semantic weight of the corresponding coding of each word in sample text and each word The feature vector of text.
Sample text is segmented, sample text is divided into several tactic phrases.Participle can use Segmentation methods carry out, such as can be using the segmentation methods based on string matching, the segmentation methods based on understanding or based on system The segmentation methods etc. of meter, herein without specifically limiting.
Textual classification model is completed after training according to the sample data of original knowledge point, is being constructed by sample data The database of text disaggregated model includes the dictionary according to constructed by sample data in the database, includes in dictionary Semantic weight corresponding to the corresponding coding of word included in sample data and word.
Semantic weight corresponding to word is for being characterized in sample text, percentage contribution of the word to sample text semanteme. Such as in " which process for handling safety car owner card has " this text, the word segmentation result obtained according to step S131 is " to handle ^ The ^ process ^ that safety ^ car owner blocks ^ have ^ which ", " " " having " " which " these three words to the semantic percentage contribution of the text not Greatly, thus three words it is corresponding in the text semantic weight it is smaller, and " handling " " safety " " car owner's card " " process " This four words are bigger to the percentage contribution of text semanteme, so that semantic weight is also opposite in the sample text where four words It is larger.Certainly the corresponding coding of each word and corresponding semantic weight are determining after training, in this application i.e. basis It is determined after the training of the sample data completion extraction of semantics layer of original knowledge point.Certainly, sample data volume is bigger, and dictionary is completeer Kind, the semantic weight of coding corresponding to word and word is also more perfect in dictionary, thus the feature of extraction of semantics layer building text The function of vector is also more perfect.
After the word segmentation is completed, which can be constructed according to the corresponding coding of each word and the corresponding semantic weight of each word The feature vector of this text.In a particular embodiment, generally with coding corresponding to digital representation word, and with real number representation word Corresponding weight, so that the feature vector of constructed input text is real vector.
Step S150 carries out classification layer according to the corresponding mark label of the feature vector of sample text and sample text Training is updated, to realize the update training of textual classification model.
The parameter of training adjustment classification layer i.e. in updating training process is updated to classification layer.According to sample text Feature vector and sample text it is corresponding mark label carry out classification layer update training after, textual classification model can be with needle Tag along sort corresponding to the relevant text output text in newly-increased knowledge point realizes the update instruction of textual classification model Practice.
By the technical solution of the application, sufficiently instructed in textual classification model according to the sample data of original knowledge point On the basis of white silk, the perfect in shape and function of the feature vector of extraction of semantics layer building text.It is needing to carry out more textual classification model When new training, the update training of classification layer is only carried out, and extracts the feature vector of layer building sample text using primitive justice, without The update training for carrying out primitive justice extract layer realizes text so as to substantially shorten the update training time of textual classification model This disaggregated model timely updates.
Especially in the customer service robot of field of artificial intelligence, before customer service robot carries out online service, customer service The sample data of the textual classification model of robot hundreds of thousands item easily, sample data volume is big, and the training time is long, text classification mould The function of the feature vector of the extraction of semantics layer building text of type is very perfect.To be updated in textual classification model needs When training, the update training of classification layer is only carried out, and without the update of extraction of semantics layer training, text classification is greatly shortened The update training time of model, and ensure that textual classification model is to original knowledge point, newly-increased knowledge point after updating training The nicety of grading of related text.It is especially less relative to the amount of original knowledge point in newly-increased knowledge point, and need to carry out text When the update training of this disaggregated model, by the technical solution of the application, timely updating for textual classification model may be implemented, and And it is also ensured that textual classification model nicety of grading.
In one exemplary embodiment, as shown in figure 5, before step S150 further include:
Step S010, according to the tag along sort of the corresponding mark label supplementary classification layer of sample text.
Step S030 updates the tag along sort set of classification layer according to the tag along sort supplemented.
It include the classification exportable all classification label of layer in tag along sort set.One mark label corresponds to classification One tag along sort of layer, at newly-increased knowledge point, due to not including the sample text of newly-increased knowledge point in original knowledge point, when So also can not the sample text to newly-increased knowledge point correctly classified.According to the corresponding mark label supplementary classification of sample text Layer tag along sort, and update classification layer tag along sort set after, thus according to sample text carry out classification layer update When training, the tag along sort of sample text can be determined from updated tag along sort set.
In one exemplary embodiment, as shown in fig. 6, step S150 includes:
Step S151 predicts to obtain classification corresponding to sample text according to the feature vector of sample text using classification layer Label.
Step S152, the consistency for carrying out obtained tag along sort and the corresponding mark label of sample text judge.
Step S153, if inconsistent, the parameter of adjustment classification layer is until obtained tag along sort and mark label one It causes.
The training of textual classification model adjusts the parameter of textual classification model in the training process, makes textual classification model The tag along sort of output is consistent with the mark label being manually labeled.If the two is consistent, adjustment text classification mould is not needed The parameter of type, if it is inconsistent, the parameter of adjustment textual classification model is until the two is consistent.In the technical solution of the application In, when updating training by textual classification model, the parameter of adjustment classification layer makes the tag along sort and mark label of sample text Unanimously.
In a particular embodiment, if passing through the tag along sort and the corresponding mark label one of sample text that layer obtains of classifying It causes, then carries out the update training of textual classification model with next sample text.
In the prior art, the either first training or update training of textual classification model, is all extraction of semantics layer It is trained with classification layer, i.e., in training process, if the tag along sort and sample text of classification the exported sample text of layer Mark label it is inconsistent, then adjust extraction of semantics layer and classify layer parameter, until the two is consistent.
Since in textual classification model, the neural network structure of extraction of semantics layer is more complicated, calculating process is more complicated, transports Calculation amount is bigger, and after the parameter for having adjusted extraction of semantics layer, extraction of semantics layer is needed according to parameter adjusted again by transporting The feature vector for calculating building text, so the time of training text disaggregated model is long.
And in this application, only the parameter of adjustment classification layer, is equivalent to and is only updated training to classification layer, so as to Substantially shorten the time that textual classification model updates training.
In actual test, tested using four public data collection, ag_news, Dbpedia, Yahoo!Answer and Safety bank FAQ knowledge base.By Experimental comparison on four data sets, instructed using the update of the textual classification model of the application Training time used in white silk method shorten to full textual classification model re -training and spends time taking 1/10.
In one exemplary embodiment, as shown in fig. 7, step S151 includes:
Step S210 predicts to obtain feature vector corresponding to updated tag along sort according to feature vector using classification layer The probability of each tag along sort in set.
Step S230 traverses the probability of each tag along sort, using tag along sort corresponding to most probable value as sample The corresponding tag along sort of text.
In one exemplary embodiment, as shown in figure 8, after step S150 further include:
Step S171 classifies to several test samples by the updated textual classification model.
The textual classification model after updating training is calculated to several surveys according to classification results in step S172 The nicety of grading of sample sheet.
Step S173 terminates the update training of textual classification model if nicety of grading reaches designated precision.
Wherein step S171-173 is used to test the nicety of grading of textual classification model after update training.Wherein test sample It may include the relevant text of the relevant text in original knowledge point and/or newly-increased knowledge point, preferably include the text of original knowledge point The text of this and newly-increased knowledge point.And test sample is labeled.In step S172, by textual classification model to each survey The tag along sort of this output of sample and the mark of each test sample compare, if the two is consistent, then it is assumed that and classification is accurate, If it is inconsistent, thinking classification error, the ratio that accurate test sample quantity of classifying accounts for total test sample, the ratio are calculated Nicety of grading of the as updated textual classification model to several test samples.
If nicety of grading reaches designated precision, terminate the update training of textual classification model, if nicety of grading is not Reach designated precision, then the update that repeatedly step S110, S130, S150 continues textual classification model is trained.
Following is embodiment of the present disclosure, can be used for executing the text classification that the above-mentioned server 200 of the disclosure executes The update training method embodiment of model.For those undisclosed details in the apparatus embodiments, the disclosure is please referred to The update training method embodiment of disaggregated model.
Fig. 9 is a kind of block diagram of the update training device of textual classification model shown according to an exemplary embodiment, should The update training of textual classification model can be used in the server 200 of implementation environment shown in Fig. 1, execute any of the above embodiment In textual classification model update training method all or part of step.As shown in figure 9, text disaggregated model is more New training device includes but is not limited to: it obtains module 110, feature vector building module 130 and updates training module 150, Middle text disaggregated model includes extraction of semantics layer and classification layer, before newly-increased knowledge point, according to the sample number of original knowledge point According to the training for completing extraction of semantics layer and layer of classifying, which includes:
Obtain module 110, be configured as executing: obtain the newly-increased corresponding sample text in knowledge point and to sample text into The mark label of rower note.
Feature vector constructs module 130, which connect with module 110 is obtained, be configured as executing: by according to sample Data complete the feature vector of the extraction of semantics layer building sample text of training.
Training module 150 is updated, which connect with feature vector building module 130, be configured as executing: according to sample The corresponding mark label of the feature vector and sample text of text carries out the update training of classification layer, to realize text classification mould The update training of type.
In one embodiment, feature vector building module 130 includes:
Participle unit is configured as executing: the extraction of semantics by completing training according to the sample data of original knowledge point Layer segments sample text.
Feature vector construction unit is configured as executing: according to the corresponding coding of each word in sample text and often The feature vector of the semantic weight building sample text of a word.
In one embodiment, the update training device of textual classification model further include:
Tag along sort complementary module is configured as executing: according to the corresponding mark label supplementary classification layer of sample text Tag along sort.
Tag along sort set update module, is configured as executing: point of classification layer is updated according to the tag along sort supplemented Class tag set.
In one embodiment, updating training module 150 includes:
Tag along sort predicting unit is configured as executing: being measured in advance using classification layer according to the feature vector of sample text To tag along sort corresponding to sample text, updated tag along sort set includes tag along sort corresponding to sample text.
Judging unit is configured as executing: carrying out mark label corresponding to obtained tag along sort and sample text Consistency judgement.
Adjustment unit is configured as executing: if inconsistent, the parameter of adjustment classification layer is until obtained tag along sort It is consistent with mark label.
In one embodiment, tag along sort predicting unit includes:
Probability prediction unit is configured as executing: being predicted to obtain feature vector correspondence according to feature vector using classification layer The probability of each tag along sort in updated tag along sort set.
Tag along sort determination unit is configured as executing: traversing the probability of each tag along sort, institute is right with most probable value The tag along sort answered is as the corresponding tag along sort of sample text.
In one embodiment, the update training device of textual classification model further include:
Class test module is configured as executing: by the updated textual classification model to several test samples Classify.
Nicety of grading computing module is configured as executing: the text after updating training is calculated according to classification results Nicety of grading of this disaggregated model to several test samples.
Training ending module is updated, is configured as executing: if nicety of grading reaches designated precision, terminating disaggregated model Update training.
Modules/unit function and the realization process of effect are specifically detailed in above-mentioned textual classification model in above-mentioned apparatus Update training method in correspond to the realization process of step, details are not described herein.
It is appreciated that these module/units can by hardware, software, or a combination of both realize.When in hardware When realization, these modules may be embodied as one or more hardware modules, such as one or more specific integrated circuits.When with soft When part mode is realized, these modules may be embodied as the one or more computer journeys executed on the one or more processors Sequence, such as the program being stored in performed by the central processing unit 270 of Fig. 2 in memory 250.
Optionally, the disclosure also provides a kind of update training equipment of textual classification model, and text sorting device can be with It is the server 200 of implementation environment shown in Fig. 1, executes the whole in the update training method embodiment of the above textual classification model Or part steps.As shown in Figure 10, the update training equipment of text disaggregated model includes:
Processor 1001;
Memory 1002 for 1001 executable instruction of storage processor;
Wherein, processor 1001 is configured as executing in the update training method any embodiment of the above textual classification model All or part of step, executable instruction can be computer-readable instruction, and processor 1001 when being executed, can pass through Communication bus/data line 1003 reads computer-readable instruction from memory 1002.
The processor of equipment in the embodiment executes the concrete mode of operation in related text disaggregated model It updates in the embodiment of training method and performs detailed description, no detailed explanation will be given here.
In the exemplary embodiment, a kind of computer readable storage medium is additionally provided, such as can be includes instruction Provisional and non-transitorycomputer readable storage medium.Storage Jie refers to the memory 250 that can be including instruction, above-mentioned finger Enabling can be executed by the central processing unit 270 of server 200 to complete the update training method of above-mentioned textual classification model.
It should be understood that the present invention is not limited to the precise structure already described above and shown in the accompanying drawings, and And various modifications and change can executed without departing from the scope.The scope of the present invention is limited only by the attached claims.

Claims (10)

1. a kind of update training method of textual classification model, the textual classification model includes extraction of semantics layer and classification layer, Before newly-increased knowledge point, the training of the extraction of semantics layer and the classification layer is completed according to the sample data of original knowledge point, It is characterised by comprising:
The mark label for obtaining the newly-increased corresponding sample text in knowledge point and the sample text being labeled;
By the feature vector for completing sample text described in the trained extraction of semantics layer building according to the sample data;
The classification layer is carried out according to the corresponding mark label of the feature vector of the sample text and the sample text Training is updated, to realize the update training of the textual classification model.
2. the method according to claim 1, wherein the institute by completing training according to the sample data Predicate justice extracts the feature vector of sample text described in layer building, comprising:
The sample text is divided by the extraction of semantics layer for completing training according to the sample data of original knowledge point Word;
The sample text is constructed according to the semantic weight of the corresponding coding of each word in the sample text and each word Feature vector.
3. the method according to claim 1, wherein the feature vector and institute according to the sample text The update training that the corresponding mark label of sample text carries out the classification layer is stated, to realize the update of the textual classification model Before training, further includes:
The tag along sort of the classification layer is supplemented according to the corresponding mark label of the sample text;
The tag along sort set of the classification layer is updated according to the tag along sort supplemented.
4. according to the method described in claim 3, it is characterized in that, the feature vector and institute according to the sample text The update training that the corresponding mark label of sample text carries out the classification layer is stated, to realize the update of the textual classification model Training, comprising:
It is predicted to obtain classification corresponding to the sample text according to the feature vector of the sample text using the classification layer Label;
The consistency for carrying out the obtained tag along sort and the corresponding mark label of the sample text judges;
If inconsistent, the parameter of the classification layer is adjusted until the obtained tag along sort and the mark label one It causes.
5. according to the method described in claim 4, it is characterized in that, it is described using the classification layer according to the sample text Feature vector is predicted to obtain tag along sort corresponding to the sample text, comprising:
Predict to obtain described eigenvector corresponding to the updated classification according to described eigenvector using the classification layer The probability of each tag along sort in tag set;
The probability for traversing each tag along sort, using tag along sort corresponding to most probable value as the sample text pair The tag along sort answered.
6. the method according to claim 1, wherein according to the feature vector of the newly-increased sample and described new Increase the corresponding mark of sample and carry out the update of the classification layer after training, further includes:
Classified by the updated textual classification model to several test samples;
Classification of the textual classification model to several test samples after updating training is calculated according to classification results Precision;
If the nicety of grading reaches designated precision, terminate the update training of the textual classification model.
7. a kind of update training device of textual classification model, the textual classification model includes extraction of semantics layer and classification layer, Before newly-increased knowledge point, the training of the extraction of semantics layer and the classification layer is completed according to the sample data of original knowledge point, It is characterised by comprising:
Module is obtained, is configured as executing: obtaining the newly-increased corresponding sample text in knowledge point and the sample text is carried out The mark label of mark;
Feature vector constructs module, is configured as executing: the extraction of semantics by completing training according to the sample data The feature vector of sample text described in layer building;
Training module is updated, is configured as executing: is corresponding according to the feature vector of the sample text and the sample text Mark label carry out it is described classification layer update training, with realize the textual classification model update training.
8. device according to claim 7, which is characterized in that described eigenvector constructs module and includes:
Participle unit is configured as executing: the extraction of semantics by completing training according to the sample data of original knowledge point Layer segments the sample text;
Feature vector construction unit is configured as executing: according to the corresponding coding of each word in the sample text and often The semantic weight of a word constructs the feature vector of the sample text.
9. a kind of update training equipment of textual classification model characterized by comprising
Processor;
Memory for storage processor executable instruction;
Wherein, the processor is configured to executing such as method described in any one of claims 1 to 6.
10. a kind of computer readable storage medium, is stored thereon with computer program, which is characterized in that the computer program Such as method described in any one of claims 1 to 6 is realized when being executed by processor.
CN201811192187.4A 2018-10-12 2018-10-12 Update training method, device and the equipment of textual classification model Withdrawn CN109241288A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201811192187.4A CN109241288A (en) 2018-10-12 2018-10-12 Update training method, device and the equipment of textual classification model
PCT/CN2018/125250 WO2020073531A1 (en) 2018-10-12 2018-12-29 Update training method, apparatus and device for text classification model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811192187.4A CN109241288A (en) 2018-10-12 2018-10-12 Update training method, device and the equipment of textual classification model

Publications (1)

Publication Number Publication Date
CN109241288A true CN109241288A (en) 2019-01-18

Family

ID=65052732

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811192187.4A Withdrawn CN109241288A (en) 2018-10-12 2018-10-12 Update training method, device and the equipment of textual classification model

Country Status (2)

Country Link
CN (1) CN109241288A (en)
WO (1) WO2020073531A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109994103A (en) * 2019-03-26 2019-07-09 北京博瑞彤芸文化传播股份有限公司 A kind of training method of intelligent semantic Matching Model
CN110472665A (en) * 2019-07-17 2019-11-19 新华三大数据技术有限公司 Model training method, file classification method and relevant apparatus
CN110633476A (en) * 2019-09-27 2019-12-31 北京百度网讯科技有限公司 Method and device for acquiring knowledge annotation information
CN110717023A (en) * 2019-09-18 2020-01-21 平安科技(深圳)有限公司 Method and device for classifying interview answer texts, electronic equipment and storage medium
CN110851546A (en) * 2019-09-23 2020-02-28 京东数字科技控股有限公司 Verification method, model training method, model sharing method, system and medium
CN111522570A (en) * 2020-06-19 2020-08-11 杭州海康威视数字技术股份有限公司 Target library updating method and device, electronic equipment and machine-readable storage medium
CN111737472A (en) * 2020-07-01 2020-10-02 携程计算机技术(上海)有限公司 Method and system for updating text classification model, electronic device and storage medium
CN112148874A (en) * 2020-07-07 2020-12-29 四川长虹电器股份有限公司 Intention identification method and system capable of automatically adding potential intention of user
WO2021119949A1 (en) * 2019-12-16 2021-06-24 深圳市欢太科技有限公司 Text classification model training method, text classification method and apparatus, and electronic device
CN116881464A (en) * 2023-09-06 2023-10-13 北京睿企信息科技有限公司 Method for model training based on newly added label and storage medium

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115186780B (en) * 2022-09-14 2022-12-06 江西风向标智能科技有限公司 Discipline knowledge point classification model training method, system, storage medium and equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108009589A (en) * 2017-12-12 2018-05-08 腾讯科技(深圳)有限公司 Sample data processing method, device and computer-readable recording medium
CN108062331A (en) * 2016-11-08 2018-05-22 南京理工大学 Increment type naive Bayesian file classification method based on Lifelong Learning
CN108090178A (en) * 2017-12-15 2018-05-29 北京锐安科技有限公司 A kind of text data analysis method, device, server and storage medium
CN108520030A (en) * 2018-03-27 2018-09-11 深圳中兴网信科技有限公司 File classification method, Text Classification System and computer installation

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104166706B (en) * 2014-08-08 2017-11-03 苏州大学 Multi-tag grader construction method based on cost-sensitive Active Learning
US9858262B2 (en) * 2014-09-17 2018-01-02 International Business Machines Corporation Information handling system and computer program product for identifying verifiable statements in text
CN108509484B (en) * 2018-01-31 2022-03-11 腾讯科技(深圳)有限公司 Classifier construction and intelligent question and answer method, device, terminal and readable storage medium

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108062331A (en) * 2016-11-08 2018-05-22 南京理工大学 Increment type naive Bayesian file classification method based on Lifelong Learning
CN108009589A (en) * 2017-12-12 2018-05-08 腾讯科技(深圳)有限公司 Sample data processing method, device and computer-readable recording medium
CN108090178A (en) * 2017-12-15 2018-05-29 北京锐安科技有限公司 A kind of text data analysis method, device, server and storage medium
CN108520030A (en) * 2018-03-27 2018-09-11 深圳中兴网信科技有限公司 File classification method, Text Classification System and computer installation

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109994103A (en) * 2019-03-26 2019-07-09 北京博瑞彤芸文化传播股份有限公司 A kind of training method of intelligent semantic Matching Model
CN110472665A (en) * 2019-07-17 2019-11-19 新华三大数据技术有限公司 Model training method, file classification method and relevant apparatus
CN110717023B (en) * 2019-09-18 2023-11-07 平安科技(深圳)有限公司 Method and device for classifying interview answer text, electronic equipment and storage medium
CN110717023A (en) * 2019-09-18 2020-01-21 平安科技(深圳)有限公司 Method and device for classifying interview answer texts, electronic equipment and storage medium
CN110851546B (en) * 2019-09-23 2021-06-29 京东数字科技控股有限公司 Verification method, model training method, model sharing method, system and medium
CN110851546A (en) * 2019-09-23 2020-02-28 京东数字科技控股有限公司 Verification method, model training method, model sharing method, system and medium
CN110633476A (en) * 2019-09-27 2019-12-31 北京百度网讯科技有限公司 Method and device for acquiring knowledge annotation information
CN110633476B (en) * 2019-09-27 2024-04-05 北京百度网讯科技有限公司 Method and device for acquiring knowledge annotation information
WO2021119949A1 (en) * 2019-12-16 2021-06-24 深圳市欢太科技有限公司 Text classification model training method, text classification method and apparatus, and electronic device
CN111522570A (en) * 2020-06-19 2020-08-11 杭州海康威视数字技术股份有限公司 Target library updating method and device, electronic equipment and machine-readable storage medium
CN111522570B (en) * 2020-06-19 2023-09-05 杭州海康威视数字技术股份有限公司 Target library updating method and device, electronic equipment and machine-readable storage medium
CN111737472A (en) * 2020-07-01 2020-10-02 携程计算机技术(上海)有限公司 Method and system for updating text classification model, electronic device and storage medium
CN112148874A (en) * 2020-07-07 2020-12-29 四川长虹电器股份有限公司 Intention identification method and system capable of automatically adding potential intention of user
CN116881464A (en) * 2023-09-06 2023-10-13 北京睿企信息科技有限公司 Method for model training based on newly added label and storage medium
CN116881464B (en) * 2023-09-06 2023-11-24 北京睿企信息科技有限公司 Method for model training based on newly added label and storage medium

Also Published As

Publication number Publication date
WO2020073531A1 (en) 2020-04-16

Similar Documents

Publication Publication Date Title
CN109241288A (en) Update training method, device and the equipment of textual classification model
US11720822B2 (en) Gradient-based auto-tuning for machine learning and deep learning models
US11488055B2 (en) Training corpus refinement and incremental updating
CN103679190B (en) Sorter, classification method and electronic equipment
CN111639516B (en) Analysis platform based on machine learning
CN111602148A (en) Regularized neural network architecture search
US11379718B2 (en) Ground truth quality for machine learning models
CN110598869B (en) Classification method and device based on sequence model and electronic equipment
US11550707B2 (en) Systems and methods for generating and executing a test case plan for a software product
CN113204938B (en) Time delay characteristic improvement method and device of integrated circuit and storage medium
CN111259647A (en) Question and answer text matching method, device, medium and electronic equipment based on artificial intelligence
CN112420125A (en) Molecular attribute prediction method and device, intelligent equipment and terminal
CN114144770A (en) System and method for generating data sets for model retraining
US20230004486A1 (en) Software testing using machine learning
CN113947086A (en) Sample data generation method, training method, corpus generation method and apparatus
CN111723222A (en) Image search and training system
CN113312899A (en) Text classification method and device and electronic equipment
CN113515625A (en) Test result classification model training method, classification method and device
CN109376243A (en) File classification method and device
Mat Jizat et al. Evaluation of the transfer learning models in wafer defects classification
CN108733702B (en) Method, device, electronic equipment and medium for extracting upper and lower relation of user query
US20220292393A1 (en) Utilizing machine learning models to generate initiative plans
CN113609018A (en) Test method, training method, device, apparatus, medium, and program product
CN106156077A (en) The method and apparatus selected for mixed model
US11609936B2 (en) Graph data processing method, device, and computer program product

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WW01 Invention patent application withdrawn after publication
WW01 Invention patent application withdrawn after publication

Application publication date: 20190118